
Got a nice kick in the butt yesterday to do something I've been putting off. Re-ranker for Arxiv articles is trained and ~industry standard
Matt
2.3K posts

@Matthewagi
agi is here but it got an email job

Got a nice kick in the butt yesterday to do something I've been putting off. Re-ranker for Arxiv articles is trained and ~industry standard

i was wrong. even if you take a post train model with already heavy post training (k2.5 base model is k2), you can do another heavy round and it's likely better than starting from the base model that said, kimi team built k2.5 starting from k2 base model (not the last ckpt) and not k2 post train model?


Are you up for a challenge? openai.com/parameter-golf


I think we're going to come up with an architecture that makes LLMs look like toys

you've heard of ozempic but have you heard of eating a diet of 25% soft plastic by volume

i've been noticing a lot of people developing LLM psychosis lately. this isn't a dunk. i find it very scary.




don't people eventually get a little bored of reading the exact same cookie cutter "insights" on gender dynamics minimally reworded every 2 days ?
