MTAN

10 posts

MTAN

MTAN

@striveforlight

Beigetreten Ocak 2025
58 Folgt7 Follower
MTAN
MTAN@striveforlight·
@yifever Spontiferaneousity
English
0
0
0
8
MTAN retweetet
Elad Gil
Elad Gil@eladgil·
Enjoyed chatting w @AndrewYNg * Bleeding Edge of Agentic AI * Will Models Bootstrap Themselves? * Vibe Coding vs. AI Assisted Coding * Successful Founder Profiles * Next Industry Transformations youtube.com/watch?v=SYisFb…
YouTube video
YouTube
English
15
59
260
106.8K
MTAN retweetet
Dwarkesh Patel
Dwarkesh Patel@dwarkesh_sp·
The most interesting part for me is where @karpathy describes why LLMs aren't able to learn like humans. As you would expect, he comes up with a wonderfully evocative phrase to describe RL: “sucking supervision bits through a straw.” A single end reward gets broadcast across every token in a successful trajectory, upweighting even wrong or irrelevant turns that lead to the right answer. > “Humans don't use reinforcement learning, as I've said before. I think they do something different. Reinforcement learning is a lot worse than the average person thinks. Reinforcement learning is terrible. It just so happens that everything that we had before is much worse.” So what do humans do instead? > “The book I’m reading is a set of prompts for me to do synthetic data generation. It's by manipulating that information that you actually gain that knowledge. We have no equivalent of that with LLMs; they don't really do that.” > “I'd love to see during pretraining some kind of a stage where the model thinks through the material and tries to reconcile it with what it already knows. There's no equivalent of any of this. This is all research.” Why can’t we just add this training to LLMs today? > “There are very subtle, hard to understand reasons why it's not trivial. If I just give synthetic generation of the model thinking about a book, you look at it and you're like, 'This looks great. Why can't I train on it?' You could try, but the model will actually get much worse if you continue trying.” > “Say we have a chapter of a book and I ask an LLM to think about it. It will give you something that looks very reasonable. But if I ask it 10 times, you'll notice that all of them are the same.” > “You're not getting the richness and the diversity and the entropy from these models as you would get from humans. How do you get synthetic data generation to work despite the collapse and while maintaining the entropy? It is a research problem.” How do humans get around model collapse? > “These analogies are surprisingly good. Humans collapse during the course of their lives. Children haven't overfit yet. They will say stuff that will shock you. Because they're not yet collapsed. But we [adults] are collapsed. We end up revisiting the same thoughts, we end up saying more and more of the same stuff, the learning rates go down, the collapse continues to get worse, and then everything deteriorates.” In fact, there’s an interesting paper arguing that dreaming evolved to assist generalization, and resist overfitting to daily learning - look up The Overfitted Brain by @erikphoel. I asked Karpathy: Isn’t it interesting that humans learn best at a part of their lives (childhood) whose actual details they completely forget, adults still learn really well but have terrible memory about the particulars of the things they read or watch, and LLMs can memorize arbitrary details about text that no human could but are currently pretty bad at generalization? > “[Fallible human memory] is a feature, not a bug, because it forces you to only learn the generalizable components. LLMs are distracted by all the memory that they have of the pre-trained documents. That's why when I talk about the cognitive core, I actually want to remove the memory. I'd love to have them have less memory so that they have to look things up and they only maintain the algorithms for thought, and the idea of an experiment, and all this cognitive glue for acting.”
Dwarkesh Patel@dwarkesh_sp

The @karpathy interview 0:00:00 – AGI is still a decade away 0:30:33 – LLM cognitive deficits 0:40:53 – RL is terrible 0:50:26 – How do humans learn? 1:07:13 – AGI will blend into 2% GDP growth 1:18:24 – ASI 1:33:38 – Evolution of intelligence & culture 1:43:43 - Why self driving took so long 1:57:08 - Future of education Look up Dwarkesh Podcast on YouTube, Apple Podcasts, Spotify, etc. Enjoy!

English
229
760
5.3K
1M
MTAN
MTAN@striveforlight·
@vedantdotrpm Incident response teams with overworked engineers
English
0
0
0
5
Vedant
Vedant@vedantdotrpm·
So, if everything depends on AWS, what does AWS depend on?
English
2K
1.3K
32.3K
1.9M
Wondermonger
Wondermonger@fireandvision·
I love words
Wondermonger tweet media
English
1
2
14
2.7K
MTAN
MTAN@striveforlight·
@jsuarez @grok It may be worthwhile to try the Ai2 Paper Finder, or you could even try building an agent which searches through arxiv. My guess is that making your own agent would be more expensive but more effective.
English
0
0
0
45
Joseph Suarez 🐡
Joseph Suarez 🐡@jsuarez·
Quick check in on LLM-land: can any models do this? "Go find me all RL papers that use a fast-slow RNN and pick out the ablation numbers with/without" @grok fails terribly
English
2
0
7
984
∿ Ropirito (0commoDTE)
∿ Ropirito (0commoDTE)@ropirito·
The team I used to work for previously has now made themselves obsolete by focusing on shitty RAG pipelines for a year instead of getting the internal safety team to approve long context models for company use. When I told them this would inevitably happen, the tech PM acted like he knew better (hasn’t coded in 5 years). I can assure you that the average corp is unable to keep up to date with AI advancements and will end up dependent on random B2B SaaS instead of just keeping a skunkworks team at hand to implement the latest tech.
∿ Ropirito (0commoDTE) tweet media
English
8
0
33
2.1K