Dalcimar Casanova Ph.D. أُعيد تغريده

🚨 Holy shit... LeCun's team just cracked world models wide open.
Everyone's obsessing over the next Claude update.
Meanwhile Yann LeCun quietly dropped a paper that could matter way more long term.
It's called LeWorldModel.
And to understand why it's a big deal, you need to understand the difference between what LLM does and what this does.
LLMs predict the next word. That's it.
They're incredibly good at language. But they don't understand reality.
They can write about a ball bouncing off a wall. They can't predict where it lands.
World models predict what happens next in the physical world. Objects moving, colliding, falling.
That's the foundation for robots that plan, self-driving cars that simulate scenarios, any AI that needs to act in reality instead of just talk about it.
The problem? World models kept collapsing.
The model would cheat by mapping every input to the same output. Like a weather app that predicts "sunny" every single day.
Technically it's predicting. It's just useless. And fixing this required 6+ loss hyperparameters, frozen pre-trained encoders, stop-gradient hacks, exponential moving averages.
A house of cards just to keep the thing from breaking.
LeCun's team (Mila, NYU, Samsung SAIL, Brown) threw all of that out. LeWorldModel uses just 2 loss terms.
A prediction loss and a regularizer called SIGReg that forces representations to stay diverse instead of collapsing into garbage.
6 hyperparameters reduced to 1.
The simplicity IS the breakthrough.
The numbers: 15M parameters. Trains on a single GPU in a few hours. Plans up to 48x faster than foundation-model-based world models.
Uses roughly 200x fewer tokens than alternatives. Competitive across 2D and 3D control tasks.
This isn't a supercomputer experiment. You could run this on your own hardware.
LeCun has been pushing JEPA as the architecture for real AI since 2022.
The criticism was always the same: "sounds nice, doesn't train stably."
LeWorldModel just removed that objection. Small model. Stable training.
No hacks. No frozen encoders. No collapse.
Two AI futures are competing right now.
Path 1: bigger LLMs, more text, more compute.
Path 2: world models that learn physics from raw pixels and plan in real time.
LeWorldModel is the strongest signal yet that Path 2 is real, getting cheaper, and closing in fast.

English






























