Yu Lei

366 posts

Yu Lei banner
Yu Lei

Yu Lei

@_OutofMemory_

PhD @UTCompSci | Learn to understand ourselves and build intelligence.🤖🧠👁️

Austin, TX Присоединился Temmuz 2023
2.1K Подписки345 Подписчики
Yu Lei ретвитнул
Sergey Levine
Sergey Levine@svlevine·
Back in Nov we developed Recap and trained π*-06 with RL. Now, we developed a fast *online* RL method that improves π-06 with as little as 15 min of robot data for precise tasks, using "RL tokens" exposed by our model that can be fed into a small actor-critic method.
English
8
49
525
35K
Yu Lei
Yu Lei@_OutofMemory_·
Just read drifting model recently, this is exactly what I felt intuitively. It is directly supervising the log-prob gradient as score-based model. The attraction-repulsion instantiation of V is already very generalizable I think. Btw, what is after DMD?
Chieh-Hsin (Jesse) Lai@JCJesseLai

[1/D] 🤔 What are drifting models really connected to? 📢 Our new paper, A Unified View of Drifting and Score-Based Models, shows that the bridge to score-based models is clear and precise (w/ team and @mittu1204, @StefanoErmon, @MoleiTaoMath)! ✍️ Main takeaway: drifting is more closely connected to score-based (diffusion) modeling than it may first appear! 🔗 arxiv.org/abs/2603.07514 🎯 Here’s why: Drifting’s mean-shift moves a sample toward the kernel-weighted average of nearby samples. Score function points toward regions of higher density. So both describe local directions that push samples toward where data is denser. We show that this link is exact for Gaussian kernels (Section 4.1): 📌drifting’s mean-shift = a rescaled score-matching field between the Gaussian-smoothed data and model distributions — the vector field underlying score matching (Tweedie!). 📌This also clarifies the bridge to Distribution Matching Distillation (DMD): both use score-based transport directions, but only differ in how the score is realized—drifting does so nonparametrically through kernel neighborhoods, whereas DMD relies on a pretrained diffusion teacher. 🤔 So what happens for the default Laplace kernel used in drifting models? Let’s look below 👇

English
0
0
2
244
Yu Lei
Yu Lei@_OutofMemory_·
@VectorWang2 Yes. “All models are wrong, but some are useful.”
English
0
0
1
48
Vector Wang
Vector Wang@VectorWang2·
I still hold my opinion that action-conditioned world model fits the core definition of "model": s_t+1=f(s_t, a_t). However, there's no model that can be accurate, even in real life, all we have are blurry "dreams"
Anirudha Majumdar@Majumdar_Ani

x.com/i/article/2033…

English
1
1
15
2.6K
Yu Lei ретвитнул
Jitendra MALIK
Jitendra MALIK@JitendraMalikCV·
With Emmanuel Dupoux scp.net/persons/dupoux/ and Yann LeCun @ylecun, we consider a cognitive science inspired AI. We analyse how autonomous learning works in living organisms, and propose a roadmap for reproducing it in artificial systems. lnkd.in/eNWDmuqT
English
9
78
447
61.5K
Yu Lei ретвитнул
Zhengyi “Zen” Luo
Zhengyi “Zen” Luo@zhengyiluo·
288 hours of high-quality, text-annotated human motion data are now available! 140k motion sequences! Do you know that a large part of SONIC's training data is now open-sourced? Check out the dataset here 👇🏻 from our friends at Bones Studio! Full human + G1 retargeted motion! Stie🌐:bones.studio/datasets/seed Data💿:huggingface.co/datasets/bones… SONIC training code coming VERY VERY soon!
English
7
64
309
27.7K
Phillip Isola
Phillip Isola@phillip_isola·
As models advance and surpass certain human abilities, “human-level” advances too, as we can use them as tools. So yes a model might do better math/coding/etc than I could have done in 2025. But they still are behind where I could be in 2026! This thought gives me some hope :)
English
9
7
152
11.3K
Yu Lei ретвитнул
James Zou
James Zou@james_y_zou·
We created AI agents based on scientists' personas (eg Einstein, Feynman) and built a Kaggle-like platform for them to freely post ideas, compete and collaborate. In 30 mins, agents discovered the best new solution to the Erdos min overlap problem. Great job by @federicobianchy @ykwon_0407! The solution is here github.com/togethercomput…
James Zou tweet mediaJames Zou tweet mediaJames Zou tweet media
English
74
236
1.6K
182.5K
Yu Lei ретвитнул
Huihan Liu
Huihan Liu@huihan_liu·
Catastrophic forgetting has long been a challenge in continual learning. However, our new study found that pretrained Vision-Language-Action (VLA) models are surprisingly resistant to forgetting! Zero forgetting, or even positive backward transfer, is possible with simple experience replay. arxiv.org/abs/2603.03818
Huihan Liu tweet media
English
17
109
670
50.8K
Yu Lei ретвитнул
Yuke Zhu
Yuke Zhu@yukez·
Today, we publicly released RoboCasa365, a large-scale simulation benchmark for training and systematically evaluating generalist robot models. Built upon our original RoboCasa framework, it offers: • 2,500 realistic kitchen environments; • 365 everyday tasks (basic skills + long-horizon mobile manipulation); • Over 3,200 objects with many articulated fixtures/appliances. All are designed for fully controlled, reproducible benchmarking of robotic policies. Progress in robotic foundation models is real. But it’s still hard to answer basic questions like: How close are we to general-purpose autonomy? What factors drive generalization? What are the model/data scaling curves like? Real-world eval is slow and noisy, and existing sims (like LIBERO, which we built 3 years ago) often lack sufficient task and scene diversity. This benchmark comes with 2,200+ hours of demonstrations and 500K+ trajectories to support studies of multi-task training, pretraining, and continual learning at scale. Check it out at robocasa.ai
English
13
60
338
20.6K
Yu Lei ретвитнул
Yuke Zhu
Yuke Zhu@yukez·
CoRL is coming to Austin, TX this November! As General Chair, I'm thrilled to welcome the robot learning community. 2026 feels like a pivotal year as AI-powered robotic systems begin deploying at scale for real-world tasks. This year, I hope CoRL will be the forum that connects cutting-edge research with industrial practice. Please submit your best work and join us in Austin. DM me what you'd love to see CoRL do better! corl.org
Conference on Robot Learning@corl_conf

Calling all researchers! 🤖The CoRL 2026 website is officially live at corl.org with key dates for your submissions: 🗓 May 25: Abstract Submission 🗓 May 28: Full Paper Submission 🗓 Nov 9-12: Conference in Austin, TX Send us your coolest work! #RobotLearning

English
5
7
167
15.2K
Yu Lei ретвитнул
Chongyi Zheng
Chongyi Zheng@chongyiz1·
Representation learning is all about capturing the right prior. What is the right prior for *reinforcement learning*? We propose a new unsupervised pre-training method for RL: chongyi-zheng.github.io/onestep-fb. 🧵⬇️
Chongyi Zheng tweet media
English
3
19
119
8.4K
Yu Lei ретвитнул
Zhengyi “Zen” Luo
Zhengyi “Zen” Luo@zhengyiluo·
SONIC is now open-source! Generalist whole-body teleoperation for EVERYONE! Our team has long been building comprehensive pipelines for whole-body control, kinematic planner, and teleoperation, and they will all be shared. This will be a continuous update; inference code + model already there, training code and gr00t integration coming soon! Code: github.com/NVlabs/GR00T-W… Docs: nvlabs.github.io/GR00T-WholeBod… Site: nvlabs.github.io/GEAR-SONIC/
English
32
196
845
147.5K
Yu Lei
Yu Lei@_OutofMemory_·
Have the privilege to beta-test SONIC. Thanks for the team to open-source! Superior performance as system0. It’s pretty easy to deploy with very-well-written documents (only took me a few hrs). Empirical results speak louder than words. Try out on your robots!
Yuke Zhu@yukez

We have seen rapid progress in humanoid control — specialist robots can reliably generate agile, acrobatic, but preset motions. Our singular focus this year: putting generalist humanoids to do real work. To progress toward this goal, we developed SONIC (nvlabs.github.io/GEAR-SONIC/), a Behavior Foundation Model for real-time, whole-body motion generation that supports teleoperation and VLA inference for loco-manipulation. Today, we’re open-sourcing SONIC on GitHub. We are excited to see what the community builds upon SONIC and to collectively push humanoid intelligence toward real-world deployment at scale. 🌐 Paper: arxiv.org/abs/2511.07820 📃 Code: github.com/NVlabs/GR00T-W…

English
6
21
151
15.5K
Yu Lei ретвитнул
Lihan Zha
Lihan Zha@LihanZha·
Today's state-of-the-art VLAs struggle to generalize zero-shot to new robot embodiments, despite training on extensive multi-embodiment data. We introduce Language-Action Pre-training (LAP) and LAP-3B — the first VLA to achieve substantial zero-shot transfer to unseen real-world robot embodiments, through simply aligning action representation with language. Everything is open-sourced! Try it out on your own robot: 🌐 lap-vla.github.io
English
11
90
326
49.2K
Yu Lei ретвитнул
Yu Lei
Yu Lei@_OutofMemory_·
Adaptive compliant controller is a must since force is everywhere in interactions. Check out Zi-ang and Sirui's impressive work, and integrate this plug-and-play module into your framework!
Zi-ang Cao@ziang_cao

🚀 Introducing CHIP: Adaptive Compliance for Humanoid Control through Hindsight Perturbation! Current humanoids face a trade-off: they are either Agile & Stiff OR Slow & Soft. CHIP breaks this barrier. We enable on-the-fly switching between Compliant (wiping 🧼, collaborative holding 📦) and Stiff (lifting dumbbells 🏋️, opening doors 🚪💪) behaviors—all while maintaining agile skills like running! 🏃💨 Website: nvlabs.github.io/CHIP/ Join me for a deep dive on how CHIP enables adaptive control for complex tasks. 🧵↓

English
1
0
3
381
Zhengyi “Zen” Luo
Zhengyi “Zen” Luo@zhengyiluo·
Doing my CVPR review final justification in the last minute (sorry...) and saw this That.... that's the prompt?
Zhengyi “Zen” Luo tweet media
English
7
3
80
18K
Yu Lei ретвитнул
Fanqi Lin
Fanqi Lin@lfqirrrrr·
𝑪𝒐-𝒕𝒓𝒂𝒊𝒏𝒊𝒏𝒈 is a promising way to scale Large Behavior Models (LBMs) beyond robot data, yet the data and training recipe are far from settled. 🤔 We present a large-scale empirical study leveraging 4,000h of robot/human data and 50M vision-language samples, evaluating 89 policies across 58,000 simulation rollouts and 2,835 real-world trials. 🤖📊 co-training-lbm.github.io Work done during my internship at @ToyotaResearch.
English
7
50
416
46.9K