Yevgen Chebotar
45 posts

Yevgen Chebotar
@YevgenChebotar
Robotic foundation models @NVIDIA 🤖 Previously @GoogleDeepMind (RT-2, VLAs, Offline RL) and @Figure_robot (Helix)


Introducing DreamZero 🤖🌎 from @nvidia > A 14B “World Action Model” that achieves zero-shot generalization to unseen tasks & few-shot adaptation to new robots > The key? Jointly predicting video & actions in the same diffusion forward pass Project Page: dreamzero0.github.io 🧵 (1/10)

Meet Helix, our in-house AI that reasons like a human Robotics won't get to the home without a step change in capabilities Our robots can now handle virtually any household item:


Our OpenX paper won best paper at ICRA! Congrats to all my co-authors! 🎉🎉 This is an ongoing effort, we recently added new datasets from the community that double the size of the OpenX dataset -- keep 'em coming! :) Check datasets & how to contribute: robotics-transformer-x.github.io


Super simple code change to get value-based deep RL scale *much* better w/ big models across the board on Atari games, robotic manipulation w/ transformers, LLM + text games, & even Chess! Just use classification loss (i.e., cross entropy), not MSE!! arxiv.org/abs/2403.03950🧵⬇️

Is language capable of representing low-level *motions* of a robot? RT-Hierarchy learns an action hierarchy using motions described in language, like “move arm forward” or “close gripper” to improve policy learning. 📜: arxiv.org/abs/2403.01823 🏠: rt-hierarchy.github.io (1/10)







RT-X: generalist AI models lead to 50% improvement over RT-1 and 3x improvement over RT-2, our previous best models. 🔥🥳🧵 Project website: robotics-transformer-x.github.io



Today, we announced 𝗥𝗧-𝟮: a first of its kind vision-language-action model to control robots. 🤖 It learns from both web and robotics data and translates this knowledge into generalised instructions. Find out more: dpmd.ai/introducing-rt2

