Omar Rayyan

105 posts

Omar Rayyan banner
Omar Rayyan

Omar Rayyan

@omarrayyann

phd @ucla

Katılım Haziran 2014
495 Takip Edilen621 Takipçiler
Sabitlenmiş Tweet
Omar Rayyan
Omar Rayyan@omarrayyann·
MolmoSpaces provides singular scale and diversity. We built a benchmark that puts that scale to use. MolmoSpaces-Bench evaluates zero-shot policies across thousands of environments previously unseen to them under systematic variation, providing insights that go beyond a success rate % More Below:
Ai2@allen_ai

Introducing MolmoSpaces, a large-scale, fully open platform + benchmark for embodied AI research. 🤖 230k+ indoor scenes, 130k+ object models, & 42M annotated robotic grasps—all in one ecosystem.

English
6
16
156
13K
Omar Rayyan retweetledi
William Shen
William Shen@WillShenSaysHi·
𝗧𝗶𝗣𝗧𝗼𝗣 𝗶𝘀 #𝟭 𝗼𝗻 𝗠𝗼𝗹𝗺𝗼𝗦𝗽𝗮𝗰𝗲𝘀! Outperforming VLAs including MolmoAct2 and π₀.₅, and WAMs like DreamZero It's the only method that uses inference-time search and 𝙯𝙚𝙧𝙤 robot data. We didn't do any benchmark-specific tuning.
William Shen tweet media
English
3
15
133
12.3K
Zu Wang
Zu Wang@zuwang95·
Happy to share what I’ve been working on since joining Genesis! GENE-26.5 is a one-of-a-kind, robotics-native multimodal foundation model that learns from diverse, in-the-wild data across modalities and outputs actions enabling a 54-DoF robot system to perform the most dexterous, long-horizon manipulation tasks to date—approaching human-level capability. This is the result of innovations across the full stack—data collection and processing, robot systems, model architecture, training strategies, and scalable evaluation infrastructure.
English
17
26
347
16.7K
Jie Wang
Jie Wang@JieWang_ZJUI·
@snasiriany My complain of robotics now: no real benchmark and frontier lab do not compete against each other publicly --- RoboArena has not receive new generalist from big tech! We should make the evaluation as a service
English
1
0
1
254
Jie Wang
Jie Wang@JieWang_ZJUI·
RoboCasa365 is amazing, interesting to find this leaderboard, where we can have ready-to-use ckpt. And results are not surprising low lol Excited to see #DreamZero, #GeminiRobotics, #Generalist appearing here, we need more scalable sim-based benchmark
Jie Wang tweet media
English
2
6
26
3.4K
Kevin Zakka
Kevin Zakka@kevin_zakka·
Gave my PhD dissertation talk on Friday! It's been an incredible journey made possible by the best advisor who believed in me and gave me the freedom and support to explore. Thank you @pabbeel! And thank you to everyone who came to support and share this milestone with me 🙏
Kevin Zakka tweet media
English
63
15
651
31.1K
Xuning Yang
Xuning Yang@xuningy·
When every generalist robot model scores 95%+ on a benchmark, the numbers become meaningless. What if we built a photorealistic benchmark that never saturates and can generate new scenes and tasks with AI Workflows in minutes? We introduce RoboLab! 🧵(1/6)
Xuning Yang tweet media
English
10
26
144
27.1K
Omar Rayyan retweetledi
Kevin Zakka
Kevin Zakka@kevin_zakka·
Just merged an amazing contribution by @omarrayyann to mjlab's viser viewer: checkpoint hot-swapping! You can now browse and load any checkpoint mid-session without restarting and it works with local checkpoints and W&B runs.
English
1
16
141
9.4K
Omar Rayyan retweetledi
RoboPapers
RoboPapers@RoboPapers·
Benchmarking, evaluating, and developing robotics code is difficult, and part of this is because no simulator really reflects the diversity and scale of real embodiments. Enter MolmoSpaces from AI2: a massive open ecosystem with a range of 230,000 handcrafted and procedurally-generated home environments, including 48,000 manipulable objects. Crucially, MolmoSpaces provides simulation environments which work for both navigation and manipulation. We talked to the team: @YejinKim4, @omarrayyann, and Max Argus, to tell us more. Watch Episode 69 of RoboPapers, with @micoolcho and @DJiafei, now!
English
1
16
74
20.8K
Omar Rayyan retweetledi
Shumo Chu
Shumo Chu@shumochu·
@VilleKuosmanen @pravsels Wouldn’t you need to reproduce pi star 06 with RECAP first? To my best knowledge there is no good OSS version of that.
English
3
0
3
611
Ville🤖
Ville🤖@VilleKuosmanen·
"RL Token" looks like a great and surprisingly simple post-training methodology for optimising robot models for dexterous tasks in the real world! Over the next few weeks, me and @pravsels will be attempting to reproduce the results (& open source the code) Stay tuned 👀
Physical Intelligence@physical_int

We developed an RL method for fine-tuning our models for precise tasks in just a few hours or even minutes. Instead of training the whole model, we add an “RL token” output to π-0.6, our latest model, which is used by a tiny actor and critic to learn quickly with RL.

English
8
5
92
22.3K
Omar Rayyan retweetledi
Mahi Shafiullah 🏠🤖
Mahi Shafiullah 🏠🤖@notmahi·
MolmoSpaces leaderboard is now open for submissions! When we created this benchmark for zero-shot real-to-sim eval in diverse homes, we didn’t expect things to heat up so quickly. But it did, thanks to @jang_yoel and team at GEAR toppling PI to take the crown on task-general category. Congrats 🎉 You can evaluate and submit your model to this leaderboard: molmospaces.allen.ai/leaderboard
Joel Jang@jang_yoel

𝐃𝐫𝐞𝐚𝐦𝐙𝐞𝐫𝐨 𝐢𝐬 #𝟏 𝐨𝐧 𝐛𝐨𝐭𝐡 𝐌𝐨𝐥𝐦𝐨𝐒𝐩𝐚𝐜𝐞𝐬 𝐚𝐧𝐝 𝐑𝐨𝐛𝐨𝐀𝐫𝐞𝐧𝐚 🏆 𝗪𝗵𝗮𝘁 𝗺𝗮𝗸𝗲𝘀 𝘁𝗵𝗶𝘀 𝗻𝗼𝘁𝗮𝗯𝗹𝗲: DreamZero-DROID is trained 𝑓𝑟𝑜𝑚 𝑠𝑐𝑟𝑎𝑡𝑐ℎ using only the DROID dataset. No pretraining on large-scale robot data, unlike competing VLAs. This demonstrates the strength of video-model backbones for generalist robot policies (VAMs/WAMs). More broadly, training 𝑜𝑛𝑙𝑦 on real data and evaluating on (1) transparent, distributed benchmarks like 𝐑𝐨𝐛𝐨𝐀𝐫𝐞𝐧𝐚 or (2) scalable sim-benchmarks like 𝐌𝐨𝐥𝐦𝐨𝐒𝐩𝐚𝐜𝐞𝐬 is an exciting step toward fairer and more reproducible evaluation of generalist policies, one that the community can hillclimb together to measure progress. Special thanks to the Ai2 MolmoSpaces team (@notmahi @omarrayyann @YejinKim4 Max Argus) and the RoboArena team (@pranav_atreya) for helping with the set-up and getting these evaluations! Special shout out to @youliangtan @NadunRanawakaA @chuning_zhu, who led these efforts from the GEAR side :) + We also release our DreamZero-AgiBot checkpoint & post-training code to enable very efficient few-shot adaptation. Post-train on just ~30 minutes of play data for your specific robot, and see the robot do basic language following and pick-and-place 🤗(See YAM experiments in our paper for more detail). ++ We also provide the entire codebase & preprocessed dataset to replicate the DreamZero-DROID checkpoint. 🌐 dreamzero0.github.io 💻 github.com/dreamzero0/dre… RoboArena: robo-arena.github.io/leaderboard MolmoSpaces: molmospaces.allen.ai/leaderboard

English
2
4
40
4.6K
Chris Paxton
Chris Paxton@chris_j_paxton·
Well dreamzero is: - a much bigger model - has this clever auxiliary loss (predicting video) which probably makes its smaller amount of data go a lot farther unfortunately not enough information yet to tell. at least from the comparisons in the paper. it seems like the aux loss stuff made a huge difference (see figure here) but we don't KNOW that pi-0.5 at 14b params wouldn't do well. although it sure seems like it made a difference. i think there's a lot of work to do on the exact best data mixture.
Chris Paxton tweet mediaChris Paxton tweet media
English
3
1
43
3.6K
Omar Rayyan
Omar Rayyan@omarrayyann·
Also thanks to @youliangtan @jang_yoel for their DreamZero API. Their world action model now leads the benchmark with zero sim data. x.com/jang_yoel/stat…
Joel Jang@jang_yoel

𝐃𝐫𝐞𝐚𝐦𝐙𝐞𝐫𝐨 𝐢𝐬 #𝟏 𝐨𝐧 𝐛𝐨𝐭𝐡 𝐌𝐨𝐥𝐦𝐨𝐒𝐩𝐚𝐜𝐞𝐬 𝐚𝐧𝐝 𝐑𝐨𝐛𝐨𝐀𝐫𝐞𝐧𝐚 🏆 𝗪𝗵𝗮𝘁 𝗺𝗮𝗸𝗲𝘀 𝘁𝗵𝗶𝘀 𝗻𝗼𝘁𝗮𝗯𝗹𝗲: DreamZero-DROID is trained 𝑓𝑟𝑜𝑚 𝑠𝑐𝑟𝑎𝑡𝑐ℎ using only the DROID dataset. No pretraining on large-scale robot data, unlike competing VLAs. This demonstrates the strength of video-model backbones for generalist robot policies (VAMs/WAMs). More broadly, training 𝑜𝑛𝑙𝑦 on real data and evaluating on (1) transparent, distributed benchmarks like 𝐑𝐨𝐛𝐨𝐀𝐫𝐞𝐧𝐚 or (2) scalable sim-benchmarks like 𝐌𝐨𝐥𝐦𝐨𝐒𝐩𝐚𝐜𝐞𝐬 is an exciting step toward fairer and more reproducible evaluation of generalist policies, one that the community can hillclimb together to measure progress. Special thanks to the Ai2 MolmoSpaces team (@notmahi @omarrayyann @YejinKim4 Max Argus) and the RoboArena team (@pranav_atreya) for helping with the set-up and getting these evaluations! Special shout out to @youliangtan @NadunRanawakaA @chuning_zhu, who led these efforts from the GEAR side :) + We also release our DreamZero-AgiBot checkpoint & post-training code to enable very efficient few-shot adaptation. Post-train on just ~30 minutes of play data for your specific robot, and see the robot do basic language following and pick-and-place 🤗(See YAM experiments in our paper for more detail). ++ We also provide the entire codebase & preprocessed dataset to replicate the DreamZero-DROID checkpoint. 🌐 dreamzero0.github.io 💻 github.com/dreamzero0/dre… RoboArena: robo-arena.github.io/leaderboard MolmoSpaces: molmospaces.allen.ai/leaderboard

English
0
0
8
360
Omar Rayyan
Omar Rayyan@omarrayyann·
You can get more insights than just the success rate (e.g. AR policies like DreamZero and pi0-Fast generate smoother trajectories) and cross-compare policy performance across objects.
Omar Rayyan tweet mediaOmar Rayyan tweet media
English
1
0
6
195
Omar Rayyan
Omar Rayyan@omarrayyann·
MolmoSpaces-Bench leaderboard is now live! Test your generalist policies to see how they compare across tasks and environments. Feel free to reach out if you need help setting it up. molmospaces.allen.ai/leaderboard
Omar Rayyan tweet media
English
2
5
35
1.8K
Yu Xiang
Yu Xiang@YuXiang_IRVL·
Has anyone built a URDF for the Panda arm + Robotiq 2F-85 gripper setup used in the DROID dataset? Thanks! 🙏
Yu Xiang tweet media
English
5
4
48
4.6K
Alberto Fuentes (e/acc)
Alberto Fuentes (e/acc)@AlberFuen·
@omarrayyann @RanjayKrishna In the middle of the teleop the camera screens Freeze for some time, then lose track of pos. Id like a way to teleop with Xbox comtroller or keyboard. Possibly the grasps are a way to go. How to use them? I was unable to do simple tasks like grab potato with the gripper
English
1
0
0
35
Ranjay Krishna
Ranjay Krishna@RanjayKrishna·
The amount and diversity of robot data we need is much higher than what we can scale. We are betting on simulation! MolmoSpaces allows you to generate seemingly unlimited amounts of robot data in large diverse environments across multiple simulators.
Ai2@allen_ai

Introducing MolmoSpaces, a large-scale, fully open platform + benchmark for embodied AI research. 🤖 230k+ indoor scenes, 130k+ object models, & 42M annotated robotic grasps—all in one ecosystem.

English
5
4
48
6.3K
Omar Rayyan
Omar Rayyan@omarrayyann·
MolmoSpaces also comes with 42M+ grasps that cover 48K+ objects across 250K+ scenes, allowing large-scale functional trajectory generation in MuJoCo and IsaacSim.
English
5
35
319
18.9K