Ruihan Yang

199 posts

Ruihan Yang

Ruihan Yang

@RchalYang

Applied Scientist @ Amazon Frontier AI & Robotics (FAR) PhD from @UCSanDiego Robot Learning / Embodied AI

San Diego, CA Katılım Temmuz 2017
1K Takip Edilen2.4K Takipçiler
Ruihan Yang
Ruihan Yang@RchalYang·
Lately I’ve been thinking about agentic embodied AI: - how to build it - what APIs agents actually need - where roboticists should focus Sharing some very early thoughts below. Curious to hear how others are thinking about this.
Yongyuan Liang@cheryyun_l

In our new blog (written w/ @RchalYang), we discuss where, when, and how to make vibe agents come alive in the real world. cheryyunl.github.io/blog/vibe-agen… We look at two interfaces for bringing AI agents into the physical world: code (composable, transparent, but whose reliability hinges on API design and feedback) and action (fluid, contact-rich, but compounds errors and forgets). The hybrid of both is emerging, but the harness that closes the loop does not yet exist. Every roboticist in 2026 will need to think about which problems or tasks are fundamentally solvable by vibe agents, where that boundary lies, and what scaffolding and harness robots need.

English
0
9
38
6.6K
Jon Barron
Jon Barron@jon_barron·
There's an engineer on YouTube building his own room-scale laundry-picking UFO catcher robot out of QR codes and string, it's one of the most compelling robotics demos I've seen in a while.
English
79
243
4K
511.9K
Ruihan Yang retweetledi
Jianglong Ye
Jianglong Ye@jianglong_ye·
How do we make dexterous hands handle both power and precision tasks with ease? 🫳👌🫰 We introduce Power to Precision (💪➡️🎯), our new paper that optimizes both control and fingertip geometry to unlock robust manipulation from power grasp to fine-grained manipulations. With simplified finger motions and augmented fingertips, the hand can perform diverse motions from pinching a nut🔩 to handling a pan🍳. Check the demos below🎥.
English
13
78
395
109.3K
Ruihan Yang retweetledi
Yongyuan Liang
Yongyuan Liang@cheryyun_l·
Unified multimodal models can generate text and images, but can they truly reason across modalities? 🎨 Introducing ROVER, the first benchmark that evaluates reciprocal cross-modal reasoning in unified models, the next frontier of omnimodal intelligence. 🌐 Project: roverbench.github.io 📄 Paper: arxiv.org/abs/2511.01163 📂 Benchmark: huggingface.co/datasets/chery…
English
6
32
201
80.3K
Ruihan Yang
Ruihan Yang@RchalYang·
@iamRezaSayar all the code has been released, please let me know if there is any questions.
English
1
0
1
26
Ruihan Yang
Ruihan Yang@RchalYang·
How can we leverage diverse human videos to improve robot manipulation? Excited to introduce EgoVLA — a Vision-Language-Action model trained on egocentric human videos by explicitly modeling wrist & hand motion. We build a shared action space between humans and robots, enabling seamless transfer. With some robot demos, EgoVLA becomes a powerful, generalizable robot policy.
English
6
72
492
58.5K
Tom Zhang
Tom Zhang@tom_jiahao·
We are the ones who are serious about robot learning. We won't be just selling the hardware. We will be selling the system, including the hand and the capabilities enabled by learning. Plus not everyone will be building their own hardware, cauz not everyone is serious enough. Our team is all full stack, with more learning folks than hardware folks...
English
1
0
1
198
Vector Wang
Vector Wang@VectorWang2·
XLeRobot 0.3.0 Showcases Open fridge, get drinks, fill ice, wipe table, clean room, take care plants and cats... All for 660$, fully open-sourced, based on HF LeRobot. Teleop with Joy-con, or RL/VLA. Assembly kit ready for purchase soon Stay tuned! github.com/Vector-Wangel/…
English
10
49
315
18.4K
Ruihan Yang retweetledi
Rocky Duan
Rocky Duan@rocky_duan·
We're hiring interns (and full-times) all year long! Please email me if interested.
Rocky Duan tweet media
English
41
81
1.8K
245.4K