Shuo Cheng

32 posts

Shuo Cheng banner
Shuo Cheng

Shuo Cheng

@ShuoCheng94

PhD student @ Georgia Tech | Robotics, Computer Vision, Machine Learning

Atlanta, GA Beigetreten Kasım 2021
332 Folgt566 Follower
Shuo Cheng retweetet
Danfei Xu
Danfei Xu@danfei_xu·
Introducing EgoVerse: an ecosystem for robot learning from egocentric human data. Built and tested by 4 research labs + 3 industry partners, EgoVerse enables both science and scaling 1300+ hrs, 240 scenes, 2000+ tasks, and growing Dataset design, findings, and ecosystem 🧵
English
34
158
856
251.6K
Shuo Cheng retweetet
Danfei Xu
Danfei Xu@danfei_xu·
New essay on robot learning from human data. I like @karpathy’s idea that LLMs are “ghosts” distilled from human knowledge. In robotics, we are attempting something similar: to summon a sensorimotor ghost. Our current ritual is teleoperation. It produces data, but strips away the reflexes, priors, and social interactions that make human behavior rich. My bet: robot learning will scale less with more robots, and more with better models of humans. Right now we lack both the systems and algorithms to model humans well. If we succeed, the result won’t just be better robots. It may be the first learned theory of how humans act in the physical world. Robots would simply be the first place we deploy it.
Danfei Xu@danfei_xu

x.com/i/article/2021…

English
6
13
150
28.2K
Shuo Cheng
Shuo Cheng@ShuoCheng94·
In evaluation, our method delivers up to 30% higher success rates than the co-training baseline and generalizes to scenarios seen only in simulation, marking a step toward scalable robot learning without large real-world datasets. (4/n)
English
1
1
3
519
Shuo Cheng
Shuo Cheng@ShuoCheng94·
Can large-scale sim data enable real-world generalization?🤔 In our new work, we introduce a generalizable domain adaptation setting, where policies must handle real-world situations never presented in the real training data. (1/n)
English
1
14
41
15.5K
Shuo Cheng retweetet
Yangcen Liu
Yangcen Liu@Randle_Liu·
What if one unified method helps robots learn from human videos across many tasks, many robots? Meet ImMimic: Cross-Domain Imitation from Human Videos via Mapping and Interpolation (CoRL 2025 Oral Presentation🏆) @ICatGT Check it here sites.google.com/view/immimic!
English
6
24
78
26.8K
Shuo Cheng retweetet
Simar Kareer
Simar Kareer@simar_kareer·
Introducing EgoMimic - just wear a pair of Project Aria @meta_aria smart glasses 👓 to scale up your imitation learning datasets! Check out what our robot can do. A thread below👇
English
10
54
239
49.4K
Shuo Cheng
Shuo Cheng@ShuoCheng94·
With large-scale simulation study, we show NOD-TAMP can solve challenging tasks with a handful of demos (4 v.s. 500 demos compared to BC) and achieves strong generalization across diverse shapes, spatial layouts, and task goals. (5/N)
English
1
1
22
2.6K
Shuo Cheng
Shuo Cheng@ShuoCheng94·
Together, NOD-TAMP flexibly integrates the adaptation of recorded trajectories with traditional motion planning to generalize across drastically different scene layouts. Here we show the full process of skill planning and adaptation for the mug sorting task. (5/N)
English
1
0
2
1K
Shuo Cheng
Shuo Cheng@ShuoCheng94·
NOD-TAMP reasons about the pre- and post-conditions of each skill in NOD space and plans skill sequences to reach different goals. For instance, it can decide whether to pick a mug by the rim or handle to hang it on rack and use tools to manipulate hard-to-reach objects. (4/N)
English
1
0
2
1.2K
Shuo Cheng
Shuo Cheng@ShuoCheng94·
For skill adaptation, our key insight is to use learned neural object descriptors (NOD) to transform skill trajectories from one task instance to others, thus being able to apply the demoed skills to manipulate unseen object shapes at novel poses. (3/N)
English
1
0
3
1.1K
Shuo Cheng
Shuo Cheng@ShuoCheng94·
NOD-TAMP is a bi-level planner that reasons about (1) what skills to use given a high-level task goal and (2) how to co-adapt each skill and compose them to form a long-horizon trajectory plan. (2/N)
Shuo Cheng tweet media
English
1
0
3
1.3K
Shuo Cheng
Shuo Cheng@ShuoCheng94·
Can we teach a robot hundreds of tasks with only dozens of demos? Introducing NOD-TAMP: A framework that chains together manipulation skills from as few as one demo per skill to compositionally generalize across long-horizon tasks with unseen objects and scenes. (1/N)
English
4
25
166
32.3K
Danfei Xu
Danfei Xu@danfei_xu·
Congratulations to @ShuoCheng94 for leading LEAGUE, 1 of 5 papers out of 1200+ to receive an RA-L best paper award honorable mention at ICRA! As the sole student author on a two-person team in a field trending towards 10+ authors/paper, Shuo's vision and technical prowess shine through. More exciting work is brewing along this line, so stay tuned. I'll be at ICRA this year. Email/DM to catch up. Check out our paper: TuAT7-CC.1, 10:30-12:00 sites.google.com/view/guidedski…
English
3
2
49
5.8K
Shuo Cheng retweetet
Danfei Xu
Danfei Xu@danfei_xu·
Since we are entering the "BC is all you need" phase of Robot Learning😜 --- Robomimic (robomimic.github.io) allows you to play with SOTA algorithms (BC-Transformer, DiffusionPolicy, etc.) on challenging tasks. Also easy to integration with physical robots!
Danfei Xu tweet media
English
2
19
99
13K
Shuo Cheng retweetet
Shangjie Xue
Shangjie Xue@ShangjieXue·
How to represent granular materials for robot manipulation? Introducing our #CoRL2023 project: Neural Field Dynamics Model for Granular Object Piles Manipulation, a field-based dynamics model for granular object piles manipulation. 🌐 arxiv.org/abs/2311.00802 👇 Thread
English
1
7
19
10.1K
Shuo Cheng retweetet
Vaibhav Saxena
Vaibhav Saxena@saxenavaibhav11·
If you're at #ICRA2023 come chat with us about our poster on "Generalizable Pose Estimation using Implicit Scene Representations!" Pod 11 at 3pm BST Read more about our paper: sites.google.com/view/generaliz…
Vaibhav Saxena tweet media
English
0
5
22
2.8K