LeCAR Lab at CMU

38 posts

LeCAR Lab at CMU banner
LeCAR Lab at CMU

LeCAR Lab at CMU

@LeCARLab

Learning and Control for Agile Robotics Lab at @CarnegieMellon @SCSatCMU @CMU_Robotics.

Pittsburgh شامل ہوئے Ağustos 2023
13 فالونگ583 فالوورز
LeCAR Lab at CMU ری ٹویٹ کیا
Guanya Shi
Guanya Shi@GuanyaShi·
My lab at CMU @LeCARLab is hiring a postdoc! (Vibe-made the poster via @ChatGPTapp)
Guanya Shi tweet media
English
1
21
173
14.7K
LeCAR Lab at CMU ری ٹویٹ کیا
Guanya Shi
Guanya Shi@GuanyaShi·
I’m so tired of writing rebuttals to this kind of “lack of novelty” review: “This paper trivially combines A, B, and C, so the algorithmic novelty is limited.” Technically, most (if not all) robotics papers are convex combinations of existing ideas. I still deeply appreciate A+B+C papers—especially when they deliver: - New capabilities: the “trivial combination” unlocks behaviors we simply couldn’t achieve before - Sensible & organic design: A+B+C is clearly the right composition—not some arbitrary A′+B+C′ - Nontrivial interactions: careful analysis of the dynamics, coupling, or failure modes between A, B, C - Rehabilitating old ideas: A was dismissed for years, but paired with modern B/C, it suddenly works—and teaches us why - System-level & "interface" insight: the contribution is not any single piece, but how the pieces talk to each other - Scaling laws or regimes: identifying when/why A+B+C works (and when it doesn’t) - Engineering clarity: making something actually work robustly in the real world is not “trivial” - New problem formulations: sometimes the real novelty is in the reformulation—only under this view does A+B+C make sense. Maybe worth keeping these in mind when reviewing the next A+B+C paper : )
English
30
119
961
110.8K
LeCAR Lab at CMU ری ٹویٹ کیا
Chaoyi Pan
Chaoyi Pan@ChaoyiPan·
Generative models (diffusion/flow) are taking over robotics 🤖. But do we really need to model the full action distribution to control a robot? We suspected the success of Generative Control Policies (GCPs) might be "Much Ado About Noising." We rigorously tested the myths. 🧵👇
English
16
94
546
109.6K
LeCAR Lab at CMU ری ٹویٹ کیا
Guanya Shi
Guanya Shi@GuanyaShi·
We tried to rethink modern robot learning policy design in this paper. TL;DR: Generative policies perform well NOT because of their distributional-learning formulation or their ability to capture “multi-modality.” Our proposed Minimal Iterative Policy (MIP) achieves similar performance to flow with much lower inference & training cost. Some interesting takeaways: 1. “Multi-modal behavior” is a very nuanced concept in robotics. Robot learning cares about the conditional distribution P(a|o), where a is the action (chunk) and o is the observation. The objective of generative modeling in vision/language is fundamentally different from the goal in control. In vision/language, we want diverse samples from the data distribution. In control, any action that leads to better downstream performance is sufficient. Also, robotics has very little (labeled) action data, and they lie on a low-dimensional manifold. As a result, even if the marginal P(a) is multi-modal in many tasks (e.g., PushT), the conditional P(a|o) is often not multi-modal in sparse-data regions. Empirically, we observe that the common explanation—“flow/diffusion policies win because demonstrations are multi-modal”—does not hold for most studied behavior cloning benchmarks. 2. Architecture and action representation matter more than flow vs. regression. We find that the strong performance of flow/diffusion policies is largely due to their modern architectures and action representations. UNet / Transformer designs and action chunking play a vital role. In fact, L1/L2 regression policies can match flow/diffusion when paired with the right architectures and chunking, except in a few high-precision tasks. 3. What actually helps in flow/diffusion? Stochasticity injection and supervised iterative computation. To isolate these factors, we performed a “surgery” on flow to design Minimal Iterative Policy (MIP), a deterministic two-step policy that keeps only these mechanisms. Surprisingly, with far lower compute, MIP is comparable to—or even better than—flow in most studied behavior cloning benchmarks. Why? We hypothesize that these mechanisms provide good inductive biases that improve manifold adherence, especially for OOD states. Flow and MIP exhibit significantly better manifold adherence than regression. I don’t think these conclusions are complete—there are still many mysteries. But I strongly feel we need a new subfield: “the physics/science of robot learning,” aimed at understanding fundamental learning mechanisms and principles for data-driven robotics. As @ilyasut said for LLMs: “We’re moving from the age of scaling to the age of research.” For robot learning, I believe we must embrace the age of research to unlock better scaling to ignite the age of scaling.
Chaoyi Pan@ChaoyiPan

Generative models (diffusion/flow) are taking over robotics 🤖. But do we really need to model the full action distribution to control a robot? We suspected the success of Generative Control Policies (GCPs) might be "Much Ado About Noising." We rigorously tested the myths. 🧵👇

English
3
30
181
14.1K
LeCAR Lab at CMU ری ٹویٹ کیا
Chaoyi Pan
Chaoyi Pan@ChaoyiPan·
🕸️ Introducing SPIDER — Scalable Physics-Informed Dexterous Retargeting! A dynamically feasible, cross-embodiment retargeting framework for BOTH humanoids 🤖 and dexterous hands ✋. From human motion → sim → real robots, at scale. 🔗 Website: jc-bao.github.io/spider-project/ 🧵 1/n
English
12
60
234
64.5K
LeCAR Lab at CMU ری ٹویٹ کیا
Guanya Shi
Guanya Shi@GuanyaShi·
Introduce SPIDER - a unified scalable dynamics-level retargeting method for both dex hand and humanoid! What does "dynamics-level retargeting" mean? Human motions in, physically feasible high-quality robot motions out (BOTH state and action). These robot motions are so good that you can directly open-loop execute the action sequence in the real world! Also, dynamics-level retargeting makes the downstream RL policy learning super easy since it only needs to figure out a "small residual" on top of the already physically feasible refs. Even better, we can easily do physics-level data augmentation & robustification, such as adding external forces and changing frictions. SPIDER can be used as a data engine for real2sim2real and learning from human data. How it works? It is based on the diffusion-inspired sampling-based optimal control framework we developed in the past (MBD, DIAL-MPC) and virtual contact guidance. Led by @ChaoyiPan and collaboration with @Meta FAIR.
Chaoyi Pan@ChaoyiPan

🕸️ Introducing SPIDER — Scalable Physics-Informed Dexterous Retargeting! A dynamically feasible, cross-embodiment retargeting framework for BOTH humanoids 🤖 and dexterous hands ✋. From human motion → sim → real robots, at scale. 🔗 Website: jc-bao.github.io/spider-project/ 🧵 1/n

English
1
9
69
8.1K
LeCAR Lab at CMU ری ٹویٹ کیا
Yitang Li
Yitang Li@li_yitang·
Meet BFM-Zero: A Promptable Humanoid Behavioral Foundation Model w/ Unsupervised RL👉 lecar-lab.github.io/BFM-Zero/ 🧩ONE latent space for ALL tasks ⚡Zero-shot goal reaching, tracking, and reward optimization (any reward at test time), from ONE policy 🤖Natural recovery & transition
English
7
76
268
79.7K
LeCAR Lab at CMU ری ٹویٹ کیا
Guanya Shi
Guanya Shi@GuanyaShi·
Excited to release BFM-Zero, an unsupervised RL approach to learn humanoid Behavior Foundation Model. Existing humanoid general whole-body controllers rely on explicit motion tracking rewards, on-policy PG methods like PPO, and distillation to one policy. In contrast, BFM-Zero directly learns an effective shared latent representation that embeds motions, goals, and rewards into a common space, which enables a single policy zero-shot perform multiple tasks: (1) natural transition from any pose to any goal pose, (2) real-time motion following, (3) optimize any user-specified reward function at test time, etc. How it works? We don't give the model any specific reward in training. It builds upon recent advances in Forward-Backward (FB) models, where a latent-conditioned policy, a deep "forward dynamics model" and a deep "inverse dynamics model"are jointly learned. In such a way, the learned representation space understands humanoid dynamics and unifies different tasks. More details: lecar-lab.github.io/BFM-Zero/
Yitang Li@li_yitang

Meet BFM-Zero: A Promptable Humanoid Behavioral Foundation Model w/ Unsupervised RL👉 lecar-lab.github.io/BFM-Zero/ 🧩ONE latent space for ALL tasks ⚡Zero-shot goal reaching, tracking, and reward optimization (any reward at test time), from ONE policy 🤖Natural recovery & transition

English
2
13
97
18.5K
LeCAR Lab at CMU ری ٹویٹ کیا
Harsh Gupta
Harsh Gupta@hgupt3·
✈️🤖 What if an embodiment-agnostic visuomotor policy could adapt to diverse robot embodiments at inference with no fine-tuning? Introducing UMI-on-Air, a framework that brings embodiment-aware guidance to diffusion policies for precise, contact-rich aerial manipulation.
English
9
33
215
60.6K
LeCAR Lab at CMU ری ٹویٹ کیا
Guanya Shi
Guanya Shi@GuanyaShi·
100% human data -> long-horizon💡insertion on an aerial manipulator Key: Embodiment-aware diffusion policy (EADP) steers UMI's embodiment-agnostic DP using the gradient of the low-level controller's tracking error. My favorite: We quantify how "UMI-able" different robots are.
Guanya Shi tweet media
Harsh Gupta@hgupt3

✈️🤖 What if an embodiment-agnostic visuomotor policy could adapt to diverse robot embodiments at inference with no fine-tuning? Introducing UMI-on-Air, a framework that brings embodiment-aware guidance to diffusion policies for precise, contact-rich aerial manipulation.

English
0
4
13
1.3K
LeCAR Lab at CMU ری ٹویٹ کیا
Guanya Shi
Guanya Shi@GuanyaShi·
Didn't get chance to attend #CoRL2025 in person, but students from @LeCARLab will present three papers: - Sampling-based system ID for legged sim2real lecar-lab.github.io/spi-active_/ at Oral5 by @NikhilSoban353 - HoldMyBeer: learning humanoid end-effector stabilization control lecar-lab.github.io/SoFTA/ at Spotlight3 by @li_yitang - Humanoid Policy ~ Human Policy human-as-robot.github.io at Spotlight1 by @chaitanya_cha Check them out!
GIF
GIF
GIF
English
1
11
60
5.3K
LeCAR Lab at CMU ری ٹویٹ کیا
Haoyang Weng
Haoyang Weng@ElijahGalahad·
We present HDMI, a simple and general framework for learning whole-body interaction skills directly from human videos — no manual reward engineering, no task-specific pipelines. 🤖 67 door traversals, 6 real-world tasks, 14 in simulation. 🔗 hdmi-humanoid.github.io
English
24
142
736
116K
LeCAR Lab at CMU ری ٹویٹ کیا
Guanya Shi
Guanya Shi@GuanyaShi·
On my way ✈️ to ATL for @ieee_ras_icra! @LeCARLab will present 8 conference papers (including DIAL-MPC as the Best Paper Finalist) and one RA-L paper. Details: lecar-lab.github.io Hope to meet old & new friends and chat about building generalist 🤖 with agility 🚀
English
0
18
106
9K
LeCAR Lab at CMU ری ٹویٹ کیا
Yuanhang Zhang
Yuanhang Zhang@Yuanhang__Zhang·
🦾How can humanoids unlock real strength for heavy-duty loco-manipulation? Meet FALCON🦅: Learning Force-Adaptive Humanoid Loco-Manipulation. 🌐: lecar-lab.github.io/falcon-humanoi… See the details below👇:
English
4
51
182
57.4K
LeCAR Lab at CMU ری ٹویٹ کیا
Guanya Shi
Guanya Shi@GuanyaShi·
I've been working on dynamics model learning in robotics for >6 years (from the Neural-Lander paper). Training a small DNN + some regularization + some online weight adaptation have proven to be effective for specific robots in specific tasks. LLMs have shown the power of transformer, large-scale pertaining, fine-tuning, and in-context adaptation. Can we do the same thing for dynamics model learning? Excited to introduce AnyCar🚗, a transformer-based universal vehicle dynamics model that adapts to various cars & envs & tasks & state estimators, via in-context adaptation. It has some interesting designs akin to LLMs: 1. A seq2seq model aware of history. Input: K-step noisy state/action history, H-step future action. Output: H-step future state prediction; 2. Large-scale pertaining using 100M sim data collected in diverse sims, including Isaac Sim, MuJoCo, and F1-level Assetto Corsa; 3. Real-world "alignment" via fine-tuning on 20K real data; 4. Online "reasoning" by sampling-based MPC that rolls out action sequences using the learned model. AnyCar shows strong adaptability in many scenarios, such as drifting in the wild with 3D-printed plastic tires. Fully open-sourced (including the 1/16 racing car design): lecar-lab.github.io/anycar/
Wenli Xiao@_wenlixiao

🚨🚨Can generalist robotics models perform agile tasks? Introducing 𝗔𝗻𝘆𝗖𝗮𝗿🏎️ 🚚 🚗, a transformer-based 𝗴𝗲𝗻𝗲𝗿𝗮𝗹𝗶𝘀𝘁 vehicle dynamics model that can adapt to various cars, tasks, and envs via in-context adaptation, 𝗼𝘂𝘁𝗽𝗲𝗿𝗳𝗼𝗿𝗺𝗶𝗻𝗴 𝘄𝗲𝗹𝗹-𝘁𝘂𝗻𝗲𝗱 𝘀𝗽𝗲𝗰𝗶𝗮𝗹𝗶𝘀𝘁𝘀 by up to 𝟱𝟰%! Key recipe: large-scale pretrain in diverse sim (100M data) + fine-tuning in real (20K) + Sampling-based MPC Website and code: lecar-lab.github.io/anycar/ 🧵👇 1/6

English
2
26
127
12.9K
LeCAR Lab at CMU ری ٹویٹ کیا
Guanya Shi
Guanya Shi@GuanyaShi·
🎙️I gave a talk "Building Generalist Robots with Agility via Learning and Control: Humanoids and Beyond" at the CMU RI Seminar and Michigan AI Symposium. It covers @LeCARLab's research and my thoughts on building agile generalist robots. Recording: youtu.be/Uym3Tr6t5TM
YouTube video
YouTube
English
1
24
101
6.8K
LeCAR Lab at CMU ری ٹویٹ کیا
LeCAR Lab at CMU ری ٹویٹ کیا
Haoru Xue
Haoru Xue@HaoruXue·
🎉 Diffusion-style annealing + sampling-based MPC can surpass RL, and seamlessly adapt to task parameters, all 𝘁𝗿𝗮𝗶𝗻𝗶𝗻𝗴-𝗳𝗿𝗲𝗲! We open sourced DIAL-MPC, the first training-free method for whole-body torque control using full-order dynamics 🧵 lecar-lab.github.io/dial-mpc/
English
10
158
743
172.3K
LeCAR Lab at CMU ری ٹویٹ کیا
Guanya Shi
Guanya Shi@GuanyaShi·
Thanks @_akhaliq! Finally, robot can do continuous, agile, autonomous, adaptive jumping over stair and stepping stone Key idea: combine the pros of model-free RL and model-based control. RL (for CoM refs) + QP (for GRF) + WBC (for torque) Open-sourced: yxyang.github.io/jumping_cod/
AK@_akhaliq

Agile Continuous Jumping in Discontinuous Terrains discuss: huggingface.co/papers/2409.10… We focus on agile, continuous, and terrain-adaptive jumping of quadrupedal robots in discontinuous terrains such as stairs and stepping stones. Unlike single-step jumping, continuous jumping requires accurately executing highly dynamic motions over long horizons, which is challenging for existing approaches. To accomplish this task, we design a hierarchical learning and control framework, which consists of a learned heightmap predictor for robust terrain perception, a reinforcement-learning-based centroidal-level motion policy for versatile and terrain-adaptive planning, and a low-level model-based leg controller for accurate motion tracking. In addition, we minimize the sim-to-real gap by accurately modeling the hardware characteristics. Our framework enables a Unitree Go1 robot to perform agile and continuous jumps on human-sized stairs and sparse stepping stones, for the first time to the best of our knowledge. In particular, the robot can cross two stair steps in each jump and completes a 3.5m long, 2.8m high, 14-step staircase in 4.5 seconds. Moreover, the same policy outperforms baselines in various other parkour tasks, such as jumping over single horizontal or vertical discontinuities.

English
4
34
168
32.1K
LeCAR Lab at CMU ری ٹویٹ کیا
Guanya Shi
Guanya Shi@GuanyaShi·
The CMU LeCAR Lab will present several papers at #RSS2024 this week, covering topics from learning & control & optimization, safe autonomy, to humanoids. Check them out!
English
0
13
102
11K