Elgce

102 posts

Elgce

Elgce

@BenQingwei

Ph.D. at MMLab, CUHK https://t.co/WpwzUeEBwi

Shanghai, China Katılım Ocak 2023
472 Takip Edilen521 Takipçiler
Sabitlenmiş Tweet
Elgce
Elgce@BenQingwei·
Introducing Gallant: Voxel Grid-based Humanoid Locomotion and Local-navigation across 3D Constrained Terrains 🤖 Project page: gallantloco.github.io Arxiv: arxiv.org/abs/2511.14625 Gallant is, to our knowledge, the first system to run a single policy that handles full-space constraints — including ground-level barriers, lateral clutter, and overhead obstacles on a humanoid robot. Instead of elevation maps or depth cameras, Gallant uses a voxel grid built directly from raw LiDAR as its perception representation, giving it inherent 3D coverage of the scene. With our custom LiDAR simulation toolkit (github.com/agent-3154/sim…), we model realistic scans, including returns from the robot’s own moving links, which is crucial for sim-to-real transfer. On the control side, we use a target-based training scheme rather than standard velocity tracking. The robot is given a goal and learns to discover its own in-path velocities and trajectories, so no external high-frequency command stream is needed during deployment. The policy itself is intentionally lightweight: just a 3-layer CNN + 3-layer MLP (~0.3M params), running onboard on the Unitree G1’s Orin NX at 50 Hz with no extra compute. Training takes about 6 hours on 8× NVIDIA RTX 4090 GPUs. The resulting policy transfers directly to the real robot and achieves >90% success rate on most tested terrain types. Gallant is our “half-way” step toward robust perceptive locomotion — a problem we believe remains fundamental for humanoid robots. We’re now working toward closing the gap to near-100% reliability and expanding the pipeline further. Code will be fully released soon. Discussion, feedback, and collaboration are very welcome! 🙌
English
3
35
207
52.7K
Elgce
Elgce@BenQingwei·
Gallant is now open source at github.com/InternRobotics… 🎉
Elgce@BenQingwei

Introducing Gallant: Voxel Grid-based Humanoid Locomotion and Local-navigation across 3D Constrained Terrains 🤖 Project page: gallantloco.github.io Arxiv: arxiv.org/abs/2511.14625 Gallant is, to our knowledge, the first system to run a single policy that handles full-space constraints — including ground-level barriers, lateral clutter, and overhead obstacles on a humanoid robot. Instead of elevation maps or depth cameras, Gallant uses a voxel grid built directly from raw LiDAR as its perception representation, giving it inherent 3D coverage of the scene. With our custom LiDAR simulation toolkit (github.com/agent-3154/sim…), we model realistic scans, including returns from the robot’s own moving links, which is crucial for sim-to-real transfer. On the control side, we use a target-based training scheme rather than standard velocity tracking. The robot is given a goal and learns to discover its own in-path velocities and trajectories, so no external high-frequency command stream is needed during deployment. The policy itself is intentionally lightweight: just a 3-layer CNN + 3-layer MLP (~0.3M params), running onboard on the Unitree G1’s Orin NX at 50 Hz with no extra compute. Training takes about 6 hours on 8× NVIDIA RTX 4090 GPUs. The resulting policy transfers directly to the real robot and achieves >90% success rate on most tested terrain types. Gallant is our “half-way” step toward robust perceptive locomotion — a problem we believe remains fundamental for humanoid robots. We’re now working toward closing the gap to near-100% reliability and expanding the pipeline further. Code will be fully released soon. Discussion, feedback, and collaboration are very welcome! 🙌

English
1
22
84
9.9K
Elgce retweetledi
Yu Lei
Yu Lei@_OutofMemory_·
Have the privilege to beta-test SONIC. Thanks for the team to open-source! Superior performance as system0. It’s pretty easy to deploy with very-well-written documents (only took me a few hrs). Empirical results speak louder than words. Try out on your robots!
Yuke Zhu@yukez

We have seen rapid progress in humanoid control — specialist robots can reliably generate agile, acrobatic, but preset motions. Our singular focus this year: putting generalist humanoids to do real work. To progress toward this goal, we developed SONIC (nvlabs.github.io/GEAR-SONIC/), a Behavior Foundation Model for real-time, whole-body motion generation that supports teleoperation and VLA inference for loco-manipulation. Today, we’re open-sourcing SONIC on GitHub. We are excited to see what the community builds upon SONIC and to collectively push humanoid intelligence toward real-world deployment at scale. 🌐 Paper: arxiv.org/abs/2511.07820 📃 Code: github.com/NVlabs/GR00T-W…

English
6
21
151
15.5K
Elgce
Elgce@BenQingwei·
Amazing Manipulation with World Model 🫡
Jiazhi Yang@jiazhi_yang2024

🧐Applying world models to improve real-world policy on challenging manipulation tasks used to be considered out of reach. 😌After sustained effort, we’re now seeing encouraging progress. 🚀Thrilled to introduce RISE: Self-Improving Robot Policy with Compositional World Model opendrivelab.com/kai0-rl/ arxiv.org/abs/2602.11075 RISE is, to our knowledge, the first work to use a world model as an effective learning environment for challenging real-world manipulation, enabling policy improvement on tasks that demand high dynamics, dexterity, and precision. Incredible teamwork with @lin_kunyang111 @francislee2020 @YueXiangyu @HaoZhao_AIRSUN @smch_1127

English
0
1
2
854
Elgce retweetledi
Haoru Xue
Haoru Xue@HaoruXue·
Reality of robotics: humanoid kung fu is solved before they can open doors with RGB. Here we are. Introducing the frontier of sim2real at NVIDIA GEAR. 100% sim data. RGB input only. Code name: 𝗗𝗼𝗼𝗿𝗠𝗮𝗻. We are opening the sim-to-real door. doorman-humanoid.github.io 🧵
English
14
82
502
353K
Elgce retweetledi
Tairan He
Tairan He@TairanHe99·
Zero teleoperation. Zero real-world data. ➔ Autonomous humanoid loco-manipulation in reality. Introducing VIRAL: Visual Sim-to-Real at Scale. We achieved 54 autonomous cycles (walk, stand, place, pick, turn) using a simple recipe: 1. RL 2. Simulation 3. GPUs Website: viral-humanoid.github.io Arxiv: arxiv.org/abs/2511.15200 Deep dive with me: 🧵
English
19
157
772
201.4K
Elgce
Elgce@BenQingwei·
Really impressive! After all the fancy yet overfit demos, it’s time to focus on the real goal: a general agent that truly helps in everyday life. So many impressive demos in dual-arm robots this month. Hoping the same breakthroughs for legged humanoid robots.
Tony Zhao@tonyzzhao

Today, we present a step-change in robotic AI @sundayrobotics. Introducing ACT-1: A frontier robot foundation model trained on zero robot data. - Ultra long-horizon tasks - Zero-shot generalization - Advanced dexterity 🧵->

English
0
0
2
436
Elgce retweetledi
Tony Zhao
Tony Zhao@tonyzzhao·
Today, we present a step-change in robotic AI @sundayrobotics. Introducing ACT-1: A frontier robot foundation model trained on zero robot data. - Ultra long-horizon tasks - Zero-shot generalization - Advanced dexterity 🧵->
English
435
652
5.5K
2M
Elgce retweetledi
The Humanoid Hub
The Humanoid Hub@TheHumanoidHub·
"Bring me the healthiest snack." The robot goes to the kitchen and gets the snack, fully autonomously. This is the first public demo of NVIDIA's Isaac GR00T N1.6 foundation model presented by Yuke Zhu at CoRL 2025. The previous versions focused only on bimanual stationary manipulation; the N1.6 unlocks the entire kinematic range of the robot. The latest release of the open Isaac GR00T N1.6 VLA model will be available soon on Hugging Face. It integrates with NVIDIA Cosmos Reason, an open, customizable reasoning vision language model that turns vague instructions into step-by-step plans.
English
25
64
423
42.7K
Elgce retweetledi
Weishuai Zeng
Weishuai Zeng@weishuaizeng·
We are excited to re-introduce our Behavior Foundation Model for Humanoid Robots, built upon a unified perspective of diverse WBC tasks, as a promising step toward a foundation model for general humanoid control. 🔗Website: bfm4humanoid.github.io 📜Paper: arxiv.org/abs/2509.13780
English
11
31
155
14.8K
Elgce
Elgce@BenQingwei·
Welcome to pay attention to @InternRobotics IROS challenge, and work together to promote manipulation and navigation through the competition! I should also be at IROS by then. Looking forward to meeting new and old friends in person.😃
Intern Robotics@InternRobotics

🚀 3 steps to ace IROS 2025 Nav Track: Setup · Develop · Submit 🦾 📺 We’ve prepared a Quickstart Guide to help you quickly grasp the task, explore the dataset, and submit your model to the leaderboard. 🥇 Winner prize: $10K 📌 internrobotics.shlab.org.cn/challenge/2025/

English
0
0
4
412
Elgce retweetledi
Weishuai Zeng
Weishuai Zeng@weishuaizeng·
Excited to share our latest progress on building Behavior Foundation Model for Humanoid Robots🎈 Forward roll, hip-pop dance, even cartwheel -- all the things you have never imagined the little G1 could do -- we have made it based on ONE model👌 Stay tuned for paper and code😉
English
12
31
229
21.5K
Elgce retweetledi
Yitang Li
Yitang Li@li_yitang·
Just a temporary helper today, but very excited to join the HOMIE squad! 👊 Here to support my amazing friends Qingwei @BenQingwei (online 👀) and Feiyu @Jia_Fei_Yu 💪 Don’t miss HOMIE at #RSS2025 — and don’t miss the chance to chat with Qingwei👇 WeChat QR here!
Elgce@BenQingwei

HOMIE will be presented at #RSS2025 today! Spotlight Talks: 4:30pm-5:30pm Poster: 6:30pm-8:00pm BoardNr: 34 @li_yitang will be there to help us present this paper And I will be online to introduce and discuss it🥳 Talk video: drive.google.com/file/d/10uYskZ…

English
1
2
15
1.2K