Wenhao Ding

60 posts

Wenhao Ding banner
Wenhao Ding

Wenhao Ding

@wenhaoding95

Research Scientist @NVIDIA | Ph.D.@CarnegieMellon | B.E. @tsinghua_uni

Santa Clara, CA เข้าร่วม Mart 2021
296 กำลังติดตาม342 ผู้ติดตาม
Wenhao Ding รีทวีตแล้ว
Yue Wang
Yue Wang@yuewang314·
Introducing Ψ₀ (psi-lab.ai/Psi0) — an open foundation model for universal humanoid loco-manipulation. 🏆 Outperforms GR00T N1.6 by 40%+ overall success rate 📉 Uses only ~10% of the pre-training data 📦 Fully open-source: model, data, code, and deployment pipeline 1/10
English
2
41
205
12.7K
Wenhao Ding
Wenhao Ding@wenhaoding95·
Our Alpamayo model has a major update! Stay tuned for the code and model release
Marco Pavone@drmapavone

Jensen today announced Alpamayo 1.5 at #NVIDIAGTC! #Alpamayo 1.5 is a major update to Alpamayo 1—@nvidia’s open 10B-parameter chain-of-thought reasoning VLA model, first introduced at #CES. Built on the #Cosmos-Reason2 VLM backbone and post-trained with RL, it adds support for navigation guidance, flexible multi-camera setups, configurable camera parameters, and user question answering. The result is an interactive, steerable reasoning engine for the AV community. We’re also releasing post-training scripts to help researchers and developers adapt the model. Additionally, we’ve significantly expanded the Alpamayo open platform across data and simulation, including releasing highly requested reasoning labels for the PhysicalAI Autonomous Vehicles dataset (huggingface.co/datasets/nvidi…), as well as our chain-of-causation auto-labeling pipeline. 🔎 Learn more about Alpamayo 1.5 and the latest extensions to the Alpamayo open platform: huggingface.co/blog/drmapavon… (please note that most of the links will become active in the next few days.) Happy building—and stay tuned for more in the coming months! @NVIDIADRIVE @NVIDIAAI

English
0
0
2
123
Wenhao Ding รีทวีตแล้ว
Marco Pavone
Marco Pavone@drmapavone·
What does it take to build autonomous vehicles that can reason about the world they drive in? Tomorrow at #NVIDIAGTC, Patrick Liu and I will take a deep dive into the #Alpamayo #reasoning model family—a family of reasoning-based vision–language–action (#VLA) models that form a core component of the Alpamayo open platform (huggingface.co/blog/drmapavon…). We’ll cover three main topics: - How reasoning-based VLA models like Alpamayo 1 are designed and built - What it takes to bring Alpamayo 1 to production, including some of our latest results - Several exciting announcements about the expansion of the Alpamayo open platform If you're working on autonomous driving, robotics, or foundation models for physical AI, this session will offer a look at where the field is heading. Session details: 📅 Monday, Mar 16 | 3:00 PM PDT 📍 #NVIDIAGTC 2026 🔗 nvda.ws/4rze5oj Looking forward to seeing many of you there. @NVIDIADRIVE @NVIDIAAI
English
0
18
72
7.1K
Wenhao Ding รีทวีตแล้ว
NVIDIA DRIVE
NVIDIA DRIVE@NVIDIADRIVE·
Alpamayo 1 is now @huggingface’s top-downloaded robotics model with 100K downloads and counting. 🎉 It helps researchers and autonomous-driving practitioners develop and evaluate vision-language-action models for complex autonomous-driving scenarios, especially rare long-tail events. 🔗 Get started with Alpamayo 1 today: nvda.ws/3OnZoWU 🎥 Watch the deep-dive: nvda.ws/4tJxvbN
NVIDIA DRIVE tweet media
English
10
25
140
33.1K
Wenhao Ding
Wenhao Ding@wenhaoding95·
Checkout the replay of the livestrem to learn more about how we design and build alpamayo 1!
NVIDIA DRIVE@NVIDIADRIVE

💨 How fast can an autonomous vehicle think? With Alpamayo 1, NVIDIA's 10B-parameter chain-of-thought reasoning model, the distilled version can reason in real time. Hear Marco Pavone (@drmapavone), Yan Wang, Yurong You, and Wenhao Ding from our AV Research team break down Alpamayo 1 and what's next for reasoning in autonomous driving. 🔁 Watch the replay: nvda.ws/3O5gKb3

English
0
1
2
193
Wenhao Ding รีทวีตแล้ว
Marco Pavone
Marco Pavone@drmapavone·
Join me and my collaborators for a *live* discussion on @nvidia Alpamayo 1 (huggingface.co/nvidia/Alpamay…), a reasoning-based vision–language–action (VLA) model for autonomous driving. 🎥 Livestream: Inside NVIDIA Alpamayo 1: Making Autonomous Vehicles Reason 🗓 February 11 ⏰ 9:00am PST 📍 Watch here: youtube.com/watch?v=V9E4GX… As NVIDIA CEO Jensen Huang put it: “The ChatGPT moment for physical AI is here — when machines begin to understand, reason, and act in the real world. Robotaxis are among the first to benefit. Alpamayo brings reasoning to autonomous vehicles, allowing them to think through rare scenarios, drive safely in complex environments, and explain their driving decisions — it’s the foundation for safe, scalable autonomy.” During the livestream, we’ll cover: - How #reasoning-based #VLA models like #Alpamayo 1 are designed and built - Applications ranging from end-to-end #autonomy to reasoning-driven auto-labeling - Key opportunities and challenges in developing reasoning models for #Physical #AI I’ll be joined by core Alpamayo 1 developers @yan_wang_9 @YurongYou @wenhaoding95, and we’ll take questions live from the community. 📖 Ahead of time, you might enjoy this overview of the Alpamayo ecosystem: huggingface.co/blog/drmapavon… And if you’re attending @NVIDIAGTC (March 16–19) and would like to meet some of the Alpamayo team in person, you can use my employee code for 25% off your conference pass: nvidia.com/gtc/?ncid=GTC-… Hope to see you at the livestream! @NVIDIAAI @NVIDIADRIVE
YouTube video
YouTube
Marco Pavone tweet media
English
2
8
38
4.1K
Wenhao Ding รีทวีตแล้ว
Marco Pavone
Marco Pavone@drmapavone·
It’s incredibly exciting to see how quickly the community is engaging with the @nvidia Alpamayo ecosystem for developing reasoning-based autonomous vehicles (huggingface.co/blog/drmapavon…)! In this instance, TIER IV is showcasing Alpamayo 1’s reasoning capabilities in Tokyo, integrated with Autoware and ROS. Fantastic work, @ShinpeiKato and the @tier_iv_global team! 👏 Quick highlights about Alpamayo: Alpamayo 1: - Among HuggingFace’s top 10 overall trending models - Among the top 3 most downloaded models on HuggingFace when filtered by 'robotics' Alpamayo PhysicalAI–Autonomous-Vehicles dataset: - Trending in HuggingFace’s top 10 overall datasets Happy developing! 🚀 #AutonomousVehicles #Robotics #AI #Reasoning #HuggingFace #Autoware #ROS #AutonomousDriving #PhysicalAI #Alpamayo #RobotLearning @NVIDIAAI @NVIDIADRIVE
Shinpei KATO (加藤真平)@ShinpeiKato

Alpamayoちゃんと学習させれば日本でも結構使えそう!自動運転も世界モデルもオープンソースの時代!

English
1
14
51
5.7K
Wenhao Ding รีทวีตแล้ว
Marco Pavone
Marco Pavone@drmapavone·
More on #reasoning in Vision-Language-Action (#VLA) models --- Traditional VLA models decide what action to take by decomposing complex situations into their most salient factors. But reasoning models can do much more. When viewed as implicit world models operating in a semantic space, they can be used counterfactually—exploring multiple “what if” scenarios before acting. In our recent paper, Counterfactual VLA (CF-VLA, arxiv.org/pdf/2512.24426), we show that counterfactual reasoning consistently improves trajectory accuracy, safety, and reasoning quality. Key contributions: - Self-reflective counterfactual reasoning: CF-VLA reflects on predicted meta-actions, anticipates consequences, and revises plans before execution—enabling causal self-correction. - Automated data pipeline: A novel data pipeline generates counterfactual data, forming a self-improving loop for reasoning and action. - Adaptive thinking in autonomous driving: CF-VLA focuses reasoning on the most challenging scenarios, improving performance while keeping test-time computation efficient. Paper: arxiv.org/pdf/2512.24426 #AI #Robotics #VisionLanguageAction #AutonomousSystems #MachineLearning #CounterfactualReasoning @NVIDIAAI @NVIDIADRIVE
Marco Pavone tweet media
English
0
21
105
7.2K
Wenhao Ding
Wenhao Ding@wenhaoding95·
Excited to witness and contribute to the incredible effort of open-sourcing driving ecosystem 🚀
Marco Pavone@drmapavone

🚀 Exciting news from #CES2026! In his keynote today, Jensen announced @nvidia Alpamayo — a *fully open* ecosystem of models, simulation tools, and datasets designed to accelerate reasoning-based autonomous vehicle (AV) architectures and advance the path to Level 4 autonomous driving. Alpamayo brings together several technologies we’ve developed to enable reasoning-based vision–language–action (VLA) models for AVs. Our goal is to provide researchers and developers with a flexible, fast, and scalable platform for evaluating and training reasoning-based AV architectures in realistic closed-loop settings. Explore Alpamayo: -- Press Release: nvidianews.nvidia.com/news/alpamayo-… -- Hugging Face Blog: huggingface.co/blog/drmapavon… -- Tech Blog: developer.nvidia.com/blog/building-… -- Alpamayo 1 reasoning model: research.nvidia.com/publication/20… -- Physical AI AV Dataset: huggingface.co/datasets/nvidi… -- AlpaSim simulator: github.com/NVlabs/alpasim I’m incredibly proud of the @nvidia AV Research team (research.nvidia.com/labs/avg/) and our many @nvidia collaborators whose contributions made this possible. More releases and features are coming soon — we can’t wait to see what the community builds with Alpamayo! 💡 Want to help grow the Alpamayo ecosystem? We’re hiring: [Sr.] Research Scientist: nvidia.wd5.myworkdayjobs.com/en-US/NVIDIAEx… [Sr.] Research Engineer: nvidia.wd5.myworkdayjobs.com/en-US/NVIDIAEx… #AutonomousVehicles #AutonomousDriving #AI #Simulation #ReasoningAI #OpenEcosystem #Alpamayo @NVIDIAAI @NVIDIADRIVE

English
0
0
2
91
Wenhao Ding รีทวีตแล้ว
Marco Pavone
Marco Pavone@drmapavone·
🚗 Imitation learning is everywhere—but is it enough? So far, imitation learning—most commonly via behavior cloning (BC)—remains the go-to approach for training real-world autonomous vehicle (AV) driving policies. Yet BC operates in an open-loop (OL) fashion, overlooking the critical interdependence among inputs, outputs, and future states that comes with closed-loop (CL) operation. The result? The notorious—but often overlooked—OL–CL gap ⚠️ To address this challenge and encourage broader adoption of CL techniques, we’ve just published a survey (research.nvidia.com/publication/20…) presenting a comprehensive taxonomy of closed-loop training methods for end-to-end driving. Our framework organizes approaches along three key axes: - Action generation - Environment response generation - Training objectives 💡 Bottom line: enabling technologies—like neural rendering, generative world models, and scalable RL—have now matured, making closed-loop AV training ready for wide-scale adoption. We’d love to hear your thoughts—drop a comment and join the discussion! 💬 And as a reminder, we are hiring for full-time research scientist and research engineer positions: 🔹 [Sr.] Research Scientist: nvidia.wd5.myworkdayjobs.com/en-US/NVIDIAEx… 🔹 [Sr.] Research Engineer: nvidia.wd5.myworkdayjobs.com/en-US/NVIDIAEx… @NVIDIADRIVE @NVIDIAAI @nvidia
English
1
16
96
9.2K
Wenhao Ding
Wenhao Ding@wenhaoding95·
Proud to contribute to Alpamayo-R1, NVIDIA’s new open reasoning VLA model. Together with the Physical AI datasets and AlpaSim, this release completes an open AV research stack. Excited to see what the community builds next.
Marco Pavone@drmapavone

We’ve just released @nvidia #DRIVE Alpamayo-R1 (AR1) — the world’s first industry-scale open #reasoning #VLA model for autonomous-vehicle (AV) research. AR1 integrates Chain-of-Causation reasoning with trajectory planning to improve decision-making in complex driving scenarios. Built on @nvidia #Cosmos #Reason, AR1 is designed as a customizable foundation for a broad range of AV applications — from instantiating an end-to-end backbone for autonomous driving to powering advanced, reasoning-based auto-labeling tools. Resources: Model: huggingface.co/nvidia/Alpamay… Inference Code: github.com/NVlabs/alpamayo Paper: research.nvidia.com/publication/20… Blog Post: blogs.nvidia.com/blog/neurips-o… A subset of the data used to train and evaluate AR1 is available in the @nvidia Physical AI Open Datasets: huggingface.co/datasets/nvidi… AR1 can be evaluated using AlpaSim (github.com/NVlabs/alpasim), @nvidia's newly released open-source AV simulation framework built specifically for research and development. (Separate post on AlpaSim coming soon.) This release completes @nvidia’s trifecta — model, data, and simulator — to accelerate research and development in the autonomous-vehicle domain. Happy developing, and stay tuned for more! Huge thanks to the phenomenal team that made this possible @NVIDIAAI @nvidia.

English
0
0
6
434
Wenhao Ding
Wenhao Ding@wenhaoding95·
@peide_huang At the same time, people are discussing AI bubbles. What a separated world
English
1
0
1
132
Peide Huang
Peide Huang@peide_huang·
It is Wednesday night. Here are what we got this week so far: Gemini 3.0 GPT 5.1 pro Grok 4.1 PI 0.6 Sunday Robotics SAM 3 What a time to be alive!
English
2
1
31
2K
Wenhao Ding
Wenhao Ding@wenhaoding95·
Excited to release our recent exploration in reasoning VLA for Autonomous Driving!
Marco Pavone@drmapavone

Excited to unveil @nvidia's latest work on #Reasoning Vision–Language–Action (#VLA) models — Alpamayo-R1! Alpamayo-R1 is a new #reasoning VLA architecture featuring a diffusion-based action expert built on top of the #Cosmos-#Reason backbone. It represents one of the core technologies driving NVIDIA’s push toward Level 4 autonomy and robotaxis (nvidianews.nvidia.com/news/nvidia-ub…), as announced by Jensen Huang at #gtc DC last week. 📄 Paper: Alpamayo-R1 research.nvidia.com/publication/20… We present: - Architecture & Design: How to transform a VLM into a driving-ready Reasoning VLA - Chain of Causation Labeling: A new framework enabling reasoning-based learning - Training Strategy: From internet-scale pre-training → AV-specific SFT → RL-based post-training - Extensive Evaluation: From closed-loop simulation to real-world, on-vehicle testing 📈 Results: Alpamayo-R1 delivers significant performance gains over end-to-end baselines — especially in rare, safety-critical scenarios — all while maintaining real-time inference (99 ms end-to-end latency). Coming soon: releases of model variants and reasoning metadata built on top of the Physical AI Dataset (huggingface.co/datasets/nvidi…)—with more updates on the way. Stay tuned! 🙌 Huge thanks to Wenjie Luo and @yan_wang_9 (project co-leads); the @nvidia AV Research team (@iamborisi, @YurongYou, @xinshuoweng, @tianran_, @wenhaoding95, and many others); collaborators across @nvidia Research (@liu_mingyu, @visualyang, @PavloMolchanov, and many others); and the @nvidia AV Product team (Sarah Tariq, Patrick Liu, Jack Huang, and many more). Full contributor list in the Appendix. @NVIDIADRIVE @NVIDIAAI

English
2
2
33
6.1K
Wenhao Ding รีทวีตแล้ว
Marco Pavone
Marco Pavone@drmapavone·
Excited to unveil @nvidia's latest work on #Reasoning Vision–Language–Action (#VLA) models — Alpamayo-R1! Alpamayo-R1 is a new #reasoning VLA architecture featuring a diffusion-based action expert built on top of the #Cosmos-#Reason backbone. It represents one of the core technologies driving NVIDIA’s push toward Level 4 autonomy and robotaxis (nvidianews.nvidia.com/news/nvidia-ub…), as announced by Jensen Huang at #gtc DC last week. 📄 Paper: Alpamayo-R1 research.nvidia.com/publication/20… We present: - Architecture & Design: How to transform a VLM into a driving-ready Reasoning VLA - Chain of Causation Labeling: A new framework enabling reasoning-based learning - Training Strategy: From internet-scale pre-training → AV-specific SFT → RL-based post-training - Extensive Evaluation: From closed-loop simulation to real-world, on-vehicle testing 📈 Results: Alpamayo-R1 delivers significant performance gains over end-to-end baselines — especially in rare, safety-critical scenarios — all while maintaining real-time inference (99 ms end-to-end latency). Coming soon: releases of model variants and reasoning metadata built on top of the Physical AI Dataset (huggingface.co/datasets/nvidi…)—with more updates on the way. Stay tuned! 🙌 Huge thanks to Wenjie Luo and @yan_wang_9 (project co-leads); the @nvidia AV Research team (@iamborisi, @YurongYou, @xinshuoweng, @tianran_, @wenhaoding95, and many others); collaborators across @nvidia Research (@liu_mingyu, @visualyang, @PavloMolchanov, and many others); and the @nvidia AV Product team (Sarah Tariq, Patrick Liu, Jack Huang, and many more). Full contributor list in the Appendix. @NVIDIADRIVE @NVIDIAAI
English
10
40
235
36.3K
Wenhao Ding
Wenhao Ding@wenhaoding95·
Check our latest release of autonomous driving dataset 😃
Marco Pavone@drmapavone

We’ve just released the @nvidia Physical AI Autonomous Vehicles Dataset! huggingface.co/datasets/nvidi… Highlights: - 1,727 hours of driving data collected by @nvidia - Spanning 25 countries and 2,500+ cities - Capturing diverse traffic, weather, and driving scenarios - Includes camera, LiDAR, and radar data This is just the beginning — features, tools, and challenges will continue to evolve. Stay tuned! @NVIDIADRIVE @NVIDIAAI

English
0
0
2
297
Wenhao Ding รีทวีตแล้ว
Marco Pavone
Marco Pavone@drmapavone·
🚗🤖 Interested in reasoning models for embodied AI? I’m excited to share that at #NVIDIAGTC in DC I’ll unveil our latest work at #NVIDIA on reasoning Vision-Language-Action (VLA) models for vehicle autonomy: I’ll cover how we’re: • Advancing reasoning in VLA models • Powering a data flywheel for AV foundation models • Making autonomous driving more human-like and safer — with real-world driving videos! 🔗 Session info: nvda.ws/3IGfvwq 📍 Tuesday, Oct 28 • 3 PM Walter E. Washington Convention Center
Marco Pavone tweet media
English
0
12
49
4.5K
Wenhao Ding รีทวีตแล้ว
Marco Pavone
Marco Pavone@drmapavone·
The Autonomous Vehicle (AV) Research Group @NVIDIAAI is looking for talented interns! Dive into cutting-edge work—from reasoning models and generative simulation to AI safety—and help shape the future of AV and embodied AI. Ready to push the limits? Apply now: nvidia.wd5.myworkdayjobs.com/en-US/NVIDIAEx…
English
2
14
62
9K
Wenhao Ding รีทวีตแล้ว
Marco Pavone
Marco Pavone@drmapavone·
We’re now accepting applications for the 2026–2027 NVIDIA Graduate Fellowships! If you’re passionate about advancing cutting-edge reasoning models for Physical AI applications 🚗🤖, apply here: research.nvidia.com/graduate-fello… — and be sure to select “Autonomous Vehicles.” @NVIDIAAI
English
1
26
105
12.3K
Wenhao Ding รีทวีตแล้ว
Marco Pavone
Marco Pavone@drmapavone·
Can we use simulation to validate Physical AI? Yes—with far fewer real-world tests. We propose a control variates–based estimation framework that pairs sim & real data to dramatically cut validation costs. #AI #Robotics #Sim2Real" Paper: arxiv.org/pdf/2506.20553 @NVIDIADRIVE
English
1
13
37
5.7K