Yunhao Luo

27 posts

Yunhao Luo banner
Yunhao Luo

Yunhao Luo

@yluo_y

PhD Student @UMichCSE | Prev @GeorgiaTech @BrownUniversity Embodied AI

Katılım Aralık 2014
667 Takip Edilen152 Takipçiler
Sabitlenmiş Tweet
Yunhao Luo
Yunhao Luo@yluo_y·
🚀Excited to share our @NeurIPSConf paper Compositional Diffuser (CompDiffuser)! CompDiffuser scales planning horizons at test time, able to construct long-horizon plans while only trained on short-horizon data. Project page: comp-diffuser.github.io Thread 👇(1/n)
English
2
17
68
33.8K
Yunhao Luo
Yunhao Luo@yluo_y·
@zhiyuanjsu Thanks. As in Section 4.3, we propose multiple objectives to optimize for each factor x^i (i.e., short-horizon videos). By doing so, we enforce consistency between factors, and hence are able to use them to construct coherent long-horizon plans.
English
0
0
1
37
Zhiyuan "Julian" Su
Zhiyuan "Julian" Su@zhiyuanjsu·
@yluo_y Interesting approach with CVP. How does the message-passing scheme specifically address the challenge of long-horizon planning beyond short-horizon diffusion models? THX!
English
1
0
1
47
Yunhao Luo
Yunhao Luo@yluo_y·
Introducing Compositional Visual Planning (CVP)🤖✨#ICLR2026 How can we scale short-horizon video diffusion models for test-time long-horizon planning? We propose a completely training-free method!🤯 We frame the problem as inference over a chain-structured factor graph and use a novel message-passing scheme on clean Tweedie estimates to enforce strict boundary agreement. 🧩
Woo Chul Shin@woochulshin1726

What if your robot could plan tasks it has never seen before without ever being retrained? Meet Compositional Visual Planning via Inference-Time Diffusion Scaling (ICLR 2026 🏆) comp-visual-planning.github.io If you are in Rio🇧🇷 visit us! Sat, 04/25/26 6:30-9:00 AM PDT Pavillion 4 #4203

English
2
9
50
5.8K
Yunhao Luo
Yunhao Luo@yluo_y·
Very cool work on generating coherent videos over pre-defined long camera trajectories. Glad to see our work CompDiffuser (comp-diffuser.github.io) played a part in inspiring this. Congrats to the team on the #ICLR2026 paper!
Chonghyuk (ND) Song@ndsong95

Check out our #ICLR2026 paper Generative View Stitching! I unfortunately couldn’t attend but @MichalStaryy will be presenting our poster tomorrow (Sat) morning at Pavillon 4 PA-#3016. Shoutout to my other collaborators @BoyuanChen0, @gkopanas, and @vincesitzmann!

English
0
1
4
268
Yunhao Luo retweetledi
Calvin Luo
Calvin Luo@calvinyluo·
How can visual planning agents 𝙨𝙚𝙡𝙛-𝙞𝙢𝙥𝙧𝙤𝙫𝙚 from their own collected experience? We present 𝗦𝗜𝗟𝗩𝗥🩶, a framework that combines offline data with online experience for concurrent zero-shot generalization and sample-efficient self-improvement capabilities!#ICLR2026
English
1
19
103
18.2K
Yunhao Luo retweetledi
Danfei Xu
Danfei Xu@danfei_xu·
Check out our #ICLR2026 oral paper CDGS! We show that modular generative models can be composed via inference-time classical search plus score composition. This enables new capabilities such as synthesizing large images from small generators and extending short-horizon video planners into long-horizon plans.
Utkarsh Mishra@utkarshm0410

Our paper "Compositional Diffusion with Guided Search (CDGS)" is an Oral at #ICLR2026! Short-horizon Foundation Models + Compositional Generative Planning + Inference-time Search = CDGS for goal-conditioned long-horizon planning! More details: cdgsearch.github.io 🧵 below

English
1
11
54
8.4K
Yunhao Luo retweetledi
William Yijiang Li
William Yijiang Li@Williamiumli·
🚨 New paper alert !! 🎥 Video VLMs are strong at high-level semantics and long-range temporal understanding. 🧠 JEPA is almost the opposite: better at dense, high-frequency dynamics, local physical consistency, and fast corrective control, but are less suited for rich semantic reasoning and long-horizon reasoning. We try to get the best of both: 🧩 A VLM as a cortex-like reasoner for semantics and long-horizon planning ⚡ A JEPA branch as a cerebellum-like controller for fine-grained dynamics, physical consistency, and rapid corrections Proudly, we present ThinkJEPA: a VLM-guided latent world model that FiLM-fuse the pyramid repr of VLMs encoding long-horizon semantic reasoning into the JEPA repr for fine-grained, physically consistent dynamics prediction. 🔗 Project: zhanghaichao.xyz/ThinkJEPA/ 📄 Paper: arxiv.org/pdf/2603.22281
William Yijiang Li tweet mediaWilliam Yijiang Li tweet media
English
8
67
359
17.2K
Yunhao Luo retweetledi
Danfei Xu
Danfei Xu@danfei_xu·
Introducing EgoVerse: an ecosystem for robot learning from egocentric human data. Built and tested by 4 research labs + 3 industry partners, EgoVerse enables both science and scaling 1300+ hrs, 240 scenes, 2000+ tasks, and growing Dataset design, findings, and ecosystem 🧵
English
34
158
856
251.8K
Yunhao Luo retweetledi
Rhoda AI
Rhoda AI@RhodaAI·
To bring generalist intelligent robots to the real world, we have to overcome the data scarcity problem. At Rhoda, we are solving it by reformulating robot policies as video generation. Today, we introduce the Direct Video-Action Model (DVA)
English
18
38
206
64.5K
Yunhao Luo retweetledi
Yinpei Dai
Yinpei Dai@YinpeiD·
Robot memory methods are growing fast, but systematic evaluation is largely lacking. 📉 Introducing RoboMME: a new benchmark for memory-augmented robotic manipulation! 🤖🧠 Featuring 16 tasks across temporal, spatial, object, and procedural memory 🔗 robomme.github.io
English
6
48
230
62.2K
Yunhao Luo
Yunhao Luo@yluo_y·
Video is a natural and scalable interface for robot planning. LVP demonstrates impressive in-the-wild planning and execution results, opening numerous research avenues for large-scale video models in robotics, such as real-time reactivity, memory, geometry, and precision.
Boyuan Chen@BoyuanChen0

Introducing Large Video Planner (LVP-14B) — a robot foundation model that actually generalizes. LVP is built on video gen, not VLA. As my final work at @MIT, LVP has all its eval tasks proposed by third parties as a maximum stress test, but it excels!🤗 boyuan.space/large-video-pl…

English
0
0
3
592
Yunhao Luo
Yunhao Luo@yluo_y·
🙏Huge thanks to my amazing collaborators: Utkarsh A. Mishra @utkarshm0410 Yilun Du @du_yilun and Danfei Xu @danfei_xu. This project wouldn’t have been possible without them. Our whole team will be at NeurIPS – feel free to reach out if you’d like to chat! (7/7)
English
0
0
6
378
Yunhao Luo
Yunhao Luo@yluo_y·
🚀Excited to share our @NeurIPSConf paper Compositional Diffuser (CompDiffuser)! CompDiffuser scales planning horizons at test time, able to construct long-horizon plans while only trained on short-horizon data. Project page: comp-diffuser.github.io Thread 👇(1/n)
English
2
17
68
33.8K