Siyin Wang @ICLR 2026

28 posts

Siyin Wang @ICLR 2026

Siyin Wang @ICLR 2026

@wang_siyin

PhD Student at Fudan University #NLProc #LLM

Katılım Aralık 2022
158 Takip Edilen45 Takipçiler
Xuning Yang
Xuning Yang@xuningy·
When every generalist robot model scores 95%+ on a benchmark, the numbers become meaningless. What if we built a photorealistic benchmark that never saturates and can generate new scenes and tasks with AI Workflows in minutes? We introduce RoboLab! 🧵(1/6)
Xuning Yang tweet media
English
9
26
144
26.8K
Siyin Wang @ICLR 2026
Siyin Wang @ICLR 2026@wang_siyin·
Made it to Rio after nearly 30 hours in the air ✈️ Excited to be at #ICLR2026! We'll be sharing our work on RoboOmni on Apr 25 (10:30 AM – 1:00 PM). If you're interested in VLAs, Omni-LLMs, World Models, or Multimodal Agentic AI, feel free to reach out! #multimodality #VLA
Siyin Wang @ICLR 2026 tweet media
English
0
1
11
388
Siyin Wang @ICLR 2026 retweetledi
机器之心 JIQIZHIXIN
机器之心 JIQIZHIXIN@jiqizhixin·
What if robots could understand what you want without being told? RoboOmni makes that possible — an omni-modal LLM that fuses speech, sound, and vision to infer human intent, confirm actions, and execute tasks. Trained on the new OmniAction dataset (140k episodes), it outperforms text- and ASR-based baselines in success rate, speed, and proactive assistance, paving the way for more intuitive human-robot collaboration. RoboOmni: Proactive Robot Manipulation in Omni-modal Context Fudan, SII, NUS Paper: arxiv.org/abs/2510.23763 Code: github.com/OpenMOSS/RoboO… Project: OpenMOSS.github.io/RoboOmni Our report: mp.weixin.qq.com/s/PXBqdEW7_Ta_… 📬 #PapersAccepted by Jiqizhixin
机器之心 JIQIZHIXIN tweet media
English
3
11
41
3.8K
Siyin Wang @ICLR 2026
Siyin Wang @ICLR 2026@wang_siyin·
I will be at EMNLP from Nov 5th to Nov 8th. If you are interested in multimodal spatial reasoning, Embodied AI (like VLA), Omni-LLMs, please feel free to chat with me!👋 📍I will also present a poster (ConvSearch-R1) with Changtai at Nov 6th 16:30-18:00. #EMNLP2025 #suzhou
English
0
0
8
372
Siyin Wang @ICLR 2026 retweetledi
DailyPapers
DailyPapers@HuggingPapers·
Unveiling the hidden fragilities of AI robotics New research introduces LIBERO-Plus, a comprehensive benchmark that systematically reveals vulnerabilities in Vision-Language-Action models under 7 real-world perturbation dimensions.
DailyPapers tweet media
English
1
3
9
1.1K
Siyin Wang @ICLR 2026
Siyin Wang @ICLR 2026@wang_siyin·
🚀Tired of Libero? Try our Libero-Plus! 🤔Libero’s at 99%, but we’ve found VLA drops points with even minor disturbances. 🤩Switch to Libero+ in just a few steps and unlock your VLA’s true generalization ability. #Embodied #VLA #Robotics
Senyu Fei@SenyuFei

🤯 Shocking findings from our new LIBERO-Plus benchmark for VLA robustness! 💡 Key Insight: High LIBERO scores ≠ strong models. 🔗 Paper: huggingface.co/papers/2510.13… 🌐 Page: sylvestf.github.io/LIBERO-plus 💻 Code: github.com/sylvestf/LIBER… ⭐ Star us & 🚀 upvote! #VLA #Robotics 1/8

English
0
2
5
656
Fei Liu
Fei Liu@feiliu_nlp·
🏆 Thrilled that our paper #PlanGenLLMs (arxiv.org/abs/2502.11221) won the SAC Award at #ACL2025!! Couldn't have done it without the amazing team: @HuiWei15, Zihao Zhang, Shenghua He, Tian Xia, and Shijia Pan. So thankful and beyond proud! 💖 #ACL2025NLP #NLProc 🧠 Planning is a core aspect of both human and artificial intelligence. LLMs/agents have been used in various planning tasks, from navigating websites and planning trips to querying databases, but most benchmarks are narrow and task-specific. That makes it difficult to compare systems across domains, or figure out which one's best for a new planning problem. That's where our paper comes in: We offer a comprehensive overview of LLM-based planning agents, highlighting gaps, challenges, and what's next. Check it out 👉 arxiv.org/abs/2502.11221
Fei Liu tweet media
English
4
9
62
7K
Siyin Wang @ICLR 2026
Siyin Wang @ICLR 2026@wang_siyin·
Excited to be in Vienna! Come by if you’re around!👋#ACL2025 📌Poster: VisuoThink (Thinking with images, Multimodal reasoning) 📌Poster: D2PO (World Modeling, Embodied) 🗓️Tue, July 29 16:00–17:30 |📍Hall 4/5 (Session 3) Presenting with @ngc7293q @JinlanFu !👏
Siyin Wang @ICLR 2026@wang_siyin

Thrilled to share our TWO papers accepted to #ACL2025 Main Conference! 🥳🎉 🎨VisuoThink: Empowering LVLM Reasoning with Multimodal Tree Search 🌏World Modeling Makes a Better Planner: Dual Preference Optimization for Embodied Task Planning #AI #MultimodalLearning #worldmodel

English
4
0
5
296
Siyin Wang @ICLR 2026
Siyin Wang @ICLR 2026@wang_siyin·
Exciting work on Embodied Agents! 🦾 By leveraging interactive reinforcement learning, we shatter the ceiling on ALFWorld (31.05 -> 97.78) and ScienceWorld (22.05 -> 79.92). 🔥 Huge thanks to amazing coauthors @ngc7293q li ji, junhao, @JingjingGong_ @xpqiu
Zhaoye Fei(ngc7293)@ngc7293q

🚀 New work: OpenMOSS Embodied Planner-R1 - A step toward AI self-improvement in interactive planning! We've developed an RL framework where LLMs learn to plan through autonomous environmental exploration - no human demonstrations needed. 🤖 🧵 Thread below 👇

English
0
0
2
228
Siyin Wang @ICLR 2026
Siyin Wang @ICLR 2026@wang_siyin·
Thrilled to share our TWO papers accepted to #ACL2025 Main Conference! 🥳🎉 🎨VisuoThink: Empowering LVLM Reasoning with Multimodal Tree Search 🌏World Modeling Makes a Better Planner: Dual Preference Optimization for Embodied Task Planning #AI #MultimodalLearning #worldmodel
Siyin Wang @ICLR 2026 tweet mediaSiyin Wang @ICLR 2026 tweet media
English
1
1
5
769
Siyin Wang @ICLR 2026
Siyin Wang @ICLR 2026@wang_siyin·
🔎 We also compared action-conditioned (predict state given state & action) vs. goal-directed world models (imagines future state from history & goal). While action-conditioned excel in familiar settings, goal-directed models generalize better to novel environments! 7/8
Siyin Wang @ICLR 2026 tweet media
English
1
0
2
98
Siyin Wang @ICLR 2026
Siyin Wang @ICLR 2026@wang_siyin·
✨ Excited to share our latest research “World Modeling Makes a Better Planner: Dual Preference Optimization for Embodied Task Planning” 🤔 Current LVLMs struggle with grounding in embodied environments, how can we make AI agents understand the physical world like humans? 1/8
Siyin Wang @ICLR 2026 tweet media
English
1
3
6
435