Michael Munje

10 posts

Michael Munje

Michael Munje

@michaelmunje

PhD Student @UTAustin | MS @GeorgiaTech | Former Intern @Microsoft @NASAJPL @IBMResearch

Austin Katılım Eylül 2025
42 Takip Edilen13 Takipçiler
Sabitlenmiş Tweet
Michael Munje
Michael Munje@michaelmunje·
[1/8] New social navigation paper + benchmark: SocialNav-SUB 🚶🤖 Recent work puts VLMs on robots for navigation, but can they really interpret scenes and extract key details for social navigation? 🔎 larg.github.io/socialnav-sub
Michael Munje tweet media
English
8
3
9
3K
Michael Munje retweetledi
Zichao
Zichao@ZichaoHu99·
How can robots follow complex instructions in dynamic environments? 🤖 Meet ComposableNav — a diffusion-based planner that enables robots to generate novel navigation behaviors that satisfy diverse instruction specifications on the fly — no retraining needed. 📄 Just accepted to CoRL 2025 🔗 Project: amrl.cs.utexas.edu/ComposableNav/ A Thread (1/8)
English
1
6
19
2.6K
Michael Munje
Michael Munje@michaelmunje·
[1/8] New social navigation paper + benchmark: SocialNav-SUB 🚶🤖 Recent work puts VLMs on robots for navigation, but can they really interpret scenes and extract key details for social navigation? 🔎 larg.github.io/socialnav-sub
Michael Munje tweet media
English
8
3
9
3K
Michael Munje
Michael Munje@michaelmunje·
[7/8] SocialNavSUB is also fully open-source, actively maintained, and easily extendable to customized prompts and/or additional VLMs! Pull requests are always welcome! github.com/LARG/SocialNav…
English
0
0
0
50
Michael Munje
Michael Munje@michaelmunje·
[6/8] 🧪 Does chain-of-thought (using spatial/spatiotemporal VQAs first) improve social reasoning? ✅Yes. Does BEV context help models? ⚖️ Model-dependent (sometimes a lot). Does better spatial(temporal) context improve social reasoning? ✅Yes.
English
0
0
1
47
Michael Munje
Michael Munje@michaelmunje·
[5/8] 📊 Do today’s VLMs agree with human judgments? We find that they still trail behind humans and simple rule-based baselines.
Michael Munje tweet media
English
0
0
0
46
Michael Munje
Michael Munje@michaelmunje·
[4/8] 👥 We collected human data from an IRB-approved human-subject study to construct our benchmark and evaluate whether models align with human judgments in social navigation scenes.
English
0
0
1
48
Michael Munje
Michael Munje@michaelmunje·
[3/8] SocialNav-SUB features real-world social navigation scenarios built from SCAND scenarios @ 4 Hz → PHALP tracking → front-view & BEV with labeled pedestrians, combining them with a set of carefully designed questions to create our VQA prompts (5k in total).
Michael Munje tweet media
English
0
0
1
59
Michael Munje
Michael Munje@michaelmunje·
[2/8] We introduce SocialNav-SUB: a VQA benchmark to evaluate spatial, spatiotemporal, and social reasoning for real-world social navigation scenarios with object-centric grounding (front view + Bird’s-Eye-View (BEV) + numbered markers) to provide rich context to VLMs.
English
0
0
1
73