Vijay Veerabadran

738 posts

Vijay Veerabadran

Vijay Veerabadran

@simple_cell_

Research @Meta @RealityLabs | Prev: Ph.D. from @CogSciUCSD

San Diego, CA Katılım Temmuz 2010
1.7K Takip Edilen493 Takipçiler
Sabitlenmiş Tweet
Vijay Veerabadran
Vijay Veerabadran@simple_cell_·
🚨 Appearing as a #NeurIPS2025 D&B spotlight(~3%) Could VLMs guess your next prompt for a wearable AI agent? We present WAGIBench, the 1st large-scale Goal Inference Benchmark for Wearable Agents w/ audiovisual, digital & longitudinal context! Paper: arxiv.org/abs/2510.22443 1/
English
1
5
14
827
Vijay Veerabadran retweetledi
David Bar
David Bar@observie·
URDF Visualizer: a vscode extension for URDF visualization/debugging, and by far the best way I've found so far
English
9
29
388
38.8K
Vijay Veerabadran retweetledi
hardmaru
hardmaru@hardmaru·
One of my favorite findings: Positional embeddings are just training wheels. They help convergence but hurt long-context generalization. We found that if you simply delete them after pretraining and recalibrate for < 1% of the original budget, you unlock massive context windows.
Sakana AI@SakanaAILabs

Introducing DroPE: Extending the Context of Pretrained LLMs by Dropping Their Positional Embeddings pub.sakana.ai/DroPE/ We are releasing a new method called DroPE to extend the context length of pretrained LLMs without the massive compute costs usually associated with long-context fine-tuning. The core insight of this work challenges a fundamental assumption in Transformer architecture. We discovered that explicit positional embeddings like RoPE are critical for training convergence but eventually become the primary bottleneck preventing models from generalizing to longer sequences. Our solution is radically simple: We treat positional embeddings as a temporary training scaffold rather than a permanent architectural necessity. Real-world workflows like reviewing massive code diffs or analyzing legal contracts require context windows that break standard pretrained models. While models without positional embeddings (NoPE) generalize better to these unseen lengths, they are notoriously unstable to train from scratch. Here, we achieve the best of both worlds by using embeddings to ensure stability during pretraining and then dropping them to unlock length extrapolation during inference. Our approach unlocks seamless zero-shot context extension without any expensive long-context training. We demonstrated this on a range of off-the-shelf open-source LLMs. In our tests, recalibrating any model with DroPE requires less than 1% of the original pretraining budget, yet it significantly outperforms established methods on challenging benchmarks like LongBench and RULER. We have released the code and the full paper to encourage the community to rethink the role of positional encodings in modern LLMs. Paper: arxiv.org/abs/2512.12167 Code: github.com/SakanaAI/DroPE

English
49
241
2.5K
345.5K
Vijay Veerabadran retweetledi
Sahithya Ravi
Sahithya Ravi@Sahithya_Ravi·
Hi👋  I am at #neurips2025 to present our position paper on world models.🌍👇 🗓️ Dec 7  📍Upper Level Ballroom 20D/LAW workshop. I’m excited to chat about VL Reasoning & video generation and am actively looking for Postdoc/ Full-time roles. DM me if you would like to chat.
Adi Chinchure@adityachinchure

How can we build more robust and safe world models? 🤔 Our position paper, World Models Must Live in Parallel Worlds 🌍, tries to answer this question. Find us at NeurIPS 2025: 🗓️ Dec 7 📍 Upper Level Ballroom 20D #neurips2025 w/ @Sahithya_Ravi @VeredShwartz, Leonid Sigal 1/5

English
0
5
12
2.5K
Vijay Veerabadran retweetledi
Sergey Levine
Sergey Levine@svlevine·
We just released results for our newest VLA from Physical Intelligence: π*0.6. This one is trained with RL, and it makes it quite a bit better: often doubles throughput, enables real-world tasks like folding real laundry and making espresso drinks at the office.
English
46
194
1.7K
296.3K
Vijay Veerabadran retweetledi
Josh McDermott
Josh McDermott@JoshHMcDermott·
New pre-print from our lab, by Lakshmi Govindarajan @lakshming92 with help from Sagarika Alavilli, introducing a new type of model for studying sensory uncertainty. biorxiv.org/content/10.110… Here is a summary. (1/n)
English
1
8
26
3.6K
Vijay Veerabadran
Vijay Veerabadran@simple_cell_·
🧠 vs 🤖We benchmarked various evaluation functions by comparing them to human raters. We find that the LLM Judge parameterized with scenario script best aligns with human raters (76.8% agreement), performing as well as any individual from a separate group. 8/
Vijay Veerabadran tweet media
English
1
0
0
71
Vijay Veerabadran
Vijay Veerabadran@simple_cell_·
🚨 Appearing as a #NeurIPS2025 D&B spotlight(~3%) Could VLMs guess your next prompt for a wearable AI agent? We present WAGIBench, the 1st large-scale Goal Inference Benchmark for Wearable Agents w/ audiovisual, digital & longitudinal context! Paper: arxiv.org/abs/2510.22443 1/
English
1
5
14
827
Vijay Veerabadran retweetledi
Nicholas Fabiano, MD
Nicholas Fabiano, MD@NTFabiano·
In 1992 Peter Ratcliffe received this rejection letter from Nature. His findings were not "a sufficient advance in our understanding". 27 years later he won the Nobel Prize for the same discovery. Don't lose faith in the things you believe in.
Nicholas Fabiano, MD tweet media
English
172
1.9K
7.8K
725.5K
Vijay Veerabadran retweetledi
Effie Li
Effie Li@_EffieLi_·
🌟To appear in the MechInterp Workshop @ #NeurIPS2025 🌟 Paper: arxiv.org/abs/2509.04466 How do language models (LMs) form representation of new tasks, during in-context learning? We study different types of task representations, and find that they evolve in distinct ways. 🧵1/7
Effie Li tweet media
English
1
15
111
20K