tyler bonnen

588 posts

tyler bonnen banner
tyler bonnen

tyler bonnen

@tylerraye

neuroscientist @berkeley_ai. NIH K00 + UC Presidential Postdoctoral Fellow

Unceded Muwekma Ohlone Land شامل ہوئے Ağustos 2020
662 فالونگ1.7K فالوورز
پن کیا گیا ٹویٹ
tyler bonnen
tyler bonnen@tylerraye·
excited to share some recent work! tldr; models trained on multi-view sensory data are the first to match human-level 3D shape perception—all zero shot, with no training on experimental data/images project page: tzler.github.io/human_multiview 1/🧠
English
3
23
126
15.3K
tyler bonnen ری ٹویٹ کیا
Qianqian Wang
Qianqian Wang@QianqianWang5·
The shell game is a fun challenge that cannot be solved by looking at a single frame. The model has to track every move, from the moment the object is hidden. Excited to share this!
Rhoda AI@RhodaAI

Here’s something we’ve never seen done before. Real-world tasks are long and ambiguous. Solving them requires visual memory and state tracking. Most robot policies only see the last few frames. Ours doesn't. We put our DVA, FutureVision, to the perfect testbed: the shell game 🐚. The DVA nails it.

English
2
7
70
10.5K
tyler bonnen ری ٹویٹ کیا
Imran Thobani
Imran Thobani@cogphilosopher·
1/ Most model-brain comparisons only ask: can the model predict the brain? — without also checking the reverse direction. When you map in both directions, differences between models emerge that were previously invisible. In prior work, we showed there's a deeper principle behind bidirectional mapping: we should compare models to brains the same way we compare real brains to each other 🧵
Jorge Bravo Abad@bravo_abad

The missing half of the neural network–brain comparison For a decade, the standard benchmark for artificial neural networks as models of the brain has been forward predictivity: learn a linear mapping from model activations to neural recordings and measure explained variance. Top models of the macaque inferior temporal (IT) cortex—central to object recognition—have plateaued near 50% regardless of architecture. Muzellec and Kar argue this plateau hides something important. Two models can score identically on forward predictivity while relying on fundamentally different internal strategies. One may have many units tightly coupled to IT responses; the other may reach the same score with a smaller aligned subset while carrying a large pool of biologically inaccessible dimensions. To expose this, they introduce reverse predictivity: instead of asking how well model features predict neurons, they ask how well IT neurons predict individual model units. A truly brain-like model should be bidirectionally predictable—just as two monkeys' IT populations predict each other symmetrically, which the authors confirm as their empirical baseline. Across 39 architectures—CNNs, transformers, self-supervised and robust models—reverse predictivity is consistently lower than forward predictivity and the two metrics are uncorrelated. Strikingly, higher ImageNet accuracy predicts lower reverse predictivity. Adversarial training helps; higher dimensionality hurts. The "common" units identified this way predict primate behavior more consistently across species and models than the "unique" ones inaccessible from neural activity. For AI in drug discovery, neurotechnology, or computational biology, this has a direct implication: forward accuracy alone does not guarantee that a model's internal representations are embedded in the biological system it claims to describe. When those representations guide mechanistic interpretations or experimental decisions, the mismatch can mislead. Paper: Muzellec et al., Nature Machine Intelligence (2026) | nature.com/articles/s4225…

English
1
27
112
15.7K
tyler bonnen ری ٹویٹ کیا
Jennifer Listgarten
Jennifer Listgarten@jlistgarten·
Our review on AI for protein engineering is out now, about this too-fast-moving field full of hype and overclaim, yet one that is having a real impact on the world and can be described in a coherent manner without histrionics science.org/eprint/666XRGR…
English
7
85
410
33.4K
tyler bonnen ری ٹویٹ کیا
Neerja Thakkar
Neerja Thakkar@neerjathakkar·
What’s the right representation for a world model? 3D, pixels, or something else? Excited to release our new paper “Forecasting Motion in the Wild” where we propose point tracks as tokens for generating complex non-rigid motion and behavior From @GoogleDeepmind @Berkeley_AI @TTIC_Connect
GIF
English
8
71
452
75.3K
tyler bonnen ری ٹویٹ کیا
Vongani Maluleke
Vongani Maluleke@vonekels·
When people share a space, their movements become intertwined. Embodied agents need to understand these social dynamics to interact effectively. Introducing MAGNet 🧲, a unified autoregressive diffusion forcing model for multi-agent motion generation that captures these interactions. MAGNet is flexible: predict the future, fill in missing motion, or have people react to each other, all while naturally scaling to N>2 people and generating ultra-long motion sequences.
English
6
57
367
64.3K
tyler bonnen ری ٹویٹ کیا
tyler bonnen ری ٹویٹ کیا
Baifeng
Baifeng@baifeng_shi·
Humans can see in high-res, high-FPS in real-time. Why can't VLMs? Introducing AutoGaze: ViTs/VLMs "gaze" only at key video regions! Up to 4-100x token savings, 19x speedup, and enables scaling to 4K-res 1K-frame videos. 📄 arxiv.org/abs/2603.12254 🌐 autogaze.github.io 🤗 huggingface.co/collections/bf… (1/n)🧵
English
47
200
1.6K
152.1K
tyler bonnen ری ٹویٹ کیا
Roy Eyono
Roy Eyono@RoyEyono·
How do neural circuits in the brain implement normalization? 🧠 In our new paper, we show that just normalizing sensory input isn't enough. Crucially, we must also normalize the error signals! 🧵👇 Paper: arxiv.org/abs/2603.17676
Roy Eyono tweet media
English
1
27
129
8.6K
tyler bonnen ری ٹویٹ کیا
Junyi Zhang
Junyi Zhang@junyi42·
𝗢𝗻𝗲 𝗺𝗲𝗺𝗼𝗿𝘆 𝗰𝗮𝗻’𝘁 𝗿𝘂𝗹𝗲 𝘁𝗵𝗲𝗺 𝗮𝗹𝗹. We present 𝗟𝗼𝗚𝗲𝗥, a new 𝗵𝘆𝗯𝗿𝗶𝗱 𝗺𝗲𝗺𝗼𝗿𝘆 architecture for long-context geometric reconstruction. LoGeR enables stable reconstruction over up to 𝟭𝟬𝗸 𝗳𝗿𝗮𝗺𝗲𝘀 / 𝗸𝗶𝗹𝗼𝗺𝗲𝘁𝗲𝗿 𝘀𝗰𝗮𝗹𝗲, with 𝗹𝗶𝗻𝗲𝗮𝗿-𝘁𝗶𝗺𝗲 𝘀𝗰𝗮𝗹𝗶𝗻𝗴 in sequence length, 𝗳𝘂𝗹𝗹𝘆 𝗳𝗲𝗲𝗱𝗳𝗼𝗿𝘄𝗮𝗿𝗱 inference, and 𝗻𝗼 𝗽𝗼𝘀𝘁-𝗼𝗽𝘁𝗶𝗺𝗶𝘇𝗮𝘁𝗶𝗼𝗻. Yet it matches or surpasses strong optimization-based pipelines. (1/5) @GoogleDeepMind @Berkeley_AI
English
63
449
3.4K
556.4K
tyler bonnen ری ٹویٹ کیا
Micha Heilbron
Micha Heilbron@m_heilb·
📢 PhD position in Developmental Language Modelling (plz RT🙏) What can human language acquisition teach us about training language models? Join us as a PhD! 4 yrs, fully funded, MPI-NL; april 3 mpi.nl/career-educati…
Micha Heilbron tweet mediaMicha Heilbron tweet mediaMicha Heilbron tweet mediaMicha Heilbron tweet media
English
3
28
106
19.2K
tyler bonnen
tyler bonnen@tylerraye·
these findings provide a computational bridge between cognitive theories and current practices in deep learning we talk about some of these connections in the manuscript, but there are so many exciting questions to explore at the intersection of cog/comp/neuro science
English
1
0
3
368
tyler bonnen
tyler bonnen@tylerraye·
excited to share some recent work! tldr; models trained on multi-view sensory data are the first to match human-level 3D shape perception—all zero shot, with no training on experimental data/images project page: tzler.github.io/human_multiview 1/🧠
English
3
23
126
15.3K