Katherine Hermann retweetledi
Katherine Hermann
385 posts

Katherine Hermann
@khermann_
Research Scientist @GoogleDeepMind | Past: PhD from @Stanford
Katılım Ağustos 2016
1.3K Takip Edilen1.7K Takipçiler
Katherine Hermann retweetledi

Thrilled to announce I'll start in 2026 as faculty in Psych & CS @UAlberta + @AmiiThinks Fellow!! 🥳 Recruiting students to develop theories of cognition in natural and artificial systems 🤖💭🧠. Find me at #NeurIPS2025 workshops (talk at @CogInterp & organising @DataOnBrainMind)
English
Katherine Hermann retweetledi

LLMs memorize a lot of training data, but memorization is poorly understood.
Where does it live inside models? How is it stored? How much is it involved in different tasks?
@jack_merullo_ & @srihita_raju's new paper examines all of these questions using loss curvature! (1/7)

English
Katherine Hermann retweetledi

🧵🎉 Our mega-paper is finally published in TMLR! We're "Getting Aligned on Representational Alignment" - the degree to which internal representations of different (biological & artificial) information processing systems agree. 🧠🤖🔬🔍 #CognitiveScience #Neuroscience #AI

English
Katherine Hermann retweetledi

I am going on tour with my new book! Green Crime: Inside the minds of the people destroying the planet and how to stop them. Get your tickets here! 💚 linktr.ee/drjuliashaw

English
Katherine Hermann retweetledi

Many representational analyses (implicitly) prioritize signals by the amount of variance they explain in the representations. However, in arxiv.org/abs/2507.22216 we discuss results from our prior work that challenge this assumption; variance != computational importance.
English
Katherine Hermann retweetledi
Katherine Hermann retweetledi

🚀 New Open-Source Release! PyTorchTNN 🚀
A PyTorch package for building biologically-plausible temporal neural networks (TNNs)—unrolling neural network computation layer-by-layer through time, inspired by cortical processing. PyTorchTNN naturally integrates into the Encoder-Attender-Decoder (EAD) architecture (Chung*, Shen* et al., 2025), which flexibly combines diverse neural networks, motivated by the fact that no single model (Transformer, SSM, RNN) dominates all sequence learning tasks.
🧵👇
GIF
English
Katherine Hermann retweetledi

Our first NeuroAgent! 🐟🧠
Excited to share new work led by the talented @rdkeller, showing how autonomous behavior and whole-brain dynamics emerge naturally from intrinsic curiosity grounded in world models and memory.
Some highlights:
- Developed a novel intrinsic drive (3M-Progress) that better matches the reliable autonomy of animals
- First task-optimized model of neural-glial computation
- Surprisingly, no linear regression needed: a simple 1-to-1 mapping was enough to pass the NeuroAI Turing Test on whole-brain zebrafish data (~130,000 recorded units), provided you have the right intrinsic drive of course!
Check it out! 👇
Reece Keller@rdkeller
1/ I'm excited to share recent results from my first collaboration with the amazing @aran_nayebi and @Leokoz8! We show how autonomous behavior and whole-brain dynamics emerge in embodied agents with intrinsic motivation driven by world models.
English
Katherine Hermann retweetledi
Katherine Hermann retweetledi

Humans can tell the difference between a realistic generated video and an unrealistic one – can models?
Excited to share TRAJAN: the world’s first point TRAJectory AutoeNcoder for evaluating motion realism in generated and corrupted videos.
🌐 trajan-paper.github.io
🧵
GIF
English

@DynamicWebPaige For CA: Mount Langley (Southern Sierras) and Mount Tallac (Desolation Wilderness) are both really nice
English

Congratulations, Lukas! 🎉
Lukas Muttenthaler@lukas_mut
This past Friday I successfully defended my PhD 🎉🙏🏼 What a journey it was! 4.5 years of many ups and many downs. Can’t believe it’s over. I am still processing… Special thanks to my wonderful committee KR Müller, @martin_hebart, @cpilab, and @scychan_brains!
English
Katherine Hermann retweetledi

Train your vision SAE on Monday, then again on Tuesday, and you'll find only about 30% of the learned concepts match.
⚓ We propose Archetypal SAE which anchors concepts in the real data’s convex hull, delivering stable and consistent dictionaries.
arxiv.org/pdf/2502.12892…

English
Katherine Hermann retweetledi

Had a lot of fun speaking with @avileddie about the practical challenges of scaling (especially in Embodied AI), NeuroAI, what to expect in the future, and advice for students getting into the field.
Check it out here!
youtube.com/watch?v=ZRo-fL…

YouTube
English
Katherine Hermann retweetledi

1/ 🧵👇
What should count as a good model of intelligence?
AI is advancing rapidly, but how do we know if it captures intelligence in a scientifically meaningful way?
We propose the *NeuroAI Turing Test*—a benchmark that evaluates models based on both behavior and internal representations.
👉The key principle: given a metric, models should be *at least as good as brains are to each other*:

English
Katherine Hermann retweetledi

Are there fundamental barriers to AI alignment once we develop generally-capable AI agents?
We mathematically prove the answer is *yes*, and outline key properties for a "safe yet capable" agent. 🧵👇
Paper: arxiv.org/abs/2502.05934

English









