Eric Todd

141 posts

Eric Todd

Eric Todd

@ericwtodd

Computer Science PhD Student at Northeastern University

Boston, MA Inscrit le Aralık 2014
444 Abonnements474 Abonnés
Tweet épinglé
Eric Todd
Eric Todd@ericwtodd·
Can you solve this algebra puzzle? 🧩 cb=c, ac=b, ab=? A small transformer can learn to solve problems like this! And since the letters don't have inherent meaning, this lets us study how context alone imparts meaning. Here's what we found:🧵⬇️
Eric Todd tweet media
English
9
50
319
54.9K
Eric Todd retweeté
David Bau
David Bau@davidbau·
In 1982, high school students in Sudbury, Mass. wrote a dungeon game called Hack. They had Atari 800s and Logo and an obsession with a Unix game called Rogue that most of them had never seen. I grew up one town over with the same computers and the same obsession.
David Bau tweet media
English
1
6
17
1.4K
Eric Todd retweeté
Rohit Gandikota
Rohit Gandikota@rohitgandikota·
I’ll be presenting our work, “Distilling Diversity and Control in Diffusion Models,” at @wacv_official this Sunday at 11 AM local time. 🔍We uncover the “secret to unlocking diversity” in diffusion models - using **interpretability**!! DM me if you’d like to connect in Tucson.
Rohit Gandikota@rohitgandikota

Why do distilled diffusion models generate similar-looking images? 🤔 Our Diffusion Target (DT) visualization reveals the secret to diversity. It is the very first time-step! And—there is a simple, training-free way to make them more diverse! Here is how: 🧵👇

English
0
4
23
2.1K
Eric Todd retweeté
Jaden Fiotto-Kaufman
Jaden Fiotto-Kaufman@jadenfk23·
NNsight 0.6 is out now! We directly address your feedback in our biggest release yet. Pain points included cryptic errors, slow traces, no remote execution of custom code, and limited vLLM support. We tackle all of these and more in this new release. 🧵 Here's what changed:
English
1
10
38
7.3K
Eric Todd retweeté
Eric Todd retweeté
Kerem Şahin
Kerem Şahin@keremsahin2210·
Are induction heads necessary for the emergence of in-context learning (ICL)? Their emergence coincides with a sharp ICL improvement, raising the hypothesis they may underlie much of ICL. However, we find that ICL beyond copying can emerge even when we suppress induction heads!
Kerem Şahin tweet media
English
3
16
124
16.6K
Eric Todd retweeté
Chris Wendler
Chris Wendler@wendlerch·
Data is plenty, knowledge is scarce. We began to close this gap thanks to deep learning <3 Neural networks can learn “programs” that often achieve superhuman performance from data alone. What insights are encoded in their weights? Here we took a first step on AI protein folding.
Kevin Lu@kevinlu4588

How do protein folding models turn sequence into structure? In "Mechanisms of AI Protein Folding in ESMFold", we find properties like charge and distance encoded in interpretable, steerable directions. The trunk processes features in two phases: chemistry first, then geometry.

English
2
10
28
1.8K
Eric Todd retweeté
Kevin Lu
Kevin Lu@kevinlu4588·
How do protein folding models turn sequence into structure? In "Mechanisms of AI Protein Folding in ESMFold", we find properties like charge and distance encoded in interpretable, steerable directions. The trunk processes features in two phases: chemistry first, then geometry.
Kevin Lu tweet media
English
4
44
205
19.9K
Eric Todd retweeté
Andrew Lee
Andrew Lee@a_jy_l·
😻New preprint! As an interp researcher, I often ask “why did the model attend to this token?” We study this by decomposing the query-key (QK) space into interpretable low-rank subspaces. When these subspaces of Qs and Ks align, the model produces high attention scores. 1/N
Andrew Lee tweet media
English
4
19
134
6.8K
Eric Todd retweeté
Grace Luo
Grace Luo@graceluo_·
We trained diffusion models on a billion LLM activations, and we want you to use them! New preprint: Learning a Generative Meta-Model of LLM Activations Joint work with @feng_jiahai, @trevordarrell, @AlecRad, @JacobSteinhardt. More in thread 🧵
English
30
170
1.3K
189.1K
Eric Todd
Eric Todd@ericwtodd·
Can you solve this algebra puzzle? 🧩 cb=c, ac=b, ab=? A small transformer can learn to solve problems like this! And since the letters don't have inherent meaning, this lets us study how context alone imparts meaning. Here's what we found:🧵⬇️
Eric Todd tweet media
English
9
50
319
54.9K
Eric Todd
Eric Todd@ericwtodd·
@jjcvip @davidbau @jannikbrinkmann @rohitgandikota In our paper we let the symbols to represent arbitrary algebraic elements, and focus mainly on groups. But you're right to point out that when less structure is assumed, these algebra puzzles become harder or even impossible to solve in closed form.
English
1
0
1
49
Eric Todd
Eric Todd@ericwtodd·
@plumsirawit Yes, in our paper we study mainly groups and find transformers learn strategies that reflect group structure like identity & cancellation rules. But as you point out, when we test models on less algebraic structure, these puzzles become harder (or sometimes impossible) to solve
English
0
0
3
98
Sirawit wants to reach the aliens from Mars 👽
ab = aac = aacb = abb Uhh I seem to be unable to solve the puzzle without any unnatural assumptions… (ofc associativity and extensionality laws are naturally assumed by notation) If one assumes group law (or even just cancellation+neutral law) then cb=c implies b=id, so ab=a.
Eric Todd@ericwtodd

Can you solve this algebra puzzle? 🧩 cb=c, ac=b, ab=? A small transformer can learn to solve problems like this! And since the letters don't have inherent meaning, this lets us study how context alone imparts meaning. Here's what we found:🧵⬇️

English
1
0
3
524
Eric Todd
Eric Todd@ericwtodd·
@AlecEBG @davidbau @jannikbrinkmann @rohitgandikota The symbols don't have to be integers (they represent arbitrary group elements), but yes you can think of them as non-zero. See this comment as well: x.com/glassala/statu…
Violaine the Alchemist@glassala

@Marshwiggle119 @ericwtodd @davidbau @jannikbrinkmann @rohitgandikota The product is a product of arbitrary group elements, and groups don’t allow for an element which behaves like 0 under multiplication. (Even the multiplicative group of e.g. the real numbers has to explicitly exclude 0.)

English
0
0
3
647
Eric Todd
Eric Todd@ericwtodd·
@octonion Yes, cool right!? This sudoku-style cancellation is one of the context-based strategies we saw the transformers learn! We also talk a bit about the model's performance on finite quasigroups (represented via Latin squares) in the paper's appendix. x.com/ericwtodd/stat…
Eric Todd@ericwtodd

Another strategy infers meaning using sets. We have seen models keep track of "positive" and "negative" sets that let it narrow its understanding of a symbol using Sudoku-style cancellation. Red bars (a) show the positive set and blue boxes (b) show the negative.

English
0
0
3
1.1K
Eric Todd
Eric Todd@ericwtodd·
@Marshwiggle119 @davidbau @jannikbrinkmann @rohitgandikota Right! For this puzzle c is not 0. While variables can (and often do) represent numbers, our setup also lets us assign them represent arbitrary group elements. If symbols were to represent elements of a dihedral group their interactions can be seen as "rotations" of a polygon.
English
1
0
2
600
Eric Todd
Eric Todd@ericwtodd·
@plain_simon @davidbau @jannikbrinkmann @rohitgandikota Glad you found it interesting! And thanks for pointing this out - I have some (appendix) updates to the preprint coming soon so in the next preprint version we can update the language to clarify that we're just talking about "groups".
English
0
0
6
1.1K
Eric Todd
Eric Todd@ericwtodd·
Takeaway: contextual reasoning can be richer than just fuzzy copying! See the paper for more results, including an analysis of learning dynamics. Code & data are available at our project website. 📜: arxiv.org/abs/2512.16902 🌐: algebra.baulab.info
English
1
2
29
1.9K
Eric Todd
Eric Todd@ericwtodd·
Another strategy infers meaning using sets. We have seen models keep track of "positive" and "negative" sets that let it narrow its understanding of a symbol using Sudoku-style cancellation. Red bars (a) show the positive set and blue boxes (b) show the negative.
Eric Todd tweet media
English
1
2
18
3.8K