Pantelis Vafidis

117 posts

Pantelis Vafidis

Pantelis Vafidis

@vafidisp

AI Research Scientist @Meta GenAI. @Caltech PhD. Ex-@Instacart, @CSHL, @bccn_berlin, ECE @Auth_University. Used to be brains, now it’s LLMs

Pasadena, CA Katılım Ağustos 2020
330 Takip Edilen349 Takipçiler
Sabitlenmiş Tweet
Pantelis Vafidis
Pantelis Vafidis@vafidisp·
Excited to share our work on disentangled/abstract representations, to appear at #ICLR2025 (@iclr_conf)! We mathematically prove and experimentally demonstrate that multi-task learning leads to disentangled representations, and propose a unifying mechanism for generalization in brains and machines: parallel processing (🧵+paper below) Our work connects to the Platonic representation hypothesis, suggests why alignment across models/organisms can occur, and shows why transformers excel at constructing world models 🤖🚀
GIF
English
1
6
27
14.4K
Pantelis Vafidis
Pantelis Vafidis@vafidisp·
@YinChaoqun @KordingLab Agreed. But this blogpost shows that the criterion used to detect “line attractors” is extremely lax: even random dynamical systems pass it 50% of the time, and get classified as “approximate line attractors”, which is clearly wrong.
English
0
0
1
53
Chaoqun Yin
Chaoqun Yin@YinChaoqun·
@vafidisp @KordingLab I agree that the model somehow appears to fit everything. But I think it results from the very low-dimensional neural data itself. Its simple dynamics (mostly slow ramping) limits the model's complexity.
English
1
0
0
44
Kording Lab 🦖
Kording Lab 🦖@KordingLab·
Attractors are usually not mechanisms.
Kording Lab 🦖 tweet media
English
8
21
132
14.6K
Pantelis Vafidis
Pantelis Vafidis@vafidisp·
@rudzinskimaciej Yes, I could certainly see inhibition helping with decorrelation. Thanks for pointing it out. Not sure about the frequencies part though. Any references?
English
1
0
1
13
Rudzinski Maciej
Rudzinski Maciej@rudzinskimaciej·
Columns seem to have few mechanisms that keep their representations different as a design choice both through inhibitory neurons and frequencies organization I'm pointing it as it's testable architecture choice that maps exactly to what you suggested Random initial weights wouldn't give different enough results, we know already that NN with random initiation but trained on the same data usually form similar representation
English
1
0
1
24
Pantelis Vafidis
Pantelis Vafidis@vafidisp·
Yes great point, I forgot to reference the 1000 brains theory here but we do in the paper. One main difference is that they require dense signals to map the world, while we show that it can be done with sparse signals With regards to differentiation of cortical columns, it may simply come about due to the random initial projections (similar to heads in transformers or filters in CNNs)
English
1
0
1
14
Pantelis Vafidis
Pantelis Vafidis@vafidisp·
If you're at #ICLR2025, and interested in how we can guarantee true out-of-distribution generalization in neural networks (extrapolation), Aman Bhargava (@ABhargava2000) and I will be presenting our work tomorrow Saturday the 26th at 3:00-5:30pm, at Hall 3 (poster number #69) We will be happy to see you there! short presentation + slides: iclr.cc/virtual/2025/p…
Pantelis Vafidis tweet media
English
0
2
11
807
Pantelis Vafidis
Pantelis Vafidis@vafidisp·
Finally, huge thanks to amazing collaborator Aman Bhargava (@ABhargava2000) for recognizing the mathematical potential of this project and doing the theory part, and advisor Antonio Rangel! This project a prime example of the amplifying effect of great collaborations. Looking forward to more! Link to top:
Pantelis Vafidis@vafidisp

Excited to share our work on disentangled/abstract representations, to appear at #ICLR2025 (@iclr_conf)! We mathematically prove and experimentally demonstrate that multi-task learning leads to disentangled representations, and propose a unifying mechanism for generalization in brains and machines: parallel processing (🧵+paper below) Our work connects to the Platonic representation hypothesis, suggests why alignment across models/organisms can occur, and shows why transformers excel at constructing world models 🤖🚀

English
0
0
1
248
Pantelis Vafidis
Pantelis Vafidis@vafidisp·
Thanks for reading this far! For an in depth view of the above, I include the paper below (it’s 40 pages long!). Tldr: it worked no matter what we threw at it! And if you happen to be in Singapore for #ICLR2025, we will be presenting at poster session 6 on Saturday the 26th, 3:30-5 pm (Hall 3 + Hall 2B #69). We will be happy to see you there! arxiv.org/abs/2407.11249
English
1
0
0
243
Pantelis Vafidis
Pantelis Vafidis@vafidisp·
Excited to share our work on disentangled/abstract representations, to appear at #ICLR2025 (@iclr_conf)! We mathematically prove and experimentally demonstrate that multi-task learning leads to disentangled representations, and propose a unifying mechanism for generalization in brains and machines: parallel processing (🧵+paper below) Our work connects to the Platonic representation hypothesis, suggests why alignment across models/organisms can occur, and shows why transformers excel at constructing world models 🤖🚀
GIF
English
1
6
27
14.4K