Timoleon (Timos) Moraitis

2.1K posts

Timoleon (Timos) Moraitis banner
Timoleon (Timos) Moraitis

Timoleon (Timos) Moraitis

@timos_m

Building brain-like AI @noemon_ai Previously @Huawei @IBMResearch @ETH_en @UZH_en @ntua

Zurich, Switzerland Katılım Ocak 2009
1.9K Takip Edilen1.4K Takipçiler
Timoleon (Timos) Moraitis
@kevinmcld "The service said it would use this principle to build the first of its Perceptron thinking machines that will be able to read and write." Admittedly, at least this part worked.
English
1
0
0
52
De myth ology
De myth ology@kevinmcld·
@timos_m It doesn't really work, it just appears to fleetingly. The binary was and still is just a toy model of reality.
English
1
0
1
37
Dane Malenfant
Dane Malenfant@dvnxmvl_hdf5·
@kalomaze It feels like everything fun was defined in the 90s and now we are just scaling
English
2
0
3
12.1K
Emmett Shear
Emmett Shear@eshear·
Learning is as much about effectively forgetting noise as it is about remembering signal.
English
29
40
412
19.7K
Nabil Iqbal
Nabil Iqbal@nblqbl·
@timos_m @noemon_ai super interesting. is it easy to understand why having a local learning rule would lead to features with these properties? (is there e.g. a toy model?)
English
1
0
2
147
Timoleon (Timos) Moraitis
Models trained with certain biologically-inspired local learning rules rather than backpropagation can be MUCH more robust to adversarial attacks, as we have shown in the past. Our view at @noemon_ai is that the locality of Hebbian plasticity in certain architectures leads to more compositional features, which in turn helps address the binding problem and adversarial robustness. In this new framework of "Perceptual Manifold" (PM) by @AleSalvatore00, @stanislavfort and @SuryaGanguli, our models would have measurably smaller PM than the backprop-trained ones.
Timoleon (Timos) Moraitis tweet media
Surya Ganguli@SuryaGanguli

Our new paper: "Solving adversarial examples requires solving exponential misalignment", expertly lead by @AleSalvatore00 w/ @stanislavfort arxiv.org/abs/2603.03507 Key idea: We all want to align AI systems to human values and intentions. We connect adversarial examples to AI alignment by showing they are a prototypical but exponentially severe form of misalignment at the level of perception. The fact that adversarial examples remain unsolved for over a decade thus serves as a cautionary tale for AI alignment, and provides new impetus for revisiting them. We shed light on why adversarial examples exist and why they are so hard to remove by asking a basic question: what is the dimensionality of neural network concepts in image space? For ResNets, and CLIP models, we show that neural network concepts (the space of images the network confidently labels as a concept) fill up almost the ENTIRE space of images (~135,000 dimensions out of ~150,000 for ImageNet & ~3000 out of 3072 for CIFAR10). In contrast natural image concepts are only ~20 dimensional. This indicates exponential misalignment between brain and machine perception (neural networks perceive exponentially many images as belonging to a concept that humans never would). This also explains why adversarial examples exist: if a concept fills up almost all of image space, ANY image will be close to that concept manifold. We further do experiments across > 20 networks showing that adversarial robustness inversely relates to concept dimensionality, though the most robust networks do not completely align machine and human perception. Overall the curse of dimensionality raises its ugly head as an impediment to both adversarial examples and alignment: if can be difficult to get AI systems to behave in accordance with human intentions, values, or perceptions over an exponentially large space of inputs. See @AleSalvatore00's excellent thread for more details: x.com/AleSalvatore00…

English
2
5
70
7.4K
Timoleon (Timos) Moraitis
@eshear If all models are overfit, does "overfitting" mean anything, and does it identify how to solve the issue? But you are technically right.
English
0
0
0
104
Emmett Shear
Emmett Shear@eshear·
@timos_m It is not a separate problem. It’s literally too many dimensions in the representation space — overfit.
English
1
0
4
223
Timoleon (Timos) Moraitis
Timoleon (Timos) Moraitis@timos_m

Models trained with certain biologically-inspired local learning rules rather than backpropagation can be MUCH more robust to adversarial attacks, as we have shown in the past. Our view at @noemon_ai is that the locality of Hebbian plasticity in certain architectures leads to more compositional features, which in turn helps address the binding problem and adversarial robustness. In this new framework of "Perceptual Manifold" (PM) by @AleSalvatore00, @stanislavfort and @SuryaGanguli, our models would have measurably smaller PM than the backprop-trained ones.

QME
0
1
30
3.4K
Alessandro Salvatore
Alessandro Salvatore@AleSalvatore00·
Why can't we solve adversarial examples? After a decade of work, neural nets still get fooled by imperceptible noise. We think we finally know the geometric reason why — and it connects to AI alignment. 🧵
Alessandro Salvatore tweet media
English
17
90
801
68.9K
Beff (e/acc)
Beff (e/acc)@beffjezos·
In 3.5 years @extropic: -reinvented how to use the transistor -reinvented architectures for probabilistic compute -reinvented deep learning for thermo compute -created our CUDA-like THRML -created our TF-like framework (coming soon) -scaled our systems 1000x yoy (3 gens of TSUs)
English
39
60
764
58K
Timoleon (Timos) Moraitis
Reminder: Long Short-Term Memory (LSTM) has been far outperformed. By Synaptic Plasticity, by us. We also showed this enables chips that burn 100x less power than GPUs. As another reminder, we have ~stopped publishing since. But the progress of the team @noemon_ai has been relentless, so feel free to extrapolate. ICML & Nature Comms paper links in the comment below 👇
Timoleon (Timos) Moraitis tweet media
Rohan Paul@rohanpaul_ai

Sam Altman just said in his new interview, that a new AI architecture is coming that will be a massive upgrade, just like Transformers were over Long Short-Term Memory. And also now the current class of frontier models are powerful enough to have the brainpower needed to help us research these ideas. His advice is to use the current AI to help you find that next giant step forward. --- From 'TreeHacks' YT Channel (link in comment)

English
1
2
17
2.3K
Timoleon (Timos) Moraitis
They don't report only median/robust. The more fine-grained analysis looks even more pessimistic to me. You do have a point, but I think what you/we are experiencing is very specific to ML experimentation. It's a rare combination of high-value long horizons on the one hand, with dense verifiability and standard public knowledge recipes on the other hand.
Timoleon (Timos) Moraitis tweet media
English
0
0
6
252
Dimitris Papailiopoulos
Dimitris Papailiopoulos@DimitrisPapail·
METR and other long-horizon eval orgs are being conservative and moderate in how they measure agent capabilities. That's reasonable as we have already enough hype and don't need more. But I think we're missing something important by only reporting median/robust performance. I've had Claude Code and Codex sustain end to end ML research tasks for days without intervention. Not robustly across all settings, but it's happening and it's incredible. We need a shameless, cherry-picked frontier eval. Not to mislead but because knowing exactly where the ceiling of capabilities lies is just as important as knowing the average. I keep seeing pessimistic long horizon results and thinking: am I in a bubble? Are MY 50-hour autonomous tasks a hallucination? I don't think they are!! AI agents can do sustained multi-day research. Not always and not for everyone, but it's real and people should know where the frontier actually is.
English
18
11
153
18.5K