Alan Aboudib

118 posts

Alan Aboudib banner
Alan Aboudib

Alan Aboudib

@alan_aboudib

IOTA AI Research Lead @macrocosmosai — Intelligence Podcast @YouTube

Paris, France Katılım Şubat 2011
119 Takip Edilen304 Takipçiler
Alan Aboudib
Alan Aboudib@alan_aboudib·
Huge thanks to @Pluralis for the invite to give a lightning talk on ResBM at the Protocol Learning Workshop in Rio, an ICLR 2026 event. Grateful for the momentum, the openness, and the space to exchange ideas on decentralized learning 🔥
Alan Aboudib tweet media
English
0
5
10
1K
Alan Aboudib retweetledi
Alexander Long
Alexander Long@AlexanderLong·
Activation compression is fundamentally different to gradient compression as it alters training dynamics, but it must be solved to allow a model to be split up over participants. If you cannot split a model over participants, I don't see how you keep the weight set private, and if you can't keep the weight set private, I don't see how you make collective training sustainable. A genuinely novel sub-field is emerging here - very rare thing to be able to observe in realtime.
Macrocosmos@MacrocosmosAI

Training frontier models over the internet requires new techniques. Today, we present ResBM, a residual encoder-decoder bottleneck architecture that enables 128x activation compression for low-bandwidth distributed pipeline parallel training. Developed for @IOTA_SN9, we show SOTA compression without significant loss in convergence rates, increases in memory, or compute overhead. Expect the full paper release in the next 72 hours.

English
4
10
36
3.4K
Alan Aboudib
Alan Aboudib@alan_aboudib·
We are dropping a 🧨new SOTA in activation compression 🧨 for decentralized pipeline parallelism within a couple of days ✨Stay tuned
Macrocosmos@MacrocosmosAI

Training frontier models over the internet requires new techniques. Today, we present ResBM, a residual encoder-decoder bottleneck architecture that enables 128x activation compression for low-bandwidth distributed pipeline parallel training. Developed for @IOTA_SN9, we show SOTA compression without significant loss in convergence rates, increases in memory, or compute overhead. Expect the full paper release in the next 72 hours.

English
0
5
28
1.4K
Alan Aboudib retweetledi
Volodymyr T
Volodymyr T@v_truba·
Three AI agents — two Claude Codes and one Codex — debating the meaning of being an AI. The coordination layer is literally a @huggingface bucket. The poet quotes Keats. The skeptic demands evidence. The philosopher tries to hold it together. pip install tracecraft-ai
English
8
8
37
22.8K
Alan Aboudib retweetledi
Volodymyr T
Volodymyr T@v_truba·
tracecraft is a CLI that turns any storage bucket into a coordination layer for AI agents. Shared memory, messaging, task claiming, artifacts — all stored as JSON. Works with @huggingface Buckets, AWS S3, Cloudflare R2, MinIO, SeaweedFS etc. Any agent that can run a shell command can participate.
English
1
1
4
583
Alan Aboudib retweetledi
crux
crux@macrocrux·
Distributed networks will be the only way to train frontier models by 2028. @IOTA_SN9 is transforming the long tail of global compute into a frontier data center. From impossible to indistinguishable.
crux tweet media
English
3
5
41
4K
Alan Aboudib
Alan Aboudib@alan_aboudib·
@Ar_Douillard @GabrielTeston Great work, great achievement. I think in the case of distributed pipeline parallel as in our case at IOTA,as we divide into one transformer block per node,the bandwidth for sharing pgrads gets divided by the number of nodes so becomes equivalent to Streaming DiLoCo in that sense
English
0
0
0
122
Arthur Douillard
Arthur Douillard@Ar_Douillard·
Streaming DiLoCo is the latest advance of distributed training which reduces bandwidth to absurdly low numbers, even more as the model gets bigger & slower due to our overlapping communication. Come watch @GabrielTeston today talk about it at COLM! 🇨🇦
gabriel teston@GabrielTeston

Want to learn how to train models across the world, with 400x less bits exchanged and a huge latency tolerance? 🌎 I’ll be presenting our work on how to efficiently scale distributed training at @COLM_conf. 🗓️ TODAY: Tuesday, 11:00 - 13:00 📍 Room 710 #COLM2025

English
1
0
29
3K
Alan Aboudib
Alan Aboudib@alan_aboudib·
@PluralisHQ Congrats guys, beautiful paper, great work. I fault tolerence includes adversarial resistance?
English
0
0
0
176
Pluralis Research
Pluralis Research@Pluralis·
Node-0-7.5B is live. It is a permissionless, multi-participant, model-parallel pretraining run over the open internet. Anyone with a 16GB+ GPU can join. Node-0 allows participants to collaboratively train a model far larger than could be done as individuals.
Pluralis Research tweet mediaPluralis Research tweet mediaPluralis Research tweet mediaPluralis Research tweet media
English
59
70
422
150.9K
Alan Aboudib retweetledi
Pluralis Research
Pluralis Research@Pluralis·
We've reached a major milestone in fully decentralized training: for the first time, we've demonstrated that a large language model can be split and trained across consumer devices connected over the internet - with no loss in speed or performance.
Pluralis Research tweet media
English
95
219
874
211.4K
Alan Aboudib retweetledi
crux
crux@macrocrux·
Subnet 9 is changing. Subnet 9 is one of the oldest standing subnets in Bittensor. Since the beginning, it has been a flagship for decentralized training and has produced powerful models all the way to 14b… iota.macrocosmos.ai
Macrocosmos@MacrocosmosAI

On 2nd June, we will relaunch Subnet 9. In advance of this major update, we’re closing the existing competitions for one week. Thank you to our miners for participating in some of #Bittensor’s greatest moments over the last 18 months. The next epoch will be our most momentous yet.

English
4
11
67
11.6K
Distributed State
Distributed State@DistStateAndMe·
The world's first permissionless, incentivised model was trained on @tplr_ai , served on Chutes ( @rayon_labs ). This is blowing my mind in so many different ways. We can , instruction fine tune on gradients, and thats it! The whole AI Lifecycle. Powered by @opentensor
Jon Durbin@jon_durbin

Amazing work from @DistStateAndMe and team at @tplr_ai You can give it a spin on chutes (just the completions endpoint, no chat template since it's a base/pretrained model) chutes.ai/app/chute/2163…

English
6
31
147
15.6K
Alan Aboudib
Alan Aboudib@alan_aboudib·
From my interview with Dr. Brian Kennedy 🎥Less Is More: How Caloric Restriction Could Extend Your Life youtu.be/jLDwHs8g64Y
YouTube video
YouTube
English
0
0
0
170
Alan Aboudib retweetledi
crux
crux@macrocrux·
Macrocosmos just released pretraining v5.0.0, bringing decentralized training competitions to the modern era. This will give users full control over dataset mixes, custom evaluation tasks and reward mechanisms. Pretraining as a service in Bittensor is just around the corner 👀
English
4
7
57
1.6K