

Alan Aboudib
118 posts

@alan_aboudib
IOTA AI Research Lead @macrocosmosai — Intelligence Podcast @YouTube



Training frontier models over the internet requires new techniques. Today, we present ResBM, a residual encoder-decoder bottleneck architecture that enables 128x activation compression for low-bandwidth distributed pipeline parallel training. Developed for @IOTA_SN9, we show SOTA compression without significant loss in convergence rates, increases in memory, or compute overhead. Expect the full paper release in the next 72 hours.

Training frontier models over the internet requires new techniques. Today, we present ResBM, a residual encoder-decoder bottleneck architecture that enables 128x activation compression for low-bandwidth distributed pipeline parallel training. Developed for @IOTA_SN9, we show SOTA compression without significant loss in convergence rates, increases in memory, or compute overhead. Expect the full paper release in the next 72 hours.




Want to learn how to train models across the world, with 400x less bits exchanged and a huge latency tolerance? 🌎 I’ll be presenting our work on how to efficiently scale distributed training at @COLM_conf. 🗓️ TODAY: Tuesday, 11:00 - 13:00 📍 Room 710 #COLM2025


Iota launched a week ago. First subnet launch in a over a year, and man, it’s fun. Data- and pipeline-parallelism with incentives is the hardest problem I’ve ever worked on. It’s our moonshot, and we are all in. Over 30 updates to the codebase in a week: relentlessly iterating every part of the design, from weight merging improvements to frenzied compression experiments and ablations, we have moved non-stop as a community. Woke up this morning and seeing that our work is paying off — weight merging is now stabilized for a 15B model split into 5. This feeling is gold. Can’t wait for Novelty Search tomorrow with @WSquires @const_reborn and @shibshib89


Subnet 9 proved decentralized LLM pretraining is viable. We are proud to release the technical primer for IOTA in advance of mainnet launch on June 2nd. IOTA comprises a series of key innovations: - Data- and Pipeline-parallel SWARM execution across heterogeneous and unreliable nodes. - 128× activation compression for home-grade links - CLASP: Contribution Loss Assessment via Sampling of Pathways - Butterfly All-Reduce for O(1) sync bandwidth Together, we believe they can combine to push the field of permissionless, performant decentralised training measurably forward.

IOTA (Incentivized Orchestration Training Architecture) is a framework for pretraining large language models across a network of heterogeneous, unreliable, permissionless and token incentivized machines. In our technical primer we report the following advances: Incentivized Data- and Pipeline-parallel training across heterogeneous and unreliable nodes 128× activation compression to enable training on memory-limited hardware CLASP: Contribution Loss Assessment via Sampling of Pathways Butterfly All-Reduce for O(1) sync bandwidth


On 2nd June, we will relaunch Subnet 9. In advance of this major update, we’re closing the existing competitions for one week. Thank you to our miners for participating in some of #Bittensor’s greatest moments over the last 18 months. The next epoch will be our most momentous yet.


Amazing work from @DistStateAndMe and team at @tplr_ai You can give it a spin on chutes (just the completions endpoint, no chat template since it's a base/pretrained model) chutes.ai/app/chute/2163…

