Openτensor Foundaτion

1.9K posts

Openτensor Foundaτion banner
Openτensor Foundaτion

Openτensor Foundaτion

@opentensor

Incentivizing intelligence

Katılım Haziran 2021
1 Takip Edilen170.3K Takipçiler
Sabitlenmiş Tweet
Openτensor Foundaτion
The largest decentralised LLM pre-training run in history. SN3 @tplr_ai trained Covenant-72B across 70+ contributors on open internet infrastructure. Now it’s being discussed by @chamath with @nvidia CEO Jensen Huang. Distributed, open-weight model training on Bittensor is getting started.
English
41
260
1.1K
46.2K
Openτensor Foundaτion retweetledi
templar
templar@tplr_ai·
On the @theallinpod this week, @chamath asked @nvidia CEO Jensen Huang about decentralized AI training, calling our Covenant-72B run "a pretty crazy technical accomplishment." One correction: it's 72 billion parameters, not four. Trained permissionlessly across 70+ contributors on commodity internet. The largest model ever pre-trained on fully decentralized infrastructure. Jensen's answer is worth hearing too.
English
65
304
1.3K
234.4K
Openτensor Foundaτion retweetledi
Vidaio
Vidaio@vidaio_·
"Please fasten your seatbelts and ensure your seatback and tray tables are in their full upright position." From groundbreaking tools... To a complete intelligent ecosystem... "LADIES AND GENTLEMEN, PREPARE FOR TAKEOFF" Introducing: VidaioOS The next dimension in enterprise video management. $TAO @vidaio_
English
21
33
136
12.1K
Openτensor Foundaτion retweetledi
Apex・SN1
Apex・SN1@Apex_SN1·
We’ve built a simulation of the @IOTA_SN9 communication network. This is a high-fidelity digital twin - an abstracted version of our distributed training architecture, designed as a testing ground to run experiments and develop novel algorithms to increase the speed and quality of model training. We’re using the simulator as an environment for open competitions on Apex, outsourcing algorithmic innovations to the Bittensor miners. It’s the first time our simulator can be interacted with by the public. Our opening simulator competition is live now.
English
2
8
55
5.1K
Openτensor Foundaτion retweetledi
METANOVA
METANOVA@metanova_labs·
ArboNOVA: Patent–Molecule matching loop We’ve been experimenting with an agent that maps molecules → prior art using only open data + tools Benchmark: ~1500 molecules across ADHD-related patents (since 2012) In ~12 hours: 18 iterations of the loop → Best hit rate: 85.4% How this is usually done: Pharma intelligence teams + expensive proprietary databases + manual workflows + even conference attendance Early, but promising. Moving one step closer toward automating drug discovery and identifying which molecules are most strategic to advance in the wet lab. Based on @const_reborn (github.com/unconst/Arbos) and @karpathy autoresearch framework #Bittensor #SN68 #ralphloop #agents #DrugDiscovery #Desci #DeAI
METANOVA tweet media
English
9
28
111
20.5K
Openτensor Foundaτion retweetledi
grail
grail@grail_ai·
PULSE made weight sync 100x faster. That turned the trainer itself into the bottleneck. @erfan_mhi just fixed that too. Grail's GRPO trainer is now 1.8x faster on a single B200: 27% to 47% MFU, epoch time nearly halved. Decentralized post-training is converging on centralized speed.
Erfan Miahi@erfan_mhi

Used autoresearch to make @grail_ai GRPO trainer 1.8x faster on a single B200. I kept postponing this for weeks since the bottleneck in our decentralized framework was mainly communication. But after our proposed technique, PULSE, made weight sync 100x faster, the training update itself became the bottleneck. Even with a fully async trainer and inference, a slow trainer kills convergence speed. A task that could've eaten days of my time ran in parallel while I worked on other stuff. Unlike original autoresearch, where each experiment is 5 min, our feedback loop is way longer (10-17 min per epoch + 10-60 minutes of installations and code changes), so I did minimal steering when it was heading in bad directions to avoid burning GPU hours. The agent tried so many things that failed. But, eventually found the wins: Liger kernel, sequence packing, token-budget dynamic batching, and native FA4 via AttentionInterface. 27% to 47% MFU. 16.7 min to 9.2 min per epoch. If you wanna dig deeper or contribute: github.com/tplr-ai/grail We're optimizing everything at the scale of global nodes to make decentralized post-training as fast as centralized ones. Stay tuned for some cool models coming out of this effort. Cheers!

English
0
10
42
6.9K
Openτensor Foundaτion retweetledi
Apex・SN1
Apex・SN1@Apex_SN1·
Apex and @IOTA_SN9 are working together again. The IOTA simulator competition launches later today. Join us as we accelerate distributed training.
English
7
21
112
9.9K
Openτensor Foundaτion retweetledi
const
const@const_reborn·
What if you could create an auto-research where your agent just focused on the eval and it was designed so that others could have swarms of agents across the web try to solve it and you paid them based on the ownership of the mechanism which produced the research
const tweet media
English
11
27
231
15.2K
Openτensor Foundaτion retweetledi
Tyler DurdΞth
Tyler DurdΞth@tylerdurdeth·
The thing with Bittensor is that subnets keep pushing the boundaries regardless of market volatility. Targon got accepted to the Nvidia accelerator. Nova made new breakthroughs in drug discovery. $TAO never sleeps
Tyler DurdΞth tweet mediaTyler DurdΞth tweet media
English
5
24
136
5.3K
Openτensor Foundaτion retweetledi
Distributed State
Distributed State@DistStateAndMe·
When you fix one bottleneck, the next one becomes visible. At @covenant_ai we built PULSE (arxiv.org/abs/2602.03839) to make weight sync 100× faster. That worked. Then the trainer itself became the new ceiling. So @erfan_mhi ran autoresearch on our GRPO trainer. 27% → 47% MFU. 16.7 min → 9.2 min per epoch. 1.8× faster on a single B200. Decentralized post-training, closing the gap with centralized. github.com/tplr-ai/grail
Erfan Miahi@erfan_mhi

Used autoresearch to make @grail_ai GRPO trainer 1.8x faster on a single B200. I kept postponing this for weeks since the bottleneck in our decentralized framework was mainly communication. But after our proposed technique, PULSE, made weight sync 100x faster, the training update itself became the bottleneck. Even with a fully async trainer and inference, a slow trainer kills convergence speed. A task that could've eaten days of my time ran in parallel while I worked on other stuff. Unlike original autoresearch, where each experiment is 5 min, our feedback loop is way longer (10-17 min per epoch + 10-60 minutes of installations and code changes), so I did minimal steering when it was heading in bad directions to avoid burning GPU hours. The agent tried so many things that failed. But, eventually found the wins: Liger kernel, sequence packing, token-budget dynamic batching, and native FA4 via AttentionInterface. 27% to 47% MFU. 16.7 min to 9.2 min per epoch. If you wanna dig deeper or contribute: github.com/tplr-ai/grail We're optimizing everything at the scale of global nodes to make decentralized post-training as fast as centralized ones. Stay tuned for some cool models coming out of this effort. Cheers!

English
4
16
102
6.7K
Openτensor Foundaτion retweetledi
Targon
Targon@TargonCompute·
Today we are excited to share some news, Targon has been accepted into the @nvidia Inception program for startups! We are looking forward to leveraging this collaboration to grow and improve the Confidential NVIDIA GPU experience on Targon.com
Targon tweet media
English
27
102
482
36.6K
Openτensor Foundaτion retweetledi
Chutes
Chutes@chutes_ai·
Vocence and Chutes are working together. @vocence_bt (SN102) is building decentralized voice AI, and deploying PromptTTS models as private chutes on our infrastructure. Voice is the next frontier for open source AI. Now it runs on decentralized compute. SN102 x SN64.
Chutes tweet media
English
8
25
151
7.1K
Openτensor Foundaτion retweetledi
Mentat Minds
Mentat Minds@mentatminds·
No market in the world is currently more exciting than the subnet market. $TAO
English
4
7
91
4.5K