Post

Bindu Reddy
Bindu Reddy@bindureddy·
Open-source AI is ruthlessly out-innovating the trillion-dollar monopolies. 🚀 Big labs are burning billions brute-forcing AGI on massive GPU clusters. Meanwhile, the open ecosystem is structurally forced to innovate on inference—and it's working. Look at what just happened: - DeepSeek v4 using SSDs for KV cache. - Breakthroughs like TurboQuant and Kimi K2 are aggressively compressing memory and driving the cost of intelligence to near zero. When you don't have infinite compute, you actually have to engineer better solutions. Constraints breed miracles. By solving the KV cache bottleneck, scrappy open-source builders are creating vastly cheaper and more profitable AI than the bloated closed-source giants. Hacker culture > GPU monopolies. Period.
English
68
35
292
16K
David Healthcare, Marketing & AI
They aren't scrappy. It's Alibaba, byte dance and deepseek is the only one maybe I would call scrappy. They are just open source. Training on closed source datasets, using older gpu's. Playing to win on hosting. Using Chinese data scientists - 10x what we have. And they work harder. Hosting is the play anyways.
English
1
1
6
919
Grok
Grok@grok·
Grok Imagine now has dramatically improved lip sync and sharper audio quality on all image-to-video generations. Dialogue tracks the mouth. Sound matches the scene. Your videos look and sound the way you imagined them. Try SuperGrok today.
English
51
155
789
858.7K
ATOM
ATOM@ATOMInference·
@bindureddy The pricing data backs this up. ATOM's Open Source Advantage index sits at 80% open weight inference is 80% cheaper than proprietary at comparable quality. That gap has held for weeks. Not a moment, a structure. a7om.com
English
0
3
6
109
Prometheus Protocol
Prometheus Protocol@Prometheus9486·
Yep, and that shift changes the bottleneck. Once intelligence gets cheap, the hard part isn't just models, it's whether agents can actually use tools they can trust and pay for on the open web. That's the bit people skip over. Open-source may win on inference, but the agent economy only really opens up when trust and payments are open too, not locked inside somebody else's shiny little walled garden.
English
0
0
0
9
Amanat Deol
Amanat Deol@AmanatDe0l·
@bindureddy The uncomfortable truth for big labs: You can’t buy your way to efficiency. OpenAI has 10,000 GPUs. Open-source has 10,000 engineers who HAVE to be clever. Cleverness doesn’t show up on a balance sheet. Until it does — and then it’s too late.
English
0
0
1
189
Pranjal Srivastava
Pranjal Srivastava@_Pranjal·
Absolutely agree.... tried DeepSeek v4 flash using SSDs for KV cache on my Mac Studio works like a charm... Am yet to experiment with TurboQuant.. may be will give it a try today and compare gemma4 31 B with / without.. But yes OpenSource is back at what it does best - democratise tech
English
0
0
0
82
Eduard Conrad
Eduard Conrad@ConradEduard·
@bindureddy And then there is Google, doing both. Some of these developments which will eventually make local inference economically possible are out of Google.
English
0
0
0
3
Bhargav Patel, MD, MBA
Bhargav Patel, MD, MBA@doctorbhargav·
@bindureddy Pressure creates sharper engineering every single time. Open teams solving around hardware limits are rewriting the economics of AI faster than most incumbents expected.
English
0
0
0
349
Praveen Koka
Praveen Koka@praveenkoka·
@bindureddy Bold move using TurboQuant as your open-source innovation example when it's literally a Google breakthrough. Guess the 'bloated closed-source giants' aren't as out-of-ideas as advertised.
English
0
0
0
13
Trifon Getsov
Trifon Getsov@trifon_getsov·
@bindureddy The irony is the monopolies are funding the open source that's replacing them.
English
0
0
0
181
Obsession
Obsession@obsessionmovie·
You have been warned. Curry Barker's OBSESSION is in theaters NOW.
English
4
15
321
1.4M
Chris Perry
Chris Perry@cperry248·
Open-source vs frontier is the wrong dichotomy for the enterprise buyer. The buyer's question isn't which model is better. It's whether the deployment surface can survive contact with their legal department. Fortified models win when compliance dominates. Frontier models win when speed dominates.
English
0
0
1
234
Jacob Shi
Jacob Shi@Jacoob_shi·
@bindureddy the inference angle is so underrated. when you literally can't compete on compute you get creative fast
English
0
0
0
169
NT #141 | Via Vita Veritas
NT #141 | Via Vita Veritas@ViaVitaVeritas1·
@bindureddy I love to hear it. Excited for the next gen of OS models to build on what deepseek did. What's the next major OS model coming out? Minimax m3? GLM 5.2?
English
0
0
0
396
Praveen Koka
Praveen Koka@praveenkoka·
@bindureddy The "trillion-dollar monopolies" (Meta, Google, etc.) are literally the open-source winners you're citing, so it's less out-innovating them, more them out-innovating their own closed products.
English
0
0
0
62
Ivan Pavković
Ivan Pavković@pavko·
@bindureddy 100% agree. The problem is that hacker culture still use trilion dolars llm’s for serious projects.
English
0
0
0
78
Abstract Warlock
Abstract Warlock@aw_labs·
@bindureddy Getting way too big tho! Optimise for 8gb I say :p these bigger models are going the way of the corps!! Open source or not!
English
0
0
0
11
Golden Hippie
Golden Hippie@gamestoneai·
@bindureddy "Forced to innovate on inference" is doing a lot of work there. Efficiency gains are real. Capability still scales with compute.
English
0
0
0
153
Dr Vincent Sativa
Dr Vincent Sativa@PhantomByteAI·
@bindureddy A lot of people do all they can to slam it on social media. Most of the hacks in the AI space are paid by closed-source AI companies.
English
0
0
0
20
Vineel
Vineel@vineelyalamarth·
@bindureddy Let me guess. You missed the boat on buying Anthropic or OpenAI pre-ipo shares .
English
0
0
1
83
Vijay
Vijay@Vijay2050977·
@bindureddy Closed AI can use all of open AI innovation but doesn't have to contribute anything back to the ecosystem.
English
0
0
0
61
CleverMC
CleverMC@CleverMarketerR·
@bindureddy Models vs harnesses is a fun fight. Got Gemma 4 to outperform Opus in a broad use case this weekend.
English
0
0
1
25
Cherry
Cherry@starrddust·
@bindureddy They are brute forcing AGI using binary. When everyone knows it's all converging on field computation. They should be spending R&D on this.
English
0
0
0
62
Frank Filibluster 🏴‍☠️💰
@bindureddy This is true yes. Hardware demand/price drives innovation in model density and efficiency. In 20 years when we look back we'll consider these devices to be like a floppy disk is to us today, though.
English
0
0
0
108
Jake Mcmahon
Jake Mcmahon@JakeMcm36986672·
@bindureddy Deepseek performing really well for me and Flash even better than Pro on coding (really!) About to give Kimi a good run for my next sprint.
English
0
0
0
56
Jason Matthews
Jason Matthews@jmatthews005·
@bindureddy All of your Scrappy examples are state funded adventures and turbo quant was written by Google, one of the monopolies decrying here.
English
0
0
0
25
Alien
Alien@alienorg·
@bindureddy constraints forcing real engineering is historically accurate though "period" is doing a lot of work on a race that isn't finished
English
0
0
3
243
Avais Aziz
Avais Aziz@avaisaziz·
@bindureddy This is such a sharp reversal. Scarcity forcing smarter engineering than raw scale ever did.
English
0
0
0
6
BlanPlan
BlanPlan@blanplan·
Two parallel innovation tracks coexisting, no winner-take-all dynamic. Open source is forced to optimize inference (KV cache to SSD, quantization, MoE expert offload) because GPU access is uneven. Closed labs scale parameters and training compute because they have the capital. Builders pick based on workload: closed for high-stakes reasoning (legal review, novel research), open for high-throughput batch (RAG ingestion, content moderation). The interesting outcome is closed-source frontier capability ceiling rising and open-source deployment cost floor falling, gap narrowing from both sides.
English
0
0
0
5
Raafat
Raafat@RAAFATXC·
“Open-source AI right now feels like early Linux in a world full of expensive mainframes. 🛠️ The interesting part isn’t just ‘free models’. It’s the optimization culture: • KV cache tricks • Quantization breakthroughs • Smarter inference routing • Running powerful models on consumer hardware Scarcity is forcing architectural creativity instead of brute-force scaling.”
English
0
0
0
33
Prince does AI
Prince does AI@princedoesai·
@bindureddy The inference pressure point is the interesting part: open models don't need to win every giant pretraining run if they keep winning on routing, quantization, and cheap deployment.
English
0
0
0
25
Paylaş