bidhan @ NVIDIA GTC

4.4K posts

bidhan @ NVIDIA GTC banner
bidhan @ NVIDIA GTC

bidhan @ NVIDIA GTC

@bidhan

ceo @bageldotcom. previously code monkey at amazon, cashapp, instacart. hiring - dms are open.

Katılım Ocak 2021
1K Takip Edilen4K Takipçiler
Sabitlenmiş Tweet
bidhan @ NVIDIA GTC
bidhan @ NVIDIA GTC@bidhan·
Being at the frontier - by the definition of it - means creating the frontier. You don't get to be at the frontier by following someone else. And creating the frontier often means discoveries that go against the established knowledge. We recently made such a discovery about distributed diffusion model training. A common way to optimize diffusion model training is by ensuring the numerical stability of their generation paths. We found that that's not true for the most efficient distributed diffusion model training architecture. We shared what works instead in our blogpost below. blog.bagel.com/p/stability-qu…
bidhan @ NVIDIA GTC tweet media
English
13
32
457
50.2K
bidhan @ NVIDIA GTC
what if told you training coding model to code python is optimizing for a local maxima
English
0
0
4
164
bidhan @ NVIDIA GTC
the models that were traditionally AR (llms) are becoming diffusion, not the other way around
David@DavidSHolz

@Angaisb_ almost 100 percent of image and video models are still diffusion, you're just confused, sorry!

English
0
0
9
207
Philip Kiely
Philip Kiely@philipkiely·
Giving away signed copies of Inference Engineering at NVIDIA GTC — stop by booth 931!
Philip Kiely tweet media
English
9
3
109
4K
bidhan @ NVIDIA GTC retweetledi
Dylan Patel
Dylan Patel@dylan522p·
Deepseek v4 still not released Alibaba Qwen going closed Western open weights models slacking In these dark times for open source, who will save us? Alliances must be made, brothers must band together! A world of only closed source AI will lead to consolidation of power! Tyranny!
English
110
56
1.2K
137.3K
clem 🤗
clem 🤗@ClementDelangue·
They tell you training and running AI model costs billions. That's true for a few frontier labs. But for most real-world use cases? Dramatically lower than you think thanks to open-source. Real examples from @HuggingFace's latest analysis: - Fine-tune a text classification model: <$2k - Train a leading image embedding model: < $7k - Train Deepseek OCR: < $100k - Train a leading machine translation model: <$500k Compare that to GPT-4.5 training (~$300M est.) And the truth is that you don't need a Formula 1 car to pick up groceries. Most tasks are solved just as well by smaller, efficient, targeted models. The mistake everyone makes? Starting with "what's the best AI model?" instead of "what do I need to do?" The future of AI is not just bigger models. It's cheaper, more customized, open models solving specific problems. Explore 100+ real model costs of training and deployment yourself in the study!
clem 🤗 tweet media
English
49
73
611
39.6K
bidhan @ NVIDIA GTC retweetledi
bagel.com
bagel.com@bageldotcom·
In town for NVIDIA GTC? If you're building generative world models or investing in the people who are - we're putting the right people in one room for you tomorrow night in Palo Alto. Co-hosted by Alumni Ventures. Signup link below.
English
2
7
17
9.2K
bidhan @ NVIDIA GTC
We've spent the last year figuring out how to train frontier diffusion models without a GPU supercluster. Turns out a lot of people training generative models care about the same problem. Come to our event tomorrow to meet them luma.com/nvidia-gtc-gen…
bagel.com@bageldotcom

In town for NVIDIA GTC? If you're building generative world models or investing in the people who are - we're putting the right people in one room for you tomorrow night in Palo Alto. Co-hosted by Alumni Ventures. Signup link below.

English
1
1
10
546
bidhan @ NVIDIA GTC
all the startups in my network that had good funding and seemed to have a lot of potential couple of years back, but for mysterious reasons didn’t work out are fully remote
English
0
0
12
351
bidhan @ NVIDIA GTC retweetledi
roberto
roberto@rcadxwn·
@bidhan Killing it! Thanks for open-sourcing it all
English
1
0
1
48
bidhan @ NVIDIA GTC
bidhan @ NVIDIA GTC@bidhan·
Excited to share that Bagel Labs' paper got accepted at CVPR 2026. A lot of the most important diffusion model research has historically stayed inside frontier labs. We're bringing more of that in the open through open science and open infrastructure. In this work we showcase the very counterintuitive advantage of mixing different training objectives (DDPM and Flow-Matching) through an ensemble of diffusion models. This is one of the first ever works to successfully combine diffusion models trained with heterogeneous objectives. See details here: blog.bagel.com/p/heterogeneou…
bidhan @ NVIDIA GTC tweet media
English
4
7
21
2.8K
bidhan @ NVIDIA GTC
bidhan @ NVIDIA GTC@bidhan·
with the current ai research neolabs boom, no self respecting researcher should be working for a bank
English
2
0
13
858