Oleksii Kuchaiev

869 posts

Oleksii Kuchaiev banner
Oleksii Kuchaiev

Oleksii Kuchaiev

@kuchaev

Sr. Director of AI model post-training @NVIDIA

in the cloud Katılım Şubat 2010
1.2K Takip Edilen3.6K Takipçiler
Sabitlenmiş Tweet
Oleksii Kuchaiev
Oleksii Kuchaiev@kuchaev·
1/4 We see no wall in post-training. Scaling RL software, infra, and data keeps yielding major capability gains. We trained across 30 RL environments with up to 4,000 instances per batch — math, code, STEM, agentic tool use, SWE, terminal, safety — all in a unified multi-environment RLVR setup.
Oleksii Kuchaiev tweet media
English
4
34
284
77K
Oleksii Kuchaiev retweetledi
clem 🤗
clem 🤗@ClementDelangue·
Nvidia just crossed Google as the biggest org on @huggingface with 3,881 team members on the hub. I'm officially calling it: Nvidia is the new American king of open-source AI!
clem 🤗 tweet media
English
36
72
627
81.7K
Oleksii Kuchaiev retweetledi
Tinker
Tinker@tinkerapi·
Nemotron 3 Nano and Super from @nvidia are now available on Tinker! We're offering a limited-time GTC 50% discount for both. The Nemotron family features open hybrid MoE models optimized for compute efficiency for agentic applications.
English
4
17
111
13.9K
Oleksii Kuchaiev retweetledi
The Wall Street Journal
Breaking: Russia is secretly providing satellite imagery and sharing drone technology to help Iran target U.S. forces in the region on.wsj.com/478OhaN
English
2K
4.6K
11.2K
947.9K
Oleksii Kuchaiev
Oleksii Kuchaiev@kuchaev·
If you are attending GTC 2026, stop by tomorrow, Tuesday, March 17  at 3:00 p.m. - 3:40 p.m. PDT to learn about how we post-train Nemotrons. Looking forward to seeing everyone! Session number: S81558
Oleksii Kuchaiev tweet media
English
0
4
19
859
Oleksii Kuchaiev
Oleksii Kuchaiev@kuchaev·
An angry suicidal teenager is not going to be able to bypass frontier models' gurdrails. And even if he does, no lab is going to synthesize it or even ship components to the same place. A bans on this exists since at least 1975 (see Asilomar Conference banning certain recombinant DNA experiments).
English
1
0
7
601
Oleksii Kuchaiev
Oleksii Kuchaiev@kuchaev·
@natolambert Adoption-wise, I agree, closed will be far ahead, especially on a consumer side. But in terms of capabilities, I do not think the gap will persist for long. Will be very much like closed OS (iOS, Mac, Windows for consumers) vs Linux/Unix (which rules the Internet).
English
0
0
3
360
Nathan Lambert
Nathan Lambert@natolambert·
World will converge on 3 types of models 1. Closed frontier (Ant, OAI, Gemini) 2. Open frontier (2-3 labs, much consolidation coming) 3. Open small / tool (fairly empty now) The open frontier will be far from the closed frontier, but way cheaper. Other statements are cope.
English
43
31
484
36.2K
Oleksii Kuchaiev retweetledi
Perplexity
Perplexity@perplexity_ai·
NVIDIA’s Nemotron 3 Super is now available in Perplexity, Agent API, and Computer.
Perplexity tweet media
English
107
124
1.8K
189.5K
Oleksii Kuchaiev retweetledi
Unsloth AI
Unsloth AI@UnslothAI·
We collaborated with @NVIDIA to teach you about Reinforcement Learning and RL environments. Learn: • Why RL environments matter + how to build them • When RL is better than SFT • GRPO and RL best practices • How verifiable rewards and RLVR work Blog: unsloth.ai/blog/rl-enviro…
Unsloth AI tweet media
English
26
249
1.7K
85.9K
Adam Schiff
Adam Schiff@SenAdamSchiff·
There are few people in the world who have derived more benefit from our war with Iran than the Russian dictator, Vladimir Putin. We explain why:
English
473
632
2.2K
54.6K
Oleksii Kuchaiev retweetledi
NVIDIA AI Developer
NVIDIA AI Developer@NVIDIAAIDev·
🦞These innovations come together to create a model that is well suited for long-running autonomous agents. On PinchBench—a benchmark for evaluating LLMs as @OpenClaw coding agents—Nemotron 3 Super scores 85.6% across the full test suite, making it the best open model in its class.
NVIDIA AI Developer tweet media
English
19
47
311
188.9K
Oleksii Kuchaiev
Oleksii Kuchaiev@kuchaev·
4/4. I'll go deep on the post-training methodology at GTC 2026 Conference — session S81558, Tuesday March 17, San Jose. Will cover RL scaling across environments, what worked in agentic training, and where we think this is headed.
English
1
1
54
2.4K
Oleksii Kuchaiev
Oleksii Kuchaiev@kuchaev·
Nemotron 3 Super is here — 120B total / 12B active, Hybrid SSM Latent MoE, designed for Blackwell. Truly open: permissive license, open data, open training infra. See analysis on @ArtificialAnlys Details in thread 🧵below:
Oleksii Kuchaiev tweet mediaOleksii Kuchaiev tweet media
English
10
45
275
28.6K