Rishabh Tiwari

87 posts

Rishabh Tiwari banner
Rishabh Tiwari

Rishabh Tiwari

@rish2k1

CS PhD @UCBerkeley | Ex-Deepmind, FAIR | Research area: Efficient LLM reasoning, scaling RL

Berkeley, CA เข้าร่วม Mayıs 2019
417 กำลังติดตาม839 ผู้ติดตาม
ทวีตที่ปักหมุด
Rishabh Tiwari
Rishabh Tiwari@rish2k1·
Post-training is no longer a final polishing step, it has become a central scaling dimension for modern LLMs. Understanding its science is essential to unlocking the full potential of foundation models. We’re excited to be organizing SPOT Workshop at @iclr_conf , with an outstanding lineup of speakers spanning academia and industry. 📢 Call for papers is open — deadline Feb 5, 2026 (AoE). Please follow workshop X page (@spoticlr) for more updates.
SPOT Workshop at ICLR 2026 🇧🇷@spoticlr

📢 Announcing 𝗦𝗣𝗢𝗧: 𝗦𝗰𝗮𝗹𝗶𝗻𝗴 𝗣𝗼𝘀𝘁-𝗧𝗿𝗮𝗶𝗻𝗶𝗻𝗴 𝗳𝗼𝗿 𝗟𝗟𝗠𝘀 Workshop at #ICLR2026 (@iclr_conf )🚀 🚨 We invite work on the principles of post-training scaling, bridging algorithms, data & systems 📅 Feb 5, 2026 for papers 🌐 spoticlr.github.io 🧵(1/5)

English
1
3
27
3.1K
Rishabh Tiwari รีทวีตแล้ว
Mert Cemri
Mert Cemri@mertcemri·
AlphaEvolve proved LLMs can discover novel algorithms, but it remains closed-source, and open-source alternatives (OpenEvolve, GEPA) rely on rigid, static search policies. Introducing AdaEvolve: a fully adaptive evolutionary algorithm that dynamically adjusts its own search strategy based on observed progress. It matches or beats AlphaEvolve and best known Human SOTA on math and systems benchmarks, and boosts Frontier-CS median scores by 33% over the best open-source baseline across 185 tasks. 🧵👇 (1/n)
Mert Cemri tweet media
English
10
68
338
74.2K
Rishabh Tiwari รีทวีตแล้ว
Mert Cemri
Mert Cemri@mertcemri·
Great work with Frontier-CS! This benchmark has hundreds of quite challenging algorithmic tasks where even the best algorithms developed at SkyDiscover (AdaEvolve and EvoX) have a lot of headroom to improve. Thinking you can design a better evolutionary system? Go and start building your own algorithm at SkyDiscover github.com/skydiscover-ai…
Mert Cemri tweet media
Hanchen Li@lihanc02

x.com/i/article/2031…

English
1
8
15
907
Rishabh Tiwari รีทวีตแล้ว
Dvij Kalaria
Dvij Kalaria@DvijKalaria·
My 🏓 robot can now learn to serve just by watching me :) Project release coming soon!
English
27
37
430
50.4K
Rishabh Tiwari รีทวีตแล้ว
Rishabh Tiwari รีทวีตแล้ว
Mert Cemri
Mert Cemri@mertcemri·
Introducing SPECS (SPECulative test time Scaling), a test-time scaling (TTS) algorithm with pareto-frontier latency/accuracy trade-off. Scaling test-time compute improves LLM reasoning but imposes a latency overhead. Prior work optimizes TTS accuracy as a function of FLOPS, we propose to further reduce latency by addressing the memory bottleneck of LLM inference through speculative drafts. See a breakdown of the method below. (1/n) 🧵 👇
Mert Cemri tweet media
English
6
24
107
21.3K
Rishabh Tiwari รีทวีตแล้ว
Sai Surya Duvvuri
Sai Surya Duvvuri@dvsaisurya·
Excited to share LUCID — a new attention mechanism that improves retrieval and reasoning in long-context LLMs! [1/9]🧵 Here's how it works:
Sai Surya Duvvuri tweet media
English
12
69
437
123.7K
Rishabh Tiwari รีทวีตแล้ว
Lakshya A Agrawal
Lakshya A Agrawal@LakshyAAAgrawal·
Excited to release @gepa_ai's optimize_anything: a universal API for optimizing any text parameter. It consistently matches or outperforms domain-specific tools optimizing code, prompts, agent harnesses, cloud policies, even visuals! If you can measure it, you can optimize it.
Lakshya A Agrawal tweet media
English
22
96
519
120.5K
Rishabh Tiwari รีทวีตแล้ว
Mayank Mishra
Mayank Mishra@MayankMish98·
We cooked!🚀🚀🚀 Releasing SonicMoE: a fast MoE implementation for NVIDIA H100s GPUs. Special thanks to my collaborators: @WentaoGuo7 @XinleC295 @istoica05 and @tri_dao from whom I learnt a lot!
Wentao Guo@WentaoGuo7

🚀SonicMoE🚀: a blazingly-fast MoE implementation optimized for NVIDIA Hopper GPUs. SonicMoE reduces activation memory by 45% and is 1.86x faster on H100 than previous SOTA😃 Paper: arxiv.org/abs/2512.14080 Work with @MayankMish98, @XinleC295, @istoica05, @tri_dao

English
7
23
168
45.8K
Rishabh Tiwari รีทวีตแล้ว
Rachit Bansal
Rachit Bansal@rach_it_·
Current LLMs support contexts with millions of tokens. However, we keep seeing failure modes due to poor long-context reasoning. Our new work shows that, for long contexts, we must perform test-time training updates rather than vanilla ICL or “thinking”! w/ @Meta & @KempnerInst
Rachit Bansal tweet media
English
17
85
472
34.7K
Rishabh Tiwari
Rishabh Tiwari@rish2k1·
I’m in San Diego attending @NeurIPSConf this week and would love to meet and chat about efficient reasoning, RL, and scaling.
English
2
1
31
2.7K
François Fleuret
François Fleuret@francoisfleuret·
NeurIPS registration line is something like 250m
English
8
2
95
22.1K
Rishabh Tiwari รีทวีตแล้ว
Divy Thakkar
Divy Thakkar@divy93t·
Thrilled to share a professional update -- I'm starting a new effort at Google DeepMind! I'll be exploring / building the next paradigm of human-A(G)I collaboration and the future of interaction with advanced models. Grateful to be spending time with incredible collaborators to figure this out! If you're building / researching / investing in this space - let's chat, DMs open (also at NeurIPS in SD).
English
33
12
476
37.9K
Rishabh Tiwari รีทวีตแล้ว
Harman Singh
Harman Singh@Harman26Singh·
Late life update 🚀 I started my PhD at @UCBerkeley after an incredible time at @GoogleDeepMind. It was exciting to work on Gemini over the past couple of years. These days I am interested in reasoning/improving RL, agents, and diffusion language models. Looking forward to contributing to open science. Also thrilled to be back in the Bay Area. Grateful to mentors, collaborators, and folks who supported me, @partha_p_t @PengchuanZ @nitish_gup @trevorcohn @xiangrenNLP @divy93t @ManishGuptaMG1, Parag Singla, friends, and family. Excited to be at @NeurIPSConf #NeurIPS2025 this week. Looking forward to meeting folks. Feel free to DM if you'd like to chat!
Harman Singh tweet mediaHarman Singh tweet media
English
23
11
617
45.9K