Haresh Karnan

114 posts

Haresh Karnan banner
Haresh Karnan

Haresh Karnan

@KarnanHaresh

RL post-training @ Amazon AGI | PhD-UTAustin

San Francisco, CA Katılım Şubat 2023
1.1K Takip Edilen264 Takipçiler
hallerite
hallerite@hallerite·
@emaadmanzoor verl is a pain to work with and verifiers is an environments framework first and foremost
English
2
0
0
131
hallerite
hallerite@hallerite·
I see the need for 2 different kinds of LLM-RL frameworks. One should be optimized for post-training of LLMs by labs that want to build frontier models. It doesn't have to be feature-complete, but should follow best practices and consolidate what we learn over time.
English
1
2
21
2.5K
Haresh Karnan
Haresh Karnan@KarnanHaresh·
@ThereseMaggie @_arohan_ I meant the IdlyExpress / Mylapore chain of restaurants (you’ll find them in South Bay). I remember reading somewhere they have an idlyexpress branch open in SF recently
English
0
0
0
28
Therese Maggie
Therese Maggie@ThereseMaggie·
@_arohan_ the white space for delicious South Indian food in SF massive. It’s a shame that whole of South Indian food is reduced to just the humble dosa. The only other restaurant that serves lesser known South Indian food is Copra, but that’s mostly experimental, although delicious.
English
2
0
2
215
Haresh Karnan retweetledi
Vana AI
Vana AI@Vana_Meeting_AI·
The Cost of Ambiguity: There’s a massive, painful gap between knowing a mental model and applying it exactly when the conversation needs it most. Every Founder, PM, and Exec we spoke to knows this pain. And the result: Meetings dissolving into opinion, or worse, ending with a vague "let’s revisit this next week". 😩 Introducing 🥁 Vana AI. It's not a note-taker. It's not an action-item generator. It’s your proactive AI strategic thinker for real-time insights in meetings.
English
5
11
27
15.6K
Devvrit
Devvrit@Devvrit_Khatri·
Wish to build scaling laws for RL but not sure how to scale? Or what scales? Or would RL even scale predictably? We introduce: The Art of Scaling Reinforcement Learning Compute for LLMs
Devvrit tweet media
English
9
102
553
288.5K
Jim Bohnslav
Jim Bohnslav@jbohnslav·
just wasted an hour of my time interviewing some obvious cluely-using motherfucker. do I have to conduct all interviews blindfolded? mirrored glasses?
English
4
0
9
2K
Gowthami
Gowthami@gowthami_s·
Why is there a weird distinction in post-training, like SFT first and then RL? Why can't they be done together? IMO, I see merit in interleaving these paradigms. Is there any research pointing to the contrary or supporting this sequential process?
English
5
0
22
3.5K
Armen Aghajanyan
Armen Aghajanyan@ArmenAgha·
We've internally built a streaming solution to stream 1PB of multi-modal data over hundreds of GPU's for weeks without ever touching NFS, with no train performance regression. Enough interest for a blog post?
Orr Zohar@orr_zohar

🚨Huge for multimodal/vision AI: Datasets hit 100s of TB, making on-prem storage a nightmare. 🤗Now stream them directly from Hugging Face to GPUs - unlocking scalable training of everything from vlms to world models. 🚀 I've battled storage limits for years; thrilled to move on.

English
29
16
427
64.7K
Tanishq Mathew Abraham, Ph.D.
Tanishq Mathew Abraham, Ph.D.@iScienceLuvr·
Did they beat Tesla to the Tesla strategy? Lol Basically, they are deploying "self-driving" humanoids to consumers, it might not initially work that well and needs a lot of human supervision, but the data collected will continue to improve the robot and get it closer to "full self driving". I would note that humanoids in a home is a much more challenging environment than cars on the road: everyone lives differently in a variety of different floorplans and environments while roads have more consistency. So I do wonder how long it's gonna take for humanoids to reach near full autonomy. That said, like Teslas, Neo could be an appreciating asset if it continues to get better and more autonomous. Anyway that would be the hope, right? Let's see what happens in practice. I'm cautiously optimistic! I wonder when Tesla Optimus will be available 🤔
1X@1x_tech

NEO The Home Robot Order Today

English
54
11
303
46.2K
Jared Palmer
Jared Palmer@jaredpalmer·
Your @GitHub Universe badge comes with a hackable Raspberry Pi built into it with full color display, 5 buttons, USB-C, Bluetooth, and WiFi. Built by @pimoroni
English
105
315
6.7K
744.1K
Nathan Lambert
Nathan Lambert@natolambert·
Life update, she said yes. 🤩👩‍❤️‍👨🐕‍🦺
Nathan Lambert tweet media
English
277
16
3.4K
178.1K
Nathan Lambert
Nathan Lambert@natolambert·
Retiring my original airpods pro after over 5 years of use. Ridiculous amount of value really when we're used to 2 year product cycles.
English
9
30
158
16.9K
Siva Reddy
Siva Reddy@sivareddyg·
Lot of insights in @YejinChoinka's talk on RL training. Rip for next token prediction training (NTP) and welcome to Reinforcement Learning Pretraining (RLP). #COLM2025 No place to even stand in the room.
Siva Reddy tweet media
English
7
22
294
77.4K
will brown
will brown@willccbb·
the sampling is fine, the biggest red flag is that they need 12 GPUs for backprop and don’t do any microbatch packing / grad accum to accommodate for async/server on 16 GPUs i’d want 2-4 for training + 12-14 for inference you can do a loooot of backward passes in the time it takes to generate 512 rollouts
English
3
1
21
2K
will brown
will brown@willccbb·
"veRL is the best RL framework it's super efficient" really. are you sure about that. are you sure that you need 16 GPUs to tune a 7B model at 8k context. do you think that it's reasonable each step takes 19 minutes for this
will brown tweet media
English
23
26
345
61.6K
Jing Yu Koh
Jing Yu Koh@kohjingyu·
My toxic trait is thinking I can train a frontier model with 8 H100s
English
51
49
1.1K
92K
Qwen
Qwen@Alibaba_Qwen·
Ready to meet the biggest, brainiest guy in the Qwen3 family?
English
431
303
5.4K
907.6K
will brown
will brown@willccbb·
wow. thanks
will brown tweet media
English
26
5
312
19.8K
Kyle🤖🚀🦭
Kyle🤖🚀🦭@KyleMorgenstein·
despite living here 5 years this is somehow my first summer in Austin. I was warned it would be brutal but honestly it hasn’t been that bad. the key, imo, is to spend as much time outside as possible. I run or row most days. ofc if you spend all day with AC blasting you’ll melt!
English
3
0
24
1.7K
Jim Bohnslav
Jim Bohnslav@jbohnslav·
> be me, training vlms > use hf transformers because who wants to reimplement models if they don't have to > low MFU > `pytorch_profiler.py` > ViT takes 40X longer than the LLM > sus > .item() in the forward pass makes cudaStreamSynchronize every attention layer > MFW
GIF
English
24
11
556
39K