Vim

25.2K posts

Vim banner
Vim

Vim

@vim_dzl

Presidential Fitness Test Award Winner | 🤘🌁🗽| Training/Finetuning @basetenco, ex-@Plaid | Professional Attender of Weddings | DM for consulting inquiries

New York, USA Katılım Temmuz 2017
1.3K Takip Edilen935 Takipçiler
Vim retweetledi
Helen Min
Helen Min@helen_min·
New ad unit just dropped
Helen Min tweet media
English
3
1
25
1.4K
Vim retweetledi
Charlie O'Neill
Charlie O'Neill@oneill_c·
This was one of our early steps to explore the characteristics of different learning algorithms. How important is it to be on-policy? How important is it to have dense rewards? We can roughly place different approaches (SFT, RL, vanilla distillation, on-policy distillation, on-policy self-distillation, iSFT) on this grid of "on-policiness vs reward density" and use constitutional alignment as the test bed to explore how they drive internalised behavioural changes in the model. We find that being on-policy AND having dense rewards both matter for internalising the values in the constitution we're training on, and that this internalisation generalises in surprising ways (eg significant improvements on BullshitBench v2). Awesome stuff from @kirkby_max!
Max Kirkby@kirkby_max

x.com/i/article/2032…

English
0
5
49
8.1K
Charlie O'Neill
Charlie O'Neill@oneill_c·
New hiring strategy is to sit in barbershops and see who can successfully prompt their barber about what haircut they want. If you can prompt that you can prompt Opus 4.6 to find your context parallelism bug
English
7
0
51
3.9K
Vim retweetledi
Baseten
Baseten@baseten·
The Posit team is behind some of the most widely used data science tools, including RStudio and Positron. Posit uses Basten Training and the Baseten Inference Stack to: -> Spin up training compute in 1 minute -> Generate code edit suggestions in <200ms (60% faster than other providers) We're excited to power @posit_pbc's newest AI features! Read the full case study here: baseten.co/resources/cust…
English
0
2
14
1.3K
Vim retweetledi
Charlie O'Neill
Charlie O'Neill@oneill_c·
This is one of the early steps we took into researching truly infinite context with repeated, smart KV cache compaction. We have some more bitter lesson continual learning/cache research coming soon!
Baseten@baseten

Long-running agents accumulate context while model memory stays fixed. This leads to a tradeoff: either discard older information or compress it. New work by @oneill_c explores repeated KV-cache compression for persistent agents using Attention Matching. Our research shows one-shot compaction preserves detailed information remarkably well with 65–80% accuracy at 2–5× compression. This far outperforms text summarization. But what happens when you compress, add more context, and compress again repeatedly? baseten.co/research/repea…

English
0
2
15
780
Vim
Vim@vim_dzl·
@HEBenjoyer @stfukhi Ah I do not have that data. But I will take your word for it.
English
0
0
0
75
J Milli
J Milli@HEBenjoyer·
@vim_dzl @stfukhi Funny how there’s only PnR problems with Sengun, but not w Reed and Capela or Adams.
English
1
0
3
388
Vim
Vim@vim_dzl·
@HEBenjoyer @stfukhi Matches exactly what we saw against the Knicks to start fourth. Every possession they attacked Reed and Sengun.
English
1
0
0
352
Vim
Vim@vim_dzl·
@philipkiely Great work PK! Can't wait to get my hands on it
English
0
0
5
158
Vim
Vim@vim_dzl·
the story is so insanely full circle i can barely believe it
English
0
0
1
35
Vim
Vim@vim_dzl·
By providing support to on-demand experimentation, scaling from a single GPU to multiple nodes immediately, Baseten's Training Platform accelerated OE's model development, helping the provide the best medical tools in the world. So proud (and thankful!) to work with them
English
1
0
2
66
Vim
Vim@vim_dzl·
In December 2025, I suffered significant head trauma from a snowboarding incident. After an ER visit, I had tons of questions about how to best recover. I went to see my PCP, and upon hearing out my case, they turned their screen around and it's @OpenEvidence 💚🧡
Baseten@baseten

"No other product lets you launch ten different training jobs on four different datasets." –Head of Clinical NLP, OpenEvidence Over 40% of U.S. physicians trust @EvidenceOpen's platform for fast, accurate medical information. Their secret: custom, specialized models built on Baseten Training. Here's how we helped them save $1.9M via model training and improved their latency 23x to power 100M+ clinical consultations per year. baseten.co/resources/cust…

English
1
0
7
309
Vim retweetledi
Baseten
Baseten@baseten·
GLM 5 is live on Baseten. Opus 4.6 level performance at 10% of the cost. It takes the intelligence from its predecessors and layers in long-horizon agentic capabilities and complex systems engineering. That means your model can now do more than just chat (or code). It can solve sticky issues, call tools, and cohesively answer the problems we face in real life use cases at work. Get access on Baseten with our Model APIs or a dedicated deployment. baseten.co/library/glm-5/
English
2
1
33
2K
Vim
Vim@vim_dzl·
So blessed to work with the most talented + humble folks I’ve ever met. Come join us at Baseten!
Charlie O'Neill@oneill_c

Come join @baseten where you get to research and be part of an index on AI to expose yourself to all of it

English
0
0
9
313