Benjamin Warner

618 posts

Benjamin Warner banner
Benjamin Warner

Benjamin Warner

@benjamin_warner

Research @SophontAI. Previously answerdotai. Vaccines save lives.

Se unió Eylül 2011
484 Siguiendo2.4K Seguidores
Tweet fijado
Benjamin Warner
Benjamin Warner@benjamin_warner·
Today we released ModernBERT, the first encoder to reach SOTA on most common benchmarks across language understanding, retrieval, and code, while running twice as fast as DeBERTaV3 on short context and three times faster than NomicBERT & GTE on long context.
Benjamin Warner tweet media
English
2
12
80
11.3K
Benjamin Warner
Benjamin Warner@benjamin_warner·
@waydegilliam Almost everything other than UX. It's better at GPU kernel writing, debugging, you name it. GPT-5.4 is better than 5.3 at inferring intent, but Codex is still more literal than Claude, so if you don't change how you prompt Codex, it'll probably seem worse then CC.
English
0
0
1
52
Benjamin Warner retuiteado
Labomen
Labomen@labomen001·
@cursor_ai Here's the graph with the same data, but plotted against the actual output cost for each (Composer 1.5 output from Cursor docs is $17.5). Although this doesn't account for >200K Opus 4.6/>272K GPT 5.4/Gemini 3.1 >200K.
Labomen tweet media
English
4
10
113
28.3K
Benjamin Warner retuiteado
Ben Clavié
Ben Clavié@bclavie·
I'm so excited to introduce this! We've worked on a million different moving parts to produce this. I'm fairly confident it's the best multimodal model that exists, period -- and it's not too shabby at pushing back the LIMITs of retrieval either...
Mixedbread@mixedbreadai

Introducing Mixedbread Wholembed v3, our new SOTA retrieval model across all modalities and 100+ languages. Wholembed v3 brings best-in-class search to text, audio, images, PDFs, videos... You can now get the best retrieval performance on your data, no matter its format.

English
37
41
410
138.3K
Benjamin Warner retuiteado
Benjamin Warner
Benjamin Warner@benjamin_warner·
GPT 5.4 in Codex needs to be able to delegate simpler coding work to 5.3-Spark.
English
0
0
1
199
Mario Filho
Mario Filho@mariofilhoml·
My timeline is split between people that believe 5.4 XHigh is the best thing ever And people that believe 5.4 High is enough A minority don't like 5.4 at all
English
1
0
0
174
Benjamin Warner retuiteado
PyTorch
PyTorch@PyTorch·
FlexAttention now has a FlashAttention-4 backend. FlexAttention has enabled researchers to rapidly prototype custom attention variants—with 1000+ repos adopting it and dozens of papers citing it. But users consistently hit a performance ceiling. Until now. We've added a FlashAttention-4 backend to FlexAttention on Hopper and Blackwell GPUs. PyTorch now auto-generates CuTeDSL score/mask modifications and JIT-instantiates FlashAttention-4 for your custom attention variant. The result: 1.2× to 3.2× speedups over Triton on compute-bound workloads. 🖇️ Read our latest blog here: hubs.la/Q045FHPh0 No more choosing between flexibility and performance. hashtag#PyTorch hashtag#FlexAttention hashtag#FlashAttention hashtag#OpenSourceAI
PyTorch tweet media
English
12
98
731
99.6K
Shizhe Diao
Shizhe Diao@shizhediao·
Time to upgrade your pretraining dataset. Instead of FineWeb-EDU / DCLM / X, try ClimbMix-400B. 📄 Paper: arxiv.org/pdf/2504.13161 📦 Data: huggingface.co/datasets/nvidi… CLIMBMix uses clustering-based iterative data mixture to improve pretraining efficiency and data quality. Would love to see the community experiment with it and push it further 🚀
Shizhe Diao@shizhediao

Nemotron-CLIMBMix is now becoming the default recipe in nanochat speedrun. During the Time-to-GPT-2 Leaderboard experiments started by @karpathy, the community revisited CLIMBMix and found that it delivers by far the single biggest improvement to nanochat’s GPT-2 speedrun time. It’s incredibly rewarding to see the idea validated and adopted by the community. Huge thanks to everyone who experimented with it and pushed it forward 🚀 #L42" target="_blank" rel="nofollow noopener">github.com/karpathy/nanoc…

English
3
15
165
27K
Benjamin Warner
Benjamin Warner@benjamin_warner·
New optimizer paper and library which looks to be an improvement on the low precision error correction I've used in optimi and quantization used for optimizers in bitsandbytes et al.
Davis Blalock@davisblalock

🚀 Today we’re releasing FlashOptim: better implementations of Adam, SGD, etc, that compute the same updates but save tons of memory. You can use it right now via `pip install flashoptim`. 🚀 arxiv.org/abs/2602.23349 A bunch of cool ideas make this possible: [1/n]

English
1
1
2
368
Benjamin Warner
Benjamin Warner@benjamin_warner·
@wightmanr @BlancheMinerva Given the referenced attack was with a previous generation of Claude and the best open models are approaching (or already reached) that level of performance, I don't see how one can come to any conclusion other then it's just a matter of time.
English
0
0
1
56
Ross Wightman
Ross Wightman@wightmanr·
There's already a number of pen test oriented LLM attack orchestration demos/tools like HexStrike (open source), Cobalt Strike + LLM. And aside from being used to orchestrate attacks. LLM coding tools are definitely at a level where they can help build attack tools, especially for those that may have been lacking in their coding ability but not their imagination.
English
2
0
12
1.2K
Stella Biderman
Stella Biderman@BlancheMinerva·
It's very common for people to claim that open LLMs will be used to commit cyber attacks at massive scale. What public evidence is there for this claim? The best (and one of the only) accounts I've seen of a cyber LLM attack was done using Claude anthropic.com/news/disruptin…
English
10
3
38
6.5K
Benjamin Warner retuiteado
Alexis Gallagher
Alexis Gallagher@alexisgallagher·
I am thrilled and honored that Sparky and I were selected winners for NVIDIA GTC Golden Ticket. Here's how he received the news.
NVIDIA GTC@NVIDIAGTC

Congratulations to our #NVIDIAGTC Golden Ticket winners 🎉: @alexisgallagher Brandon I. Hans B. Julia S. Lluís D. Marco D. Tarique S. You’re headed to GTC! We’ll be reaching out soon with next steps to claim your prize. Thank you to our partners for collaborating with NVIDIA on the 2026 Golden Ticket Developer Contest: @huggingface / @pollenrobotics, @ollama, @ethroboticsclub, and @googlecloud. Stay tuned for one more winner reveal 👀

English
13
14
106
14.7K