Jeffrey Li

32 posts

Jeffrey Li

Jeffrey Li

@jeffwpli

@AnthropicAI | Prev: PhD at @uwcse

Seattle, WA Katılım Ocak 2022
46 Takip Edilen91 Takipçiler
Jeffrey Li retweetledi
Etash Guha
Etash Guha@etash_guha·
OpenThoughts is going to be an Oral Presentation at ICLR! It's my first oral presentation so super excited! See y'all in Brazil! :)
Etash Guha tweet media
English
5
4
92
11.7K
Jeffrey Li retweetledi
Anas Awadalla
Anas Awadalla@anas_awadalla·
We're releasing🍨Gelato-30B-A3B, a state-of-the-art computer grounding model that delivers immediate performance gains for computer-use agents! Trained on our open-source🖱️Click-100k dataset, Gelato achieves 63.8% on ScreenSpot-Pro and 69.1% on OS-World-G. It outperforms specialized models like GTA1-32B and VLMs ~8× its size like Qwen3-VL-235B. (1/N) 🧵
Anas Awadalla tweet media
English
7
41
236
34.2K
Jeffrey Li retweetledi
Alex Shaw
Alex Shaw@alexgshaw·
Today, we’re announcing the next chapter of Terminal-Bench with two releases: 1. Harbor, a new package for running sandboxed agent rollouts at scale 2. Terminal-Bench 2.0, a harder version of Terminal-Bench with increased verification
Alex Shaw tweet media
English
25
74
393
142.5K
Jeffrey Li retweetledi
Guilherme Penedo
Guilherme Penedo@gui_penedo·
> we've hit a data wall > pretraining is dead Is it? Today we are releasing 📄 FinePDFs: 3T tokens of new text data for pre-training that until now had been locked away inside PDFs. It is the largest permissively licensed corpus sourced exclusively from PDFs.
Guilherme Penedo tweet media
English
26
154
1.5K
133.7K
Jeffrey Li retweetledi
Fartash Faghri
Fartash Faghri@FartashFg·
📢Submissions are now open for #NeurIPS2025 CCFM workshop. Submission deadline: August 22, 2025, AoE. Website: sites.google.com/view/ccfm-neur… Call for papers: sites.google.com/view/ccfm-neur… Submission Link: openreview.net/group?id=NeurI…
Fartash Faghri@FartashFg

Is your AI keeping Up with the world? Announcing #NeurIPS2025 CCFM Workshop: Continual and Compatible Foundation Model Updates When/Where: Dec. 6-7 San Diego Submission deadline: Aug. 22, 2025. (opening soon!) sites.google.com/view/ccfm-neur… #FoundationModels #ContinualLearning

English
0
6
11
10.7K
Jeffrey Li retweetledi
David Mizrahi
David Mizrahi@dmizrahi_·
Excited to share our new work: “Language Models Improve When Pretraining Data Matches Target Tasks” Yes, it sounds obvious (and it is!), but typically this only happens implicitly and indirectly: intuitively select data → benchmark → refine → repeat. We wondered: what happens if we explicitly match pretraining data to benchmarks? The result is a dead simple approach that yields 2x+ compute multipliers over strong baselines and gives us a principled way to study how benchmark choices shape (and constrain!) model capabilities. Bonus: extensive scaling laws from training 500+ models that reveal how optimal data selection evolves as models scale. 🧵 (1/14)
David Mizrahi tweet media
English
7
47
402
55.2K
Jeffrey Li retweetledi
Mike A. Merrill
Mike A. Merrill@Mike_A_Merrill·
Terminal-Bench and @warpdotdev @zachlloydtweets in TechCrunch today :) (link in replies) I firmly believe that the future of LLM-Computer interaction is through something that looks like a terminal interface. Great to see this picking up steam.
English
3
4
15
2K
Jeffrey Li
Jeffrey Li@jeffwpli·
Excited to share TiC-LM (Oral at #ACL2025)! LLMs can become outdated ⏲️ and re-training from scratch is costly💰. Ideally, we'd keep reusing and updating models on newer data ♻️. We study continual training as 114 CC months are revealed one-at-a-time. arxiv.org/abs/2504.02107
Jeffrey Li tweet media
English
2
8
60
7K
Jeffrey Li retweetledi
Andy Konwinski
Andy Konwinski@andykonwinski·
Today, I’m launching a deeply personal project. I’m betting $100M that we can help computer scientists create more upside impact for humanity. Built for and by researchers, including @JeffDean & @jpineau1 on the board, @LaudeInstitute catalyzes research with real-world impact.
Andy Konwinski tweet media
English
55
119
1.2K
337.3K