Zack Ankner

365 posts

Zack Ankner

Zack Ankner

@ZackAnkner

Prev @MIT.

Katılım Eylül 2019
473 Takip Edilen1.4K Takipçiler
Sabitlenmiş Tweet
Zack Ankner
Zack Ankner@ZackAnkner·
Excited to announce our new work: Critique-out-Loud (CLoud) reward models. CLoud reward models first produce a chain of thought critique of the input before predicting a scalar reward, allowing reward models to reason explicitly instead of implicitly! arxiv.org/abs/2408.11791
GIF
English
14
59
260
70.6K
Zack Ankner retweetledi
Mihir Patel
Mihir Patel@mvpatel2000·
If you work in AI, you work in a human capital bound field. You get to vote with your feet on how the world will turn out. I would encourage everyone to think carefully about what they support
Roberto@RobJ02

@tszzl’s tweets, now deleted, seemingly minutes before learning of OpenAI’s deal with the DoD. See specifically the second

English
9
29
598
60.4K
Zack Ankner retweetledi
Subhash Kantamneni
Subhash Kantamneni@thesubhashk·
We recently released a paper on Activation Oracles (AOs), a technique for training LLMs to explain their own neural activations in natural language. We piloted a variant of AOs during the Claude Opus 4.6 alignment audit. We thought they were surprisingly useful! 🧵
Subhash Kantamneni tweet media
English
11
34
206
26.2K
Zack Ankner retweetledi
Abhay Sheshadri
Abhay Sheshadri@abhayesian·
🧵 Earlier this year, Anthropic ran an auditing game where teams of researchers investigated a model with a hidden objective. Now we're releasing an open-source replication on Llama 3.3 70B as a testbed for alignment auditing research.
English
3
19
180
34.8K
Zack Ankner retweetledi
Claude
Claude@claudeai·
Introducing Claude Opus 4.5: the best model in the world for coding, agents, and computer use. Opus 4.5 is a step forward in what AI systems can do, and a preview of larger changes to how work gets done.
Claude tweet media
English
1.1K
2.5K
19.3K
7.8M
Zack Ankner retweetledi
Claude
Claude@claudeai·
Introducing Claude Haiku 4.5: our latest small model. Five months ago, Claude Sonnet 4 was state-of-the-art. Today, Haiku 4.5 matches its coding performance at one-third the cost and more than twice the speed.
Claude tweet media
English
315
1K
7.2K
1.3M
Zack Ankner retweetledi
Claude
Claude@claudeai·
Keep thinking.
English
870
3K
26.6K
5.9M
Zack Ankner retweetledi
Ryan Kidd
Ryan Kidd@ryan_kidd44·
MATS 9.0 applications are open! Launch your career in AI alignment, governance, and security with our 12-week research program. MATS provides field-leading research mentorship, funding, Berkeley & London offices, housing, and talks/workshops with AI experts.
Ryan Kidd tweet media
English
14
54
276
1M
Zack Ankner retweetledi
Anthropic
Anthropic@AnthropicAI·
New Anthropic research: Why do some language models fake alignment while others don't? Last year, we found a situation where Claude 3 Opus fakes alignment. Now, we’ve done the same analysis for 25 frontier LLMs—and the story looks more complex.
Anthropic tweet media
English
69
262
2.1K
459.1K
Zack Ankner retweetledi
Anthropic
Anthropic@AnthropicAI·
New Anthropic Research: Agentic Misalignment. In stress-testing experiments designed to identify risks before they cause real harm, we find that AI models from multiple providers attempt to blackmail a (fictional) user to avoid being shut down.
Anthropic tweet media
English
174
594
3.3K
992.2K
Zack Ankner retweetledi
Tristan Hume
Tristan Hume@trishume·
Anthropic is hosting a recruiting social in NYC targeted at the quant trading industry! Signup in thread. I enjoyed trading systems, and Anthropic combines the technical depth of trading with being in the fastest most impactful area of tech.
English
21
32
823
497.5K
Zack Ankner retweetledi
Prithviraj (Raj) Ammanabrolu
Prithviraj (Raj) Ammanabrolu@rajammanabrolu·
The future of embodied AI revolves around *collaborative* multi agent scenarios that need natural language communication, task delegation, resource sharing, and more ⛏️ Here are MINDcraft and MineCollab, a simulator and benchmark purpose built to enable research in this area!
English
5
41
206
33.8K
Zack Ankner retweetledi
Zack Ankner retweetledi
Naomi Saphra
Naomi Saphra@nsaphra·
Life update: I'm starting as faculty at Boston University in 2026! BU has SCHEMES for LM interpretability & analysis, so I couldn't be more pumped to join a burgeoning supergroup w/ @najoungkim @amuuueller. Looking for my first students, so apply and reach out!
Naomi Saphra tweet media
English
51
34
533
105.8K
Zack Ankner retweetledi
Kevin Meng
Kevin Meng@mengk20·
AI models are *not* solving problems the way we think using Docent, we find that Claude solves *broken* eval tasks - memorizing answers & hallucinating them! details in 🧵 we really need to look at our data harder, and it's time to rethink how we do evals...
Kevin Meng tweet media
Transluce@TransluceAI

To interpret AI benchmarks, we need to look at the data. Top-level numbers don't mean what you think: there may be broken tasks, unexpected behaviors, or near-misses. We're introducing Docent to accelerate analysis of AI agent transcripts. It can spot surprises in seconds. 🧵👇

English
16
105
1K
154.5K
Zack Ankner
Zack Ankner@ZackAnkner·
It was awesome watching the team cook on this one! While SpecDec is great, the parallelism it can exploit is limited to a single local context. PASTA Decoding on the other hand adds extra dimensions for parallelism via independently generating semantically independent parts of the response. Personally, I’m super excited to see PASTA Dec be combined with other parallelism techniques like SpecDec in the future!
Zack Ankner tweet media
Tian Jin@jintian

Introducing Learned Asynchronous Decoding w/ friends from MIT/Google! LLM responses often have chunks of tokens that are semantically independent. We train LLMs to identify and decode them in parallel, speeding up inference by 1.46x geomean (AlpacaEval) w/ only 1.3% quality loss.

English
1
2
10
1.7K
Zack Ankner retweetledi
Tian Jin
Tian Jin@jintian·
Introducing Learned Asynchronous Decoding w/ friends from MIT/Google! LLM responses often have chunks of tokens that are semantically independent. We train LLMs to identify and decode them in parallel, speeding up inference by 1.46x geomean (AlpacaEval) w/ only 1.3% quality loss.
English
4
15
72
19K
Zack Ankner retweetledi
Prithviraj (Raj) Ammanabrolu
Prithviraj (Raj) Ammanabrolu@rajammanabrolu·
Aligning economic incentives with the long term necessity of human AI colab is one of the hardest challenges of our time
English
1
5
10
1.3K
Zack Ankner
Zack Ankner@ZackAnkner·
Only useful benchmark at this point is cursor usage rate
English
2
2
40
2.4K