Andreas Tande
584 posts

Andreas Tande
@andreastande
Be kind, and the world will be kind back



We've published a paper that explains our views on AI competition between the US and China. The US and democratic allies hold the lead in frontier AI today. Read more on what it’ll take to keep that lead: anthropic.com/research/2028-…

Today we release Token Superposition Training (TST), a modification to the standard LLM pretraining loop that produces a 2-3× wall-clock speedup at matched FLOPs without changing the model architecture, optimizer, tokenizer, or training data. During the first third of training, the model reads and predicts contiguous bags of tokens, averaging their embeddings on the input side and predicting the next bag with a modified cross-entropy on the output side. For the remainder of the run, it trains normally on next-token prediction. The inference-time model is identical to one produced by conventional pretraining. Validated at 270M, 600M, and 3B dense scales, and at 10B-A1B MoE. The work on TST was led by @bloc97_, @gigant_theo, and @theemozilla.

In three years we went from AI being unable to do simple arithmetic to AI threatening to put almost all professional mathematicians out of a job. Pretty impressive!

one feeling i get from talking with both openai/anthropic alignment is a lot of people believe we’re on a good trajectory and also that the next generation of models will be much better alignment researchers than any human is. not everyone obviously

@So8res Here, Jan Leike speaks on how alignment research is getting more and more automated. Anthropic is the most out-spoken on this, I’m surprised OpenAI is so quiet — are they not experiencing the same? Albeit alignment research, we’re not far off all research aligned.substack.com/p/alignment-is…


I write about this in more detail in a blog post with a guest contribution from Isaac Rajagopal, a student at MIT on whose work ChatGPT built, who gives his assessment of the level of mathematical ability displayed by the model. gowers.wordpress.com/2026/05/08/a-r…






Introducing GPT-Realtime-2 in the API: our most intelligent voice model yet, bringing GPT-5-class reasoning to voice agents. Voice agents are now real-time collaborators that can listen, reason, and solve complex problems as conversations unfold. Now available in the API alongside streaming models GPT-Realtime-Translate and GPT-Realtime-Whisper — a new set of audio capabilities for the next generation of voice interfaces.



this was very obvious to anyone with 10 brain cells im ngl






GPT-5.5 & Opus 4.7 on ARC-AGI-3 - GPT-5.5: 0.43% - Opus 4.7: 0.18% We found 3 failure modes: - True local effect, false world model - Wrong level of abstraction from training data - Solved the level, didn’t reinforce the reward See our full analysis 🧵













