


Dan Hendrycks
1.6K posts

@hendrycks
• Center for AI Safety Director • xAI and Scale AI advisor • GELU/MMLU/MATH/HLE • PhD in AI • Analyzing AI models, companies, policies, and geopolitics








It is imperative that the Mexican state re-establish their monopoly on violence before AGI.




Different people seem to mean radically different things by "AGI." More concrete & vivid milestones can better highlight underlying disagreements. In a new post, I define *self-sufficient AI:*









It really seems like ARC-AGI is only derailing the quest for AGI, mainly due to its popularity. I actually think it's good for what it is - a necessary (but not sufficient) condition for AGI. That is, if your model can't solve it, it's probably not AGI. But so much effort is now being poured into approaches designed to solve colored grid puzzles, and I can't see most of those methods being good for anything else.







The term “AGI” is currently a vague, moving goalpost. To ground the discussion, we propose a comprehensive, testable definition of AGI. Using it, we can quantify progress: GPT-4 (2023) was 27% of the way to AGI. GPT-5 (2025) is 58%. Here’s how we define and measure it: 🧵


The GDM mechanistic interpretability team has pivoted to a new approach: pragmatic interpretability Our post details how we now do research, why now is the time to pivot, why we expect this way to have more impact and why we think other interp researchers should follow suit