
Reflection AI
37 posts

Reflection AI
@reflection_ai
Frontier open intelligence accessible to all.


Reflection is partnering with Shinsegae Group to build a 250-megawatt sovereign AI factory for the Republic of Korea. Open intelligence. Built on trust between allies. Owned by the nations that need it most. The future of sovereign AI. Read more in the @WSJ.



Today, we're joined by @achowdhery, member of technical staff at @reflection_ai, to explore the fundamental shifts required to build true agentic AI. While the industry has largely focused on post-training techniques to improve reasoning, Aakanksha draws on her experience leading pre-training efforts for Google’s PaLM and early Gemini models to argue that pre-training itself must be rethought to move beyond static benchmarks. We explore the limitations of next-token prediction for multi-step workflows and examine how attention mechanisms, loss objectives, and training data must evolve to support long-form reasoning and planning. Aakanksha shares insights on the difference between context retrieval and actual reasoning, the importance of "trajectory" training data, and why scaling remains essential for discovering emergent agentic capabilities like error recovery and dynamic tool learning. 🗒️ For the full list of resources for this episode, visit the show notes page: twimlai.com/go/759. 📖 CHAPTERS =============================== 00:00 - Introduction 02:26 - Reflection 04:54 - Limitations of post-training for building agents 07:31 - Rethinking pre-training in agents 10:51 - Scaling 11:27 - Evolving attention mechanisms for agentic capabilities 12:39 - Memory as a tool 14:13 - Loss objectives and training data 15:50 - Fine-tuning loss in agent performance 19:37 - Training data 21:29 - Augmenting dominant training data source 24:11 - Overcoming challenges in training on synthetic data 25:47 - Benchmarks 30:44 - Scaling laws in large models versus small models 33:20 - Long-form versus short-form reasoning 37:57 - Agent’s ability to recover from failure 40:15 - Hallucinations and failure recovery 43:53 - Tool use in agents 46:38 - Coding agents 48:37 - How researchers can contribute to agentic AI

An update: I have left Meta Superintelligence Labs and joined @reflection_ai in NYC!! Today is my first day. I started in the Fundamental AI Research (FAIR) lab at Meta, then Facebook, over six (!) years ago as my first job out of the PhD. They were some formative years. The group is full of exceptionally talented people that have profoundly shaped my perspective on life and research. I am grateful for everything we have shared and proud of everything we created together. I have decided it's time to try to build a startup and new frontier models with Reflection. Superintelligence will be one of the most significant advancements of our lifetimes, resulting in a computational reflection of ourselves. We believe it should be safe, open, and accessible to all. I am excited to be jumping into the post-training and reinforcement learning pipelines to advance capabilities and alignment. And we are hiring! Please get in touch.







Hi friends, after three incredible years at OpenAI I am excited to share that I am starting a new chapter at @reflection_ai, where I will be leading the Science of Scaling team. Our mission is to deepen the scientific understanding of large scale learning and to turn compute into intelligence as efficiently and predictably as possible.






🔴 Sam Altman LIVE on TBPN x.com/i/broadcasts/1…









