Keyon Vafa

1.4K posts

Keyon Vafa

Keyon Vafa

@keyonV

Postdoctoral fellow at @Harvard_Data | Former computer science PhD with @Blei_Lab at @Columbia University | Researching AI + implicit world models

เข้าร่วม Ağustos 2011
882 กำลังติดตาม4.7K ผู้ติดตาม
ทวีตที่ปักหมุด
Keyon Vafa
Keyon Vafa@keyonV·
Can an AI model predict perfectly and still have a terrible world model? What would that even mean? Our new ICML paper formalizes these questions One result tells the story: A transformer trained on 10M solar systems nails planetary orbits. But it botches gravitational laws 🧵
English
209
991
6.7K
1.4M
Keyon Vafa รีทวีตแล้ว
liz tan
liz tan@liztansz·
Say you have trained your deep learning model. It works. But do you know what it has actually learned? 🚀 We’ve built SymTorch: a library that translates deep learning models into human-readable equations. I've attached here a quick video demonstrating how SymTorch works.
English
51
373
2.7K
133.7K
Keyon Vafa รีทวีตแล้ว
Alex Imas
Alex Imas@alexolegimas·
Super interesting paper from @AnthropicAI Fellows Program on model breakdown as task complexity increases. The longer the model has to reason, the more unpredictable it becomes: not consistently wrong, not completely random, just pursuing strange goals that are neither systematically aligned nor misaligned. Reminds me of @keyonV and co. human generalization function paper. This research suggests that human beliefs about model performance will be increasingly miscalibrated at longer reasoning length.
Alex Imas tweet mediaAlex Imas tweet media
Anthropic@AnthropicAI

New Anthropic Fellows research: How does misalignment scale with model intelligence and task complexity? When advanced AI fails, will it do so by pursuing the wrong goals? Or will it fail unpredictably and incoherently—like a "hot mess?" Read more: alignment.anthropic.com/2026/hot-mess-…

English
6
10
109
15.2K
Keyon Vafa รีทวีตแล้ว
Alex Imas
Alex Imas@alexolegimas·
New post with @AndreyFradkin on the obstacles for AI agentic commerce. Imagine an AI that could optimize your credit card points across airlines and hotels to book a trip to Tokyo. The technology exists, at least in theory. So why can't you use an agent to book a flight, or shop for groceries? There are two main obstacles: 1) The interface of the internet was optimized for human interactions. What's needed is a machine-readable internet twin. Problem: the platforms best positioned to develop this twin often have an incentive not to. 2) The regulatory framework for agentic commerce is a mess. In many cases, it's legally ambiguous whether you can "bring your own" agent to shop for you (and agent providers get sued); in other cases, a platform can simply revoke permission and deny service. Firms have incentive to develop their own agents that you interact with, rather than allow you to BYO an agent aligned with your interests. To foster competition and improve consumer welfare, a new regulatory framework is needed. We outline what one might look like. aleximas.substack.com/p/why-cant-you…
English
6
17
64
34.3K
Keyon Vafa
Keyon Vafa@keyonV·
@BenSManning @alexolegimas @arpitrage Thanks Ben! I agree with your questions. Something I've been thinking a bit about is the behavioral work showing where heuristics work for people/where they don't. For LLMs, beyond the yes/no world model question, can we characterize where the lack of world models work vs don't?
English
1
1
3
381
Keyon Vafa รีทวีตแล้ว
Benjamin Manning
Benjamin Manning@BenSManning·
This is an awesome post @alexolegimas and @arpitrage - I've been thinking a lot about this stuff too. Screenshot is from a revision section I’ve been writing with @johnjhorton and @apo_filippas. The core point is pretty obvious, almost to the point of being boring: you don’t need a perfectly “true” model of the world to be useful. Newtonian mechanics is wrong, quantum is more right, Newton is still incredibly predictive over a huge domain. But what feels newly salient with LLMs is that we don’t actually have good ways to measure the differences that matter. How incomplete is a model’s world model in the dimensions that show up under intervention? Where does predictive competence stop being reliable when you push on the system? Work like @keyonV NYC navigation + planets papers makes this very concrete: strong surface-level prediction can coexist with brittle or incoherent internal structure that only reveals itself under the “wrong” perturbations (I'm obsessed with these papers - they're so good)
Benjamin Manning tweet media
Alex Imas@alexolegimas

New post : Can a Transformer “Learn” Economic Relationships? Revisiting the Lucas Critique in the age of Transformers. with @arpitrage We simulated data from an NK model, fit a transformer, and tested out of sample fit How did it do? Pretty well! Link and thread in reply:

English
9
18
101
36.3K
Keyon Vafa รีทวีตแล้ว
Owain Evans
Owain Evans@OwainEvans_UK·
New paper: We train Activation Oracles: LLMs that decode their own neural activations and answer questions about them in natural language. We find surprising generalization. For instance, our AOs uncover misaligned goals in fine-tuned models, without training to do so.
Owain Evans tweet media
English
24
79
597
99K
Keyon Vafa รีทวีตแล้ว
Jessica Hullman
Jessica Hullman@JessicaHullman·
Many think LLM-simulated participants can transform behavioral science. But social scientists lack accessible discussion of what it means to validate these models. Under what conditions can we trust LLMs to learn about human parameters? Our paper maps the validation landscape. 1/
Jessica Hullman tweet media
English
2
25
73
15.3K
Keyon Vafa รีทวีตแล้ว
Edgar Dobriban
Edgar Dobriban@EdgarDobriban·
I wrote a review paper about statistical methods in generative AI; specifically, about using statistical tools along with genAI models for making AI more reliable, for evaluation, etc. See here: arxiv.org/abs/2509.07054! I have identified four main areas where statistical thinking can be helpful. These are just a subset of what is out there; other topics have been well-covered in other reviews. 1. Designing "statistical wrappers" around a model, for instance, changing behavior of a trained model (e.g., abstaining), where a score, e.g., an "unsafety score" is too high. The key connection to statistics is to use the quantiles of the loss (on a calibration set) to set the critical threshold, thus enabling conformal-type high probability guarantees. 2. Closely related, methods for uncertainty quantification, which enable the model to express uncertainty in an answer. A crucial component here is "calibration", whereby the uncertainty is required to reflect reality. 3. Statistical methods for AI evaluation: Specifically, tools for statistical inference (e.g., confidence intervals) on model performance. Exciting recent work proposes careful statistical models for leveraging a very small high-quality dataset, possibly combined with much larger low-quality datasets, for accurate evaluation. 4. Experiment design and interventions. Careful AI experiments to understand and steer models may require interventions such as modifying experimental settings in a controlled manner. This brings up connections to classical experimental design in statistics. This connection has largely remained implicit so far, and my review aims to make it more explicit; hoping that experimental design principles will become useful here. This review references the work of many, including @HamedSHassani @obastani @tatsu_hashimoto @yuekai_sun @CsabaSzepesvari @ml_angelopoulos @stats_stephen @yaniv_romano @yaringal @KilianQW @_onionesque +their teams, and some work that I was also involved in. Hopefully, my review will be helpful to orient yourself in this exciting area. Nonetheless, since the area is rapidly expanding, it is possible that I missed important references. Please feel free to let me know of anything that I should add/change!
Edgar Dobriban tweet mediaEdgar Dobriban tweet mediaEdgar Dobriban tweet mediaEdgar Dobriban tweet media
English
11
96
482
33.9K
Pontus Rendahl
Pontus Rendahl@pontus_rendahl·
Why not just simulate, say, and NK model and use the data to form a neural net (or whatever). If it uncovers the deep parameters, the discussion is settled no?
Alex Imas@alexolegimas

Many disagreed with post below, saying Lucas critique is not about simpler models per se, but importance of modeling the system structurally to account for responses to policy changes. Even with a “large n”, a model can’t learn about missing data and things like adverse selection. This amounts to: reduced form regressions (which is what Lucas was critiquing) do not recover the data generating process, and so will not account for responses to a policy change. My interpretation of the post was consistent with the response below to @JesusFerna7026: one of the lessons we learned from LLMs is that training neural nets with enough data can lead to emergent properties, one of which seems to be that the model *sometimes* seems to “learns” the structure of the DGP despite not being exposed to all of the potential comparative statics in the training data. Will this always be the case with enough data? Will a model “learn” to account for adverse selection and other types of “missing data”? Given what we’ve seen so far, I think it’s worth considering. But I could be wrong!

English
7
1
38
43.7K
Keyon Vafa รีทวีตแล้ว
Damien Teney
Damien Teney@DamienTeney·
Can vision transformers learn without images?🤔👀 Our latest work shows that pretraining ViTs on procedural symbolic data (eg sequences of balanced parentheses) makes subsequent standard training (eg on ImageNet) more data efficient! How is this possible?! ⬇️🧵
Damien Teney tweet media
English
12
88
603
49.8K
Keyon Vafa รีทวีตแล้ว
Melanie Sclar
Melanie Sclar@melaniesclar·
Our final invited speaker Keyon Vafa @keyonV shares insights on Evaluating Implicit World Models! Right after this we'll have six contributed talks and our best paper awards 🏆
Melanie Sclar tweet media
English
1
2
15
876
Keyon Vafa
Keyon Vafa@keyonV·
I'll be speaking at the Language, Agent, and World Models (LAW) Workshop this afternoon at 2:15 in Upper Level Ballroom 20D!
LAW Workshop@NeurIPS 2025@LAW2025_NeurIPS

🔥 We are pleased to announce the talk title for our #LAW2025 #NeurIPS workshop Join us on 📅Sunday, Dec 7th in📍Upper Level Ballroom 20D Full schedule here: sites.google.com/view/law-2025/… We look forward to seeing you all this Sunday! #NeurIPS2025 #AI #ML #WorldModel #Agents #LLMs

English
1
0
7
1.1K
Keyon Vafa รีทวีตแล้ว
Zhen Wang
Zhen Wang@zhenwang9102·
The San Diego party ain’t over yet🌴 LAW 2025 is happening today at @NeurIPSConf 🥳 in📍Ballroom 20D This is the first year of LAW workshop and we are honored to have a world-class lineup of speakers to reshape the directions on the unification of language, agent and world models: @sherryyangML @chelseabfinn @fchollet @philipjohnball Stephen Spencer @ericxing @danijarh Ying Nian Wu @keyonV 🚀Huge thanks to our sponsors: @LambdaAPI, nexovaai.io Also, thanks to our amazing co-organizers: @ziqiao_ma @realJessyLin @melaniesclar @jianwen_xie @KelseyRAllen @alsuhr @jacobandreas @tianminshu @ZhitingHu sites.google.com/view/law-2025 neurips.cc/virtual/2025/l… #NeurIPS #NeurIPS2025 #LLMs #AgenticAI #WorldModels #AI #ML #RL #Reasoning #EmbodiedAI #VLMs
LAW Workshop@NeurIPS 2025@LAW2025_NeurIPS

🔥 We are pleased to announce the talk title for our #LAW2025 #NeurIPS workshop Join us on 📅Sunday, Dec 7th in📍Upper Level Ballroom 20D Full schedule here: sites.google.com/view/law-2025/… We look forward to seeing you all this Sunday! #NeurIPS2025 #AI #ML #WorldModel #Agents #LLMs

English
0
6
14
3.7K
Alex Imas
Alex Imas@alexolegimas·
Read Wolfram's excellent "What is ChatGPT Doing..." (h/t @danielrock). He writes that we learned a lot about how language works from fact that GPT3, with only 175 billion weights, is able to emulate it so well. This implies it's computationally a lot simpler than we may have thought. But what about math? At time this was written (2023), GPT was still very bad at math. The models became very (very) good at math when the first reasoning model came out (o1), which relied a lot more on reinforcement learning rather than just brute force pretraining. Wonder what this says about math? Conceptually, language is a lot "fuzzier" than math: multiple words can sound "right" in the same spot in a sentence. This is what makes the probabilistic LLM architecture work. Math is less fuzzy. This is perhaps why the more "rule based" RL step was crucial. But this also implies formal math is less computationally complex than we thought. Thoughts? @littmath @alz_zyd_
Alex Imas tweet media
English
26
101
800
64.9K
Keyon Vafa
Keyon Vafa@keyonV·
In San Diego for NeurIPS! Today: Presenting a poster for our paper on steerability from 4:30-7:30 in Exhibit Hall CDE (#3715). Paper w/ Sarah Bentley, Jon Kleinberg, and @m_sendhil. Sunday: Giving a talk at the Language, Agent, and World Model workshop on testing world models.
Keyon Vafa tweet mediaKeyon Vafa tweet mediaKeyon Vafa tweet media
English
1
3
11
885