Michael Bernstein

768 posts

Michael Bernstein banner
Michael Bernstein

Michael Bernstein

@msbernst

@stanford Professor of Computer Science, @simile_ai co-founder, nationally bestselling author. I build interactive, social, and societal tech.

Stanford, CA Katılım Kasım 2007
1.9K Takip Edilen18.7K Takipçiler
Michael Bernstein retweetledi
Nilou Salehi
Nilou Salehi@nilou_salehi·
It was standing room only at the kick-off for our research series on continual learning. Thank you to @NikzadAfshin (@across_ai ) @sarahookr (@adaption_ai) and @mralbertchun (AI Circle) for hosting! @oshaikh13 shared his research on human grounding in continual learning. It was so cool to be reminded of the old Apple Knowledge Navigator and how close we are to it and yet how far we still are :) how much easier some questions have gotten and how some remain so hard. Omar, you reminded me of my PhD defense where at some point I annoyed Maneesh so much he said: you can't keep saying "depends on the user context" in response to every question 😅 youtu.be/umJsITGzXd0?si… Stay tuned for the next meetup next month and check out Omar's research with @msbernst and @Diyi_Yang : •⁠ ⁠Creating General User Models from Computer Use (arxiv.org/abs/2505.10831): an architecture for a model that learns about you by observing any interaction with your computer, building confidence-weighted propositions about preferences and intent. •⁠ ⁠Learning Next Action Predictors from Human-Computer Interaction (arxiv.org/abs/2603.05923): predicting a user's next action from their full multimodal interaction history (screenshots, clicks, sensor data) rather than just typed prompts.
YouTube video
YouTube
Nilou Salehi tweet mediaNilou Salehi tweet mediaNilou Salehi tweet media
English
2
5
30
8.1K
Michael Bernstein retweetledi
Serina Chang
Serina Chang@serinachang5·
🎉 Thrilled to have two papers accepted to ACL 2026 main! 1. Graph-based models match LLMs on close-ended human simulation tasks with far less compute & greater transparency 2. (oral) How to allocate human samples towards fine-tuning vs post-hoc rectification in simulation
Serina Chang tweet mediaSerina Chang tweet media
English
4
19
135
14K
Dora Zhao
Dora Zhao@dorazhao9·
Excited to be sharing three papers at #CHI2026! 1⃣ Value Alignment of Social Media Ranking Algorithms 2⃣ Mapping the Spiral of Silence: Surveying Unspoken Opinions in Online Communities 3⃣ Whose Knowledge Counts? Co-Designing Community-Centered AI Auditing Tools with Educators in Hawai`i
Dora Zhao tweet media
English
2
20
89
18K
Michael Bernstein retweetledi
Omar Shaikh
Omar Shaikh@oshaikh13·
This is one of my fav figures in our paper. You can: 1. Identify a user's objective by observing general interaction with their computer. 2. Use it to construct a "just in time" rubric. 3. Sample bunch from model and SCALE TEST TIME COMPUTE ON LITERALLY ANY OPEN-ENDED TASK?!?
Michelle Lam@michelle123lam

Once you have JIT objectives, you can embed them into various LLM architectures via existing generators and evaluators. Evaluations on N=205 participant-provided inputs show that JIT objectives produce user-preferred outputs, whether generating experts, tools, or feedback.

English
0
5
34
9K
Robb Willer
Robb Willer@RobbWiller·
Very grateful to receive a Guggenheim Fellowship. I’ve been so lucky to have such incredible students, collaborators, and mentors whose coattails I’ve hitched myself to for many years. Above all, I feel very lucky that I get to work on topics I care deeply about with collaborators who are also my friends.✊❤️#guggfellows2026
Robb Willer tweet media
English
18
8
99
4.2K
Michael Bernstein
Michael Bernstein@msbernst·
This paper was a fascinating experience where, when we first submitted it, reviewers refused to believe that we could create contentious social media content with LLMs. This time around, they saw it as the main point of novelty. As time passed, our work got _more_ novel?
Dora Zhao@dorazhao9

2. Mapping the Spiral of Silence: Surveying Unspoken Opinions in Online Communities w/ @Diyi_Yang @msbernst We introduce a human–AI pipeline to measure the spiral of silence across political subreddits, revealing how community design shapes when people choose to stay silent online. Preprint: arxiv.org/abs/2502.00952

English
1
0
24
2.9K
Michael Bernstein retweetledi
Michelle Lam
Michelle Lam@michelle123lam·
Most of what I actually need help with, I never think to tell a model. But why is it on me to remember? Our new paper asks: what if AI could proactively specialize to individuals and the tasks they’re carrying out at this very moment? 🧵
English
13
44
260
42.5K
Percy Liang
Percy Liang@percyliang·
Academic titles are funny. After 14 years, I finally have the official title that people might have always assumed I had.
English
93
22
1.3K
115.4K
Michael Bernstein retweetledi
Mitchell Gordon
Mitchell Gordon@mitchellgordon·
MIT postdoc opportunity! We're hiring a human-AI interaction postdoc (HCI+ML/RL) to train agents that deepen how people think and collaborate - rewarded by how humans actually build skill together. With @arvindsatya1 @ZanaBucinca, me & more! Apply by May 1 tinyurl.com/4jsr8ee9
English
3
18
121
12.8K
Michael Bernstein retweetledi
Omar Shaikh
Omar Shaikh@oshaikh13·
Passive interaction data is super underrated. Recruit users, observe what they're already doing (and willing to share!), and label those trajectories with a VLM! It's free lunch!!!! We're releasing an open-source package (NAPsack) to do this, tested on 1.9M screenshots. 🧵
Omar Shaikh tweet media
English
6
30
137
23.1K
Michael Bernstein retweetledi
Joon Sung Park
Joon Sung Park@joon_s_pk·
We trained our models to solve problems with objective answers. But can we build models that solve problems where success is subjective, messy, and human? The latter is even more impactful imo. Agree with Percy: simulation is the next frontier for AI.
Percy Liang@percyliang

I think it’s pretty clear that simulation is the next frontier for AI. The most impressive feats of AI to date are when we have a clear environment + reward, whether it be beating Le Sedol at Go, winning an IMO gold medal, or writing entire apps from scratch. In these cases, the RL algorithm can try different actions, and observe the well-defined consequences in the safety of a docker container. But what about messy real-world situations involving people? The rewards are unclear, the stakes are high, and you can’t experiment in the real world. But these situations are precisely where the next big opportunity in AI is. To crack this, we need to *simulate* society (“put society into a docker container”). Concretely, this means building a model that can predict what will happen in any given situation (real or hypothetical). If we can do this, we are only limited by our imagination: predict the future, optimize for better outcomes, answer hypothetical (“what if”) questions. Ultimately, this goes beyond making better decisions, but it’s about giving us a better understanding of ourselves and the world. Simulation is the whole enchilada. And this is exactly the research that @simile_ai is working on. Read more here: simile.ai/blog/simulatio…

English
6
4
94
15.8K
Michael Bernstein retweetledi
Diyi Yang
Diyi Yang@Diyi_Yang·
Current AI is reactive. You prompt, it responds. True proactivity requires predicting what you'll do before you ask. Our new work done by @oshaikh13 formalizes this as Next Action Prediction (NAP ): given a user's computer use, predict their next action. We annotated 360K actions across 1 month of continuous computer use from 20 users and open-sourced a pipeline for private-infra labeling. LongNAP combines parametric + in-context learning to reason over long interaction traces. This is one step closer to an assistant that proactively anticipates, not just reactively responds 🚀
Omar Shaikh@oshaikh13

What’s the point of a “helpful assistant” if you have to always tell it what to do next? In a new paper, we introduce a reasoning model that predicts what you’ll do next over long contexts (LongNAP 💤). We trained it on 1,800 hours of computer use from 20 users. 🧵

English
8
28
290
51.8K
Michael Bernstein
Michael Bernstein@msbernst·
Predict what the user will do next: a task that underlies a huge number of goals. Better assistants, better user models, the next generation of operating system metaphors...
Omar Shaikh@oshaikh13

What’s the point of a “helpful assistant” if you have to always tell it what to do next? In a new paper, we introduce a reasoning model that predicts what you’ll do next over long contexts (LongNAP 💤). We trained it on 1,800 hours of computer use from 20 users. 🧵

English
0
3
20
2.6K
Michael Bernstein retweetledi
Omar Shaikh
Omar Shaikh@oshaikh13·
What’s the point of a “helpful assistant” if you have to always tell it what to do next? In a new paper, we introduce a reasoning model that predicts what you’ll do next over long contexts (LongNAP 💤). We trained it on 1,800 hours of computer use from 20 users. 🧵
English
17
81
308
120.5K