Sabitlenmiş Tweet
Ryan Keisler
370 posts

Ryan Keisler
@RyanKeisler
AI weather @brightbandtech. Previously @KoBold_Metals, @DescartesLabs, cosmology @UChicago & @Stanford. Sensors, simulation, & ML.
New Mexico, USA Katılım Haziran 2013
778 Takip Edilen1.8K Takipçiler

@jcamdr70 @benedictk__ Agreed, it's a good and under-explored direction. Here they used an AI weather model to discover a problem with the gold-standard record of the atmospheric state: arxiv.org/abs/2601.04701
English

@RyanKeisler @benedictk__ Very interesting, congratulations!
This raises a fascinating question: how do we identify what the AI model possesses that the physical model lacks?
English

@benedictk__ It has skill, i.e. better than the avg weather, out to 9 or 10 days. When I released the paper in 2022, this was the SOTA AI model. Fast forward to 2026, and the top (open) AI models are from Google DeepMind and ECMWF, and are generally better than the best physics-based model.
English

@TeksEdge Good question! I haven't profiled it, but I bet it will work on most laptops. I would start with CPU only (the default). You'll need to first install git and uv. If you try it and run into any issues, please LMK via a github issue. #installation" target="_blank" rel="nofollow noopener">github.com/rkeisler/keisl…
English

@RyanKeisler Congratulations! What are the minimum requirements for model/GPU support? I'd love to predict my weather.
English

@dcxStep Yea, you've got it. The only detail I would add is red = positive derivative = a perturbation to U at some point x reinforces a same-signed perturbation to U at the cursor point 24h later; blue = negative derivative = reverse-sign perturbation at cursor point.
English

@RyanKeisler This paper was such a breakthrough! Reading it was the first time I believed that SOTA pure-AI weather prediction was possible.
Thanks for sharing, Ryan.
English

@paul_skeie I knew that I wanted to highlight the forecast sensitivities because they're so fast to compute (I think this was 5 minutes?) and under-visualized. Then I happened to land on d(u500)/d(u500) as a nice one.
English

@RyanKeisler It is very cool what you did there. How did you come up with the idea to do that?
English

My hope is that, as a relatively lightweight model, this can serve as a starting point for anyone interested in ML weather prediction.
github.com/rkeisler/keisl…
English

@shoyer I thought OmniCast by @tungnd_13 and collaborators was an interesting step away from autoregressive training. arxiv.org/abs/2510.18707
English

Genuine question for those who train LLMs -- are there any effective strategies for going beyond next-few-token prediction for pre-training?
This is serious issue for building AI climate models, which are currently trained on weather forecasting (predicting forward a few days).
Reflection@reflection_ai
Most approaches to “agentic AI” focus on post-training fixes. In this conversation, member of our technical staff, @achowdhery argues the bottleneck is pre-training itself. Drawing on her work on PaLM and early Gemini, she explains why next-token prediction breaks down for long-horizon planning -- and how objectives, attention, and training data must evolve to support true agentic behavior.
English
Ryan Keisler retweetledi

Our internship program at Polymathic is open for opportunities from now through fall 2025!
I believe our program provides an opportunity to work alongside some of the best researchers and engineering experts in the world — exploring the unknown of building foundation models for science, together.
These are full-time and paid positions in the Big Apple!
Interested? 👇
forms.gle/Jm9v4VxJrna7VR…
Deadline Nov 5!
English
Ryan Keisler retweetledi

I'm incredibly proud to share NeuralGCM, our new AI and physics based approach to weather and climate modeling with state-of-the-art accuracy, published today in @Nature:
nature.com/articles/s4158…
GIF
English

@shoyer @yuqirose @PierreGentine @DWatsonParris @AstroVivi @StephanMandt @SciPritchard Yes! I remember seeing these window-to-window jumps most strongly in the low-pressure levels of ERA5. I was initially surprised but it makes sense in hindsight since there is no overlap in the underlying obs data (except indirectly, as it informs initial guess for the next state)
English

@yuqirose @PierreGentine @DWatsonParris @AstroVivi @StephanMandt @SciPritchard Amusingly ERA5 fails this test -- if you look at high enough resolution with hourly output you can see the forecasts reset every 12 hours (corresponding to a new data assimilation window).
English
Ryan Keisler retweetledi

New open source release from my team at Google: Dinosaur, a differentiable dynamical core for global atmospheric modeling, written in JAX: github.com/google-researc…
Dinosaur is a core component of NeuralGCM and we hope it is useful for the weather/climate research community.
English

The idea would be, when the DM generates samples conditioned on current state, there is higher sample variance in the nodes that are more difficult to predict (eg Fig 5 of arxiv.org/abs/2309.01745 @thuereyGroup). If you squint, this starts to look like a well calibrated ensemble.
English




