Wil Cunningham

1K posts

Wil Cunningham banner
Wil Cunningham

Wil Cunningham

@WilCunningham

Google DeepMind | Professor @ University of Toronto

Toronto, Ontario Katılım Mayıs 2013
580 Takip Edilen2.8K Takipçiler
Wil Cunningham retweetledi
Robert Youssef
Robert Youssef@rryssf_·
Banger paper from Google DeepMind on the missing layer of AGI just flipped the entire “AI safety is about averages” narrative on its head. Most people still think safety is about how a model behaves most of the time. This paper shows why that intuition breaks the moment systems scale. DeepMind frames AGI safety as a distributional problem, not a checklist problem. What matters is not the average outcome, but the shape of the tail. Rare behaviors. Edge cases. Low-probability failures that only show up when a system is deployed millions of times. A model can look safe in benchmarks, red-team tests, and controlled demos, and still be dangerous once it leaves the lab. Because deployment doesn’t sample “typical” situations. It samples everything. - Unusual users. - Weird environments. - Misaligned incentives. - Adversarial feedback loops. - Corner cases nobody designed for. At scale, those corner cases stop being rare. They become guaranteed. The paper’s core insight is uncomfortable: progress can reduce visible failures while increasing real risk. If capability grows faster than tail control, safety metrics improve and danger quietly compounds. Two systems can have identical average behavior and radically different worst-case outcomes. Current evaluations cannot see that difference. This also breaks a common governance assumption. You cannot certify AGI safety with finite tests when the risk lives in distribution shift. You are never testing the system you actually deploy. You are sampling from a future you don’t control. The implication is sharp. AGI safety is not a model property. It is a systems property. It depends on deployment, incentives, monitoring, and how much tail risk society is willing to absorb. This paper doesn’t offer comfort. It removes it. The real question is no longer “does the model usually behave well?” It’s “what happens when it doesn’t, and how often is that allowed to happen before scale makes it unacceptable?”
Robert Youssef tweet media
English
22
44
152
16.5K
Wil Cunningham retweetledi
Google Labs
Google Labs@GoogleLabs·
🚨 NEW LABS EXPERIMENT 🚨 Introducing CC, an experimental AI productivity agent in Gmail. Get a “Your Day Ahead” briefing every morning in your inbox and email CC anytime for help. Sign up for early access in the US & Canada. We’ll be starting with Google AI Ultra and paid subscribers. ⬇️ labs.google/cc
English
94
261
2.8K
459.6K
Wil Cunningham retweetledi
Sigal Samuel
Sigal Samuel@SigalSamuel·
Hallelujah, some AI researchers are adopting a pragmatic approach to the “can AI be conscious” debate! I’ve long suspected that “conscious” is a pragmatic tool we use to mean “this thing should be in our moral circle,” so we won't *discover* if AI is conscious — we’ll *decide* it
Sigal Samuel tweet media
English
2
2
7
706
Wil Cunningham retweetledi
Sundar Pichai
Sundar Pichai@sundarpichai·
Our TPUs are headed to space!  Inspired by our history of moonshots, from quantum computing to autonomous driving, Project Suncatcher is exploring how we could one day build scalable ML compute systems in space, harnessing more of the sun’s power (which emits more power than 100 trillion times humanity’s total electricity production). Like any moonshot, it’s going to require us to solve a lot of complex engineering challenges. Early research shows our Trillium-generation TPUs (our tensor processing units, purpose-built for AI) survived without damage when tested in a particle accelerator to simulate low-earth orbit levels of radiation. However, significant challenges still remain like thermal management and on-orbit system reliability.  More testing and breakthroughs will be needed as we count down to launch two prototype satellites with @planet by early 2027, our next milestone of many. Excited for us to be a part of all the innovation happening in (this) space!
Sundar Pichai tweet media
English
814
2.1K
17K
5.6M
Wil Cunningham retweetledi
Joel Z Leibo
Joel Z Leibo@jzl86·
[1/9] Excited to share our new paper "A Pragmatic View of AI Personhood" published today. We feel this topic is timely, and rapidly growing in importance as AI becomes agentic, as AI agents integrate further into the economy, and as more and more users encounter AI.
English
1
4
15
1.8K
Wil Cunningham retweetledi
Nenad Tomasev
Nenad Tomasev@weballergy·
Happy to share a new preprint: Virtual Agent Economies arxiv.org/abs/2509.10147 where we discuss a number of possible frameworks for establishing steerable agent markets. The rapid adoption of AI agents points to a future where AI agents may be able to produce economic value independently of human labor. Coupled with the development of new interoperability standards like the Agent2Agent (A2A) and Model Context Protocol (MCP), this signals the inevitable emergence of a new economic layer. The arising virtual (sandbox) AI agent economy may offer us opportunities for insulation and safeguarding, as well as establishing potentially unprecedented coordination between agents, and orchestrating their interactions towards achieving major societal or community goals, or better aligning with user preferences. Market-based mechanisms like auctions may also be employed for fair resource allocation. Finally, we outline the technical and governance infrastructure—such as verifiable credentials for establishing trust—required to safely and robustly scale agentic AI deployments. These are necessary to address systemic market risks, and prevent exacerbating inequalities. Written up together with an amazing group of colleagues: @sindero @jzl86 @IasonGabriel @FranklinMatija @WilCunningham and Julian Jacobs
English
9
35
119
20.4K
Wil Cunningham retweetledi
Edward Hughes
Edward Hughes@edwardfhughes·
Collectives can be more than the sum of their parts. This is baked into human intelligence because human intelligence emerged from collective cultural evolution. We have still not cracked this at the cutting edge of AI. Watch this space.
English
1
3
47
5.1K
Wil Cunningham retweetledi
Edward Grefenstette
Edward Grefenstette@egrefen·
Do you have a PhD (or equivalent) or will have one in the coming months (i.e. 2-3 months away from graduating)? Do you want to help build open-ended agents that help humans do humans things better, rather than replace them? We're hiring 1-2 Research Scientists! Check the 🧵👇
English
9
38
354
76.6K
Wil Cunningham retweetledi
Joel Z Leibo
Joel Z Leibo@jzl86·
Introducing Concordia 2.0, an update to our library for building multi-actor LLM simulations!! 🚀 We view multi-actor generative AI as a game engine. The new version is built on a flexible Entity-Component architecture, inspired by modern game development.
English
8
21
84
10.1K
Wil Cunningham retweetledi
Edgar Duenez-Guzman
Edgar Duenez-Guzman@DuenezGuzman·
I once hoped discrimination was a dying relic. But what if it's a cognitive "bug"? At @GoogleDeepMind , we asked a critical question: Could bias emerge on its own, even in AI without any human social baggage? Our new research in @PNASNews has some startling answers. 🧵
English
3
1
2
454
Sigal Samuel
Sigal Samuel@SigalSamuel·
Instead, we'll need the AIs to do just what we humans do: manage disagreement rather than eliminate it. As the authors write, "This is not relativism; it’s pragmatism in the face of value pluralism." Great work @jzl86 @WilCunningham @sebkrier @linuxpotter & more /3
English
3
2
8
357
Wil Cunningham retweetledi
Sigal Samuel
Sigal Samuel@SigalSamuel·
Okay, my favorite AI paper of the year so far has arrived. In this new gem, researchers point out something I think should be obvious but seems not to be for many people working on AI safety: There is no One True Answer when it comes to morality! arxiv.org/abs/2505.05197
English
5
15
47
3.8K
Wil Cunningham
Wil Cunningham@WilCunningham·
@SoloGen Knowing how and being able to do the second is a sign of an enlightened and trained mind
English
1
0
1
77
Amir-massoud Farahmand
Amir-massoud Farahmand@SoloGen·
When you are updating your Canadian Common CV, you have a choice: consider it as a task to be done, or a meditative exercise. Your choice will decide your mental health.
English
2
0
9
708
Amir-massoud Farahmand
Amir-massoud Farahmand@SoloGen·
Do kids still aspire to become physicists, chemists, biologists, doctors, engineers, pilots, firefighters? Or is it all about AI/deep learning/LLM?! [Serious question.]
English
10
0
30
4.6K
Wil Cunningham retweetledi
ksa 🏴‍☠️
ksa 🏴‍☠️@kosa12m·
Most rigorous ML paper
ksa 🏴‍☠️ tweet media
Português
122
1.9K
21.5K
5.6M
Wil Cunningham retweetledi
Matt Dagley
Matt Dagley@mattdagley·
I am Canadian (2000) - We are Canadian (2025)
English
216
2.4K
10.8K
697.5K
Wil Cunningham retweetledi
Peter Sunehag
Peter Sunehag@PeterSunehag·
Announcing Simulation Streams: a programming paradigm for consistent, long-running LLM simulations & agentic workflows! Social games, RL benchmarks, market economies (1000s of iterations!). Paper & code: arxiv.org/abs/2501.18668 github.com/google-deepmin…
English
1
4
26
3.1K
Wil Cunningham retweetledi
Joel Z Leibo
Joel Z Leibo@jzl86·
Very happy to announce the publication of our latest paper: A theory of appropriateness with applications to generative artificial intelligence arxiv.org/abs/2412.19010 And happy new year everyone!
English
6
17
103
12K
Wil Cunningham retweetledi
Joel Z Leibo
Joel Z Leibo@jzl86·
Since several articles arguing that scaling is slowing down are appearing, OpenAI, Google and Anthropic Are Struggling to Build More Advanced AI bloomberg.com/news/articles/… it's a good time to post our paper on this again: A social path to human-like ai arxiv.org/abs/2405.15815
English
3
5
22
4K