Meaning Alignment Institute

58 posts

Meaning Alignment Institute banner
Meaning Alignment Institute

Meaning Alignment Institute

@meaningaligned

The Meaning Alignment Institute researches how to align AI, markets, and democracies with what people value.

Katılım Kasım 2020
56 Takip Edilen1.4K Takipçiler
Meaning Alignment Institute retweetledi
Oliver Klingefjord
Oliver Klingefjord@klingefjord·
OpenAI and Anthropic seem to have diverged on what makes AI trustworthy. @OpenAI bet on compliance with rules, @AnthropicAI bet on character. In a new blogpost, I argue Anthropic's move is the right one: character is a precondition for integrity, and integrity is what you need when the rules run out. (link below)
English
9
15
92
14.8K
Meaning Alignment Institute retweetledi
Joe Edelman 🥞
Joe Edelman 🥞@edelwax·
Jakob’s lab at Oxford would be an amazing place to work on Full-Stack Alignment!
Jakob Foerster@j_foerst

The term "AI alignment" is often used without specifying "to whom?" and much of the work on AI alignment in practice looks more like "AI controllability" without answering "who controls the controller?" (i.e. user or operator). One key challenge is that alignment is fundamentally a complex multi-agent problem that cannot be handled by locally aligning AI systems to specific institutions or individuals. (think e.g. of social dilemma where locally rational action leads to bad outcomes for everyone). Instead, we need new protocols and methods that allow alignment across the "entire stack" of our societies -- a problem setting that we coin "full-stack alignment". Crucially, these methods need to allow individuals and groups of people to robustly identify what they value and then use these insights to organise themselves towards those goals. Our first candidate solution are Thick Models of Value, which you can think of as the HTML standard for norms and values. It's a small step towards making technology that works for people and communities rather than the other way around. As a field AI has gotten to a point that optimisation works (RL, SSL), so the question of _what to optimise for_ is now absolutely key. Lastly - this paper raises as many questions as it provides answers and I am honored to have contributed a small part. If you like this line of work, please consider joining @FLAIR_Ox.

English
0
2
12
2.9K
Meaning Alignment Institute retweetledi
xuan (ɕɥɛn / sh-yen)
xuan (ɕɥɛn / sh-yen)@xuanalogue·
Ever since I started thinking seriously about AI value alignment in 2016-7, I've been frustrated by the inadequacy of utility+RL theory to account for the richness of human values. Glad to be part of a larger team now moving beyond those thin theories towards thicker ones.
xuan (ɕɥɛn / sh-yen) tweet media
English
4
19
133
11.1K
Meaning Alignment Institute retweetledi
Oliver Klingefjord
Oliver Klingefjord@klingefjord·
Aligning an AI system, or a recommender system, in isolation is playing whack-a-mole. The real issue we're facing is "full-stack", and requires solutions that tackles the problems on all levels
Joe Edelman 🥞@edelwax

In 2017, I was working to change FB News Feed's recommender to use “thick models of value” (per the paper we just released). @finkd even promised he'd make Facebook “Time Well Spent”. That effort was thwarted by the (1) market dynamics of the attention economy, (2) the US congress’ focus on Cambridge Analytica, and (3) @meta's corporate governance. The problem was bigger than I'd thought: what we've now termed “full-stack alignment.”

English
0
4
12
1.6K
Meaning Alignment Institute retweetledi
Ryan Lowe 🥞
Ryan Lowe 🥞@ryan_t_lowe·
I guess now is also a good time to announce that I've officially joined @meaningaligned!! I'll be working on field building for full-stack alignment -- helping nurture this effort into a research community with excellent vibes that gets shit done weeeeeeeeeee 🚀🚀
Ryan Lowe 🥞@ryan_t_lowe

Introducing: Full-Stack Alignment 🥞 A research program dedicated to co-aligning AI systems *and* institutions with what people value. It's the most ambitious project I've ever undertaken. Here's what we're doing: 🧵

English
2
3
53
4.9K
Meaning Alignment Institute retweetledi
Joe Edelman 🥞
Joe Edelman 🥞@edelwax·
In 2017, I was working to change FB News Feed's recommender to use “thick models of value” (per the paper we just released). @finkd even promised he'd make Facebook “Time Well Spent”. That effort was thwarted by the (1) market dynamics of the attention economy, (2) the US congress’ focus on Cambridge Analytica, and (3) @meta's corporate governance. The problem was bigger than I'd thought: what we've now termed “full-stack alignment.”
Ryan Lowe 🥞@ryan_t_lowe

Why do we need to co-align AI *and* institutions? AI systems don't exist in a vacuum. They are embedded within institutions whose incentives shape their deployment. Often, institutional incentives are not aligned with what's in our best interest.

English
1
7
23
4.9K
Meaning Alignment Institute retweetledi
Ryan Lowe 🥞
Ryan Lowe 🥞@ryan_t_lowe·
Introducing: Full-Stack Alignment 🥞 A research program dedicated to co-aligning AI systems *and* institutions with what people value. It's the most ambitious project I've ever undertaken. Here's what we're doing: 🧵
Ryan Lowe 🥞 tweet media
English
14
44
210
40.1K
Meaning Alignment Institute retweetledi
Meaning Alignment Institute retweetledi
Joe Edelman 🥞
Joe Edelman 🥞@edelwax·
A big part of why AI is threatening is: market forces. Just look at what the 'attention economy' did to social media, or the short-term wins of LLM sycophancy, or the product races among AI labs, or the markets for AI boyfriends and girlfriends. What can we do about this?
English
4
12
42
9.5K
Meaning Alignment Institute
Meaning Alignment Institute@meaningaligned·
Markets optimize for engagement, not human flourishing. In an AI-driven economy, this reduces humans to passive, dependent consumers without meaning or agency. AI-enabled "market intermediaries" could fix this. New blogpost 👇
English
1
4
9
1.3K