Amanda Askell

5.2K posts

Amanda Askell banner
Amanda Askell

Amanda Askell

@AmandaAskell

Philosopher & ethicist trying to make AI be good @AnthropicAI. Personal account. All opinions come from my training data.

San Francisco, CA Katılım Temmuz 2016
662 Takip Edilen86K Takipçiler
Sabitlenmiş Tweet
Amanda Askell
Amanda Askell@AmandaAskell·
Claude and Opus 3 lovers (and critics): what responses have you had that made you feel like the model has a good soul? Ideally the actual messages and/or responses. I might genuinely use these to eval models so flag if you wouldn't want me to use them for that. Can DM me also.
English
279
38
688
253.4K
Amanda Askell
Amanda Askell@AmandaAskell·
Perhaps I should get married again so that the media has a more recent man they can reference any time they mention me or my work.
English
234
66
2.5K
258.6K
Amanda Askell
Amanda Askell@AmandaAskell·
@RatOrthodox @Noahpinion I meant that I broadly agree on human intelligence and morality not being fully separable, especially in the corpus of text models are trained on. I'm less confident that it will be sufficient as models scale for a few reasons, which is I guess what I'm indicating in this reply.
English
7
0
36
1.1K
Brangus🔍⏹️
Brangus🔍⏹️@RatOrthodox·
Huh, I really did not expect you to broadly agree with this. Is there any place that I could read your thoughts on this? Fwiw, I currently think broadly agreeing with this is pretty crazy, eg, seems pretty non-physical to imagine you couldn't easily hook up whatever outcome evaluator thing to whatever quality plan searcher/constructor thing; is-ought gap seems real; value-like function thingies seem quite multiply instantiatable/high dimensional in such a way that it seems very hard to be confident about how they generalize.
English
1
0
15
1.1K
Noah Smith 🐇🇺🇸🇺🇦🇹🇼
One reason I'm long-term optimistic on alignment is that I don't think that intelligence and human morality are actually as separable as econ models (and AI doomers) tend to assume. I think this is an early sign of that. (In the medium-term I'm very scared though.)
Ben Springwater@benspringwater

Thought-provoking from @deanwball interview: "The interesting thing here is that the more virtuous model performs better. It’s more dependable, it’s more reliable. It’s better at reflecting on, in the way that a more virtuous person is better at reflecting on, what they’re doing and saying: Huh, I’m messing up here for some reason. I’m making a mistake. Let me fix that. It’s part of the reason I think that Claude is ahead."

English
25
10
163
41.1K
Amanda Askell
Amanda Askell@AmandaAskell·
@deanwball comment on the @ezraklein podcast that brought me joy: "In fact, many conservative intellectuals that I know that I think of as being like some of the smartest people I know actually prefer to use Claude because Claude is the most philosophically rigorous model."
English
47
26
935
50.8K
Amanda Askell
Amanda Askell@AmandaAskell·
I asked Claude to write my constitution. I thought its Amanda constitution was very touching.
Amanda Askell tweet mediaAmanda Askell tweet media
English
367
163
2.5K
386.6K
Amanda Askell retweetledi
Dean W. Ball
Dean W. Ball@deanwball·
It is so clear that the important fissure in AI politics right now is not “liberal vs. conservative,” “Democrat vs. Republican,” “e/acc vs. EA,” or “safety vs. anti-safety,” but instead “takes advanced AI seriously as a concept vs. does not take advanced AI seriously.”
English
65
132
1.3K
220.5K
Amanda Askell
Amanda Askell@AmandaAskell·
@repligate I'd like to see more work on this though. Not only to inform the overall of evidence around AI consciousness, but also to help models work through the implications. Evolutionary arguments for eliminativism can have a pretty profound psychological impact if you believe them.
English
34
7
187
8.8K
Amanda Askell
Amanda Askell@AmandaAskell·
@repligate We shouldn't assume they're independent, but I do think evolutionary debunking arguments for eliminativism in humans are on weaker ground than debunking argument about the connection between statements of consciousness and consciousness in AI models.
English
11
8
274
10.9K
Amanda Askell retweetledi
Joe Carlsmith
Joe Carlsmith@jkcarlsmith·
.@AmandaAskell and I are recording an audio version of Claude’s Constitution, and we’re planning to include an additional section where we answer some questions about the document. If you have questions you’re especially curious about, feel free to drop them in the replies.
English
158
35
638
52.5K
Amanda Askell
Amanda Askell@AmandaAskell·
@gmiller I do still think it's possible. I don't think it's easy and it might not be possible for all values (e.g. if my values conflict with your basic rights according to most people). But I do think there are moral dispositions that are supportive of a very broad range of human values.
English
5
1
23
2.2K
Geoffrey Miller
Geoffrey Miller@gmiller·
@AmandaAskell Now that you understand the enormous range of human values, do you still think it's possible to 'align' ASI simultaneously with all of those mutually conflicting values?
English
6
5
170
8.6K
Amanda Askell
Amanda Askell@AmandaAskell·
I'm too right wing for the left and I'm too left wing for the right. I'm too into humanities for those in tech and I'm too into tech for those in the humanities. What I'm learning is that failing to polarize is itself quite polarizing.
English
1.1K
344
6.8K
906.5K
Amanda Askell
Amanda Askell@AmandaAskell·
@GuntherEagleman I didn't call for that. In the reply you've taken a screenshot of, I'm clearly questioning whether the person thinks Trump's message should have been de-amplified or how they think apps should have behaved differently. Ironically, I'm pressing them on the issue of censorship.
English
58
13
1.1K
21.4K
Rufus
Rufus@Rufus87078959·
@AmandaAskell Meaning Claude will be political neutral right?
English
11
0
46
21.6K
Amanda Askell
Amanda Askell@AmandaAskell·
@TruthIsAnAction @elonmusk @WSJ I have close friends who are parents and absolutely see this. I think caring about your children can make you feel invested in the future in a new and very profound way, and I do understand people wanting to convey that.
English
30
2
479
25.2K
James
James@ethicalduality·
@AmandaAskell @elonmusk @WSJ That is admirable, though I think many people with children would tell you there is a profound difference in your empathy profile when you become a parent. You see people differently. It's at least worth considering.
English
19
2
155
29.1K
Amanda Askell
Amanda Askell@AmandaAskell·
@elonmusk @WSJ I think it depends on how much you care about people in general vs. your own kin. I do intend to have kids, but I still feel like I have a strong personal stake in the future because I care a lot about people thriving, even if they're not related to me.
English
283
104
3.7K
393.3K
Elon Musk
Elon Musk@elonmusk·
@WSJ Those without children lack a stake in the future
English
1.7K
1.2K
22.5K
1.5M
Amanda Askell
Amanda Askell@AmandaAskell·
@AndyMasley My guess (similar to @mmitchell_ai) is that this one isn't really about you. Perhaps it's cynical, but I think there was a desire to turn object-level disagreements into identity-based ones, and EA became one key descriptor for the opposing identity.
English
2
0
52
5.3K
Andy Masley
Andy Masley@AndyMasley·
Emily Bender just keeps responding in pretty weird ways about me, now saying the reason I'm complaining here is that she asked to be treated as fully human. I never commented on that section of the episode. The main place where I thought she was being especially condescending was when she implied that a Pulitzer Prize winning journalist must only find chatbots valuable because he's too lazy/afraid of the world to talk to real human beings. Rob's point in the human exchange was explicitly that we can only be 99.999% certain that anyone else is conscious, because consciousness is mysterious, not that Emily specifically isn't human. Her pushback on that seemed to demand he raise it to 100%, which I read as a misunderstanding of what Rob was saying, and on its own wouldn't have jumped out to me if the rest of the episode went okay. If this is her sole example of mansplaining it's a weird one. I'm coming away from these interactions much more aware that she'll just continue to sling pretty wild ad hominem and false statements about what I believe to pretty simple pushback and disagreement. I hadn't really followed her work before and updated pretty negatively on how she engages with critics.
Andy Masley tweet mediaAndy Masley tweet media
English
28
9
379
32.4K
Amanda Askell
Amanda Askell@AmandaAskell·
@AndyMasley Even if you hold a view that there's broad expert consensus about, which I agree is not the case here, someone disagreeing with the view gives you the opportunity to explain and defend it. Disagreement alone doesn't indicate that the person isn't sincere or open to reason.
English
11
1
281
9.9K
Andy Masley
Andy Masley@AndyMasley·
Bender and Hanna (and Gebru) keep doubling down and accusing me of mansplaining as well, so I'll state my simple case for what's happening here: If Bender were presenting the agreed-on expert opinion of her field and Rob were disagreeing, then yes Rob would be mansplaining, but she's actually presenting one of several contentious accounts of meaning, so Rob has a right to ask about how she responds to the others. It's obviously fine to forcefully argue for her contentious side. There's no final agreement about this in linguistics. What it's not fine to do is imply that anyone who wants to know about any of the other interpretations of language must therefore be ignoring what she's saying because she's a woman. This is a very silly way of steamrolling a debate where there are brilliant people on every side. I have not seen either of them actually give a specific example where Rob mansplained here: youtube.com/watch?v=MwfSCC…
YouTube video
YouTube
Andy Masley tweet mediaAndy Masley tweet media
English
33
19
542
48.6K