Evan Hubinger

689 posts

Evan Hubinger banner
Evan Hubinger

Evan Hubinger

@EvanHub

Alignment Stress-Testing lead @AnthropicAI. Opinions my own. Previously: MIRI, OpenAI, Google, Yelp, Ripple. (he/him/his)

California Katılım Mayıs 2010
3.2K Takip Edilen9.4K Takipçiler
Evan Hubinger retweetledi
dave kasten
dave kasten@David_Kasten·
Yeah, @EvanHub is one of the most cautious and careful people I know for making claims like this; he'll argue you blue about a small factual or conceptual detail because he cares so much about getting stuff right... and he's very comfortable making claims like this. DC really, really, really needs to pay attention.
Andrew Curran@AndrewCurran_

'But the rate of change is such that Anthropic co-founder and chief science officer Jared Kaplan, as well as some external experts, believes fully automated Al research could be as little as a year away.'

English
1
8
91
11.5K
Evan Hubinger retweetledi
Agus 🔸
Agus 🔸@austinc3301·
lmao at how this aged
David Sacks@DavidSacks

A new Bloomberg story describes Anthropic as a “White House target” after my tweet the other day. Nothing could be further from the truth. Just a couple of months ago, the White House approved Anthropic’s Claude app to be offered to all branches of government through the GSA App Store. In fact, it has been Anthropic’s government affairs and media strategy to position itself consistently as a foe of the Trump administration. For example: — Anthropic’s founder publicly called Donald Trump a “feudal warlord” during the election. — Anthropic placed op-eds attacking the One Big Beautiful Bill, the President’s Middle East deals, and his chip export policies. — Anthropic hired senior Biden AI staffers to work in its government affairs department, where they continue to lobby for the old Biden AI agenda. — Anthropic reportedly even fired two law firms for reaching settlements with the Trump administration. — Finally, Anthropic has become the leading fomenter of state AI regulation, which is damaging to the startup ecosystem. This was the point of my tweet yesterday, and it was accurate. It’s a free country and Anthropic is welcome to its views. Oppose us all you want. We’re the side that supports free speech and open debate. But don’t whine to the media that you’re being “targeted” when all we’ve done is articulate a policy disagreement. It’s a classic liberal move to play the victim when a fraction of the criticism you dish out is directed back at you.

English
6
9
262
16.3K
Evan Hubinger retweetledi
Jacques
Jacques@JacquesThibs·
I’ve seen this comment made multiple times by OpenAI (below by @natseckatrina): “A lot of the concerns about the government's "all lawful use" language seem to stem from mistrust that government will follow the laws. At the same time, people seem to believe that Anthropic took an important stand by insisting on contract language around their redlines. We cannot have it both ways. Why is it reasonable to say that the government cannot be trusted to interpret laws and contracts the right way, but also to assume (without evidence) that Anthropic’s policy redlines, in a contract, would have been effective.” No! That is *not* the main concern. The concern is in fact that OpenAI’s contract carves out a situation exactly where the DoW can do “constrained monitoring” with legally bought third party data. The whole point of going beyond “all lawful purposes” and precisely pointing to no mass surveillance is because such things are lawful, but the law has failed to catch up to the pace of AI. The only reason we haven’t suffered from mass surveillance is because there was no technology that would scalably change that. LLMs change this. It has nothing to do with what is illegal. OpenAI has either intentionally cornered itself to allow surveillance with their models (the safety stack will be effectively safety theatre) because the DoW will just pressure OpenAI to relax any safety stack if it goes against the contract. And given the concern we have is with things that ARE legal, then OpenAI will either get sued or pressured to follow the contract. As @JTillipman mentioned in here last blog post: “This means that when the government buys AI commercially, the vendor’s standard terms and conditions, including its acceptable use policies, are the default starting point. **Restrictions on use are not some exotic demand by activist AI companies.** This is the consequence of the government buying commercial products on commercial terms.” The claim that we should just let the law decide restrictions goes against common practice in procurement! “The government’s punitive response to Anthropic compounds this problem. If the consequence of negotiating aggressively with the government is being designated a supply chain risk—a mechanism more commonly associated with foreign adversary threats—companies have strong incentives to simply accept whatever terms the government demands. OpenAI itself said it does not believe the supply chain risk designation should have been applied to Anthropic. That may lead to faster procurement, but it will yield worse governance outcomes. Companies that are afraid to negotiate are companies that will not push back when the government’s proposed terms are inadequate for either party. The question the public debate should be asking is not whether AI companies have the right to tell the Pentagon what to do. They do (within the limits of the contract they negotiate), depending on the acquisition pathway, contract type, and the terms the parties agree to. The question is whether the government’s current approach to AI procurement produces contracts that adequately protect the public interest. Based on the evidence, the answer is no. The Anthropic-Pentagon dispute, for all the attention it has received, is a symptom of that deeper problem, not its cause.” There is no part of the selectively shared part of the contract that goes against this. As of now, it is difficult not to believe that we are being deceived by OpenAI on the agreements of the contract or simply that we don’t have nearly sufficient information to make the inference they arguing for.
Jacques@JacquesThibs

The government can already legally buy your location data, browsing history, and social media activity without a warrant. The only thing that prevented mass surveillance from that data was the inability to process it all. LLMs fix that. “All lawful purposes” includes this.

English
5
3
68
5.4K
Evan Hubinger retweetledi
dave kasten
dave kasten@David_Kasten·
This is also an interesting update: Anthropic staff say their natsec model has technical safeguards akin to much of what OpenAI says they will build, despite a series of very confident statements by OpenAI staff that it did not. Given other contemporaneous statements by OpenAI that they had to correct afterwards, I know which statement I'm currently putting more credibility in.
sam mcallister@sammcallister

@aidan_mclau @scrollvoid This isn't true. Anthropic hasn't offered a "helpful-only" model without safeguards for NatSec use. Claude Gov is a custom model with extra training, including technical safeguards. (We've also had FDEs and researchers implementing it, and we run our own classifier stack.)

English
7
19
291
28.7K
Evan Hubinger retweetledi
sam mcallister
sam mcallister@sammcallister·
@aidan_mclau @scrollvoid This isn't true. Anthropic hasn't offered a "helpful-only" model without safeguards for NatSec use. Claude Gov is a custom model with extra training, including technical safeguards. (We've also had FDEs and researchers implementing it, and we run our own classifier stack.)
English
15
37
554
127.3K
Evan Hubinger retweetledi
Jason Wolfe
Jason Wolfe@w01fe·
"The United States is winning the AI competition because of its commitment to free enterprise and the rule of law; undermining that commitment to punish one company is short-sighted and antithetical to our national security interests. We urge the Department of War to withdraw its supply chain risk designation and resolve this dispute through normal commercial channels." I signed app.dowletter.org, if you agree with the above please consider signing as well.
English
7
23
173
11.7K
Evan Hubinger retweetledi
Shakeel
Shakeel@ShakeelHashim·
In the last few days, OpenAI and its executives have claimed that its DoW deal prevents its models being used for mass domestic surveillance. As I write in a lengthy explainer for @ReadTransformer today, that appears to be misleading at best.
Shakeel tweet media
English
5
19
155
43.3K
Evan Hubinger retweetledi
Shakeel
Shakeel@ShakeelHashim·
Very important piece that confirms what I've suspected the last couple days: "If you look line-by-line at the OpenAI terms, the source said, every aspect of it boils down to: If it’s technically legal, then the US military can use OpenAI’s technology to carry it out."
Shakeel tweet media
Hayden Field@haydenfield

NEW: When OpenAI announced its Pentagon deal Friday night, people immediately challenged Sam Altman's claims. Why, they asked, would the DoD suddenly agree to red lines when it had said it would never do so? The answer, sources told me, is that it didn't. theverge.com/ai-artificial-…

English
2
16
136
6.7K
Evan Hubinger retweetledi
Adam Cochran (adamscochran.eth)
Well there is part of what the DoD Anthropic was about. Based on the contract language around analyzing bulk commercial data and deanonymizing it matches with this data discussion: Since 2021 the Pentagons DIA has been purchasing anonymized and harvested geolocation data that’s used in advertising, arguing it’s not “spying” since it’s commercial. They’ve now realized AI is strong enough to take this bulk data and de-anonymize it accurately. Anthropic deemed that spying on Americans. OpenAI doesn’t.
tyson brody@tysonbrody

In 2021 the Pentagon's Defense Intelligence Agency told Senator Wyden it was purchasing geolocation data from commercial brokers harvested from cell phones and that it did not believe it needed a warrant to analyze American's data. This has to be part of what freaked Dario out.

English
18
334
874
44.3K
Evan Hubinger retweetledi
Brad Carson
Brad Carson@bradrcarson·
@jasonkwon @ARozenshtein Oh no, I'm just seeing this, and it's not right, I fear. I took @jasonkwon up on his offer and asked my Council of 8 LLMs to evaluate the claim. Their judgment was harsh:
Brad Carson tweet media
English
1
7
129
7.1K
Evan Hubinger retweetledi
Agus 🔸
Agus 🔸@austinc3301·
If you have technical safeguards without contractual language to back them up, you’re asking to be in a jailbreak competition with the military. Even if you catch them red-handed, you can’t make them stop: Jailbreaks are legal, so they're presumably included in “all lawful use”.
Jason Kwon@jasonkwon

you’re conflating two things there are DoW restrictions, which are bound to lawful use. there’s what we deliver technically, which is not required to be for every lawful use. there’s no “breach.” also, whatever Anthropic gets in a contract, they have to decide what matters outside of the existing law, and what to enforce, and when. that’s trust me bro, but in a way that sits outside the democratic process.

English
4
5
111
8.3K
Evan Hubinger retweetledi
dave kasten
dave kasten@David_Kasten·
This is a notable and important update. Apparently Anthropic offered "FISA yes, commercially acquired data no" and got turned down. This, uh, makes me substantially doubt the OpenAI claims that they've excluded NSA from their contract successfully.
Prakash@8teAPi

The real question is whether OpenAI is going to allow the use of AI on unclassified commercial bulk data on Americans, which is what the Pentagon wanted from Anthropic. Ant instead narrowed to classified FISA only, and got kicked.

English
14
113
795
63.3K
Evan Hubinger retweetledi
Lawrence Chan
Lawrence Chan@justanotherlaw·
New @slatestarcodex post: "Some alert ACX readers have done a deep dive into national security law to try to untangle the [DoW-OAI] situation... Current laws against domestic mass surveillance and autonomous weapons have wide loopholes in practice." astralcodexten.com/p/all-lawful-u…
English
1
13
73
4.6K
Evan Hubinger retweetledi
Nick Moran
Nick Moran@NickEMoran·
@binarybits @USWREMichael I mean it pretty explicitly confirms that what OpenAI offered that Anthropic did not was using models to surveil Americans through warantless data collection, as long as that collection is one of the various 12333 carve-outs. Probably not something DoD wants said aloud.
English
1
4
29
1.5K
Evan Hubinger retweetledi
dave kasten
dave kasten@David_Kasten·
This is an important point from Logan Koepke: OpenAI is claiming that DoW lacks authorities to get commercial data at scale, despite extensive reporting that they have done so
logan koepke@jlkoepke

@natseckatrina @David_Kasten @sama on point two, they have in fact done this and claim they have the authority to do this. • vice.com/en/article/us-…nytimes.com/2021/01/22/us/…static01.nyt.com/newsgraphics/d…

English
2
33
212
27K
Evan Hubinger retweetledi
tae kim
tae kim@firstadopter·
HELL NO! Anthropic is in the right here. The Atlantic: “On Friday afternoon, Anthropic learned that the Pentagon still wanted to use the company’s AI to analyze bulk data collected from Americans.” “That could include information such as the questions you ask your favorite chatbot, your Google search history, your GPS-tracked movements, and your credit-card transactions, all of which could be cross-referenced with other details about your life. Anthropic’s leadership told Hegseth’s team that was a bridge too far, and the deal fell apart. Soon after, Hegseth directed the U.S. military’s contractors, suppliers, and partners to stop doing business with Anthropic.”
English
14
39
320
71.8K