Nitarshan

1.6K posts

Nitarshan banner
Nitarshan

Nitarshan

@nitarshan

International policy @anthropic, PhD @cambridge_cl. Co-creator of @aisecurityinst, AI Safety Summit, UK AI Research Resource, EU AI Code of Practice.

SF / London Katılım Mayıs 2012
2.3K Takip Edilen2.3K Takipçiler
Sabitlenmiş Tweet
Nitarshan
Nitarshan@nitarshan·
The West has a closing window to win on AI. In our @JoinFAI article, @saroshnagar, @scott_r_singer and I argue that our leadership in AI requires "full-stack diffusion" to promote our entire AI stack globally. 1/6
English
5
9
62
31.3K
Nitarshan retweetledi
kamilė
kamilė@kamilelukosiute·
AI models' cyber capabilities keep getting meaningfully better, and fast. To determine how AI capabilities will impact cybercrime, we first need a baseline for global cybercrime damages. In a new @GovAIOrg technical report with John Halstead and @lucafrighetti, we arrive at a baseline estimate of global cybercrime damages: $500B (with 90% CI of $100B-$1T) per year. Existing estimates of global cybercrime damages range from tens of billions to tens of trillions of dollars. Most have serious problems: they rely on reported damages only (missing the vast majority of incidents that go unreported), or they don't publish their methodology at all. We tried to do better by extrapolating mostly from survey data, which captures unreported incidents, and by being transparent about every assumption we make. Our total estimate: ~$500B a year. This includes direct losses to individuals, direct + response costs to businesses, and defensive spending. Notably, this does not include costs that are even harder to quantify, such as IP theft, espionage, and national security costs, so the real yearly damages are presumably higher. As AI gets better at cyber, even a modest additive effect on the volume of cybercrime is a big deal. A 20% increase would mean ~$100B in additional yearly damages. Our estimates have extremely high uncertainty ranges. If we want to understand how AI is shaping cybercrime, we'll need to build new ways of measure the effects by looking at real world indicators of threat actor AI usage. Read the full report here: governance.ai/research-paper…
kamilė tweet media
English
3
13
56
6.5K
Nitarshan retweetledi
NASA Administrator Jared Isaacman
NASA does not have a top-line problem. We receive roughly $25 billion in annual appropriations, including more than a $10 billion plus-up from President Trump’s One Big Beautiful Bill. If that is not enough to run a lunar exploration program and do all the other things across science and discovery, then what is the right number? We don’t need to blame budgets or continuity of decision-making as the common excuse, as if a billion dollars is somehow not a billion dollars and troubled programs should perpetually stay troubled programs. NASA, like the federal government, cannot spend our way out of every problem, nor can we perpetuate bad decisions. That means not getting spread thin across too many imposed endeavors or jumping straight to the “dream state,” which is how everything becomes over budget and behind schedule. Instead, we concentrate on the needle-moving objectives, the reason NASA exists in the first place. We execute with urgency, in an iterative and safe way, and empower the workforce and our partners to get the job done. That is how we changed the world on July 20, 1969, and it is how we will do it again. Expect more from NASA and start believing again.
English
92
231
3.1K
107.7K
Nitarshan retweetledi
Andrew Gordon Wilson
Andrew Gordon Wilson@andrewgwils·
There was a new language. A lot of signaling, branding, tribalism, politics. A fresh wave of people who controlled immense resources but didn't know much about what came before. And some research that felt narratively driven, but only loosely connected with the stated goals. 5/6
English
1
4
82
7.8K
Nitarshan retweetledi
Xander Davies
Xander Davies@alxndrdavies·
The Red Team at @AISecurityInst is hiring! We work with frontier AI companies to red team their misuse safeguards, control measures, and alignment techniques. As the stakes rise, we need much stronger red teaming and many more talented researchers working within gov 🧵
Xander Davies tweet media
English
3
35
232
66.9K
Nitarshan retweetledi
Anthropic
Anthropic@AnthropicAI·
We partnered with Mozilla to test Claude's ability to find security vulnerabilities in Firefox. Opus 4.6 found 22 vulnerabilities in just two weeks. Of these, 14 were high-severity, representing a fifth of all high-severity bugs Mozilla remediated in 2025.
Anthropic tweet media
English
485
1.4K
15.2K
3.2M
Nitarshan retweetledi
roon
roon@tszzl·
@memeticweaver @tautologer > the USG can in general do whatever they want the founders of this great nation fought several bloody wars to make sure this is not true
English
22
38
938
51.3K
Nitarshan retweetledi
Dean W. Ball
Dean W. Ball@deanwball·
the word of the week is “alas”
English
1
2
44
3.2K
Nitarshan retweetledi
pamela mishkin
pamela mishkin@manlikemishap·
:party-parrot: slow to reply due to issues of national security :pray:
English
0
1
22
1.2K
Nitarshan retweetledi
Nathan Calvin
Nathan Calvin@_NathanCalvin·
Not sure without seeing full text, but it seems to me there are two options here: 1. This updated deal does not protect anthropics redlines (which are the same as OAIs) 2. This deal does protect them. if it does, then why was Anthropic treated so much worse by the admin?
Nathan Calvin tweet media
Sam Altman@sama

Here is re-post of an internal post: We have been working with the DoW to make some additions in our agreement to make our principles very clear. 1. We are going to amend our deal to add this language, in addition to everything else: "• Consistent with applicable laws, including the Fourth Amendment to the United States Constitution, National Security Act of 1947, FISA Act of 1978, the AI system shall not be intentionally used for domestic surveillance of U.S. persons and nationals. • For the avoidance of doubt, the Department understands this limitation to prohibit deliberate tracking, surveillance, or monitoring of U.S. persons or nationals, including through the procurement or use of commercially acquired personal or identifiable information." It’s critical to protect the civil liberties of Americans, and there was so much focus on this, that we wanted to make this point especially clear, including around commercially acquired information. Just like everything we do with iterative deployment, we will continue to learn and refine as we go. I think this is an important change; our team and the DoW team did a great job working on it. 2. The Department also affirmed that our services will not be used by Department of War intelligence agencies (for example, the NSA). Any services to those agencies would require a follow-on modification to our contract. 3. For extreme clarity: we want to work through democratic processes. It should be the government making the key decisions about society. We want to have a voice, and a seat at the table where we can share our expertise, and to fight for principles of liberty. But we are clear on how the system works (because a lot of people have asked, if I received what I believed was an unconstitutional order, of course I would rather go to jail than follow it). But 4. There are many things the technology just isn’t ready for, and many areas we don’t yet understand the tradeoffs required for safety. We will work through these, slowly, with the DoW, with technical safeguards and other methods. 5. One thing I think I did wrong: we shouldn't have rushed to get this out on Friday. The issues are super complex, and demand clear communication. We were genuinely trying to de-escalate things and avoid a much worse outcome, but I think it just looked opportunistic and sloppy. Good learning experience for me as we face higher-stakes decisions in the future. In my conversations over the weekend, I reiterated that Anthropic should not be designated as a SCR, and that we hope the DoW offers them the same terms we’ve agreed to. We will host an All Hands tomorrow morning to answer more questions.

English
7
8
84
4.4K
Nitarshan retweetledi
Dean W. Ball
Dean W. Ball@deanwball·
It is so clear that the important fissure in AI politics right now is not “liberal vs. conservative,” “Democrat vs. Republican,” “e/acc vs. EA,” or “safety vs. anti-safety,” but instead “takes advanced AI seriously as a concept vs. does not take advanced AI seriously.”
English
64
131
1.3K
223.1K
Nitarshan retweetledi
Élysée
Élysée@Elysee·
L’esprit français. 🇫🇷
Français
1.6K
3.8K
37.7K
6.8M
Nitarshan retweetledi
Andrew Curran
Andrew Curran@AndrewCurran_·
Claude has fans in DHHS.
Andrew Curran tweet media
English
12
18
165
59.5K
Nitarshan retweetledi
AJ Kourabi
AJ Kourabi@aj_kourabi·
@globeandmail I don’t think I could’ve come up with a worse idea if I tried
English
1
1
19
1.2K
Nitarshan retweetledi
Dean W. Ball
Dean W. Ball@deanwball·
I think this one needs no further explanation.
Dean W. Ball tweet media
English
33
113
811
445.1K
Nitarshan retweetledi
Peter Henderson
Peter Henderson@PeterHndrsn·
Apropos of nothing, some great researchers recently showed that you can use LLMs with internet access to successfully de-anonymize data at scale.
Peter Henderson tweet media
English
14
143
606
68.4K
Nitarshan retweetledi
Nitarshan retweetledi
roon
roon@tszzl·
there is no contractual redline obligation or safety guardrail on earth that will protect you from a counterparty that has its own secret courts, zero day retention, full secrecy on the provenance of its data etc. every deal you make here is a trust relationship
English
157
43
1.4K
127.1K
Nitarshan retweetledi
Ethan Mollick
Ethan Mollick@emollick·
I know it is a small thing, but, in these dying days of the open web, it is lovely that such a large proportion of famous poetry is online, mostly due to a $100M gift from Ruth Lily, who loved poetry (even though she never got any of her own published) poetryfoundation.org/poems/guides
English
19
169
1.8K
127.9K