Center for AI Safety

247 posts

Center for AI Safety banner
Center for AI Safety

Center for AI Safety

@CAIS

Reducing societal-scale risks from AI. https://t.co/5I9YG8IZa7 https://t.co/u91FCIyeSV

San Francisco Katılım Ağustos 2022
2 Takip Edilen9K Takipçiler
Center for AI Safety
The fellowship includes regular guest speaker events by professors at Stanford, Penn, Johns Hopkins, and more. Further details and a link to the application can be here: safe.ai/fellowship
English
2
3
17
1.8K
Center for AI Safety
Applications are now open for the AI and Society Fellowship at the Center for AI Safety: a fully funded, 3-month summer fellowship in San Francisco for scholars in economics, law, IR, and adjacent fields to research the societal impacts of advanced AI. Apply by March 24.
English
4
60
247
22.2K
Center for AI Safety
@ProfArbel Thank you for sharing. The fellowship also welcomes scholars in economics, IR, and adjacent fields to study the societal implications of advanced AI. It will also host speakers from Stanford, Penn, JHU, and more. Apply by March 24. For more information: safe.ai/fellowship
English
0
0
1
143
Yonathan Arbel
Yonathan Arbel@ProfArbel·
Hi law profs/fellows: Interested in the big social implications of AI? Want a funded summer fellowship in SF by some of the best in the business? @CAIS holds a call for summer fellowships that allow you to pursue your research tracks with close feedback loops
Yonathan Arbel tweet media
English
2
29
164
14.8K
Center for AI Safety
Center for AI Safety@CAIS·
While Gemini 3.1 Pro performs better than 3.0 Pro on a variety of safety benchmarks, its risk index remains high due to its susceptibility to jailbreaks.
Center for AI Safety tweet media
English
2
0
2
1.3K
Center for AI Safety
Center for AI Safety@CAIS·
New Gemini 3.1 Pro results: Large improvements in text reasoning, and the Gemini 3 series continues to dominate vision benchmarks. However, the model still has a higher risk index (worse safety) than each of its frontier peers.
Center for AI Safety tweet mediaCenter for AI Safety tweet mediaCenter for AI Safety tweet media
English
5
7
80
8.2K
Center for AI Safety
Center for AI Safety@CAIS·
Opus 4.6 is the most capable model on our text benchmarks, but less safe than its predecessor Text: #1 overall, strongest on coding and abstract reasoning Vision: #4, well behind Gemini 3 Safety: Worse than Opus 4.5 on bioweapons assistance, deception, and overconfidence
Center for AI Safety tweet mediaCenter for AI Safety tweet mediaCenter for AI Safety tweet media
English
4
5
58
11.9K
Center for AI Safety
AI agents are getting good at coding, but how close are they to automating all digital labor? New Remote Labor Index results: Opus 4.5 is able to automate 3.75% of remote labor projects, with GPT-5.2 in second place.
Center for AI Safety tweet media
English
12
48
405
110.4K
Center for AI Safety
Last week, Humanity’s Last Exam was published in @Nature. In just over a year, model scores on HLE have risen from under 5% to nearly 40%. Thank you to @scale_AI and the 1000+ HLE co-authors for helping policymakers and the public track these rapid advances in AI capabilities.
Center for AI Safety tweet mediaCenter for AI Safety tweet media
English
9
42
162
26.6K
Center for AI Safety
Humanity's Last Exam is now published in Nature. Since its release, HLE has become a leading frontier benchmark, used by OpenAI, Anthropic, DeepMind, and xAI. Thank you to our partners at @scale_AI and the 1,000+ co-authors who made this benchmark possible.
Center for AI Safety tweet media
English
3
15
94
6.9K