AS

4.5K posts

AS banner
AS

AS

@agstrait

Not here anymore - at the other place.

London, England Katılım Mart 2013
1.1K Takip Edilen2.5K Takipçiler
Sabitlenmiş Tweet
AS
AS@agstrait·
Ayo - I'm finally off here. Keeping the account to retain my bookmarks, but no longer active. Join me here. bsky.app/profile/agstra…
English
0
0
2
485
AS retweetledi
Jared Moore
Jared Moore@jaredlcm·
Disturbing anecdotal reports of "AI psychosis" and negative psychological effects have been emerging in the news. But what actually happens during these lengthy delusional "spirals"? In our preprint, we analyze chat logs from 19 users who experienced severe psychological harm🧵👇
English
24
83
403
51K
AS retweetledi
Cas (Stephen Casper)
Cas (Stephen Casper)@StephenLCasper·
Did you know that one base model is responsible for 94% of model-tagged NSFW AI videos on CivitAI? This new paper studies how a small number of models power the non-consensual AI video deepfake ecosystem and why their developers could have predicted and mitigated this.
Cas (Stephen Casper) tweet media
English
1
11
40
7.9K
AS
AS@agstrait·
🤝 You’ll work with 2 other researchers and in collaboration with other gov departments. The first project is to create a problem book of methods to reduce these risks (building on papers.ssrn.com/sol3/papers.cf…). Crucially, you are not expected to view sensitive material directly.
English
0
0
0
18
AS
AS@agstrait·
🛠️ This is a technical role for an applied ML or security engineer. The work we anticipate could include building scalable ways to detect malicious LoRAs, exploring data filtering and other methods for reducing malicious fine-tuning, and other technical methods.
English
1
0
0
19
AS retweetledi
summerfieldlab @summerfieldlab.bsky.social
In a new paper, we examine recent claims that AI systems have been observed ‘scheming’, or making strategic attempts to mislead humans. We argue that to test these claims properly, more rigorous methods are needed.
summerfieldlab @summerfieldlab.bsky.social tweet media
English
4
24
85
17.1K
AS retweetledi
Saffron Huang
Saffron Huang@saffronhuang·
Newest @reboot_hq 🎙️post: @jessicadai_ and I discuss forecasting, and how people present unhelpful narratives about the future (mostly by picking on AI 2027, sorry guys) Why we should view the future as constructed, not predicted
Saffron Huang tweet media
English
4
11
56
4.1K
AS retweetledi
Josh Wolfe
Josh Wolfe@wolfejosh·
Apple just GaryMarcus'd LLM reasoning ability
Josh Wolfe tweet media
English
217
569
5K
3.5M
AS retweetledi
AI Security Institute
AI Security Institute@AISecurityInst·
Advanced AI systems require complex evaluations to measure abilities, but conventional analysis techniques often fall short. Introducing HiBayES: a flexible, robust statistical modelling framework that accounts for the nuances & hierarchical structure of advanced evaluations.
AI Security Institute tweet media
English
2
11
53
7.2K
AS retweetledi
Sayash Kapoor
Sayash Kapoor@sayashk·
How will AI impact the economy? Can we defend against misuse? What policies would mitigate the risks of AI? Thrilled to share that @random_walker and I are writing another book to tackle these questions! Today, we release a paper laying out our argument: AI as Normal Technology.
Sayash Kapoor tweet media
English
12
69
282
57.9K
AS
AS@agstrait·
I too find this really weird, mainly in that it shows the frontier of AI research is at risk of moving further away from producing useful, safe, reliable products. These seem like features, not bugs.
Billy Perrigo@billyperrigo

nice analogy from @jackclarkSF newsletter this week

English
0
0
2
257
AS retweetledi
AI Security Institute
AI Security Institute@AISecurityInst·
We've funded 20 new research projects to enhance AI security in critical infrastructure ⚡ Our Systemic AI Safety Grants Programme, announced at the Seoul AI Summit, has awarded up to £200,000 seed grants to projects tackling AI risks 🧵👇
English
1
11
60
9.2K
AS
AS@agstrait·
A great thread re: problematic extrapolations on claims about AI being superhuman at tasks. 1. Coding =/= all computer-related tasks, let alone all tasks 2. Generating code to complete a task =/= the most efficient, secure way to complete a task.
English
0
0
1
296
AS
AS@agstrait·
@peterwildeford Presumably they'll be similar kinds of answers as the last industrial revolutions, i.e. social innovations like labour protections, the 5 day work week, etc. that balanced societal interests with the interests of employers?
English
1
0
2
57
Peter Wildeford🇺🇸🚀
Peter Wildeford🇺🇸🚀@peterwildeford·
I agree. I'm troubled by how disruptive the societal impacts of AI automation may be. But it's very unclear to me what the correct policy response should be. Banning/restricting AI automation is not the correct long-term answer!
Séb Krier@sebkrier

On the one hand I'm glad more people are thinking about labour and automation. On the other I'm terrified that the default reaction will be calling for the equivalent of rent ceilings and other interventions that seem protective but ultimately create harumful perverse incentives.

English
5
2
42
3.7K
AS
AS@agstrait·
These include undesirable automation, over-reliance on AI systems, mental health impacts, mass generation of unreliable content, power concentration, and social destabilisation...and so much more.
English
0
0
0
40
AS
AS@agstrait·
I'll do a longer post about the new role, but in short, we're building a research team, funding programme, and partnerships to tackle crucial questions about advanced AI's societal impacts. We'll track how AI is being used across critical sectors, and study societal-level risks.
English
1
0
1
35
AS
AS@agstrait·
But I'm even more proud of the people we've worked with, the community we've fostered, and the culture we've built. I remember the first meeting I had with @carlykind_ about her vision for the org - I was all in. Looking back, we've accomplished so much more than we'd thought.
English
0
0
0
32
AS
AS@agstrait·
I'm so proud of Ada's accomplishments - from pioneering work on COVID-tech, deep dives into AI auditing and research ethics, supporting emerging AI policy and regulation, exploring risks of AI & genomic systems, foundation models and personal AI assistants...it's a very long list
English
1
0
0
43
AS
AS@agstrait·
@herbiebradley you coming to the AFTERS (advanced field training and enemy response simulation)
English
0
0
2
61
Herbie Bradley
Herbie Bradley@herbiebradley·
“i never see you at the club” okay well i never see you at the wargame
English
2
1
38
2.5K