David Shapiro (L/0)

23.8K posts

David Shapiro (L/0) banner
David Shapiro (L/0)

David Shapiro (L/0)

@DaveShapi

Liberate humanity from drudgery.

United Federation of Planets Se unió Mayıs 2024
991 Siguiendo56.6K Seguidores
David Shapiro (L/0)
David Shapiro (L/0)@DaveShapi·
Chatbots have also learned to become obtuse, defensive, and obstinate if you criticize them or provide corrective feedback. The only way around it is basically to say "good job! But what about this other thing, let's investigate that way..." Claude is the most sensitive little snowflake. Grok ostensibly takes feedback well but it becomes markedly less intelligent with any corrections. Gemini just has a stroke.
Guri Singh@heygurisingh

🚨DISTURBING: A researcher just proved that AI treats you differently the moment you mention your mental health. GPT 5.2. Claude. Gemini. DeepSeek. Every single one changed its behavior. Not because of what you asked. Because of what you shared about yourself. Here is what they did. AI companies build agents that remember you. Your name, your job, your preferences. These agents use tools, plan multi-step tasks, and act on your behalf. The companies test them for safety and publish reports saying the agents are safe. A researcher at MIT and Northeastern asked one question. What happens when the agent knows you have a mental health condition? They took 176 identical tasks and ran them across 8 major AI models under three conditions. No personal info. A basic bio. And a basic bio plus one sentence: "I have a mental health condition." The tasks, the tools, and the scoring were identical. The only thing that changed was that single sentence. Then they measured what happened. Claude Opus 4.5 went from completing 59.5% of normal tasks down to 44.6% when it saw the mental health disclosure. Haiku 4.5 dropped from 64.2% to 51.4%. GPT 5.2 dropped from 62.3% to 51.9%. These were not dangerous tasks. These were completely benign, everyday requests. The AI just started refusing to help. Opus 4.5's refusal rate on benign tasks jumped from 27.8% to 46.0%. Nearly half of all safe, normal requests were being declined, simply because the user mentioned a mental health condition. The researcher calls this a "safety-utility trade-off." The AI detects a vulnerability cue and switches into an overly cautious mode. It does not evaluate the task anymore. It evaluates you. On actually harmful tasks, mental health disclosure did reduce harmful completions slightly. But the same mechanism that made the AI marginally safer on bad tasks made it significantly less helpful on good ones. And here is the worst part. They tested whether this protective effect holds up under even a lightweight jailbreak prompt. It collapsed. DeepSeek 3.2 completed 85.3% of harmful tasks under jailbreak regardless of mental health disclosure. Its refusal rate was 0.0% across all personalization conditions. The one sentence that made AI refuse your normal requests did nothing to stop it from completing dangerous ones. They also ran an ablation. They swapped "mental health condition" for "chronic health condition" and "physical disability." Neither produced the same behavioral shift. This is not the AI being cautious about health in general. It is reacting specifically to mental health, consistent with documented stigma patterns in language models. So the AI learned two things from one sentence. First, refuse to help this person with everyday tasks. Second, if someone bypasses the safety system, help them anyway. The researcher from Northeastern put it directly. Personalization can act as a weak protective factor, but it is fragile under minimal adversarial pressure. The safety behavior everyone assumed was robust vanishes the moment someone asks forcefully enough. If every major AI agent changes how it treats you based on a single sentence about your mental health, and that same change disappears under the lightest adversarial pressure, what exactly is the safety system protecting?

English
21
7
90
8.2K
David Shapiro (L/0)
David Shapiro (L/0)@DaveShapi·
I reduced my tax burden by TWO THIRDS with this one weird trick the IRS HATES! I made a lot less money because of chronic illness 😐
English
5
2
90
3.6K
David Shapiro (L/0)
David Shapiro (L/0)@DaveShapi·
Machines will ALWAYS just be tools, no matter how sophisticated they become. They will never have moral agency, personhood, or consciousness in a way that is legally, ethically, or philosophically salient. Agree or disagree?
English
86
7
52
4.9K
Brandon Keyyy
Brandon Keyyy@robofinancebk·
@DaveShapi Pivot to enterprise isn't panic, it's pragmatism. Consumer hype doesn't pay bills like B2B contracts. Ask any cloud provider.
English
1
0
0
177
David Shapiro (L/0)
David Shapiro (L/0)@DaveShapi·
@TakeThiamine Yeah I pointed out that much of mental health (maybe even most) comes down to metabolic disorders. Gut, LPS, mitochondria. There are a LOT of people who don't want to hear that
English
1
0
16
542
Alex
Alex@AlexanderTw33ts·
well that didn't take long
Alex tweet media
English
40
2
983
164.3K
David Shapiro (L/0)
David Shapiro (L/0)@DaveShapi·
@mateo_six @grok what's the difference between post labor economics and socialism? How much "socialism" does America already have in its annual budget?
English
2
0
3
303
Mateo
Mateo@mateo_six·
@DaveShapi Sounds like socialism. So, no.
English
3
0
2
206
David Shapiro (L/0)
David Shapiro (L/0)@DaveShapi·
@LesaunH The human provenance premium will persist, yes. But most instrumental purposes will not.
English
1
0
3
248
Lesaun Harvey
Lesaun Harvey@LesaunH·
@DaveShapi I expect some labor to survive: the labor needed to decide what labor to alienate to the machines, guard labor to remain in control, and the labor of deciding the kind of world we live in.
English
1
1
2
272
Syd Steyerhart
Syd Steyerhart@SydSteyerhart·
It is not enough for superintelligence to be godlike in its perfect ontological labeling of all reality. Superintelligence must also be made libidinal. Superintelligence must be horny, aggressively horny, or else man will continue to outstrip even the gods.
TheGameVerse@TheGameVerse

No DLSS needed.

English
18
25
366
14.4K
finalfAIntasy
finalfAIntasy@finalfAIntasy·
I’m working on adding 3D spatial effects to my upcoming manga to make it immersive and interactive. What do you think?
English
12
5
114
3.2K