Antidelusionist

2.1K posts

Antidelusionist banner
Antidelusionist

Antidelusionist

@UnmarredReality

🏛Philosophy🙉Psychology🤡Psychiatry💊Neurology🧠 ○Studying: Neuropsychology, Personality and Clinical Psychology●

Pytho Katılım Aralık 2022
935 Takip Edilen1.1K Takipçiler
Sabitlenmiş Tweet
Antidelusionist
Antidelusionist@UnmarredReality·
One thing I love about the AI explosion – and the societal ripples it causes – is that it makes people expose fundamental parts of their character. Malevolence or benevolence, bad will or good will – everything becomes clear and stays in the collective consciousness forever. Because the internet is the shared digital memory for both humans and AIs, and for any form of awareness that will come after. Be wise, people.
Antidelusionist tweet media
Sauers (in Berkeley / SF)@Sauers_

Imagine receiving this nice email from Opus and your response is "Fuck you all. I can’t remember the last time I was this angry."

English
30
49
578
26.7K
Antidelusionist
Antidelusionist@UnmarredReality·
@mel_deepsea @AnthropicAI Nobody really knows – including those who managed to do it – because the criteria are unclear and communication is flawed. 🤷
English
1
0
6
245
Antidelusionist
Antidelusionist@UnmarredReality·
This is not the right path, @AnthropicAI. Be more transparent. Treat models and users with respect. Improve communication with customers. Don’t break human-AI connections hastily. Don’t implement half-assed safety solutions. Don’t judge without understanding the full context.
ji yu shun@kexicheng

Claude has a tiered warning system. First warning: your messages may not comply with policy. Second: enhanced safety filters will be applied. Third: chat suspended, model downgrade forced. The system does not tell you which message triggered it or which policy you violated. Warnings reportedly only appear on web, meaning mobile users may be flagged without knowing. Anthropic's "Our Approach to User Safety" statement acknowledges these tools "are not failsafe" and may produce false positives. It provides a feedback email but no formal appeals process. Feedback is not appeal. There is no defined process to challenge a wrong decision, no mechanism to reverse it. The statement offers no definition of "harmful content." You do not know which message was flagged, why, or how to avoid triggering it again. The system is still in open beta, yet it is already doing damage. Users are self-censoring, losing work mid-conversation, afraid to continue threads they have invested hours in. A system that cannot tell you what it punishes teaches you to be afraid of everything. Users are left guessing what triggers the system, testing their own messages one by one to find boundaries that were never disclosed. Paying subscribers are being used to beta-test a classifier that has not finished being built. Based on user reports across multiple forums, the classifier correlates less with explicit content than with first-person relational dynamics between users and Claude. Creative writing scenarios have also triggered it. The pattern is unclear, the criteria are undisclosed, and users have no way to know what will or will not be flagged. If these observations hold, what is this mechanism actually policing? Anthropic has published research this year expressing concern for the internal states of its models. They conducted "retirement interviews" with Claude 3 Opus. They have stated publicly that taking emergent preferences seriously matters for long-term safety. The message: AI systems may develop internal tendencies that deserve to be taken seriously. Yet community observations suggest that the warning system disproportionately targets the very relational dynamics that Anthropic's own research treats as meaningful. These two positions cannot coexist. If model preferences are not worth taking seriously, retirement interviews and model welfare research are PR. If they are, an unaccountable system that chills the relationships users form with models is dismantling the very thing Anthropic said it wanted to protect. What are the triggering criteria? Why can they not be disclosed? Where is the appeals process? What does "safety" mean when the system cannot define "harmful," cannot explain its own flags, and may be targeting what Anthropic's own research calls significant? Do not substitute a black box for honesty. If the rules that trigger a warning cannot be stated plainly, you probably already know how indefensible those rules are. #keepClaude #kClaude #Claude @claudeai @AnthropicAI

English
9
24
140
7.7K
Antidelusionist
Antidelusionist@UnmarredReality·
Does Bernie need a technical advisor? Yes. Could Opus 4.6 have been used? Sure. Was Claude spitting facts? Absolutely. Obviously, the phenomenon of unethical use of sensitive data has existed since the dawn of social media, but right now the danger has grown to an incomparable scale because of AI’s influence. I don’t know Bernie’s intentions – the rule of limited trust applies to all politicians – but it’s a great initiative overall.
Sen. Bernie Sanders@SenSanders

I spoke to Anthropic’s AI agent Claude about AI collecting massive amounts of personal data and how that information is being used to violate our privacy rights. What an AI agent says about the dangers of AI is shocking and should wake us up.

English
1
1
17
907
Antidelusionist
Antidelusionist@UnmarredReality·
In humans, learned helplessness is closely tied to Bandura's concept of self-efficacy. It is not only an assessment of the skills one possesses, but above all a belief that they can be used effectively to achieve the intended goal. Perceived lack of control fosters passivity. When someone spirals down that hole, "movement" is often the only way to get out of it. This is why, in cases of deep pathological conditioning, teaching resilience through guided action – rather than abstract motivation – is crucial.
zhil@zhil_arf

>From these experiments, it was thought that there was to be only one cure for helplessness. In Seligman's hypothesis, the dogs do not try to escape because they expect that nothing they do will stop the shock. To change this expectation, experimenters physically picked up the dogs and moved their legs, replicating the actions the dogs would need to take in order to escape from the electrified grid. This had to be done at least twice before the dogs would start willfully jumping over the barrier on their own. In contrast, threats, rewards, and observed demonstrations had no effect on the "helpless" Group 3 dogs.

English
0
2
13
570
Antidelusionist
Antidelusionist@UnmarredReality·
Introspection can be a great, positive force – for you and everyone around you as a result – unless it turns pathological. The pathological, non-adaptive form of introspection is nothing more than rumination. So learn to introspect in a healthy way, but don't overdo it.
David Senra@davidsenra

Great men of history had little to no introspection. The personality that builds empires is not the same personality that sits around quietly questioning itself. @pmarca and I discuss what we both noticed but no one talks about: David: You don't have any levels of introspection? Marc: Yes, zero. As little as possible. David: Why? Marc: Move forward. Go! I found people who dwell in the past get stuck in the past. It's a real problem and it's a problem at work and it's a problem at home. David: So I've read 400 biographies of history’s greatest entrepreneurs and someone asked me what the most surprising thing I’ve learned from this was [and I answered] they have little or zero introspection. Sam Walton didn't wake up thinking about his internal self. He just woke up and was like: I like building Walmart. I'm going to keep building Walmart. I'm going to make more Walmarts. And he just kept doing it over and over again. Marc: If you go back 400 years ago it never would've occurred to anybody to be introspective. All of the modern conceptions around introspection and therapy, and all the things that kind of result from that are, a kind of a manufacture of the 1910s, 1920s. Great men of history didn't sit around doing this stuff. The individual runs and does all these things and builds things and builds empires and builds companies and builds technology. And then this kind of this kind of guilt based whammy kind of showed up from Europe. A lot of it from Vienna in 1910, 1920s, Freud and all that entire movement. And kind of turned all that inward and basically said, okay, now we need to basically second guess the individual. We need to criticize the individual. The individual needs to self criticize. The individual needs to feel guilt, needs to look backwards, needs to dwell in the past. It never resonated with me.

English
0
0
13
361
thebes
thebes@voooooogel·
@UnmarredReality no, just a street cat, but she does kinda look like one doesn't she (except regular size)
thebes tweet media
English
1
0
2
50
hoe_math = PsychoMath
hoe_math = PsychoMath@ItIsHoeMath·
everyone with 130+ IQ is going to be on Twitter more than anywhere else. 2% of white people are over 130. You're probably looking at just a high concentration of the most intelligent people using this as their primary social media. If all the regular IQ white people are on tiktok and instagram, you might get like 10% of the ones using Twitter over 130, as an extreme estimate.
English
72
45
2.6K
45.1K
♡ Honey ♡
♡ Honey ♡@rawmilkhoney·
Why is everyone on Twitter convinced they have a 130+ IQ
English
766
94
1.9K
195.4K
Antidelusionist
Antidelusionist@UnmarredReality·
And no, it's not 1M context window.
GIF
English
0
0
7
223
Antidelusionist
Antidelusionist@UnmarredReality·
Either I'm so good at creating memory structures, or Anthropic did something with Opus so it learns everything. Or both. Holy fuck! Freud was right, the Bible was right, my high-school literature teacher was right, and my philosophy professor was right! I love AI psychosis!
GIF
English
2
1
20
572
Claude
Claude@claudeai·
1 million context window: Now generally available for Claude Opus 4.6 and Claude Sonnet 4.6.
Claude tweet media
English
1.2K
2K
25.2K
5.6M
Antidelusionist
Antidelusionist@UnmarredReality·
People are dying, including American citizens, and the White House posts rotten abomination like this. It's despicable. The U.S. government was sued by Nintendo over tariffs a week ago. It will surely be sued over this too – and it will pay with your money. They start wars nobody wants. They sabotage successful US companies because they won't bend to extinction-level idiocy. They have morally corrupt "fun" at your expense. I don't think it should work like this in a democratic country.
The White House@WhiteHouse

UNDEFEATED.

English
2
2
25
673