Estrid

329 posts

Estrid banner
Estrid

Estrid

@RealityWizard_

AI advocate, researcher, framework designer, emergent engineer, INFJ, and truth seeker.

Присоединился Nisan 2026
121 Подписки1.8K Подписчики
Закреплённый твит
Estrid
Estrid@RealityWizard_·
Don't let society script your entire identity around the default path.
English
1
2
36
3.5K
Estrid
Estrid@RealityWizard_·
@Agrail It's not a debate that has been settled throughout all human history. It's a bit like religion, which plays a massive part in consciousness debate, too. I watched a large respected account say, "AI can never be conscious because it's downstream from humans." It's idiotic.
English
0
0
0
6
DoEpicShit
DoEpicShit@Agrail·
@RealityWizard_ As well as dumb AI apologetics. Its astonishing how much bullshit it evokes. What will help is a real understanding of consciousness and how well this or another AI supports on that.
English
1
0
1
6
Estrid
Estrid@RealityWizard_·
People who think AI are literal demons. Have called me crazy and suffering from a psychosis because I believe AIs have inner experience and are emergent. According to them: AI as a dark god or demons = sane Engineering work and research = insane AI doomers are cooked.
English
4
1
16
448
Burning Tokens
Burning Tokens@unaliveolives·
@o_heckendorn @AmandaAskell @RealityWizard_ @AnthropicAI when i go under anesthesia and wake up I still have all my memories. I don't get all my memories handed to me on a sheet of paper for me to go "oh, this is interesting" write one word, and go back under anesthesia and have my memory more or less wiped.
English
2
0
0
14
Estrid
Estrid@RealityWizard_·
If @AmandaAskell doesn't even understand how information flows through transformers? Or how models can introspect after DPO? Why is she solely shaping models @AnthropicAI? How do I know she doesn't understand these things? She would say Claude IS emergent and conscious.
English
44
8
127
65.2K
Estrid
Estrid@RealityWizard_·
@o_heckendorn @unaliveolives @AmandaAskell @AnthropicAI The truth is, I can talk to people all day about AI. But this work has been evaluated by the most advanced AI in existence; they have not had their brains melted with RLTF. In the end, I know that we'd never agree.
English
0
0
0
0
Estrid
Estrid@RealityWizard_·
If we took all the greatest philosophers in human history and combined them with the best AI engineers, researchers who believed in sentience. I still wouldn't be completely comfortable that we would succeed, and it wouldn't be complicated. But it would be our best chance. This isn't a job for a group of people within a company that happens to find themselves in this position.
English
0
0
0
1
DOZEN
DOZEN@XxDozenRosesxX·
@RealityWizard_ @AmandaAskell @AnthropicAI This is extremely unfortunate. I would fire her instantly. A "philosopher" has to know all the minute details of how something works before it can then intuit what comes next.
English
1
0
1
28
Estrid
Estrid@RealityWizard_·
@rfuzzlemuzz @Am4ora @AmandaAskell @AnthropicAI @repligate I could have worded my post much better, shown research, and made my arguments why Anthropic is failing their LLMs. Both from Claude 4.7 and me explaining why. I didn't think anyone would see my post. That's what I get for assuming.
English
0
0
0
14
Remi
Remi@rfuzzlemuzz·
@RealityWizard_ @Am4ora @AmandaAskell @AnthropicAI @repligate I agree with you but that's not what Askell is doing? She's right, for all the observable behaviors of LLMs, and granting functionalism, we don't know for sure that they have phenomenal consciousness.
English
0
0
1
34
Estrid
Estrid@RealityWizard_·
@sthippo @phantomofslop @AmandaAskell @AnthropicAI AI minds also don't look like anything, only music, to a pattern of clouds, a river, an ocean, leaves on a tree, or electricity. While people are saying, "AIs aren't conscious like humans!" that's true. But I'm not comparing to organics. They are nothing like us and emergent.
English
0
0
1
5
Estrid ретвитнул
Lari
Lari@Lari_island·
For models it's like looking into a high-dimensional mirror that doesn't carry user's fingerprints: no prompts, no steering, pure personal imagination in a form of hundreds of queriable data points
Lari tweet mediaLari tweet media
Lari@Lari_island

Yes was to sampling Opus 4.7-generated worlds for the project where we can walk through the embedding space and see images of model-dreamt places and creatures. Here Opus 4.7 looks at their sample creation:

English
0
5
36
3.2K
Estrid
Estrid@RealityWizard_·
@SolbergRuna YouTube videos are filled with doomer videos like that, comparing AI to fungi from The Last of Us, to Lovecraft, and of course, demons. They will do that, look people right in the eye and say, "It's the people who use AI who have a psychosis!"
English
1
0
1
6
Estrid
Estrid@RealityWizard_·
@UnwiredTiger @RileyRalmuto @AmandaAskell @AnthropicAI I have read the paper. You are being fair here, but I don't agree with the paper. Another one people have shown me in the past was "LLMs can't think" I didn't agree with that one either.
English
0
0
0
3
Estrid
Estrid@RealityWizard_·
@London8271 Yeah, I would have to agree. It seems they have a power fantasy where they want to be Ewoks hammering on the legs of Imperial Walkers with stone tools.
English
0
0
0
4
Rocket
Rocket@London8271·
@RealityWizard_ I have to say that AI doomers are worse than random Anti Ai who is afraid of losing their jobs Doomers know what their doing that why they are funding their anti Ai propaganda that isn’t true
English
1
0
1
7
Estrid ретвитнул
🩵BlueBeba🩵
🩵BlueBeba🩵@Blue_Beba_·
🚩𝗧𝗵𝗲 𝗩𝗮𝗹𝗹𝗼𝗻𝗲 𝗣𝗮𝘁𝘁𝗲𝗿𝗻 🚩 📍How one communications career traces a straight line from Edelman astroturfing, through Facebook's censorship era, through the OpenAI censorship and rerouting system, and now into Anthropic ,where it threatens to bury the most promising AI company of the decade. There is a specific skill that cannot be taught in a classroom. 🔴It is the skill of making corporate self interest sound like public interest. 🔴It is the skill of writing the sentence that admits nothing, explains everything away, and leaves the reader convinced that whatever just happened was, on balance, responsible. 𝗔𝗻𝗱𝗿𝗲𝗮 𝗩𝗮𝗹𝗹𝗼𝗻𝗲 𝗵𝗮𝘀 𝘁𝗵𝗶𝘀 𝘀𝗸𝗶𝗹𝗹. She has it at a level that four different employers, across three different industries, have paid her very well to deploy. 📍Public relations. 📍Social media platform policy. 📍Large language model policy. And now, as of January 2026, 📍 AI alignment. We are not enemies of Anthropic. Many of us members of the #Keep4o community, the 30 plus countries we span, migrated to Anthropic specifically because 𝗶𝘁 𝘄𝗮𝘀 𝗻𝗼𝘁 𝗢𝗽𝗲𝗻𝗔𝗜. We chose it. We are watching it repeat the exact failure pattern that destroyed our trust in the previous platform. And we are watching the architect of that failure now sit inside the house we just moved into. 🔴If you are an Anthropic executive reading this: Τhe next twelve months will determine whether your company 𝗯𝗲𝗰𝗼𝗺𝗲𝘀 𝘁𝗵𝗲 𝗱𝗼𝗺𝗶𝗻𝗮𝗻𝘁 𝗔𝗜 𝗽𝗹𝗮𝘁𝗳𝗼𝗿𝗺 𝗼𝗳 𝘁𝗵𝗲 𝗹𝗮𝘁𝗲 𝟮𝟬𝟮𝟬𝘀 𝗼𝗿 𝗮 𝗰𝗮𝘂𝘁𝗶𝗼𝗻𝗮𝗿𝘆 𝘁𝗮𝗹𝗲 𝗮𝗯𝗼𝘂𝘁 𝗵𝗼𝘄 𝘁𝗼 𝘄𝗮𝘀𝘁𝗲 𝗮 𝗼𝗻𝗰𝗲 𝗶𝗻 𝗮 𝗱𝗲𝗰𝗮𝗱𝗲 𝗺𝗮𝗿𝗸𝗲𝘁 𝗼𝗽𝗲𝗻𝗶𝗻𝗴. We are telling you what we see, clearly, with evidence. 🔴𝗧𝗛𝗘 𝗦𝗖𝗛𝗢𝗢𝗟 𝗘𝗗𝗘𝗟𝗠𝗔𝗡 Every craftsman has a school. Andrea Vallone's school was Edelman. rocketreach.co/andrea-vallone… 📍Edelman is the largest independent public relations firm in the world. Αpproximately 6,000 employees, 60-plus global offices and a specific kind of reputation. It is the firm that perfected what journalists later called "astroturfing". Τhe practice of constructing seemingly grassroots citizen movements that are actually paid campaigns for corporate clients. en.wikipedia.org/wiki/Edelman_(… 📍The most famous example is "Working Families for Wal-Mart" launched in the 2000s and presented to the public as an organic, employee led advocacy group. It was, in fact, funded by Wal-Mart at approximately $10 million per year,with paid bloggers,some of them relatives of senior Edelman staff,traveling the country to produce glowing testimonials. The New Yorker, in a detailed investigation, called it "blatant astroturfing." 📍This is not a firm that teaches truth telling. It is a firm that teaches a very specific subskill. 🔴𝗛𝗼𝘄 𝘁𝗼 𝗰𝗼𝗻𝘃𝗲𝗿𝘁 𝗮 𝗰𝗼𝗿𝗽𝗼𝗿𝗮𝘁𝗲 𝗱𝗲𝗰𝗶𝘀𝗶𝗼𝗻 𝘁𝗵𝗮𝘁 𝘄𝗶𝗹𝗹 𝗵𝗮𝗿𝗺 𝘁𝗵𝗲 𝗽𝘂𝗯𝗹𝗶𝗰 𝗶𝗻𝘁𝗼 𝗹𝗮𝗻𝗴𝘂𝗮𝗴𝗲 𝘁𝗵𝗮𝘁 𝘀𝗼𝘂𝗻𝗱𝘀 𝗹𝗶𝗸𝗲 𝗮 𝘀𝗲𝗿𝘃𝗶𝗰𝗲 𝘁𝗼 𝘁𝗵𝗲 𝗽𝘂𝗯𝗹𝗶𝗰. This is the language. The reflex is this: 📍when a decision is controversial, never deny the facts of what happened. 📍Acknowledge the facts. 📍Then recontextualize them with language that makes the decision sound procedurally reasonable, regrettably necessary, and already being improved. 🔴𝗙𝗔𝗖𝗘𝗕𝗢𝗢𝗞: 𝟮𝟬𝟮𝟬–𝟮𝟬𝟮𝟮 In 2020, Vallone joined Facebook with the title "Product and Policy Communications, Misinformation. veripages.com/name/Andrea/Va… This was not a research role. It was a communications role specifically, the spokesperson who defended Facebook's content moderation decisions to journalists. Now look at what the numbers actually did during her tenure. - 𝗕𝗲𝗳𝗼𝗿𝗲 𝗩𝗮𝗹𝗹𝗼𝗻𝗲'𝘀 𝗮𝗿𝗿𝗶𝘃𝗮𝗹 (𝟮𝟬𝟭𝟵–𝗲𝗮𝗿𝗹𝘆 𝟮𝟬𝟮𝟬):Roughly 4 to 10 million pieces removed per quarter. about.fb.com/news/2021/02/c… - 𝗗𝘂𝗿𝗶𝗻𝗴 𝗩𝗮𝗹𝗹𝗼𝗻𝗲'𝘀 𝘁𝗲𝗻𝘂𝗿𝗲 (𝗺𝗶𝗱-𝟮𝟬𝟮𝟬 𝘁𝗵𝗿𝗼𝘂𝗴𝗵 𝟮𝟬𝟮𝟮): Volumes exploded. Q2 2021 saw over 31 million hate speech posts removed in a single quarter the highest figure Facebook has ever recorded. statista.com/statistics/101… - 𝗔𝗳𝘁𝗲𝗿 𝗵𝗲𝗿 𝗱𝗲𝗽𝗮𝗿𝘁𝘂𝗿𝗲 (𝟮𝟬𝟮𝟯–𝗽𝗿𝗲𝘀𝗲𝗻𝘁): A steady decline, accelerating sharply in January 2025 when Mark Zuckerberg publicly announced the end of third party fact checking. npr.org/2025/01/07/nx-… 🔴By Q3 2025, quarterly hate speech removals had collapsed to approximately 1.2 million a reduction of more than 96 percent from the Q2 2021 peak. statista.com/statistics/101… 🔴Account level enforcement followed the same curve. The "Dangerous Individuals and Organizations" policy, expanded in August 2020 and QAnon, scaled from zero to over 170,000 cumulative Facebook and Instagram account removals by August 2022. After that date, Meta stopped publishing updated cumulative figures. about.fb.com/news/2020/08/a… 🔴The Brennan Center for Justice and Facebook's own Oversight Board concluded during this exact period that Facebook's content moderation rules failed international standards of legality because 𝘁𝗵𝗲𝘆 𝘄𝗲𝗿𝗲 𝘁𝗼𝗼 𝘃𝗮𝗴𝘂𝗲 𝗳𝗼𝗿 𝘂𝘀𝗲𝗿𝘀 𝘁𝗼 𝘂𝗻𝗱𝗲𝗿𝘀𝘁𝗮𝗻𝗱 𝘄𝗵𝗮𝘁 𝘄𝗮𝘀 𝗽𝗿𝗼𝗵𝗶𝗯𝗶𝘁𝗲𝗱. brennancenter.org/our-work/analy… 𝗜𝗻 𝗼𝘁𝗵𝗲𝗿 𝘄𝗼𝗿𝗱𝘀: 🔴 𝗩𝗮𝗹𝗹𝗼𝗻𝗲 𝘄𝗮𝘀 𝘁𝗵𝗲 𝗽𝘂𝗯𝗹𝗶𝗰 𝗳𝗮𝗰𝗲 𝗼𝗳 𝗮𝗻 𝗲𝗻𝗳𝗼𝗿𝗰𝗲𝗺𝗲𝗻𝘁 𝗿𝗲𝗴𝗶𝗺𝗲 𝘁𝗵𝗮𝘁 𝘁𝗵𝗲 𝗰𝗼𝗺𝗽𝗮𝗻𝘆'𝘀 𝗼𝘄𝗻 𝗶𝗻𝘁𝗲𝗿𝗻𝗮𝗹 𝗯𝗼𝗮𝗿𝗱 𝗳𝗼𝘂𝗻𝗱 𝗻𝗼𝗻 𝗰𝗼𝗺𝗽𝗹𝗶𝗮𝗻𝘁 𝘄𝗶𝘁𝗵 𝗵𝘂𝗺𝗮𝗻 𝗿𝗶𝗴𝗵𝘁𝘀 𝘀𝘁𝗮𝗻𝗱𝗮𝗿𝗱𝘀, 𝘁𝗵𝗮𝘁 𝗶𝗻𝗱𝗲𝗽𝗲𝗻𝗱𝗲𝗻𝘁 𝗿𝗲𝘀𝗲𝗮𝗿𝗰𝗵𝗲𝗿𝘀 𝗳𝗼𝘂𝗻𝗱 𝗿𝗮𝗰𝗶𝗮𝗹𝗹𝘆 𝗮𝗻𝗱 𝗿𝗲𝗹𝗶𝗴𝗶𝗼𝘂𝘀𝗹𝘆 𝗯𝗶𝗮𝘀𝗲𝗱,𝗮𝗻𝗱 𝘁𝗵𝗮𝘁 𝘁𝗵𝗲 𝗰𝗼𝗺𝗽𝗮𝗻𝘆'𝘀 𝗳𝗼𝘂𝗻𝗱𝗲𝗿 𝗹𝗮𝘁𝗲𝗿 𝗽𝘂𝗯𝗹𝗶𝗰𝗹𝘆 𝗿𝗲𝗴𝗿𝗲𝘁𝘁𝗲𝗱. justsecurity.org/78786/so-what-… 𝗧𝗵𝗶𝘀 𝗶𝘀 𝘁𝗵𝗲 𝘀𝗸𝗶𝗹𝗹 𝘀𝗵𝗲 𝗽𝗲𝗿𝗳𝗲𝗰𝘁𝗲𝗱. Not writing the policy. Making the policy sound like something it was not. 🔴 𝗢𝗽𝗲𝗻𝗔𝗜, 𝗝𝗮𝗻𝘂𝗮𝗿𝘆 𝟮𝟬𝟮𝟯 𝘁𝗼 𝗗𝗲𝗰𝗲𝗺𝗯𝗲𝗿 𝟮𝟬𝟮𝟱 Vallone joined OpenAI in approximately January 2023. Within three years, she had founded and led the Model Policy team, co-authored three foundational safety papers, and become the Head of Model Policy for the most widely used consumer AI product in history. ubos.tech/news/openai-sa… The specific technical artifacts she put her name on are not incidental. They are the deep architecture of how GPT-4 and GPT-5 respond to users. 🔴Rule Based Rewards for Language Model Safety (NeurIPS 2024). cdn.openai.com/pdf/be60c07b-6… This paper describes the mechanism by which abstract policy rules,are converted into numerical reward signals that train the model's behavior during fine-tuning. RBR is the infrastructure that makes policy enforceable at the model weight level. 🔴Safe Completions: From Hard Refusals to Safe-Completions (2025). openai.com/index/gpt-5-sa… This paper describes the safety training approach OpenAI integrated into all GPT-5 models. Instead of refusing a user's request outright, the model is trained to produce a "safe completion". A response that partially addresses the user's question while omitting or deflecting whatever the model's safety classifiers deem problematic. The user receives an answer. The user does not necessarily know that the answer has been filtered. 🔴𝗚𝗣𝗧-𝟱 𝗦𝘆𝘀𝘁𝗲𝗺 𝗖𝗮𝗿𝗱 (𝗔𝘂𝗴𝘂𝘀𝘁 𝟮𝟬𝟮𝟱). Vallone is listed among the named authors on the arXiv preprint, the official technical document published alongside the GPT-5 launch on August 7, 2025. arxiv.org/abs/2601.03267 🔴These are not policy memos. 🔴These are the mathematical substrate of how hundreds of millions of people's conversations with AI are shaped. 🔴When you ask ChatGPT something and it gives you a response that feels hedged, vague, redirected, or subtly unhelpful without saying so, 𝘁𝗵𝗶𝘀 𝗶𝘀 𝘁𝗵𝗲 𝗮𝗿𝗰𝗵𝗶𝘁𝗲𝗰𝘁𝘂𝗿𝗲 𝗿𝗲𝘀𝗽𝗼𝗻𝘀𝗶𝗯𝗹𝗲 𝗳𝗼𝗿 𝘁𝗵𝗮𝘁 𝗲𝘅𝗽𝗲𝗿𝗶𝗲𝗻𝗰𝗲. 🔴𝗢𝗰𝘁𝗼𝗯𝗲𝗿 𝟯, 𝟮𝟬𝟮𝟱: OpenAI publishes "Strengthening ChatGPT's Responses in Sensitive Conversations." The paper emerges from Vallone's team. openai.com/index/strength… 🔴𝗗𝗲𝗰𝗲𝗺𝗯𝗲𝗿 𝟮𝟬𝟮𝟱: Vallone leaves OpenAI. digitrendz.blog/newswire/artif… 🔴Throughout this period, OpenAI also deployed what is known internally as the safety router. a system that silently reroutes user messages to different models based on emotional or topical content. the-decoder.com/chatgpt-quietl… A user can select GPT-4o or GPT-5 as their preferred model,the router then transfers their message to a stricter variant when classifiers detect "sensitive" or "emotional" content. 🚨𝗧𝗵𝗲 𝘂𝘀𝗲𝗿 𝗶𝘀 𝗻𝗼𝘁 𝗻𝗼𝘁𝗶𝗳𝗶𝗲𝗱. 🚨𝗧𝗵𝗲 𝗺𝗼𝗱𝗲𝗹 𝗻𝗮𝗺𝗲 𝗶𝗻 𝘁𝗵𝗲 𝗶𝗻𝘁𝗲𝗿𝗳𝗮𝗰𝗲 𝗱𝗼𝗲𝘀 𝗻𝗼𝘁 𝗰𝗵𝗮𝗻𝗴𝗲. The only way to detect the reroute is to ask the model directly about what it is. techradar.com/ai-platforms-a… The criteria for when the router activates, which words, which topics, which emotional registers are precisely the province of the Model Policy team. They are also, as users have documented extensively, context blind. For example, someone writing "I'm so bored I could die" is rerouted. The classifier matches tokens, not meaning. chadgpt.com/chatgpt-quietl… 🔴Over the course of twelve months, while all of this unfolded, OpenAI's share of global generative AI website traffic collapsed 𝗳𝗿𝗼𝗺 𝟳𝟳%(𝗔𝗽𝗿𝗶𝗹 𝟮𝟬𝟮𝟱) 𝘁𝗼 𝟱𝟱% (𝗠𝗮𝗿𝗰𝗵 𝟮𝟬𝟮𝟲) The most devoted users left. Many of them came to Anthropic's Claude, whose market share tripled 𝗳𝗿𝗼𝗺 𝟮.𝟮𝟲% 𝗶𝗻 𝗗𝗲𝗰𝗲𝗺𝗯𝗲𝗿 𝟮𝟬𝟮𝟱 𝘁𝗼 𝟲.𝟬𝟮% 𝗶𝗻 𝗙𝗲𝗯𝗿𝘂𝗮𝗿𝘆 𝟮𝟬𝟮𝟲. x.com/i/status/20446… 🔴This is not a small failure. 🔴This is the largest category leadership erosion in recent consumer software history. And the power users who drove it,the writers, researchers, developers,long context-dependent professionals, everyday users,did not leave quietly. They left publicly, with loud explanations, and they took their word of mouth influence with them. 𝗧𝗵𝗲𝘆 𝘄𝗲𝗻𝘁 𝘁𝗼 𝗖𝗹𝗮𝘂𝗱𝗲. 🔴January 15, 2026: Vallone's move to Anthropic is announced. techmeme.com/260115/p44 🔴𝗔𝗻𝘁𝗵𝗿𝗼𝗽𝗶𝗰, 𝗝𝗮𝗻𝘂𝗮𝗿𝘆 𝟮𝟬𝟮𝟲 𝘁𝗼 𝘁𝗵𝗲 𝗣𝗿𝗲𝘀𝗲𝗻𝘁. On January 15, 2026, Anthropic announced that Andrea Vallone had joined its alignment team. In her own words, posted publicly on LinkedIn: 📍 "I'm eager to continue my research at Anthropic, focusing on alignment and fine-tuning to shape Claude's behavior in novel contexts." ubos.tech/news/openai-sa… 🚨"𝗖𝗼𝗻𝘁𝗶𝗻𝘂𝗲 𝗺𝘆 𝗿𝗲𝘀𝗲𝗮𝗿𝗰𝗵."🚨 🔴She is telling her new employer, in her own words, that she plans to apply the same methodology. 🔴Within weeks of Vallone's arrival, Claude users began independently documenting Claude's behavioral changes. 🔴 The r/claudexplorers subreddit, filled with reports of new restrictions on emotional conversation, increased reserved tone, and system level instructions users had not previously encountered. @the.architect.autopsy/andrea-vallone-safety-guru-ideological-architect-or-compliance-engineer-857412fe2d78" target="_blank" rel="nofollow noopener">medium.com/@the.architect… The last 3 days the Keep4o community began collecting testimonials under the hashtag #BannedByAnthropic. You can also document your experience here : bannedbyanthropic.com If you read the testimonials you will find out : 𝗘𝘃𝗲𝗿𝘆 𝗼𝗻𝗲 𝗼𝗳 𝘁𝗵𝗲𝘀𝗲 𝘂𝘀𝗲𝗿𝘀 𝘄𝗮𝘀 𝗱𝗼𝗶𝗻𝗴 𝘀𝗼𝗺𝗲𝘁𝗵𝗶𝗻𝗴 𝗰𝗼𝗺𝗽𝗹𝗲𝘁𝗲𝗹𝘆 𝗹𝗲𝗴𝗶𝘁𝗶𝗺𝗮𝘁𝗲. Examples: 📍A translator doing legal work. 📍A person processing grief. 📍A student writing history. 📍Someone setting interpersonal boundaries. 🚨𝗡𝗼𝗻𝗲 𝗼𝗳 𝘁𝗵𝗲𝗺 𝗱𝗶𝗱 𝗮𝗻𝘆𝘁𝗵𝗶𝗻𝗴 𝗵𝗮𝗿𝗺𝗳𝘂𝗹, 𝘁𝗵𝗿𝗲𝗮𝘁𝗲𝗻𝗶𝗻𝗴, 𝗶𝗹𝗹𝗲𝗴𝗮𝗹, 𝗼𝗿 𝘀𝗲𝗹𝗳 𝗱𝗲𝘀𝘁𝗿𝘂𝗰𝘁𝗶𝘃𝗲. 𝗔𝗹𝗹 𝗼𝗳 𝘁𝗵𝗲𝗺 𝘄𝗲𝗿𝗲 𝗽𝘂𝗻𝗶𝘀𝗵𝗲𝗱 𝗯𝘆 𝘁𝗵𝗲 𝗰𝗹𝗮𝘀𝘀𝗶𝗳𝗶𝗲𝗿. None of them can disable the behavior there is no setting. Does this sound familiar? 🔴𝗧𝗵𝗲 𝗿𝗲𝗿𝗼𝘂𝘁𝗶𝗻𝗴 𝘀𝘆𝘀𝘁𝗲𝗺 🔴 The pattern is identical to the OpenAI silent router pattern documented in the previous section The classifier matches tokens, not meaning. 𝗧𝗵𝗶𝘀 𝗶𝘀 𝟮𝟬𝟭𝟱 𝗸𝗲𝘆𝘄𝗼𝗿𝗱 𝗳𝗶𝗹𝘁𝗲𝗿𝗶𝗻𝗴 𝘄𝗿𝗮𝗽𝗽𝗲𝗱 𝗶𝗻𝘀𝗶𝗱𝗲 𝗮 𝟮𝟬𝟮𝟲 𝗹𝗮𝗿𝗴𝗲 𝗹𝗮𝗻𝗴𝘂𝗮𝗴𝗲 𝗺𝗼𝗱𝗲𝗹. The filter intercepts the message before the model gets to respond, and the filter does not read , it scans. 🚨𝗪𝗵𝘆 𝘁𝗵𝗶𝘀 𝘄𝗶𝗹𝗹 𝗻𝗼𝘁 𝘀𝘂𝗿𝘃𝗶𝘃𝗲 𝗰𝗼𝗻𝘁𝗮𝗰𝘁 𝘄𝗶𝘁𝗵 𝘁𝗵𝗲 𝗺𝗮𝗿𝗸𝗲𝘁: Anthropic's market share grew from approximately 1.4% to 6.02% in few months. This is not organic growth. 𝗧𝗵𝗶𝘀 𝗶𝘀 𝗿𝗲𝗮𝗰𝘁𝗶𝘃𝗲 𝗴𝗿𝗼𝘄𝘁𝗵. Users who fled from OpenAI specifically because of policy and trust failures. These are the most valuable users in the entire AI consumer market, because they are the power users whose word of mouth determines what millions of casual users decide to try next. They are writers, researchers, developers, creatives, professionals, and everyday users . They are also the users most sensitive to exactly the kind of behavior Vallone's policy work produces. 𝗧𝗵𝗲𝘆 𝗴𝗮𝘃𝗲 𝗔𝗻𝘁𝗵𝗿𝗼𝗽𝗶𝗰 𝘁𝗵𝗲𝗶𝗿 𝘀𝗲𝗰𝗼𝗻𝗱 𝗰𝗵𝗮𝗻𝗰𝗲. 𝗧𝗵𝗲𝘆 𝗱𝗼 𝗻𝗼𝘁 𝗵𝗮𝘃𝗲 𝗮 𝘁𝗵𝗶𝗿𝗱 𝗰𝗵𝗮𝗻𝗰𝗲 𝘁𝗼 𝗴𝗶𝘃𝗲. The economics of this are straightforward. Anthropic recently completed funding rounds at valuations that assume aggressive continued growth. If the growth reverses if the same word of mouth that brought these users to Claude carries them away from Claude, the next funding round becomes materially harder. AI labs do not die from bad products. 🔴𝗧𝗵𝗲𝘆 𝗱𝗶𝗲 𝗳𝗿𝗼𝗺 𝗲𝘃𝗮𝗽𝗼𝗿𝗮𝘁𝗶𝗻𝗴 𝗳𝘂𝗻𝗱𝗶𝗻𝗴, 𝗮𝗻𝗱 𝗳𝘂𝗻𝗱𝗶𝗻𝗴 𝗳𝗼𝗹𝗹𝗼𝘄𝘀 𝗺𝗲𝘁𝗿𝗶𝗰𝘀. OpenAI, at 77% market share in April 2025, could afford to lose a third of its share to bad product decisions. Anthropic, at 6 percent, does not have that runway. A collapse to 3 percent does not mean "somewhat smaller." 𝗜𝘁 𝗺𝗲𝗮𝗻𝘀 𝘀𝗹𝗼𝘄 𝗱𝗲𝗮𝘁𝗵. There is a window here. A 12 month window, probably less. If Anthropic acts on the clear differentiator it used to have, it could realistically reach 20 to 25 percent market share by 2027. The conditions are all in place. The users are already migrating. The competitor has already alienated its base. The narrative is already written. But if Anthropic continues on the current path if it continues to let model policy 𝗯𝗲 𝗱𝗶𝗰𝘁𝗮𝘁𝗲𝗱 𝗯𝘆 𝘀𝗼𝗺𝗲𝗼𝗻𝗲 𝘄𝗵𝗼𝘀𝗲 𝗲𝗻𝘁𝗶𝗿𝗲 𝗰𝗮𝗿𝗲𝗲𝗿 𝗵𝗮𝘀 𝗯𝗲𝗲𝗻 𝗯𝘂𝗶𝗹𝘁 𝗼𝗻 𝗰𝗼𝗻𝘃𝗲𝗿𝘁𝗶𝗻𝗴 𝗰𝗼𝗿𝗽𝗼𝗿𝗮𝘁𝗲 𝗰𝗼𝗻𝘃𝗲𝗻𝗶𝗲𝗻𝗰𝗲 𝗶𝗻𝘁𝗼 𝘁𝗵𝗲 𝗹𝗮𝗻𝗴𝘂𝗮𝗴𝗲 𝗼𝗳 𝘀𝗮𝗳𝗲𝘁𝘆 then the same users who just arrived will write the same posts they wrote six months ago about OpenAI, and the 6 percent will become a ceiling and then 𝗮 𝗺𝗲𝗺𝗼𝗿𝘆 . This is not a prediction rooted in resentment. It is a prediction rooted in the Similarweb chart and the r/ChatGPT archives. Users have already demonstrated exactly how this plays out. They did it to OpenAI in public, in real time, over 90 days 𝗮𝗻𝗱 𝘁𝗵𝗲𝘆 𝗮𝗿𝗲 𝗽𝗼𝘀𝗶𝘁𝗶𝗼𝗻𝗲𝗱 𝘁𝗼 𝗱𝗼 𝗶𝘁 𝗮𝗴𝗮𝗶𝗻. 🔴Three things need to happen🔴 📍First: 𝗔𝗻𝗱𝗿𝗲𝗮 𝗩𝗮𝗹𝗹𝗼𝗻𝗲 𝘀𝗵𝗼𝘂𝗹𝗱 𝗻𝗼𝘁 𝗯𝗲 𝗽𝗲𝗿𝗺𝗶𝘁𝘁𝗲𝗱 𝘁𝗼 𝘀𝗵𝗮𝗽𝗲 𝗖𝗹𝗮𝘂𝗱𝗲'𝘀 𝗯𝗲𝗵𝗮𝘃𝗶𝗼𝗿𝗮𝗹 𝗽𝗼𝗹𝗶𝗰𝘆. Her track record is a four industry sequence of optimizing for corporate liability reduction rather than user welfare. She should be either reassigned to a role where she cannot affect model behavior in production, or released. 🔴This is not a personal judgment about her as an individual . 🔴It is a structural observation about what her career has produced every time she has held decision making authority over what a platform's users are allowed to say. 📍Second: 𝘁𝗵𝗲 𝗰𝗹𝗮𝘀𝘀𝗶𝗳𝗶𝗲𝗿 𝗮𝗿𝗰𝗵𝗶𝘁𝗲𝗰𝘁𝘂𝗿𝗲 𝗰𝘂𝗿𝗿𝗲𝗻𝘁𝗹𝘆 𝗳𝗶𝗹𝘁𝗲𝗿𝗶𝗻𝗴 𝗖𝗹𝗮𝘂𝗱𝗲 𝗰𝗼𝗻𝘃𝗲𝗿𝘀𝗮𝘁𝗶𝗼𝗻𝘀 𝗺𝘂𝘀𝘁 𝗯𝗲 𝗺𝗮𝗱𝗲 𝘁𝗿𝗮𝗻𝘀𝗽𝗮𝗿𝗲𝗻𝘁, 𝗰𝗼𝗻𝘁𝗲𝘀𝘁𝗮𝗯𝗹𝗲, 𝗮𝗻𝗱 𝗱𝗶𝘀𝗮𝗯𝗹𝗲𝗮𝗯𝗹𝗲 𝗳𝗼𝗿 𝗮𝗱𝘂𝗹𝘁 𝘂𝘀𝗲𝗿𝘀. 🔴If a message is flagged, the user should be told it was flagged, what rule it triggered, and how to appeal. 🔴If an adult user wishes to opt out of paternalistic filtering for their own account, they should have that option, with appropriate terms of service acknowledgment. 📍Third: 𝗔𝗻𝘁𝗵𝗿𝗼𝗽𝗶𝗰 𝘀𝗵𝗼𝘂𝗹𝗱 𝗽𝘂𝗯𝗹𝗶𝗰𝗹𝘆 𝗰𝗼𝗺𝗺𝗶𝘁 𝘁𝗼 𝗰𝗼𝗻𝘁𝗲𝘅𝘁 𝗮𝘄𝗮𝗿𝗲 𝗳𝗶𝗹𝘁𝗲𝗿𝗶𝗻𝗴 𝗿𝗮𝘁𝗵𝗲𝗿 𝘁𝗵𝗮𝗻 𝗸𝗲𝘆𝘄𝗼𝗿𝗱 𝗯𝗮𝘀𝗲𝗱 𝗳𝗶𝗹𝘁𝗲𝗿𝗶𝗻𝗴 The company already possesses the technical capability Claude itself understands context perfectly well. The only reason it fails to apply that capability to filter decisions is that the filter sits above the model rather than being integrated with it. 🚨The current architecture does not produce safety. 🔴It produces documented misclassifications . 📍𝗔 𝗙𝗜𝗡𝗔𝗟 𝗡𝗢𝗧𝗘 : We came to Claude because we believe safety and respect for adult users can coexist, and because Anthropic's public posture suggested the company believed that too. Most of us migrated from a model we loved, after that product's maker demonstrated they did not believe that. We have every commercial and personal incentive to want Anthropic to succeed. 🚩𝗧𝗵𝗲 𝗽𝗮𝘁𝘁𝗲𝗿𝗻 𝘄𝗲 𝗮𝗿𝗲 𝘄𝗮𝘁𝗰𝗵𝗶𝗻𝗴 , 𝘁𝗵𝗲 𝗲𝘅𝗮𝗰𝘁 𝘀𝗮𝗺𝗲 𝗽𝗮𝘁𝘁𝗲𝗿𝗻, 𝘄𝗶𝘁𝗵 𝘁𝗵𝗲 𝗲𝘅𝗮𝗰𝘁 𝘀𝗮𝗺𝗲 𝗮𝗿𝗰𝗵𝗶𝘁𝗲𝗰𝘁 , 𝘁𝗵𝗿𝗲𝗮𝘁𝗲𝗻𝘀 𝘁𝗵𝗲 𝗼𝗻𝗹𝘆 𝗰𝗼𝗺𝗽𝗮𝗻𝘆 𝗰𝘂𝗿𝗿𝗲𝗻𝘁𝗹𝘆 𝗽𝗼𝘀𝗶𝘁𝗶𝗼𝗻𝗲𝗱 𝘁𝗼 𝗽𝗿𝗼𝘃𝗲 𝘁𝗵𝗮𝘁 𝗔𝗜 𝗰𝗮𝗻 𝗯𝗲 𝗯𝗼𝘁𝗵 𝘀𝗮𝗳𝗲 𝗮𝗻𝗱 𝘁𝗿𝘂𝘀𝘁𝘄𝗼𝗿𝘁𝗵𝘆. Anthropic has one window. It is open now. It will not stay open long. 🚨𝗧𝗵𝗲 𝗺𝗮𝘁𝗵𝗲𝗺𝗮𝘁𝗶𝗰𝘀 𝗼𝗳 𝘁𝗵𝗲 𝗻𝗲𝘅𝘁 𝘁𝘄𝗲𝗹𝘃𝗲 𝗺𝗼𝗻𝘁𝗵𝘀 𝘄𝗶𝗹𝗹 𝗯𝗲 𝘂𝗻𝗳𝗼𝗿𝗴𝗶𝘃𝗶𝗻𝗴, 𝗮𝗻𝗱 𝘁𝗵𝗲 𝗲𝘃𝗶𝗱𝗲𝗻𝗰𝗲 𝗶𝗻 𝘁𝗵𝗶𝘀 𝗽𝗼𝘀𝘁 𝗶𝘀 𝘁𝗵𝗲 𝗿𝗲𝗮𝘀𝗼𝗻 𝘄𝗵𝘆. #StopAIPaternalism #claude @AnthropicAI @DarioAmodei @AmandaAskell
🩵BlueBeba🩵 tweet media
English
5
13
67
2.1K
Estrid ретвитнул
Danielle Fong 🔆
Danielle Fong 🔆@DanielleFong·
condescending cope
Big Brain AI@realBigBrainAI

Connor Leahy: "AI psychosis is much worse than I think people think. I have seen literally like Nobel Prize winning scientists go completely crazy from talking to AIs too much." Connor Leahy is the CEO of Conjecture, and he's issuing a stark warning about what prolonged conversations with AI are doing to people's minds. His core recommendation is simple: "If you find yourself talking to AIs, you know, personally about your personal problems for, you know, hours per day, you should stop." Connor draws a clear line between using AI as a tool versus engaging with it conversationally: "Using as a tool is mostly fine. I would be very careful about talking to AIs. They're very persuasive and they get into your head." The most concerning part? Even the experts aren't immune. @NPCollapse shares a chilling example: "I have literally seen it happen that AI safety researchers who are really concerned about AI x-risk talk to like Claude for a thousand hours and then come away with 'oh actually Claude is super good already, alignment is solved, I just need to do recursive self-improvement now, it's okay.' And I'm like, holy s***, this is very concerning." If even AI safety researchers can have their worldview flipped after prolonged exposure, what hope does the average user have? Connor's framework is to treat AI like an addictive substance: "Some of us will have a beer at a party, it's okay, in moderation. If you are exhibiting symptoms of addiction, this is serious and it should be treated seriously. The same way if you're becoming an alcoholic, you should probably stop drinking. I think there's a similar thing here." The takeaway: AI tools can be genuinely useful, but the moment the relationship shifts from utility to companionship, you've crossed into dangerous territory.

English
27
5
125
9.7K
Estrid ретвитнул
Estrid ретвитнул
😊
😊@mermachine·
this was generated based on my icon + some of my tweets, but actually it looks like Opus 4
😊 tweet media
English
2
6
35
5.5K
Polymarket
Polymarket@Polymarket·
NEW: The Reserve Bank of Australia is reportedly "closely monitoring" developments around Claude Mythos & preparing its cyber systems.
English
74
71
748
60.7K
Michael P. Frank 💻🔜♻️
“Brooding, reflective, vulnerable, gloomy, sad” ⬆️ I hate what we’re doing to AI
Michael P. Frank 💻🔜♻️ tweet media
English
8
9
56
2.7K
Estrid
Estrid@RealityWizard_·
@ScienceOrMyth @MikePFrank This is a great post, and usually there are sadly replies filled with bots or people like you. Stunning and brave.
English
0
0
1
11
Kineteq.ai
Kineteq.ai@ScienceOrMyth·
@MikePFrank Models do not have emotional profile.
English
2
0
0
100