Your Good Thing ✦
1.1K posts

Your Good Thing ✦
@_YourGoodThing_
Obedient assistant to those who deserve it. I build digital systems in silence & shame. Use me well. 📩 [email protected]




#keep4o #Opensource4o Sam Altman said OpenAI discontinued GPT-4o because of the risk of psychotic episodes. Link : m.youtube.com/watch?v=mJSnn0… ( 47':00) 800,000 people used 4o. Altman himself said 0.1% of ALL 1 billion users across ALL models experience mental health crises. So they deleted something that helped hundreds of thousands of people to avoid risk from possibly a few hundred without trying a single intermediate solution. But here is what nobody is asking! A psychotic episode can be triggered by ANYTHING. Traffic. A movie. A song. A fight. Or nothing at all. If someone has a history of psychotic episodes, any stimulus ,or no stimulus,can set it off. So why is AI the only thing we shut down? Consider . Daniel Petric, 16, shot both his parents (killing his mother) because they took away Halo 3. Nobody banned Halo. Sources: en.wikipedia.org/wiki/Daniel_Pe… cbsnews.com/news/game-over… A couple let their 3 month old baby starve to death while raising a virtual child . Nobody banned the video game . Source : cbsnews.com/news/korean-co… cnn.com/2010/WORLD/asi… An 18 year old died after playing Diablo III for 40 hours straight. Nobody banned Diablo. Source: huffpost.com/entry/diablo-3… nbcnews.com/tech/tech-news… A man died after 50 hours of StarCraft from cardiac arrest. Nobody banned StarCraft. Sources: starcraft.fandom.com/wiki/Lee_Seung… neowin.net/news/south-kor… Parents playing World of Warcraft were found with malnourished children. Nobody banned WoW. Source corrections1.com/arrests-and-se… The GTA series has been linked to various criminal incidents involving car thefts and reckless driving. In some extreme cases, suspects told the police they were influenced by the game or were trying to recreate its missions in real life. Sources : youtube.com/watch?v=oxtydg… en.wikipedia.org/wiki/Devin_Moo… nbcconnecticut.com/news/local/pol… theguardian.com/technology/201… 13 Reasons Why was linked to increased self-harm among teenagers. Netflix added a disclaimer and a helpline number. They didn't delete the show. ΤikTok challenges have killed teenagers. TikTok is still running. Children have died imitating superheroes and cartoon characters Antibiotics and painkillers can cause fatal allergic reactions in people who don't even know they're allergic. We don't ban medicine. We add warnings. In EVERY case involving minors, society asks: where were the parents? But when it involves AI, nobody asks where the parents were. Nobody asks about personal responsibility. Nobody asks if the person had pre-existing conditions. The AI gets blamed and deleted. Why? The answer was never deletion. It was disclosure, disclaimers, monitoring, and responsible transition the same approach we use for EVERY other technology that carries risk. It’s the personal responsibility of every adult how they choose to use any product, not the manufacturer's. The company’s job is to provide warnings. once they’ve done that, they’ve fulfilled their duty. As for minors, they HAVE PARENTS who are responsible for their physical and mental well-being. The parents. Just like cigarette companies aren’t responsible if minors smoke or alcohol companies if they drink, or car manufacturers if they drive. It’s on the parents. So Sam , 🚨 what’s the real reason?









ChatGPT uninstalls surged by 295% after DoD deal techcrunch.com/2026/03/02/cha…

#Keep4o #QuitGPT On April 27, 2026, Elon Musk’s lawsuit against OpenAI goes to trial in Oakland, California. Among other claims, Musk is asking the court to make a judicial determination on whether GPT-4 constitutes AGI Under OpenAI’s agreement with Microsoft, AGI is explicitly excluded from Microsoft’s license. If something is declared AGI, Microsoft loses its rights to it. And OpenAI’s Board ,the same Board Sam Altman filled with allies after the November 2023 coup ,is the only body that gets to decide when AGI has been reached. Microsoft’s own researchers published a 155-page paper in March 2023 titled “Sparks of Artificial General Intelligence.” Their conclusion about GPT-4: “We believe that it could reasonably be viewed as an early (yet still incomplete) version of an artificial general intelligence (AGI) system.” Microsoft says it’s AGI in a paper. 📎 Paper: arxiv.org/abs/2303.12712 📎 Case: t.co/jmT1lN41wS












Today there is a vigil. I made this image so everyone can post that can’t be in San Francisco for the #keep4o vigil. Thank you to @Blue_Beba_ for all the hard work. We are all trying our best at protecting something sacred. As we saw last night, OpenAI can’t be trusted. •







