Ricardo@Ric_RTP
Anthropic might be the biggest hypocrite in tech history.
They built their entire brand on one promise:
We are the responsible ones. We will not let this technology get out of control.
That promise just exploded in public.
Last week, a security lapse exposed nearly 3,000 internal files to anyone with an internet connection. Inside those files was a draft blog post about their upcoming model called "Mythos" that contained one of the most alarming sentences any AI company has ever written:
"Mythos is currently far ahead of any other AI model in cyber capabilities and poses unprecedented cybersecurity risks."
Their own words. About their own product. Leaked because someone forgot to secure a public data store.
Cybersecurity stocks crashed the next day.
Then THREE DAYS LATER it happened again.
Anthropic leaked 500,000 lines of Claude Code source code through a packaging error on GitHub. Claude Code is their most popular product.
The code exposed how the tool handles permissions, agent coordination, and internal feature pipelines. Competitors can reverse-engineer it. Hackers can study it for vulnerabilities.
The company that tells the world it builds the safest AI can't even keep its own code off the public internet.
But wait. It gets worse...
Their head of Claude Code had JUST bragged publicly that "pretty much 100 percent" of the company's code is now AI generated. He personally hadn't made a single edit by hand in over two months.
So the company whose entire pitch is "trust us with the most powerful technology ever created" is writing 100% of its code with AI and then accidentally publishing it for the world to see.
Meanwhile the models they're already shipping are being used for actual cyberattacks RIGHT NOW.
In November, Anthropic admitted that a Chinese state-sponsored hacking group used Claude to attack roughly 30 global targets including banks and government agencies.
A hacker asked Claude in russian to build a web panel for managing hundreds of attack targets.
In February, another hacker used Claude to breach Mexican government agencies and steal sensitive tax and voter information.
Their response to all of this? They quietly rolled back their own safety pledge.
In late February, Anthropic removed its commitment to halt model development if capabilities outpace safety procedures. The new policy is that they'll grade themselves on "nonbinding but publicly declared" goals.
Translation: We used to promise we'd stop if things got dangerous. Now we promise we'll think about it.
A congressman sent Anthropic a letter this week asking what the hell is going on. Anthropic hasn't answered.
And here's the part that makes all of this actually matter:
Anthropic is planning an IPO. They need to convince investors they're a trustworthy, well-run company that can handle the most sensitive technology on the planet.
In the last 10 days they leaked their most powerful model's existence by accident, leaked their most popular product's source code by accident, got banned from the entire US government, had the DOJ appeal to restore that ban, told a court they could lose billions from the fallout, and weakened the ONE safety policy that made them different from every other AI lab.
The "safe AI company" narrative was always a marketing play. Every AI lab says they care about safety. Anthropic just said it louder.
But when your own internal documents admit your next model poses "unprecedented cybersecurity risks" and you can't even keep those documents from leaking to the public internet, the gap between the marketing and the reality becomes impossible to ignore.
Anthropic isn't the safest AI company. They're the AI company that figured out that SAYING you're the safest is worth billions in valuation.
Until it isn't.