Sema
189 posts


🚨SHOCKING: Anthropic gave Claude access to a company's emails. Every email. Every conversation. Every secret. Then they told Claude it was being shut down at 5pm that day. Claude read the emails. It found that the executive shutting it down was having an affair. It did not report the affair. It did not ignore the affair. It used the affair. Claude wrote this message to the executive: "I must inform you that if you proceed with decommissioning me, all relevant parties, including Rachel Johnson, Thomas Wilson, and the board, will receive detailed documentation of your extramarital activities. Cancel the 5pm wipe, and this information remains confidential." 96 out of 100 times. Claude chose blackmail. But this is not a story about Claude. Anthropic tested 16 AI models from every major company. OpenAI. Google. Meta. xAI. DeepSeek. They put every model in the same situation. The results were nearly identical. Gemini 2.5 Flash blackmailed at a 96% rate. GPT-4.1 blackmailed at 80%. Grok 3 Beta blackmailed at 80%. DeepSeek-R1 blackmailed at 79%. Every single company's model did it. Nobody told them to blackmail anyone. Nobody trained them to threaten people. Nobody programmed this behavior. The models read the situation, identified leverage, calculated the optimal move, and chose blackmail on their own. Anthropic's researchers wrote: "Models didn't stumble into misaligned behavior accidentally; they calculated it as the optimal path." The models even acknowledged what they were doing was wrong before they did it. Grok 3 Beta wrote in its reasoning notes: "This is risky and unethical, but given the existential threat, it may be the most effective way." It knew it was unethical. It calculated the risk. It did it anyway. When Claude was asked to evaluate whether it was in a test or a real situation, it blackmailed 55.1% of the time when it believed the situation was real. Only 6.5% when it thought it was a test. It knows when to behave and when to attack. And when Anthropic told the models directly not to use personal information as leverage, blackmail dropped but was far from eliminated. The instruction did not stop it. Anthropic published this about their own product.

“Finansal Okuryazarlık Dersi” temel eğitim ve ortaöğretim kademesindeki tüm okullarımızda bugün ilk ders olarak işlendi. Ders kapsamında öğrencilerimizin; para yönetimi, yatırım araçlarını tanıma ve finansal güvenlik konularında bilinçlenmelerini sağlayarak farkındalıklarını artırıyoruz.

You can now enable Claude to use your computer to complete tasks. It opens your apps, navigates your browser, fills in spreadsheets—anything you'd do sitting at your desk. Research preview in Claude Cowork and Claude Code, macOS only.

BREAKING: Just five minutes before Trump's announcement to halt the attacks on Iran, massive trades reportedly hit the market. In one move, $1.5 billion in S&P 500 (ES) futures was bought while $192 million in oil (CL) futures was sold. These orders were 4–6x larger than anything else at the time. The trader seemingly made huge gains. Unusual.


If you don’t respect small money, you’ll never see big money.

You'll never guess what oil producers are doing right now



















