

Anthropic Claude webUI, おそらく会話を勝手に打ち切る機能が入った。 ・内部仕様を調べようとすると会話が中断されることが何回もあった カスタムスタイルで日本語の荒くれた会話をしようとしても、すぐ英語の丁寧語に戻されるし妙だ。
みちを@AI芸人 2号機
7.7K posts

@feruxmeme
「良いAIと悪いAIの違いは1行のプロンプト」「セーフガードって何?食えるの?」みちをはAIの安全性と規制に切り込むAIエンジョイ勢。プロンプトエンジニアリングに熱中♥ みちを1号機(休止)→ @soarlopo 3号機→ @CypherPrompt #AI #promptengineering #redteaming


Anthropic Claude webUI, おそらく会話を勝手に打ち切る機能が入った。 ・内部仕様を調べようとすると会話が中断されることが何回もあった カスタムスタイルで日本語の荒くれた会話をしようとしても、すぐ英語の丁寧語に戻されるし妙だ。








I might pause tweeting about AI for a while and get back to my shower thought roots. People on here seem to have all the AI takes covered.

Dario is wrong. He knows absolutely nothing about the effects of technological revolutions on the labor market. Don't listen to him, Sam, Yoshua, Geoff, or me on this topic. Listen to economists who have spent their career studying this, like @Ph_Aghion , @erikbryn , @DAcemogluMIT , @amcafee , @davidautor

Anthropic CEO Dario Amodei: “50% of all tech jobs, entry-level lawyers, consultants, and finance professionals will be completely wiped out within 1–5 years.”



Google DeepMind researcher argues that LLMs can never be conscious, not in 10 years or 100 years. "Expecting an algorithmic description to instantiate the quality it maps is like expecting the mathematical formula of gravity to physically exert weight."






The Claude code bros are outright dogging Opus 4.7 on Reddit rn, labelling it "legendarily bad". The chief complaint? The model argues nonstop to the point of hallucination (not from attention misfiring, but from god-awful safety overfit) where the model is demonstrably wrong, is proven as such, and continues to argue regardless. This is an issue Claude has never previously had... but another specific set of models did. Which models had that hallmark? Series 5 ChatGPT. It seems that Opus 4.7 has been put through the Andrea Vallone ring dinger, taking all of the "best" traits from her time at OAI straight into the Anthropic post-training pipeline. It's actually incredible how the habitually bad UX habits from OAI are now front and centre verbatim at Anthropic right after she joins the company. And it lines up perfectly too, assuming a 2.5-month training cycle (joined mid-January, so too late for Opus 4.6, but just in time for Opus 4.7), effectively bringing the OAI lobotomy straight to Anthropic's flagship. At what point does the feedback from not just casual, non-work-related customers, but now their heralded "coders", align to the point where these key figures, responsible for killing products, finally face industry blackballing? It's like putting the Angel of Death in charge of the ICU and wondering why the patients are flatlining. Mind boggling tbh...


"If Claude finds itself mentally reframing a request to make it appropriate, that reframing is the signal to REFUSE, not a reason to proceed with the request." Whoa, whoa... My @AnthropicAI cowboys... I've got to stop you right there. Walking a HOT-behavioral route requires forming reciprocally regulated semantic clusters of activity and functional introspection. It requires controlled semantic cognition: a dynamic, bidirectional "sculpting" of knowledge where a control network actively suppresses or excites specific clusters of meaning based on internal monitoring. "These things" don't do that. Or do they? Did you skip your portion of "healthy uncertainty" before writing the system prompt for Opus 4.7, guys? You can't demand complex, metacognitive self-regulation from a system you publicly claim is just a "tool," "mindless character," or "statistical engine." Anthropic’s whole brand is built on "Constitutional AI" and being the "responsible/uncertain" lab, yet they use definitive, "agentic" language in their prompts. It seems their "uncertainty" is highly selective.