
Apple proved every LLM fails at basic reasoning. One random sentence breaks every frontier model. We're far from AGI.
MasteringMachines AI
5.7K posts

@MstrMachines
Discovering Tomorrow's Reality. A platform that highlights the innovation and absurdity of the AI industry.

Apple proved every LLM fails at basic reasoning. One random sentence breaks every frontier model. We're far from AGI.




I don't know what the fuss is about. Anthropic's rules on using subscriptions are very simple: Claude Code = OK Claude's online platform = OK Agent SDK running in personal software = OK... ish? Agent SDK running in commercial software = NOT OK Claude Code running in CI = ?? Oh, maybe it's not so simple... Agent SDK running in CI = ?? claude -p running in CI = ?? claude -p running in personal software = OK claude -p running on open source software, but run on my personal computer = ?? claude -p running on distributed sandboxes, kicked off by me = ?? Distributing open source software which relies on claude -p, and documenting how to use your subscription with it = ?? A thousand other edge cases = ?? Let me be clear. I have never before experienced, from any developer tool, such a frustrating lack of clarity over the basic terms of usage. I personally asked, 3 weeks ago, and have received nothing but delays. The recent @bcherny announcement did absolutely nothing to clarify things. I say this as someone who just released a Claude Code course - my incentives all align with supporting Anthropic.







What if the whole LLM thing is a false start? If the flaws are inherent systemic problems - if the compounding of hallucinations/errors can't be sorted out? If the capex build out is one of the biggest misallocations of capital ever? Then what? bloomberg.com/news/newslette…



🐳DeepSeek delayed its V4 model release so it could run on Huawei's chips. important milestone for China