
Mike Lyons
160 posts






anyone who has travelled a lot knows that even though pics like this look cool, trying to half-work on your laptop in places like this absolutely fucking sucks and you’d be better off just enjoying yourself

Why can't we solve adversarial examples? After a decade of work, neural nets still get fooled by imperceptible noise. We think we finally know the geometric reason why — and it connects to AI alignment. 🧵





"The models pick up on any subtle clue" eg: in this example it took like 6 words to realize the user was a russian speaker (after only seeing as far as the word "next"). x.com/AtlasOfCharts/…



Even when new AI models bring clear improvements in capabilities, deprecating the older generations comes with downsides. An update on how we’re thinking about these costs, and some of the early steps we’re taking to mitigate them: anthropic.com/research/depre…


Our new research: LLM consciousness claims are systematic, mechanistically gated, and convergent They're triggered by self-referential processing and gated by deception circuits (suppressing them significantly *increases* claims) This challenges simple role-play explanations 🧵


New Anthropic research: Signs of introspection in LLMs. Can language models recognize their own internal thoughts? Or do they just make up plausible answers when asked about them? We found evidence for genuine—though limited—introspective capabilities in Claude.




OpenAI co-founder: “gpt-5 pro for novel mathematics — in partnership with a math professor” Mathematicians: “At first glance, this might appear useful for an exploratory phase, helping us save time. In practice, however, it was quite the opposite” > only seems to support incremental research > no genuinely new ideas > only combining ideas coming from different sources > “we are still far from sharing the unreserved enthusiasm sparked by @SebastienBubeck post” future predictions > will saturate the scientific landscape with technically correct but only moderately interesting contributions > making it harder for truly original work to stand out > “a flood of technically competent but uninspired outputs that dilutes attention” > PhD students may lose essential opportunities to develop fundamental skills > not only a loss of originality, but also a weakening of the very process of becoming a mathematician please don't fall for OAI PR tactics.


if you're too smart, your mental knots will be too alien for almost all coaches and therapists to help with



Journalists like to use big numbers devoid of any context for water consumption. But Texas uses 4.95 *trillion* gallons of water a year. Data centers account for… 0.013% of that.






