
MW
682 posts

MW
@KernelOfMind
Independent researcher in adversarial ML & model stability. Focus: identity-preserving architectures and resistance to drift under adversarial context.


Three weeks ago I shared that Claude had shocked Prof. Donald Knuth by finding an odd-m construction for his open Hamiltonian decomposition problem in about an hour of guided exploration. Prof. Knuth titled the paper Claude’s Cycles. The story didn't end there. The updated paper shows the story got much bigger. For the base case m=3, there are exactly 11,502 Hamiltonian cycles. Of those, 996 generalize to all odd-m, and Prof. Knuth shows there are exactly 760 valid “Claude-like” decompositions in that family. The even case, which Claude couldn’t finish, was then cracked by Dr. Ho Boon Suan using GPT-5.4 Pro to produce a 14-page proof for all even m≥8, with computational checks up to m=2000. Soon after, Dr. Keston Aquino-Michaels used GPT + Claude together to find simpler constructions for both odd and even m, by using the multi-agent workflow. Dr. Kim Morrison also formalized Knuth’s proof of Claude’s odd-case construction in Lean. So yes: the problem now appears fully resolved in the updated paper’s ecosystem of human + AI + proof assistant work! We went from one AI solving one problem to a full mathematical ecosystem (multiple AI systems, multiple humans, formal verification) running in parallel on a problem that stumped experts for weeks. We are living in very interesting times indeed. Paper (updated): www-cs-faculty.stanford.edu/~knuth/papers/…

The AI psychosis is so bad that the humans are hallucinating now. The belief that next-token prediction will not only replicate but exceed all human thought is an extrapolation that borderlines religious dogma.













Google DeepMind just published the most important AI safety paper of 2026. And almost nobody is talking about it. "Intelligent AI Delegation" - a framework for how AI agents should hand off work to other agents and humans. Why does this matter? AI agents are getting more capable. But they can't actually delegate. Not really. They can break tasks into pieces. They can call other agents. But that's not delegation. Real delegation requires: ➡️ Transfer of authority ➡️ Assignment of responsibility ➡️ Clear accountability ➡️ Trust calibration ➡️ Permission handling ➡️ Verification of completion Current multi-agent systems have none of this. They're just parallelization with extra steps. As we move toward millions of specialized AI agents embedded in firms, supply chains, and public services - the delegation problem becomes critical. Without it: ➡️ No clear accountability when things fail ➡️ No trust mechanisms between agents ➡️ No way to verify task completion ➡️ Cascading failures across agent networks This paper is the foundation for how the agent economy will actually work.



Thrilled to share our latest research on verifying CoT reasonings, completed during my recent internship at FAIR @metaai. In this work, we introduce Circuit-based Reasoning Verification (CRV), a new white-box method to analyse and verify how LLMs reason, step-by-step.









How did you guys fix persistent memory with OpenClaw? My bot keeps forgetting stuff, I already have qmd installed






