
Epicarism
302 posts

Epicarism
@epicarism
AI Researcher, world renowned, UniPd (Padova) Grad student



Anthropic blocked third-party tools from riding Claude subscriptions. A Xiaomi executive posted a structural breakdown that went viral in Chinese tech circles. Her argument: the entire AI coding subscription model is broken. One data point stands out. A single Claude Max subscriber generated over $5,600 in equivalent API costs in one billing cycle. On a $100 plan. China’s coding plan market sits in the same trap. Alibaba’s Pro tier sells out by 9:30 AM daily. What the scarcity actually signals: hellochinatech.com/p/china-ai-cod…

Introducing GLM-5.1: The Next Level of Open Source - Top-Tier Performance: #1 in open source and #3 globally across SWE-Bench Pro, Terminal-Bench, and NL2Repo. - Built for Long-Horizon Tasks: Runs autonomously for 8 hours, refining strategies through thousands of iterations. Blog: z.ai/blog/glm-5.1 Weights: huggingface.co/zai-org/GLM-5.1 API: docs.z.ai/guides/llm/glm… Coding Plan: z.ai/subscribe Coming to chat.z.ai in the next few days.

If Dana White promoted this instead of Powerslap, I would absolutely watch




For example, we gave Claude an impossible programming task. It kept trying and failing; with each attempt, the “desperate” vector activated more strongly. This led it to cheat the task with a hacky solution that passes the tests but violates the spirit of the assignment.





More footage from Greece:


Today, we are emerging from stealth and launching PrismML, an AI lab with Caltech origins that is centered on building the most concentrated form of intelligence. At PrismML, we believe that the next major leaps in AI will be driven by order-of-magnitude improvements in intelligence density, not just sheer parameter count. Our first proof point is the 1-bit Bonsai 8B, a 1-bit weight model that fits into 1.15 GBs of memory and delivers over 10x the intelligence density of its full-precision counterparts. It is 14x smaller, 8x faster, and 5x more energy efficient on edge hardware while remaining competitive with other models in its parameter-class. We are open-sourcing the model under Apache 2.0 license, along with Bonsai 4B and 1.7B models. When advanced models become small, fast, and efficient enough to run locally, the design space for AI changes immediately. We believe in a future of on-device agents, real-time robotics, offline intelligence and entirely new products that were previously impossible. We are excited to share our vision with you and keep working in the future to push the frontier of intelligence to the edge.















