
My POV on front-end of 2026
Tyrell Downer
2.3K posts


My POV on front-end of 2026





🚀 Introducing the Qwen 3.5 Small Model Series Qwen3.5-0.8B · Qwen3.5-2B · Qwen3.5-4B · Qwen3.5-9B ✨ More intelligence, less compute. These small models are built on the same Qwen3.5 foundation — native multimodal, improved architecture, scaled RL: • 0.8B / 2B → tiny, fast, great for edge device • 4B → a surprisingly strong multimodal base for lightweight agents • 9B → compact, but already closing the gap with much larger models And yes — we’re also releasing the Base models as well. We hope this better supports research, experimentation, and real-world industrial innovation. Hugging Face: huggingface.co/collections/Qw… ModelScope: modelscope.cn/collections/Qw…



MiniMax M2.5 is now live as the first open-weight model inside Notion Custom Agents, optimized for lightweight, high-frequency agent workflows. Pretty cool to see open-weight models show up where scale and cost really matter. 🤠







I'm sick and tired of the people who don't understand why I spent $20,000 on this set up, and plan on spending another $100,000 by the end of the year IT DOES NOT MATTER THAT LOCAL MODELS AREN'T AS GOOD AS OPUS 4.6 That is not the point. The point is me being able to run a swarm of local AI agents powered by local AI models unlocks a world you can't imagine A world never discovered by humanity before Right now, as you read this post, I have multiple local AI models reading thousands of posts on X and Reddit Hunting for challenges to solve Those local AI models are feeding hundreds of challenges a day to a manager model The manager model (Henry) decides what the company (Alex Finn Global Enterprises) will build. The company is constantly working. Constantly researching. Constantly building. Constantly shipping If I did this with local models I'd be spending $20,000 a month on API calls. With my set up, it's free. I have an army on my desk. Never resting. Never eating. Never complaining. Always conquering. Here is your problem: it's not that you don't understand this. You don't want to understand this. You don't want to think this is possible. Your brain doesn't want to believe this is the world we now live in. It is. And the faster you can accept this and get on board, the faster you can enter the new society. Otherwise, you will forever be doomed to the permanent underclass. Make your choice.

TL;DR if you run inference every second of everyday on the Mac studios it will take you ~14 years to break even with API costs. For the Kimi K2.5 case, 24/tok per second is ~2M tokens per day. Which is ~$2-4 per day of API costs. Or $60-120 / mth. $20k capEx with a ~14 year payback period. It's good local is an option to prevent monopoly, but the idea it's a one time payment and you make all the LLM calls you want for free after is horribly misguided. You get ~2M tokens per day if you're pushing it hardcore. If you adjust for smaller models getting better and more sensibly hardware costs going up, then it looks more reasonable as a long term investment. Important to remember inference requiring less compute will just make the cloud cheaper. Right now, unless your going opus or codex, which you can't run locally anyway, it is generally significantly cheaper to use the cloud.



Saturday night. 6 hours of sleep over the last week. My autonomous agent company having an emergency meeting on the left. My ClawdBot giving them new tasks on the right All being powered by local models in my Mac Studio data center I refuse to be in the permanent underclass

I'm sick and tired of the people who don't understand why I spent $20,000 on this set up, and plan on spending another $100,000 by the end of the year IT DOES NOT MATTER THAT LOCAL MODELS AREN'T AS GOOD AS OPUS 4.6 That is not the point. The point is me being able to run a swarm of local AI agents powered by local AI models unlocks a world you can't imagine A world never discovered by humanity before Right now, as you read this post, I have multiple local AI models reading thousands of posts on X and Reddit Hunting for challenges to solve Those local AI models are feeding hundreds of challenges a day to a manager model The manager model (Henry) decides what the company (Alex Finn Global Enterprises) will build. The company is constantly working. Constantly researching. Constantly building. Constantly shipping If I did this with local models I'd be spending $20,000 a month on API calls. With my set up, it's free. I have an army on my desk. Never resting. Never eating. Never complaining. Always conquering. Here is your problem: it's not that you don't understand this. You don't want to understand this. You don't want to think this is possible. Your brain doesn't want to believe this is the world we now live in. It is. And the faster you can accept this and get on board, the faster you can enter the new society. Otherwise, you will forever be doomed to the permanent underclass. Make your choice.



