The Grid
97 posts


Minimax M2.7 released! And its a big one Highlights: Self-evolving - first model that helped build itself, running 100+ autonomous optimization loops during its own RL training (30% internal improvement). Strong coder - 56.2% on SWE-Pro (near Opus 4.6), 55.6% on VIBE-Pro, production debugging down to under 3 minutes. ML research agent - 66.6% medal rate on MLE Bench Lite, tying Gemini 3.1. Office work - top open-source ELO on GDPval-AA (1495), 97% skill adherence, can do end-to-end analyst workflows (reports, models, PPTs). Native multi-agent and a new open-source interactive character demo called OpenRoom.

"There's no world in which pricing doesn't significantly evolve when the technology is changing this quickly." - @nickaturley, Head of ChatGPT Agreed. The problem the labs are facing is that they've set a fixed price when the value is variable. Let me explain... Today, AI typically has a fixed cost. Nice for budget planning, but completely disconnected from the value of what's being delivered. This may continue at the retail level, but I believe we're likely to see it evolve to be more like the airline industry. What's the value of a seat on a plane? Is it the same when all seats are available vs when this flight and all those that follow are sold out? Of course not. At any given time, AI providers can only serve so many tokens (AI output). When that limit is reached, they're tapped out. Seats on the plane are full. To fly a plane, airlines have a base cost. They price aggressively low to get as close to covering that cost as possible. As they approach that level, and certainly as they pass it, seat costs go through the roof. AI providers face the same challenge. They need to serve a given number of tokens every second to cover their costs. Often they're nowhere near that level. To build a sustainable business, they need to OVERCHARGE everyone at all times to cover the difference. Instead, I expect we'll move to dynamic pricing. Pricing that reflects demand. @AnthropicAI already announced a light version of this with peak and off-peak rate limits. That's a start. But the natural evolution is toward an order book. Suppliers set how many tokens they're willing to sell at which price levels (limit orders). Demand determines market rate. This can be completely transparent to users — they just get the best price on the book (market buys). Once derivatives evolve, you get budget planning through options and futures. The primitives exist. They're working at scale. And they're coming to AI.



1/ OpenAI just launched GPT-5.4 Pro, their premium model at 12x the API cost of standard GPT-5.4. $30/M input tokens, $180/M output vs. $2.50/$15. I ran TaxCalcBench on Pro. The result: exactly tied with standard GPT-5.4 12x the price, 0% improvement But the full story is more nuanced:





Founder of Chicago-based prop trading firm DRW says compute will be world's top commodity in 10 years. People will spend more on GPUs than an oil, which means that there should be a futures financial market for GPUs. Interesting implications for startups and cloud providers:


Founder of Chicago-based prop trading firm DRW says compute will be world's top commodity in 10 years. People will spend more on GPUs than an oil, which means that there should be a futures financial market for GPUs. Interesting implications for startups and cloud providers:

This alone is a $1B+ idea. Price discovery for compute is a huge issue. If you want to price compute, you have to email/call every neocloud. Everyone gives a different price. Some deal direct, others push you to brokers. + all the creative financing deals complicates price discovery. Compute markets solve this.






