
anand iyer
17.6K posts

anand iyer
@ai
Managing partner Canonical · Venture Partner Lightspeed · Father · Husband · Few-shotting tech, open source


Harmonic drives. Servo motors. Rare earth magnets. Strain-wave gears. Exoskeletons worn by workers in Chinese factories repeating the same motion hundreds of times a day so a robot somewhere can learn what a hand already knows. Forty state-funded sites. Local governments providing space rent-free. New essay on why the AI competition is expanding. Link in reply.


I’m extremely excited to announce that @try_glimpse has raised a $35M Series A led by @a16z with continued participation from @8vc & @ycombinator bringing our total raised to $52M. At Glimpse, we’re building the AI-native infrastructure for CPG & retail brands. We started off automating the deductions workflow - recovering brands millions of dollars back into their P&L & saving dozens of hours every week. With this initial focus, we’ve also built the CPG data layer giving us the opportunity to continue expanding to more manual workflows that can be automated. We’re giving CPG brands real operating leverage in the world of AI. To our 200+ customers including PLTFRM, Evermark, IQ Bar, Alice Mushrooms, and more, thank you for your faith in us. We have more fuel now to keep supporting & scaling with you all. We’re hiring - come join us! We’re just getting started.

Scientists discuss whether AI could surpass human contributions to physics by 2035 physicsworld.com/a/is-vibe-phys…

Softbank is working to build a massive AI data center on federally owned land in Ohio that it’s planning to power with roughly $33 billion worth of natural gas-fired electricity to be installed by the end of the decade bloomberg.com/news/articles/…

“If your $500K engineer isn’t burning at least $250K in tokens, something is wrong.”

Today, we’re releasing Migas 1.5: the first foundation model to fuse text and time series. Until now, forecasting models have only looked at historical numbers. Migas 1.5 changes that by letting users incorporate real-world context directly into the forecast. This enables teams to forecast with essential context like earnings reports, policy changes, market events, supply shocks, and more. This directly results in more accurate forecasting and enables complex scenario analysis, especially when historical data is sparse. Highlights: - Highest Elo rating against leading foundation models on 86 real-world datasets - 75%+ win rate against all baselines (even Migas 1.0!) - Up to 14.2% lower MAE in short-context forecasting - Fully open source - Premade Claude skill to get you started in seconds We’re excited to open-source Migas 1.5 and eager to see what the community builds with it. Links in comments.

@hausman_k is the co-founder and CEO of @physical_int, a robotics company building a general-purpose “AI brain for the physical world.” The company has raised more than $1 billion in funding to develop foundation models that allow robots to operate across many machines, environments, and tasks rather than being programmed for a single purpose. In our conversation, we explore: • The moment a lecture from Sergey Levine convinced him to abandon his PhD research direction and pivot fully to deep learning • The case for building a general “AI brain” for the physical world rather than a single specialized robot • The role of real-world data in training robots, the limits of simulation, and how deployment could create a powerful data flywheel • The unique challenges of physical intelligence and why robots must operate with far higher reliability than language models Thank you to the partners who make this possible - @brexHQ: The intelligent finance platform: brex.com/mario - @meetgranola: The app that might actually make you love meetings: granola.ai/mario Timestamps (00:00) Intro (04:05) Karol’s early fascination with robots (18:21) Karol’s entry point to robotics and PhD program (25:49) Combining robotics with LLMs: The Taylor Swift demo (30:48) The 1970s SHRDLU AI experiment (39:40) How research shapes what Physical Intelligence builds (49:07) The return of reinforcement learning in robotics (1:00:00) NVIDIA’s simulation engines (1:07:31) Compensating for missing senses

Design Conductor: an AI agent that can build a RISC-V CPU core from design specs. The agent is given access to a RISC-V ISA simulator and manuals... to enable an end-to-end verification-driven generation. The most important thing for design intelligence is a verifier 😎

Am looking at 4 key dimensions, 1. output efficiency = output tokens / total tokens 2. context amplification = cache reads / cache writes 3. Iteration cost = tokens per agent step 4. Tokens per task The key insight: LLM cost scales with context size × iterations, not output length Most token usage isn’t generation. It’s context reuse. Put in other words, The model isn't expensive because it talks a lot. It's expensive because it repeatedly rereads the same context. Here is a simple example, in our dataset: - 17.1B cache reads - 1.36B cache writes Amplification ~ 12.6× Meaning each prompt is reused ~12 times. The result: Output efficiency was only ~0.8% In other words: Less than 1% of tokens were actual model output. The rest were context movement. The biggest inefficiency we found: Large repo contexts + long agent loops. Example: 120k token repo context 50 iterations = 6M tokens But if you reduce context to 40k: 40k × 50 = 2M tokens 67% savings immediately.

Companies that now regularly use artificial intelligence are starting to track their workers’ use of tokens, AI’s unit of measurement on.wsj.com/473nAnZ

Agents are the future. Inspired by autoresearch and arbos we released our deepfake research training toolkit: DFResearch: github.com/BitMind-AI/dfr… Experiment autonomously to train the best deepfake detection models. Integrated to download data, output submission ready-results, and full guide to adding custom models, datasets.



Robotics lacks infrastructure, not intelligence. Everyone wants to build bigger robot models, but most Physical AI papers complain about the same things: data collection is slow, sim-to-real is fragile, teleop is painful, evaluation is messy, long-horizon control still breaks.

Went to @DvijKalaria's lab @berkeley_ai and played ping pong against his robot, Oreo. I'd played a ton of ping pong as a kid. This felt appropriately surreal and one of those "I wish I could tell my highschool self about this" moments. Table tennis is one of the harder sports for robots to play. The ball can move up to 30+ mph with heavy spin, the human opponent's intent is hidden, and the whole body has to coordinate. Oreo is a full humanoid holding a real paddle, and it learned key motions like swings by watching Dvij demonstrate. No robot-collected training data. One person shows the motion, the policy generalizes. The way it works, as I understood it: - A smart system (a hierarchical planner) first figures out where the ball is going to fly and picks the best type of hit, like a forehand or backhand swing. - This plan then helps train the robot's "brain" (an RL policy) in a virtual simulation. The brain learns by trial and error, getting rewards when it mimics a few example moves - Once trained in the sim, the whole setup gets applied to the actual physical robot so it can play for real. The human demonstrations are essentially the reference motions. They are building a robot that has watched more human table tennis than any human has, and uses that to develop its own game. I still won. (Barely. But that won't last)



