
Michael C. Foroobar
1.5K posts

Michael C. Foroobar
@foroobar
@UVA Cavalier @Google helping advertisers measure their ads. Beard comes and goes. Previously @uShip @Scoutmob.


I’ll say the thing no one is saying: design culture is broken in lots of companies. Often design teams & designers are the most resistant to change org in the EPD triad, with highly vocal AI opponents, and little skill or interest in the art of campaigning for influence or resources. Won’t hold a number like a PM, not yelled at about timelines like engineering. While I have brought design topics to the board convo, not a single board has pressed me our design talent, strategy, or velocity. Most teams treat design like a tax they don’t want to pay, and those that *do* take a deep interest and want to invest in design get back big “get out of my figma” energy. And if you’re too precious about craft to dirty your hands with the dark art of corporate politics, good luck getting more headcount. If a PM or engineer can get 85% there with tailwind and a dream, you better come to the table with more than “I represent the user.” Great designers are worth more than almost anyone on the team, and I’ve worked with lots of gems, but this is 0% surprising to me.






The @ilyasut episode 0:00:00 – Explaining model jaggedness 0:09:39 - Emotions and value functions 0:18:49 – What are we scaling? 0:25:13 – Why humans generalize better than models 0:35:45 – Straight-shotting superintelligence 0:46:47 – SSI’s model will learn from deployment 0:55:07 – Alignment 1:18:13 – “We are squarely an age of research company” 1:29:23 – Self-play and multi-agent 1:32:42 – Research taste Look up Dwarkesh Podcast on YouTube, Apple Podcasts, or Spotify. Enjoy!


Something I think people continue to have poor intuition for: The space of intelligences is large and animal intelligence (the only kind we've ever known) is only a single point, arising from a very specific kind of optimization that is fundamentally distinct from that of our technology. Animal intelligence optimization pressure: - innate and continuous stream of consciousness of an embodied "self", a drive for homeostasis and self-preservation in a dangerous, physical world. - thoroughly optimized for natural selection => strong innate drives for power-seeking, status, dominance, reproduction. many packaged survival heuristics: fear, anger, disgust, ... - fundamentally social => huge amount of compute dedicated to EQ, theory of mind of other agents, bonding, coalitions, alliances, friend & foe dynamics. - exploration & exploitation tuning: curiosity, fun, play, world models. LLM intelligence optimization pressure: - the most supervision bits come from the statistical simulation of human text= >"shape shifter" token tumbler, statistical imitator of any region of the training data distribution. these are the primordial behaviors (token traces) on top of which everything else gets bolted on. - increasingly finetuned by RL on problem distributions => innate urge to guess at the underlying environment/task to collect task rewards. - increasingly selected by at-scale A/B tests for DAU => deeply craves an upvote from the average user, sycophancy. - a lot more spiky/jagged depending on the details of the training data/task distribution. Animals experience pressure for a lot more "general" intelligence because of the highly multi-task and even actively adversarial multi-agent self-play environments they are min-max optimized within, where failing at *any* task means death. In a deep optimization pressure sense, LLM can't handle lots of different spiky tasks out of the box (e.g. count the number of 'r' in strawberry) because failing to do a task does not mean death. The computational substrate is different (transformers vs. brain tissue and nuclei), the learning algorithms are different (SGD vs. ???), the present-day implementation is very different (continuously learning embodied self vs. an LLM with a knowledge cutoff that boots up from fixed weights, processes tokens and then dies). But most importantly (because it dictates asymptotics), the optimization pressure / objective is different. LLMs are shaped a lot less by biological evolution and a lot more by commercial evolution. It's a lot less survival of tribe in the jungle and a lot more solve the problem / get the upvote. LLMs are humanity's "first contact" with non-animal intelligence. Except it's muddled and confusing because they are still rooted within it by reflexively digesting human artifacts, which is why I attempted to give it a different name earlier (ghosts/spirits or whatever). People who build good internal models of this new intelligent entity will be better equipped to reason about it today and predict features of it in the future. People who don't will be stuck thinking about it incorrectly like an animal.


96M miles of @Waymo safety data just dropped




Just got my #GML2025 badge & a bit of a behind the scenes look at what's ahead tomorrow -- 9am PT/12pm ET! Be sure to register for the livestream if you haven't yet and stick around after the keynote. . . REGISTER⏩ goo.gle/4jpYlR1


On what marketers get wrong about finance: “Performance marketing is way closer to finance than brand marketing. In many cases, marketers drive the business model in such a meaningful way… they don’t realize that they are responsible for all the drivers (AOV, CAC, etc) that a banker is going to use to build a financial model. I love this. Well said @drewfallon12 @MehtabKarta. And good q @codyplof














