Paul Modderman
1.2K posts

Paul Modderman
@PaulModderman
Principal Nerd @boringnerds, founding engineer @novaintel_ai, author. @paulmodderman.bsky.social



"LLMs writing about the experience of being an LLM" is a moving and fresh genre of writing. This is my favorite example so far.


congrats to the very special team @cognition (@ScottWu46 @stevenkplus1 @walden_yan) on their newest fundraise -- and their big deployments turns out competitive programming (and density of intelligence) does translate well to company-building



We are opening up a new role at Quora: a single engineer who will use AI to automate manual work across the company and increase employee productivity. I will work closely with this person.


strong agree but also believe we'll see a surge of more people using code for creative expression–without explicitly needing to *see* that code, or learn languages, frameworks, design patterns programming languages aren't The Thing, they're the thing that gets you to The Thing


we should be teaching vim motions in elementary school


Another late-night Claude Code post. First, if you've just arrived here at the party, Claude Code is NOT the same thing as Claude 3.7 Sonnet, Claude.ai, nor any other Claudey thing. It is its own new experimental thing, also from Anthropic, makers of Claude and Other Claudey Things. Claude Code (CC) is a new coding assistant, one which, strangely enough, only runs in a terminal. Like an xterm, or a bash shell. Or any of six WSL shells that don't quite work. As a result, CC looks comically retro-futuristic: A late-1980s vision of what AI might become. And here we are. CC is what we all thought Devin was going to be last year. When Devin came out in December, people Muntz-laughed and moved on. But CC is a bona-fide AI software engineer. It deserves the title. And the funny thing is, they appear to have eschewed RAG completely, and just told Claude to go figure stuff out on its own. I am here to say that I am addicted to Claude Code. I can't put it down. I don't mean that figuratively. I mean I literally do not know how to put my computer down and go to sleep. Because Claude Code keeps doing stuff. It keeps solving massive problems, one after another. I throw larger and larger things at it, and it is unfazed. Chomp. Chomp. Chomp. It's like that old Assassin's Creed game, Rome maybe, when you had that big network of spies working for you towards the end of the game, and you just sent them out on missions while you sat on your fat ass, and it was absolutely just as much fun as "running" around the game world? Well I remember. This, is that. You know what? We can't be more than 2-3 months away from being able to say, "Yo, CC, just... go make tests. For everything. All the stuff I failed to test over the past 2 decades, go redeem me. Write tests for it all, and make sure they are clever and meaningful, and follow our testing patterns." And then you deposit like, I dunno, $5000 into its gaping maw. It just goes off for a week or two, doing its thang on a branch somewhere, mostly I/O bound waiting on your builds. And one day you get the email you've been waiting for. It says: "Send Money". After a few more weeks of this, it finally takes you from 10% to 90% test coverage, so that when you die you will be admitted into Good Engineer Heaven. All other coding assistants will follow CC's approach, in some form factor. They are all falling over themselves right now, as we speak. Because yes, to answer all your exact same FAQs: Claude Code is that much better. The race is on!

I wish more people were taking seriously the possibility that @ezraklein and the leadership of the AI labs are raising: that AGI is a real possibility in the near future. You don't have to buy it yourself, but leaders & policymakers need to consider the possibility it is true.

I claimed the inference from X="LLMs are next token predictors" to Y="LLMs lack understanding, etc." is fallacious. Marcus claims that I'm saying not-X and not-Y. So I guess I'll point out that the inference "Y doesn't follow from X" to "not-X and not-Y" is also fallacious.

Interesting! Based on “The @AnthropicAI Economic Index” (millions of Claude conversations) AI usage is currently concentrated in software development and technical writing. TL;DR: 💻 Computer & Mathematical jobs dominate AI usage (37.2% of queries) despite being only 3.4% of workforce 📝 Writing/editing tasks are second most common (10.3% of queries) 🤝 AI augments human work (57%) more than automates it (43%) 💼 36% of jobs use AI for at least 25% of tasks, only 4% use it for 75%+ of tasks 💰 Mid-to-high wage jobs show highest AI adoption, both lowest and highest wage jobs show minimal usage 🔄 Tasks are being enhanced rather than jobs being fully automated 📊 Analysis based on real usage data from millions of Claude ai conversations 🎯 Usage concentrated in specific tasks rather than entire occupations

How AI Has Changed SAP's ABAP | With CR analyst Holger Mueller x.com/i/broadcasts/1…




o1 is mind-blowing when you know how to use it. it's really not a chat model -- you have to think of it more like a "report generator" (link to article below)


I spoke to / heard from a few senior AI labs people this weekend and I now feel even more confident in these predictions from a month ago. Hearing these same ideas come out of their mouths was quite strange in how real it became. There is a sequence of future events that if you think about from first principles seems inevitable. We really do now have a test-time compute paradigm that is a straight shot to AGI (no matter how you define AGI). The bottleneck is gathering good data for RL and then all the little product details like improving speed, integrations into our workflows, etc. Just assume that by end of 2025, at minimum, we'll have multimodal agentic phd AIs doing complex tasks for you on your computer at o1-mini speeds. I highly recommend people update their worldviews and company plans to assume this. This is not hype. The world will still feel similar bc it will take long for Bank of America and the government of France and everyone to integrate AGI and there will be a compute shortage. But the AGI will exist nonetheless. Robotics will also take a bit longer and timelines are a bit less clear there but I also talked to people who are doing robotics, and seems we're only talking a couple years later for fully functional humanoid robots. I believe there is a major gap in our information ecosystem about these topics. There are a couple thousand researchers and engineers thrusting us into the next era of humanity at breakneck speed, but they're not allowed to talk about it bc of NDAs. This is the most important story of the century -- far more important than covid -- yet there is far less media and discussion about it, beside hype tweets from AI influencers and coy tweets from AI researchers. That's because there are very few people who: a) understand how these systems work b) are able to suppress the "this sounds too crazy" part of their brain c) are able to suppress the "go full hype" part of their brain d) are in SF so hear things on the ground beyond twitter/the news e) actually care about helping the world and are not just trying to sell you something f) recognize the gravity of the situation and treat it as such g) are reasonably competent at converting thoughts into words I believe I am one of these people. As a person who's been running an AI company that trains real models, but one that is not an AI lab (so I can say whatever I want), and someone who enjoys writing about things that are helpful, I believe I can bring an important perspective to this quiet but immensely important conversation. I feel a duty to write more on X, and plan to do so. You have my word that I will always strive to be as correct as possible and will not write things with hype, though I won't be scared of being provocative. I also highly encourage anyone else out there who passes the criteria above to write more, create more. The world is desperately in need of media about the world that is coming. We need realistic AI short stories, new films beyond Terminator, mocked video demos of what future systems will look like. We all need visceral examples to shake us out of apathy so that we don't just predict the AGI but feel it. That way we can prepare accordingly.


