
Trevor Gurgick
1.4K posts

Trevor Gurgick
@TGurgick
Teaching machines + humans to work better, together 🤝 @Fanatics | AI, Data, Robotics | Product | Startups | @MIT @NSF I-Corps adjunct | Fmr @Amazon @AOL @Yahoo



I've never felt this much behind as a programmer. The profession is being dramatically refactored as the bits contributed by the programmer are increasingly sparse and between. I have a sense that I could be 10X more powerful if I just properly string together what has become available over the last ~year and a failure to claim the boost feels decidedly like skill issue. There's a new programmable layer of abstraction to master (in addition to the usual layers below) involving agents, subagents, their prompts, contexts, memory, modes, permissions, tools, plugins, skills, hooks, MCP, LSP, slash commands, workflows, IDE integrations, and a need to build an all-encompassing mental model for strengths and pitfalls of fundamentally stochastic, fallible, unintelligible and changing entities suddenly intermingled with what used to be good old fashioned engineering. Clearly some powerful alien tool was handed around except it comes with no manual and everyone has to figure out how to hold it and operate it, while the resulting magnitude 9 earthquake is rocking the profession. Roll up your sleeves to not fall behind.

At @Google, we are moving from a writing‑first culture to a building‑first one. Writing was a proxy for clear thinking, optimized for scarce eng resources and long dev cycles - you had to get it right before you built. Now, when time to vibe-code prototype ≈ time to write PRD, PMs can SHOW not tell. Role profiles are blurring, creativity and building are happening in parallel.




The real shiptober (plus one day) was at Anthropic: • 11/1 - Token counting API • 11/1 - Multimodal PDF support across claude and the API • 10/31 - Voice dictation in Claude mobile apps • 10/31 - Claude desktop app • 10/29 - Claude in Github Copilot • 10/24 - Analysis tool • 10/22 - New Claude 3.5 Sonnet • 10/22 - Computer use API • 10/18 - Financial analyst quickstart • 10/17 - Mobile app design overhaul • 10/9 - Remove message order restrictions in API • 10/8 - Message Batches API • 10/4 - Artifacts errors auto-fix Btw we are able to ship this much because we use Claude all the time

We've built an API that allows Claude to perceive and interact with computer interfaces. This API enables Claude to translate prompts into computer commands. Developers can use it to automate repetitive tasks, conduct testing and QA, and perform open-ended research.

Excellent article by Sonya Huang and Pat Grady of @Sequoia, "The Agentic Reasoning Era Begins", and the $10 trillion opportunity with service-as-a-software: sequoiacap.com/article/genera… "Thanks to agentic reasoning, the AI transition is service-as-a-software. Software companies turn labor into software. That means the addressable market is not the software market, but the services market measured in the trillions of dollars."

Stripe data shows that top AI startups in 2024 (ex: OpenAI, Anthropic, Mistral, Midjourney) are making money faster than equivalent SaaS companies in 2018. Al startups that hit at least $2.5M/mo rev achieved the milestone in 20 months — 5x faster than past SaaS startups. Do we think that’s because VC money is more concentrated? TikTok? Actual higher interest? OAI skewing everything? More here from FT: ft.com/content/a9a192…



NotebookLM is quite powerful and worth playing with notebooklm.google It is a bit of a re-imagination of the UIUX of working with LLMs organized around a collection of sources you upload and then refer to with queries, seeing results alongside and with citations. But the current most new/impressive feature (that is surprisingly hidden almost as an afterthought) is the ability to generate a 2-person podcast episode based on any content you upload. For example someone took my "bitcoin from scratch" post from a long time ago: karpathy.github.io/2021/06/21/blo… and converted it to podcast, quite impressive: notebooklm.google.com/notebook/ba017… You can podcastify *anything*. I give it train_gpt2.c (C code that trains GPT-2): github.com/karpathy/llm.c… and made a podcast about that: notebooklm.google.com/notebook/2585c… I don't know if I'd exactly agree with the framing of the conversation and the emphasis or the descriptions of layernorm and matmul etc but there's hints of greatness here and in any case it's highly entertaining. Imo LLM capability (IQ, but also memory (context length), multimodal, etc.) is getting way ahead of the UIUX of packaging it into products. Think Code Interpreter, Claude Artifacts, Cursor/Replit, NotebookLM, etc. I expect (and look forward to) a lot more and different paradigms of interaction than just chat. That's what I think is ultimately so compelling about the 2-person podcast format as a UIUX exploration. It lifts two major "barriers to enjoyment" of LLMs. 1 Chat is hard. You don't know what to say or ask. In the 2-person podcast format, the question asking is also delegated to an AI so you get a lot more chill experience instead of being a synchronous constraint in the generating process. 2 Reading is hard and it's much easier to just lean back and listen.









