

Feng Peng
706 posts

@feng
Engineer. Ex-Ask/Twitter Data Infra, building @leettools





Total token use as a measure of AI literacy is wrong headed. In my experience, after some baseline, more token use is inversely correlated with competency using AI.


We've reached an agreement to acquire Astral. After we close, OpenAI plans for @astral_sh to join our Codex team, with a continued focus on building great tools and advancing the shared mission of making developers more productive. openai.com/index/openai-t…

I built an app to simulate the 2026 NCAA tournament! It uses historical data, KenPom rankings, game locations, and more to determine the win probability. ...but then has an AI model review the results and prompt for the reality of March Madness, unpredictable!

The day has finally come that any open source repo can be just rewritten in another language or just refactored into another repo without much human effort. I predicted this a while ago that the open source model has to be adjusted to the new era. Community may be the last moat an OSS project can have, but it is also much weaker now. I also predicted that some infra project will be rewritten to get rid of old jank and then only reuse the good parts to provide better and more efficient solutions. phoronix.com/news/Chardet-L…


prediction re the end of spreadsheets AI code gen means that anything that is currently modeled as a spreadsheet is better modeled in code. You get all the advantages of software - libraries, open source, AI, all the complexity and expressiveness. think about what spreadsheets actually are: they're business logic that's trapped in a grid. Pricing models, financial forecasts, inventory trackers, marketing attribution - these are all fundamentally *programs* that we've been writing in the worst possible IDE. No version control, no testing, no modularity. Just a fragile web of cell references that breaks when someone inserts a row. The only reason spreadsheets won is that the barrier to writing real software was too high. A finance analyst could learn =VLOOKUP in an afternoon but couldn't learn Python in a month. AI code gen flips that equation completely. Now the same analyst describes what they want in plain English, and gets a real application - with a database, a UI, error handling, the works. The marginal effort to go from "spreadsheet" to "software" just collapsed to near zero. this is a massive unlock. There are ~1 billion spreadsheet users worldwide. Most of them are building janky software without realizing it. When even 10% of those use cases migrate to actual code, you get an explosion of new micro-applications that look nothing like traditional software. Internal tools that used to live in a shared Google Sheet now become real products. The "shadow IT" spreadsheet that runs half the company's operations finally gets proper infrastructure. The interesting second-order effect: the spreadsheet was the great equalizer that let non-technical people build things. AI code gen is the *next* great equalizer, but the ceiling is 100x higher. We're about to see what happens when a billion knowledge workers can build real software.

Introducing the new /crawl endpoint - one API call and an entire site crawled. No scripts. No browser management. Just the content in HTML, Markdown, or JSON.




When it comes to AI agents / AI tooling + coding, I hear an awful lot of talk about: Efficiency Iteration speed / PR output rate / lines of codes produced I hear zero mentions about: Quality Customer obsession This will bite back, and it probably already is...



Pure software is rapidly becoming un-investable.

. Please reason about the abstraction thinking process and provide a good abstraction of the agent harness model." The results are not bad at all. From Codex 5.3:
Main lore around the origin of MapReduce was that we were rewriting our indexing pipeline for the search system, and we realized that lots of the different phases were conceptually simple but required large scale processing (extract link text from each page, identify language for each page, compute checksum of contents to identify duplicates, etc). Each phase needed to be parallelized, made robust to machine failures, etc. Squinting at each of the phases we came up with MapReduce as an abstraction where we could have an implementation that would do all the complex work under the abstraction boundary, and where the expression of the operations could be nice and simple.


