Alex Kemper
247 posts

Alex Kemper
@AJKemps
cofounder // fde @ampcode

This is wild. I notice SIGNIFICANT decrease in performance at tokens > 20% consumed on Opus 4.6. It degrades INSANELY, like the 1M context doesn't matter. The model just starts being delusional and unusable. 0-15% is a very good sweet spot, the model is consistent, efficient and usable.


Since 2023, the top quartile of AI spenders on @tryramp have more than doubled their revenue. Bottom quartile? Flat A roofing company in Texas. A window installer in Utah. A construction firm in Florida that grew 65% The gap is accelerating and most companies don't feel it yet







coming right up but TLDR Cut out all setup overhead (prev 30s, now about 1s) - ovh amd epic turin box with 128 cores and 256gb ram (1k/mo) - golden image of main with all cache loaded - zfs for instant copy of golden image (this is magic) - git fetch all every second for local mirror - golden image of database so only last migration runs (as Postgres template) - turbo cache locally For actual suites - much higher sharding since now no overhead to each shard - use @bunjavascript tests where possible to avoid typescript compilation - incremental typechecking with local cache Bypassing GitHub actions in favor of custom check suites - a few seconds of queue time saved - no action minutes billed (we hit 36k minutes in 3 days) For preview apps - JiT full stack preview apps (not deployed on each commit) - 2-3s cold start on any commit sha to a fully deployed full stack preview app - zfs clone of golden firecracker vm and then check out latest commit etc














