carboxydev
1.4K posts

carboxydev
@carboxydev
I write code and enjoy staring at pixels. interests: tech, AI and futurology. https://t.co/RgX9PKvf8T




Inspired by @theo, I organically coded a @raycast extension that blocks Twitter unless you have work currently running in your favorite agent! - works with any agent that supports hooks - blocks not only Twitter but also all social media sites, news, and more (customizable) - tracks your merits, allowing you to earn 15 million so you can participate in the "Hot Shot" game show! As you probably can tell, the name was inspired by @blackmirror 😊 S1E2 Here's the link for the extension: Day 132 of #100DaysOfRaycast 1 of 2




PSA: If you've been running out of Claude session quotas on Max tier, you're not alone. Read this. Some insane Redditor reverse engineered the Claude binaries with MITM to find 2 bugs that could have caused cache-invalidation. Tokens that aren't cached are 10x-20x more expensive and are killing your quota. If you're using your API keys with Claude this is even worse. This is also likely why this isn't uniform, while over 500 folks replied to me and said "me too", many (including me) didn't see this issue. There are 2 issues that are compounded here (per Redditor, I haven't independently confirmed this) : 1s bug he found is a string replacement bug in bun that invalidates cache. Apparently this has to do with the custom @bunjavascript binary that ships with standalone Claude CLI. The workaround there is to use Claude with `npx @anthropic-ai/claude-code` 2nd bug is worse, he claims that --resume always breaks cache. And there doesn't seem to be a workaround there, except pinning to a very old version (that will miss on tons of features) This bug is also documented on Github and confirmed by other folks. I won't entertain the conspiracy theories there that Anthropic "chooses" to ignore these bugs because it gets them more $$$, they are actively benefiting from everyone hitting as much cached tokens as possible, so this is absolutely a great find and it does align with my thoughts earlier. The very sudden spike in reporting for this, the non-uniform nature (some folks are completely fine, some folks are hitting quotas after saying "hey") definitely points to a bug. cc @trq212 @bcherny @_catwu for visibility in case this helps all of us.


Matt Maher tested frontier models in Cursor v. other harnesses. Cursor boosted model performance by 11% on average: Gemini: 52% → 57% GPT-5.4: 82% → 88% Opus: 77% → 93% His benchmark measures how well models implement a 100-feature PRD. @cursor_ai consistently outperformed.


ok i know my site looks like complete 💩 someone sent me a message that they could improve it and sent this mock.. am I crazy because that does not look any better at all i dont wanna be a jerk but i think they just used a single prompt, and not even w claude or gemini, but probably codex


What if the voices we love didn't have to disappear? They don't have to.

















