
gerred
11.3K posts

gerred
@devgerred
senior principal mts. model whisperer. chasing the speed of light.
가입일 Mart 2020
1.2K 팔로잉2.5K 팔로워

@lumpenspace if you go by russia offering a sub to new zealand to pay off debts, you could just put the whole nuclear sub in a lake and a datacenter on top
theguardian.com/world/2013/oct…
English

*kantians reasoning nervously amongst themselves*
bubble boi@bubbleboi
We killed the Kantian leader of Iran.. next up is a Hegelian.
English

@HanchungLee @samuelcolvin you know someone will have written autodiff in the type system itself
English

PYTHON + RUST.
Python inside the sandbox, Python (and some Rust) outside the sandbox.
Typescript for frontend developers trying to stay relevant, like Mastra.
By the way, the SF bubble is the one place where TS is popular for AI: `openai` package has 53m weekly downloads on PyPI, and 10m on NPM.
AND THAT GAP IS WIDENING - used to be 4x, now 5x.
jason liu@jxnlco
Future of AI
English

@thdxr tbqh maybe it's the editor that's skeumorphic. what you're wanting in "using it less" is real, but much like spreadsheets mirror their analogue equivalents, I think a lot about post-intelligence age UX. i don't think we're going back to the editor to solve the supervision problem
English

@jxnlco @jekbradbury You could totally do a pseudo prefix+(simulated, I suppose) radix caching from there to go further but that's really a profiling question from there. With hierarchical KV cache though, it's not out of the question.
English

I obviously don't have knowledge of what they're doing over there but it's a reasonable way to structure it, much like caching compilation graphs, pre compiling and shipping versioned KV caches is a pretty obvious optimization.
Ant has a lot of caching options in their docs that would lead to this has at least entered their minds vs the less sophisticated pre-caching everyone else outside of Google does. I used to use GCP's provisioned caching as a very cost effective, high context limit docs oracle because I could provision and expire it at my leisure.
English

I'm betting the Anthropic ban of OpenCode is as technical and cost-saving as it is political. I've long argued there's a moat to be had by closing third party tools to subs. CC can rely on KV caching across every instance, and have KV caches on a per-organization basis for further customization for their largest customers.
They can, across their entire fleet, pre-compute 1/3-1/2 (if not more) of every CC user's system prompt. By encouraging baking this into MDM and enterprise plans too, they can further negotiate that out in these large contracts. It also potentially lets them do some more clever things than just pure prefix caching and make specific tradeoffs you don't just get by allowing anybody to use those endpoints.
At least that's how I'd do it. It surprised me it took THIS long.
English

Olympian Gold Medalist Alysa Liu, recently went viral for her Teen Vogue rant on OpenAI Codex.
“I can see why Sam Altman open sourced Codex. Clearly the experience is significantly worse than Claude Code. I was unable to feel the AGI using Codex. As oppose to using Claude Code, I felt the enlightenment coming and support UBI ”


English

@stochasticchasm I can't wait to rent your instance in cortical labs cloud.
English

This is a research preview that we'll be expanding more on.
Read more in our docs on Channels here: code.claude.com/docs/en/channe…
English
gerred 리트윗함

@_BILLDING_ They can also better avoid shenanigans like quantized KV caching for their own products. So then there's even a quality edge CC can have.
I'm like the only inferencing expert that actually puts a product hat on it feels like sometimes.
English

@_BILLDING_ Yep now imagine that cached at mass scale across every instance for a long time, instead of letting that go cold (like any API user's would). Hierarchical caching is expensive but the cost savings are very worth it.
English

not shocking at all; the models don't want to write in byzantine esoteric languages instead of python or rust or whatever
Lossfunk@lossfunk
🚨 Shocking: Frontier LLMs score 85-95% on standard coding benchmarks. We gave them equivalent problems in languages they couldn't have memorized. They collapsed to 0-11%. Presenting EsoLang-Bench. Accepted to the Logical Reasoning and ICBINB workshops at ICLR 2026 🧵
English





