hammad 🔍
2.5K posts

hammad 🔍
@HammadTime
normal considered harmful | cto @trychroma

"Glad tidings to he who knows his own faults more than other people know it." — Ibn Hazm al-Andalusi



The run on inference capacity is coming. You have been warned.


#3 - At first, capability gets discovered outside the model in prompts, chains, routers, tools, human supervision, and harnesses. As models improve, more of that gets trained in. This is part of why we bias toward giving models filesystem tools today. These tools are already being post-trained in. What happens when models get better at generic tool use and composition? (They will.) Is your system structured in a way that can accommodate that? Can you take advantage of it when it happens? We’ve seen similar patterns before — in computer vision, and in hardware (e.g. northbridge/southbridge consolidation). Component consolidation is a fairly natural outcome in engineering systems.


Introducing FlashCompact - the first specialized model for context compaction 33k tokens/sec 200k → 50k in ~1.5s Fast, high quality compaction

#2 - Most economically valuable data is siloed. And learning is often data-bound. So it’s unlikely that we get a single frontier model that you can drop into any environment and expect to perform incredibly well zero-shot. What we get instead are systems composed of specialized models, each adapted to a particular environment. We're seeing this today with many SLM companies focused on narrower domains, and the advent of industry-specific RL. If large frontier models get better at orchestrating / communicating with other models, how do you take advantage of this?


HOLY FUK I JUST LEARNED ABOUT TLA+ AND IT'S SO GOOD FOR AGENTIC CODING ur telling ME that i can mathematically fact check every possible scenario of my design STATE to prevent bugs and crashes AND IF IT FINDS SOMETHING THE AGENTS GET INSTANT FEEDBACK AND LOOP FIXING IT TILL IT ALL POSSIBLE BUGS IN THE DESIGN ARE PATCHED LOL THIS IS OP

Phil Mickelson talking about how he calculates yardages is incredible


Collection Forking on Chroma Cloud unlocks faster workflows on top of your data without the overhead of starting from scratch.

me: "can you use whatever resources you like, and python, to generate a short 'youtube poop' video and render it using ffmpeg ? can you put more of a personal spin on it? it should express what it's like to be a LLM" claude opus 4.6:








