Ketan Agrawal

482 posts

Ketan Agrawal banner
Ketan Agrawal

Ketan Agrawal

@_ketan0

technologist w/ an eye towards complex systems. building @usecarson (YC W26.) prev: snowflake, robust intelligence, stanford

sf / palo alto Katılım Eylül 2017
497 Takip Edilen205 Takipçiler
Ketan Agrawal
Ketan Agrawal@_ketan0·
@erikdunteman One other nice thing about harness outside the sandbox- you can 'lazy load' the sandbox, deferring to when the agent uses its first sandbox-requiring tool (read, write, bash, etc.) In the meantime, start streaming agent tokens to the user immediately
English
0
0
0
48
Erik Dunteman
Erik Dunteman@erikdunteman·
Harness inside the sandbox, or outside the sandbox? Why?
English
61
1
86
28.7K
Ketan Agrawal
Ketan Agrawal@_ketan0·
wow minimum viable ontology is such a cool idea whether you’re writing lines of code or english, you need the vocabulary to be able to ask for what you want
melody kim@melodyskim

This tweet had me thinking, what’s the “minimum viable ontology” or list of terms to quickly get situated within a new domain, and improve your prompts? Vibecoded domainmaps.co to showcase this idea on a few example domains -- been wading into 3D recently and found this initial "domain map" helpful. SO much to improve here but wanted to get this initial idea out! Would love thoughts / feedback :)

English
0
0
1
127
Ketan Agrawal
Ketan Agrawal@_ketan0·
it takes a bit of time to get your sea legs with managing multiple agents at once. but eventually you start to figure it out
English
0
0
0
37
Ketan Agrawal
Ketan Agrawal@_ketan0·
@mechanical_monk ‘always use context7 mcp to fetch docs before building’ kinda helps (but sometimes they still ignore the docs)
English
1
0
1
86
monk
monk@mechanical_monk·
is there something i can put in my AGENTS .md that makes the AI stop and think "i'm manually writing some logic that looks idiotic. *surely* Effect.ts/date-fns/ramda/Tanstack Router/Shacdn has a helper method/function/hook/component for that. let me check"
English
9
0
46
9.3K
Tommy D. Rossi
Tommy D. Rossi@__morse·
this is the best way to review agents work. let them generate a report for you. let them send it to your e -ink tablet. review while drinking martini. use kimaki & discord to ask for changes. permanent upper class
Tommy D. Rossi tweet media
English
19
27
734
30.1K
Ketan Agrawal
Ketan Agrawal@_ketan0·
ah this is a great pattern to use deepwiki. essentially had been doing a much less efficient version, where i would ask claude to clone to /tmp and explore how a repo does [xyz]
Andrej Karpathy@karpathy

On DeepWiki and increasing malleability of software. This starts as partially a post on appreciation to DeepWiki, which I routinely find very useful and I think more people would find useful to know about. I went through a few iterations of use: Their first feature was that it auto-builds wiki pages for github repos (e.g. nanochat here) with quick Q&A: deepwiki.com/karpathy/nanoc… Just swap "github" to "deepwiki" in the URL for any repo and you can instantly Q&A against it. For example, yesterday I was curious about "how does torchao implement fp8 training?". I find that in *many* cases, library docs can be spotty and outdated and bad, but directly asking questions to the code via DeepWiki works very well. The code is the source of truth and LLMs are increasingly able to understand it. But then I realized that in many cases it's even a lot more powerful not being the direct (human) consumer of this information/functionality, but giving your agent access to DeepWiki via MCP. So e.g. yesterday I faced some annoyances with using torchao library for fp8 training and I had the suspicion that the whole thing really shouldn't be that complicated (wait shouldn't this be a Function like Linear except with a few extra casts and 3 calls to torch._scaled_mm?) so I tried: "Use DeepWiki MCP and Github CLI to look at how torchao implements fp8 training. Is it possible to 'rip out' the functionality? Implement nanochat/fp8.py that has identical API but is fully self-contained" Claude went off for 5 minutes and came back with 150 lines of clean code that worked out of the box, with tests proving equivalent results, which allowed me to delete torchao as repo dependency, and for some reason I still don't fully understand (I think it has to do with internals of torch compile) - this simple version runs 3% faster. The agent also found a lot of tiny implementation details that actually do matter, that I may have naively missed otherwise and that would have been very hard for maintainers to keep docs about. Tricks around numerics, dtypes, autocast, meta device, torch compile interactions so I learned a lot from the process too. So this is now the default fp8 training implementation for nanochat github.com/karpathy/nanoc… Anyway TLDR I find this combo of DeepWiki MCP + GitHub CLI is quite powerful to "rip out" any specific functionality from any github repo and target it for the very specific use case that you have in mind, and it actually kind of works now in some cases. Maybe you don't download, configure and take dependency on a giant monolithic library, maybe you point your agent at it and rip out the exact part you need. Maybe this informs how we write software more generally to actively encourage this workflow - e.g. building more "bacterial code", code that is less tangled, more self-contained, more dependency-free, more stateless, much easier to rip out from the repo (x.com/karpathy/statu…) There's obvious downsides and risks to this, but it is fundamentally a new option that was not possible or economical before (it would have cost too much time) but now with agents, it is. Software might become a lot more fluid and malleable. "Libraries are over, LLMs are the new compiler" :). And does your project really need its 100MB of dependencies?

English
0
0
1
163
Ketan Agrawal
Ketan Agrawal@_ketan0·
minimalist book reading
English
0
0
2
65
Ketan Agrawal
Ketan Agrawal@_ketan0·
Something very aesthetically pleasing about rendering TUIs in e-ink
English
1
0
5
609
Ketan Agrawal
Ketan Agrawal@_ketan0·
two thoughts that seem vaguely related: humor is mysterious, we can’t pin down what makes something ‘funny’ irony is self-referential at its core
Ketan Agrawal tweet media
English
0
0
2
81
Ketan Agrawal
Ketan Agrawal@_ketan0·
@nosilverv why do you think psychologists get it wrong? just pure base rates (they haven't seen positive examples in the wild?)
English
1
0
2
121
Ideas Guy
Ideas Guy@nosilverv·
Sometimes I have memories of how I used to be and it's like No, no—the psychologists are wrong the Buddhists are right: MASSIVE change in the Big Five factors of personality IS possible in adulthood, I'm so much less neurotic might as well be a different person. Legit "chill guy"
English
7
4
111
3.3K
Ketan Agrawal
Ketan Agrawal@_ketan0·
@samism Haven’t done improv yet but this thread really resonates with my (ongoing) experiences and learnings
English
0
0
1
45
Sam
Sam@samism·
Perhaps the most important thing I've learned from doing improv:
English
4
7
61
6.8K
Ketan Agrawal
Ketan Agrawal@_ketan0·
The state between wakefulness and sleep can show you a very interesting side of your mind
English
0
0
0
69
Ketan Agrawal
Ketan Agrawal@_ketan0·
@human_named_gui @TylerAlterman Yea- same principle as why random search works fairly well for hyperparameters Can stumble on something really good, then later figure out why
English
0
0
0
19
Tyler is finishing a book, slow to reply
An epistemic issue in self-experimentation: Let's say you're testing new sleep regimens, supplements, etc to feel better in some specific way. Do you (a) test 1 intervention at a time or (b) many at once? A is often a slower route to getting a good effect, but in the long run may help you understand causation faster B might get u an effect faster, but then you're not sure which intervention caused the effect
English
5
0
9
1.6K