Victor Barsoum

128 posts

Victor Barsoum banner
Victor Barsoum

Victor Barsoum

@VBarsoum

Automation Therapist: I help AI agents overcome fear of sucking @ AI Profit Architects. Co-founder Streetwise Mortgages making underwriting less meh with AI

Toronto Katılım Mart 2013
171 Takip Edilen34 Takipçiler
himanshu
himanshu@himanshustwts·
and here is the full architecture of the LLM Knowledge Base system covering every stage from ingest to future explorations.
himanshu tweet media
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
94
557
5.7K
619.6K
Thariq
Thariq@trq212·
@jarredsumner not sure if there’ll be a second date but at least she knows Claude Code is a general purpose agent and that writing code is the foundation of agentic capability
English
8
1
225
8.6K
Jarred Sumner
Jarred Sumner@jarredsumner·
me at nopa on a first date her: how’d you pick this restaurant? me: I asked claude [waitress compliments her pants, clothes come up] her: where do you shop? me: lost a bunch of weight, had to rebuy everything. sent claude a pic of me and it picked [few min later, looking at menu] her: wait how do you track the calories me: believe it or not, also claude. I send it a pic of the menu I’m such a shill
English
111
12
1.3K
164.2K
Victor Barsoum
Victor Barsoum@VBarsoum·
2) Creating your own public endpoints from HF is a crazy fun feature and needs to be way more intuitive than it currently is. 3) Once a public endpoint gets created, I lose the ability to edit or delete it. So there's a lot of garbage on the platform — some of it is mine, sorry. That doesn't make sense. I can make a GitHub repo public but I still have full edit rights.
English
0
0
0
23
Victor Barsoum
Victor Barsoum@VBarsoum·
1. Discoverability — the Explore page is super unintuitive. With new models every day, it would be awesome to be able to see and try the latest models (image generation, image editing, video, OCR). The "I Want To" section on the Explore page comes close, but the models it surfaces are ancient in AI timelines. I hit OpenRouter and fal landing pages once or twice a day just to see what just dropped.
English
0
0
0
47
Victor Barsoum
Victor Barsoum@VBarsoum·
@dok2001 @Cloudflare Oh thanks for asking — all quality of life improvements. I think the platform is way underappreciated and doesn't get the public attention it deserves. 🧵
English
0
0
0
16
Andrej Karpathy
Andrej Karpathy@karpathy·
Wow, this tweet went very viral! I wanted share a possibly slightly improved version of the tweet in an "idea file". The idea of the idea file is that in this era of LLM agents, there is less of a point/need of sharing the specific code/app, you just share the idea, then the other person's agent customizes & builds it for your specific needs. So here's the idea in a gist format: gist.github.com/karpathy/442a6… You can give this to your agent and it can build you your own LLM wiki and guide you on how to use it etc. It's intentionally kept a little bit abstract/vague because there are so many directions to take this in. And ofc, people can adjust the idea or contribute their own in the Discussion which is cool.
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
761
2.1K
21.1K
4.4M
OpenRouter
OpenRouter@OpenRouter·
Qwen 3.6 Plus from @Alibaba_Qwen is officially the first model on OpenRouter to break 1 Trillion tokens processed in a single day! At ~1,400,000,000,000 tokens, it’s the strongest full day performance of any new model dropped this year. Congrats to the Qwen team!
OpenRouter tweet media
English
139
360
4.2K
392.2K
Victor Barsoum
Victor Barsoum@VBarsoum·
@garrytan Haven't logged in to salesforce in 2 weeks, all queries and changes are natural language from the ide.
English
0
0
0
26
Victor Barsoum
Victor Barsoum@VBarsoum·
@garrytan MCP done right = API skills. MCP as a concept is dead. API skills, much more token efficient, on demand, and can get very specific when needed
English
2
0
3
389
Garry Tan
Garry Tan@garrytan·
I am coming around to the fact that MCP, done right, can be magic.
English
183
46
1.1K
131.9K
Victor Barsoum
Victor Barsoum@VBarsoum·
@danielfoch hahaha... didn't have enough VRAM on my watch, trying to load it to my shoe next. Will let you know.
English
0
0
1
81
Daniel Foch
Daniel Foch@danielfoch·
If you’re not running Gemma 4 locally on a BlackBerry you are NGMI in 2026
English
3
1
8
2.2K
Andrej Karpathy
Andrej Karpathy@karpathy·
LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.
English
2.3K
5.5K
47.5K
14.6M
Victor Barsoum
Victor Barsoum@VBarsoum·
@cline I know it's April fool don't care... ship it .. ship it...
English
0
0
2
218
Cline
Cline@cline·
New setting just dropped: Lazy Teammate Mode. Turns Cline into the most useless engineer alive. 8 categories of excuses. Zero lines of code.
English
9
8
129
17.9K
Victor Barsoum
Victor Barsoum@VBarsoum·
@garrytan @adjohu Try antigravity... Claude code inside antigravity and then pass to antigravity agent for testing..
English
0
0
0
72
Garry Tan
Garry Tan@garrytan·
@adjohu There are a lot of limitations and google and nytimes don't work - lots of anti-bot websites don't work with Chrome for Testing
English
3
0
6
2.4K
Victor Barsoum
Victor Barsoum@VBarsoum·
Can we stop talking about war, politics, real estate, and even AI for a second? Humanity is heading back to the Moon for the first time in half a century. Godspeed, Artemis II. Remind us what it means to dream bigger than ourselves.
Victor Barsoum tweet media
English
0
0
0
63
Garry Tan
Garry Tan@garrytan·
Has anyone cracked this yet with Claude Code? Trivially easy customization to make your own browser with auto-loaded browser extension?
Garry Tan tweet media
English
78
9
191
222.8K
Victor Barsoum
Victor Barsoum@VBarsoum·
@codyschneiderxx Love how this is the exact playbook you shared a few days ago. Full transparency, full build in public. 100% dogfooding all the way. Respect 👊
English
0
0
0
25
Cody Schneider
Cody Schneider@codyschneiderxx·
have an AI agent run your google ads aiformarketingautomation .com
Cody Schneider tweet media
English
2
0
3
780
Cody Schneider
Cody Schneider@codyschneiderxx·
I just built an AI agent that's completely running my google ads it's connected to my google ads data, my google analytics data, my posthog data, and my stripe data it does the keyword research, builds the ads, analyzes the conversion data, analyzes the stripe data, allocates ad spend budget entirely autonomous you just plug your company in and growth starts happening if you want early access to this lmk below
English
27
2
52
7.3K