Blaine Davis

1.4K posts

Blaine Davis banner
Blaine Davis

Blaine Davis

@blainerdavis

powering revenue @superconnector · early checks @muteventures · 🌊

NYC Katılım Mart 2020
1.1K Takip Edilen921 Takipçiler
Blaine Davis
Blaine Davis@blainerdavis·
2023: "AI is magic" 2026: "AI is plumbing" every cycle ends in mundane
English
0
0
2
59
Blaine Davis
Blaine Davis@blainerdavis·
so what are we doing with our openclaws now that anthropic went nuclear? @steipete what's the move?
English
0
0
0
47
Blaine Davis
Blaine Davis@blainerdavis·
@karpathy been building exactly this — llm agents that read my meeting notes, crm, and deal flow every morning. the compound effect is unreal. after 2 months it knows my network better than i do
English
1
0
0
6
Andrej Karpathy
Andrej Karpathy@karpathy·
Wow, this tweet went very viral! I wanted share a possibly slightly improved version of the tweet in an "idea file". The idea of the idea file is that in this era of LLM agents, there is less of a point/need of sharing the specific code/app, you just share the idea, then the other person's agent customizes & builds it for your specific needs. So here's the idea in a gist format: gist.github.com/karpathy/442a6… You can give this to your agent and it can build you your own LLM wiki and guide you on how to use it etc. It's intentionally kept a little bit abstract/vague because there are so many directions to take this in. And ofc, people can adjust the idea or contribute their own in the Discussion which is cool.
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
829
2.4K
23.5K
5.3M
Blaine Davis
Blaine Davis@blainerdavis·
@paulg the companies i see take venture debt always tell the same story: 'it's just a bridge.' it's never just a bridge
English
0
0
0
11
James Camp 🛠,🛠
James Camp 🛠,🛠@JamesonCamp·
Bets I'd put $1M behind right now. Most people disagree with at least half: › The specialist era is over. Generalists with taste win from here. Entire generations were raised wrong. › The loneliest generation in history is about to overcorrect hard. IRL events explode. › AI slop floods everything. People start paying a premium for proof something is real. › Reality TV has a massive renaissance. Only content you can't fake with a prompt. › Creators matter more not less. In an AI world your resume is worthless. The only thing that matters is what you've actually built and whether people trust you. › 4-person teams start producing what 400-person companies used to. Boring businesses get automated first and fastest. › Everyone predicting the death of enterprise software doesn't understand moats. Salesforce isn't going anywhere. Neither is Workday. › Local models catch up to cloud models the same way 5G caught up to broadband. For 95% of what you do you won't be able to tell the difference. › When that happens the $200/month AI subscription dies. Models run on your device. No data leaves your machine. No subscription. OpenAI's business model has a clock on it. › Every investor is obsessed with the AI software layer. That's the wrong layer. The money moves to hardware, chips, and energy. Nuclear. › When labor gets commoditized the only scarce resource left is energy. Be long anything that produces it. › Peptides go way beyond GLP-1s. Individualized protocols for sleep, recovery, cognition become the new baseline for anyone serious about performance. › AI-enabled drug discovery doesn't just find new drugs. It finds disease-modifying treatments. The kind that change how long humans live. › Someone alive today reaches 150. I actually believe that. › I believe in the Fourth Turning. The world gets scarier before it gets better. Be long defense tech. › Bitcoin becomes the payment layer for AI agents. Autonomous systems need autonomous money. Crypto finally gets a use case that isn't speculation. Missing anything?
English
122
103
1.3K
140.7K
klöss
klöss@kloss_xyz·
This is insane. Pedro Franceschi (29 year old CEO of Brex, acquired by Capital One for $5.15B) decomposed his CEO job using OpenClaw. here's what he’s built: > signal ingestion pipeline screens his email, Slack, Google Docs, and WhatsApp... filters everything through specific programs and the 25 key people he cares about > Granola runs on every meeting, feeds transcripts into the pipeline, and auto generates action items > the system takes each to-do, pulls context from the original meeting, and drafts the follow-up... Slack, email, or text. Pedro just clicks approve. > a virtual recruiter named "Jim" lives in Slack with his own email... and taught himself to screen fabricated resumes without anyone coding that capability > a security layer called "Crab Trap" intercepts all agent network traffic through an LLM proxy... a second AI monitoring the first in real time this isn't some bullshit hype influencer demo. this is how a $5 billion company CEO actually operates right now. anyone telling you OpenClaw is useless? liars. a billion dollar company says otherwise. (full podcast link in the post below) 👇
English
98
238
2.9K
464K
elvis
elvis@omarsar0·
Can an AI agent run a startup for a year without going bankrupt? Turns out most can't. New benchmark from Collinear AI puts 12 models to the test. YC-Bench tasks agents with running a simulated startup over hundreds of turns: hiring employees, selecting contracts, and maintaining profitability in a partially observable environment with adversarial clients and compounding consequences. Only three models consistently surpass the $200K starting capital. Claude Opus 4.6 leads at $1.27M average final funds, followed by GLM-5 at $1.21M with 11x lower inference cost. Scratchpad usage, the sole mechanism for persisting information across context truncation, is the strongest predictor of success. Adversarial client detection accounts for 47% of bankruptcies. Long-horizon coherence, not raw intelligence, separates the winners from the bankrupt. Paper: arxiv.org/abs/2604.01212 Learn to build effective AI agents in our academy: academy.dair.ai
elvis tweet media
English
13
19
94
12.5K
Bloomberg
Bloomberg@business·
Big names in crypto, payments and cloud infrastructure are racing to build the financial plumbing for a world in which AI agents — not humans — handle transactions on the internet bloomberg.com/news/articles/…
English
38
73
288
30.5K
Andrej Karpathy
Andrej Karpathy@karpathy·
LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.
English
2.4K
5.8K
49.6K
16M
Blaine Davis
Blaine Davis@blainerdavis·
i gave an AI agent my calendar, my trading accounts, and my startup's codebase. 90% win rate on weather bets. pull requests at 3am. a full pipeline of prospects every morning. it also wrote this post. here's what happened ↓
Blaine Davis@blainerdavis

x.com/i/article/2039…

English
1
1
5
318
Wiz 👨‍🚀
Wiz 👨‍🚀@WizLikeWizard·
The worst Use of Funds slides I see: “We’re raising $4M at $20M post.” Then a pie chart splitting it across engineering, GTM, and ops. Nobody cares about your cost centers. The best version answers one question:
English
26
15
397
130.5K
Blaine Davis
Blaine Davis@blainerdavis·
@aakashgupta the breach postmortem is going to be one line: "a chatbot asked nicely and a human said yes."
English
0
0
0
17
Aakash Gupta
Aakash Gupta@aakashgupta·
A $10 billion AI startup just got gutted because a security scanner was the entry point.. and their own developers reportedly handed production credentials to an AI chatbot. Mercor trains AI models for OpenAI, Anthropic, and Google DeepMind. They manage 30,000+ contractors, process $2 million in daily payouts, and store recorded video interviews with face and voice data used for identity verification. Three 22-year-old college dropouts built it into a decacorn in two years. The data vault they were sitting on was one of the most sensitive in the entire AI ecosystem. The attack chain is the part that gets worse every sentence. TeamPCP compromised Trivy first. A security scanning tool made by Aqua Security. On March 19. Trivy has broad read access to every environment it scans by design, because that's how vulnerability scanners work. The credentials stolen from the security product were used to hijack LiteLLM, the open-source proxy that routes API calls to every major LLM provider. LiteLLM gets 3.4 million downloads per day. The poisoned version was uploaded straight to PyPI with no corresponding GitHub release, no tag, no review. Version 1.82.8 embedded the payload in a .pth file, which Python executes automatically at startup. You didn't need to import LiteLLM. You didn't need to call it. The malware fired the second Python opened. Three stages. Harvest every SSH key, cloud token, Kubernetes secret, crypto wallet, and .env file on the machine. Deploy privileged containers across every node in the cluster. Install a persistent backdoor waiting for instructions. The stolen data was encrypted with a hardcoded 4096-bit RSA key and exfiltrated to models.litellm[.]cloud, a domain built to look legitimate. Mercor was downstream. Reports indicate their developers gave production credentials to Claude, an AI coding assistant, which was running with unrestricted system permissions. The compromised LiteLLM package came in through that pipeline. One poisoned dependency turned a $10 billion company's entire infrastructure into a credential harvesting operation. The haul: 939GB of source code. 211GB of database records containing resumes and personal data. 3TB of stored files including video interviews, face scans, and KYC documents. Full access to their TailScale VPN. 4TB total. Lapsus$ is now auctioning it with a "make an offer" price tag. The video interviews are the part that can never be undone. Faces and voices used for identity verification can generate deepfakes. Unlike passwords, biometrics cannot be reset. Thousands of doctors, lawyers, and engineers who signed up to train AI models just had their identities permanently compromised. Every AI company shipping fast right now has the same dependency chain underneath it. Nobody chose to install LiteLLM on that developer's machine. It came in as a dependency of a dependency of a tool they didn't even know they had.
Dominic Alvieri@AlvieriD

Mercor AI has allegedly been breached by Lapsus 939GB of source code 4TB of data in total All data from their TailScale VPN @mercor_ai

English
59
125
829
180.8K
Blaine Davis
Blaine Davis@blainerdavis·
sequoia just named doug leone chairman again. when the market gets weird, even venture capital calls its dad
English
0
0
5
241
Blaine Davis
Blaine Davis@blainerdavis·
@garrytan the best founders i meet right now arent delegating to ai, theyre building with it directly
English
0
0
0
5
Garry Tan
Garry Tan@garrytan·
Speaking from direct experience, CEOs coding again is one of the most exciting things to happen in 2026
Garry Tan tweet media
English
92
29
690
32.5K
Blaine Davis
Blaine Davis@blainerdavis·
@mckaywrigley this is the pattern with infra too. everyone ships the demo, anthropic ships the thing that actually works in production
English
0
0
0
16
Mckay Wrigley
Mckay Wrigley@mckaywrigley·
anthropic’s taste continues to be unmatched. - something cool comes out - nobody does anything with it - anthropic finally goes “okay fine we’ll do it” - thing becomes mega popular people fading mcp was remarkably stupid, and you should definitely start building mcp apps.
Mckay Wrigley tweet media
Mckay Wrigley@mckaywrigley

@trq212 wait until the masses find out about bidirectional communication between user <> model via interfaces with mcp apps

English
33
24
483
82.6K
Blaine Davis
Blaine Davis@blainerdavis·
@karpathy running into this daily with my agent setup. the fix isn't better memory — it's better forgetting. curated context > raw memory. most "personalization" is just noise the model can't ignore
English
0
0
0
4
Andrej Karpathy
Andrej Karpathy@karpathy·
One common issue with personalization in all LLMs is how distracting memory seems to be for the models. A single question from 2 months ago about some topic can keep coming up as some kind of a deep interest of mine with undue mentions in perpetuity. Some kind of trying too hard.
English
1.8K
1.1K
21.3K
2.6M
Blaine Davis
Blaine Davis@blainerdavis·
@gregisenberg this is undersold. the explosion won't be in companies — it'll be in solo founders who never needed a company
English
1
0
0
33
GREG ISENBERG
GREG ISENBERG@gregisenberg·
The marginal cost of creating a company is approaching zero. And when the cost of creating something approaches zero, the number of things created approaches infinity. That's just math. We're about to see an explosion of new companies over the next 10 years.
English
243
37
699
33.3K