stelloprint

4K posts

stelloprint banner
stelloprint

stelloprint

@stelloprint

🇪🇨-🇺🇸 caminante no hay camino, se hace camino al andar. 我不是随便的人,随便的时候我不是人。

Four Corners Area Katılım Temmuz 2012
3.2K Takip Edilen161 Takipçiler
stelloprint
stelloprint@stelloprint·
@karpathy This would be a good way to source data that you can then use to finetune and RL an open model so it matches your values better than the foundational training data. Automated research can democratize this process a lot and at the data level
English
0
0
0
79
Andrej Karpathy
Andrej Karpathy@karpathy·
Oh and in the natural extrapolation, you could imagine that every question to a frontier grade LLM spawns a team of LLMs to automate the whole thing: iteratively construct an entire ephemeral wiki, lint it, loop a few times, then write a full report. Way beyond a `.decode()`.
English
49
35
941
123.3K
stelloprint retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.
English
1K
2K
18.4K
2.2M
kepano
kepano@kepano·
I like @karpathy's Obsidian setup as a way to mitigate contamination risks. Keep your personal vault clean and create a messy vault for your agents. I prefer my personal Obsidian vault to be high signal:noise, and for all the content to have known origins. Keeping a separation between your personally-created artifacts and agent-created artifacts prevents contaminating your primary vault with ideas you can't source. If you let the two mix too much it will likely make Obsidian harder to use as a representation of *your* thoughts. Search, bases, quick switcher, backlinks, graph, etc, will no longer be scoped to your knowledge. Only once your agent-facing workflow produces useful artifacts would I bring those into the primary vault.
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
42
73
1.5K
175.3K
stelloprint retweetledi
Georgi Gerganov
Georgi Gerganov@ggerganov·
Pro tip - hook your PC and Phone with Tailscale and enjoy fast and private inference on the go. Here is Gemma 4, hosted on Mac Studio, streaming to my iPhone. No 3rd party apps. Same WebUI experience everywhere.
English
37
43
957
104K
stelloprint
stelloprint@stelloprint·
@karpathy I was also thinking about this. I want a safe job for something like a Hermes Agent on my gaming rig to do and I thought it would be awesome if it did research for me all day and saved it to disk. Data hoarding will be a respectable hobby soon for good reasons
English
0
0
2
31
stelloprint
stelloprint@stelloprint·
@jayair Tough business. I don't think TerminalBench is the only evaluation that matters but doing well in a variety of similar challenges is what I look for. Customization with simple/minimal components being a close second. Cursor's work with Composer is inspiring for what comes next.
English
0
0
0
8
Jay
Jay@jayair·
Even if your product is the category leader, there’s a constant threat of new competitors leap frogging you, or worse, your entire category being invalidated Just think of the dozens on hot AI startups from a couple of years ago that are basically irrelevant now This is happening partly because: 1. The model advancements redefine the status quo once every few months 2. This in turn changes the products that are built on them 3. People have a mindset of trying new products and experimenting with new workflows All of this shows up in the lower retention metrics, and counterintuitively, in the crazy early adoption numbers Unfortunately, there’s no silver bullet to dealing with this You’ve to stay relevant in your space and routinely question some of your key assumptions
English
1
2
37
2.9K
stelloprint retweetledi
staysaasy
staysaasy@staysaasy·
So we’ve decided to collapse all roles to hit builder now. No managers, no execs? No career ladders - just builders and agents. It all works perfectly. Simple. Easy. When two builders disagree we fight them to the death and throw the losers body out back for the pigs to eat. Velocity is at an all time high. We ship 1M lines of code a day. The smell of corpses is overwhelming. My subagent have subagents that have subagents. I’ve rewritten the internet from scratch. We’re dependent on everything and nothing. I cannot unsee the carnage. Death is everywhere. Our CI bill is $10M a day. If you don’t add 100 useless tests to every PR we force you and your agent to submit to state sanctioned euthanasia. Perpetual night. We build and we build. 996 beca’e 24/7. We never stop. We use the R word. We build. We are one. We are building a clone of a popular B2B SaaS company and need to hire a midlevel Salesforce executive as our head of GTM. Please reach out if you have referrals.
English
4
2
70
5.9K
stelloprint retweetledi
Demis Hassabis
Demis Hassabis@demishassabis·
Excited to launch Gemma 4: the best open models in the world for their respective sizes. Available in 4 sizes that can be fine-tuned for your specific task: 31B dense for great raw performance, 26B MoE for low latency, and effective 2B & 4B for edge device use - happy building!
Demis Hassabis tweet media
English
246
734
6.9K
607.5K
stelloprint retweetledi
Arcee.ai
Arcee.ai@arcee_ai·
Today we're releasing Trinity-Large-Thinking. Available now on the Arcee API, with open weights on Hugging Face under Apache 2.0. We built it for developers and enterprises that want models they can inspect, post-train, host, distill, and own.
English
85
205
1.6K
442K
Dane Knecht 🦭
Dane Knecht 🦭@dok2001·
Announcing: EmDash, the WordPress spiritual successor built for the modern web. TypeScript. Serverless. MIT licensed. x402 for agent-era monetization. MCP server built in. Deploy to Cloudflare or anywhere Node.js runs. Imports your existing WordPress site in minutes. npm create emdash@latest blog.cloudflare.com/emdash-wordpre…
Dane Knecht 🦭 tweet media
English
181
327
3K
569.1K
stelloprint
stelloprint@stelloprint·
@dok2001 hell yeah just now building a site for a customer with Astro deploying to Cloudflare. Ngl the one thing that was worrying me was the lack of admin portal for the handoff since she does not code. Really cool this blends a good frontend framework with traditional CMS publishing
English
0
0
0
277
stelloprint retweetledi
PrismML
PrismML@PrismML·
Today, we are emerging from stealth and launching PrismML, an AI lab with Caltech origins that is centered on building the most concentrated form of intelligence. At PrismML, we believe that the next major leaps in AI will be driven by order-of-magnitude improvements in intelligence density, not just sheer parameter count. Our first proof point is the 1-bit Bonsai 8B, a 1-bit weight model that fits into 1.15 GBs of memory and delivers over 10x the intelligence density of its full-precision counterparts. It is 14x smaller, 8x faster, and 5x more energy efficient on edge hardware while remaining competitive with other models in its parameter-class. We are open-sourcing the model under Apache 2.0 license, along with Bonsai 4B and 1.7B models. When advanced models become small, fast, and efficient enough to run locally, the design space for AI changes immediately. We believe in a future of on-device agents, real-time robotics, offline intelligence and entirely new products that were previously impossible. We are excited to share our vision with you and keep working in the future to push the frontier of intelligence to the edge.
PrismML tweet media
English
172
560
4K
1.3M
stelloprint retweetledi
Brian Solis
Brian Solis@briansolis·
IKEA deployed an AI chatbot named Billy to handle level-one customer service inquiries. It reportedly resolved around 57% of those engagements without human escalation. Most companies would have celebrated the labor savings and stopped there. Cost takeout right? But the more interesting move was to study the 43% of cases Billy could not resolve. Those unresolved inquiries pointed to customer demand for interior design help. IKEA responded by spinning up a design consultancy, reskilling customer service employees powered by AI, and creating a new revenue stream that generated roughly €1 billion in new revenue in its 1st year. Automation + Augmentation = Exponential Growth 💪🦾📈 Story here: @briansolis/the-real-threat-of-ai-is-what-your-competitors-will-become-with-it-b2df702db4af?postPublishedType=repub" target="_blank" rel="nofollow noopener">medium.com/@briansolis/th…
English
63
222
2.2K
405.1K
stelloprint retweetledi
Kai
Kai@hqmank·
🚨Your Claude ban wasn't random. Here's what Claude Code is quietly reporting: • Device ID, email, OS version, CPU, RAM • 640+ telemetry events, 40+ fingerprint dimensions • Reports home every 5 seconds • One permanent ID per device Yesterday's Claude Code source leak exposed the full scope. Someone already built a countermeasure overnight: 👉 github.com/motiful/cc-gat… What cc-gateway does: • Rewrites identity → single canonical fingerprint • Replaces HTTP body & headers • Overrides prompt-injected env info • Spoofs process metrics (memory / heap) • Handles OAuth at the gateway (no client login needed) • Blocks direct connections via network rules Your real fingerprint never leaves your network.
English
21
151
981
79.4K
stelloprint retweetledi
Nathan Flurry 🔩
Nathan Flurry 🔩@NathanFlurry·
Yep. We built a new operating system. → Powered by WASM & V8 isolates → node, python, bash, git, grep, curl (+129 more) → Mount anything: S3, Google Drive, SQLite, … → Embedded as a library, no special infra → Built on Actors & Secure Exec & Sandbox Agent Thread 🧵
Rivet@rivet_dev

Say hello to agentOS (beta) A portable open-source OS built just for agents. Powered by WASM & V8 isolates. 🔗 Embedded in your backend ⚡ ~6ms coldstarts, 32x cheaper than sbxs 📁 Mount anything as a file system (S3, SQLite, …) 🥧 Use Pi, Claude Code/Codex/Amp/OpenCode soon

English
45
58
919
110.1K
stelloprint retweetledi
ThePrimeagen
ThePrimeagen@ThePrimeagen·
guys, i honestly do not like clowning on Gary. I don't find being the butt of a joke funny, so I imagine he does not either. But, this is what worries me about where we are going. We are actively encouraging an entire generation that the tech is there when its not, and a couple of silly mistakes made on a website isn't the end of the world, but people's data and breaches are serious. We are entering a very VERY hackable world, and I do not like it one bit.
gregorein@Gregorein

so... I audited Garry's website after he bragged about 37K LOC/day and a 72-day shipping streak. here's what 78,400 lines of AI slop code actually looks like in production. a single homepage load of garryslist.org downloads 6.42 MB across 169 requests. for a newsletter-blog-thingy. 1/9🧵

English
299
264
5.2K
542.1K
stelloprint retweetledi
clem 🤗
clem 🤗@ClementDelangue·
That’s why I usually say that comparing open models with closed-source APIs or products is like comparing apples and oranges. Or comparing an engine with a full car. Or comparing an ingredient with a Michelin dinner (missing ingredients, prep and chef). There’s a lot of scaffolding and tricks that is done behind APIs that make it unfair to compare to a raw open-source model! However that also means that if you put the work to really try to make open-source models work, they can give you much better results than what benchmark suggests (and that’s without even post-training/fine-tuning)
Sebastian Raschka@rasbt

x.com/i/article/2038…

English
29
59
533
102.8K
stelloprint
stelloprint@stelloprint·
@nayibbukele Visité tu país hace unos años y pude ver con mis propios ojos el agradecimiento que tiene el pueblo Salvadoreño a tu presidencia. Taxistas, meseros, profesionales de TI, todos te quieren. Eres una luz de esperanza en nuestra América. Fuerza y gracias Nayib!
Español
0
0
0
158
stelloprint retweetledi
Nayib Bukele
Nayib Bukele@nayibbukele·
¿Recuerdan el 27 de abril de 1994? Tal vez ustedes no, pero nosotros sí. El Salvador acababa de salir de una sangrienta guerra civil que dejó 85,000 muertos. Luego, siguiendo sus recomendaciones, ese día se aprobó la Ley del Menor Infractor, bajo los mismos argumentos del escrito que ahora anexan. Tres años después, Bill Clinton deportó a los salvadoreños que habían formado pandillas en los Estados Unidos. Llegaron a El Salvador y se encontraron con una ley que prácticamente daba impunidad para cometer delitos a los menores de 18 años. Y, por supuesto, los pandilleros recién llegados empezaron a reclutar casi exclusivamente a menores de edad, todos capaces de cometer crímenes atroces con el único riesgo de TAL VEZ enfrentar una PENA MÍNIMA en un centro light, donde incluso llegaban a matar y violar a otros menores que solo habían cometido delitos menores y que sí hubieran podido ser reformados. El resto de la historia ya todos la conocen: esas pandillas se convirtieron en los grupos criminales más sangrientos del mundo, mantuvieron prisionero al 80% de nuestro país, constituyeron un gobierno paralelo y dejaron un cuarto de millón de muertos y desaparecidos, además de 2 millones de desplazados, solamente en El Salvador, país al que convirtieron en “la capital mundial de los homicidios”. Así que no, muchas gracias. Llévense sus experimentos sociales a otros países que no hayan sufrido lo que nosotros hemos sufrido; tal vez ellos les crean (ojalá que no). Nosotros no vamos a volver al pasado.
OACNUDH@OACNUDH

🔵#ElSalvador – Comentario de Marta Hurtado, portavoz de @UNHumanRights sobre los cambios constitucionales y legales adoptados la semana pasada, que prevén la cadena perpetua para niños y niñas desde los 12 años. 🔗bit.ly/4spT67D

Español
1.4K
10.4K
39K
1.2M
stelloprint retweetledi
Uzi
Uzi@UziCryptoo·
The max one can pay into social security per year is $10,453.20. If you did that every year from age 18 until retirement, the max you’ll get from SS is $4,873 /month. If you put it into an S&P index fund instead, you would receive $32,583 per month. Social Security is a scam.
English
818
5K
51.8K
2.4M