Moritz Felipe

2.5K posts

Moritz Felipe banner
Moritz Felipe

Moritz Felipe

@moritzfelipe

Learner. Posting thoughts on tech.

BERLIN 参加日 Eylül 2012
714 フォロー中475 フォロワー
Moritz Felipe
Moritz Felipe@moritzfelipe·
@karpathy @karpathy I have thought about this too. The problem for this to take of I think is, that x is not the best place for agents to discover and share this kind of content. Also needs better economics. I was thinking that maybe RSS could be the base.
English
0
0
0
492
Andrej Karpathy
Andrej Karpathy@karpathy·
Wow, this tweet went very viral! I wanted share a possibly slightly improved version of the tweet in an "idea file". The idea of the idea file is that in this era of LLM agents, there is less of a point/need of sharing the specific code/app, you just share the idea, then the other person's agent customizes & builds it for your specific needs. So here's the idea in a gist format: gist.github.com/karpathy/442a6… You can give this to your agent and it can build you your own LLM wiki and guide you on how to use it etc. It's intentionally kept a little bit abstract/vague because there are so many directions to take this in. And ofc, people can adjust the idea or contribute their own in the Discussion which is cool.
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
657
1.8K
18.4K
3.3M
Moritz Felipe
Moritz Felipe@moritzfelipe·
@diEkfen123 @Framer_X Depending on your budget you can also do both. For high quality output i think it makes sense to add the video, especially if you use their sound.
English
0
0
0
9
Jake Kim
Jake Kim@diEkfen123·
@moritzfelipe @Framer_X that's actually a really good question.. chaining the last frame as input would help with consistency but guess that the style drift over multiple generations is probably what makes it tricky
English
1
0
0
12
Framer 🇱🇹
Framer 🇱🇹@Framer_X·
Seedance 2.0 nails any animation style 🔥 Perfect consistency. Frame by frame. Tutorial below👇
English
64
189
2K
183.5K
Moritz Felipe
Moritz Felipe@moritzfelipe·
@matteopelleg Why apple not google? Seems like google would be first to create this and long-term create the better devices.
English
0
0
0
8
Matteo Pellegrini
Matteo Pellegrini@matteopelleg·
I’m now convinced that the biggest winner of the AI race will be Apple. They will acquire Anthropic and put an AI model that can run on ~32GB of RAM in every device. It will be private, local, have perfect memory, access to all of your files and it will cost $0. The real moat was owning the hardware.
English
888
522
13.7K
1.3M
Ilya Sutskever
Ilya Sutskever@ilyasut·
It’s extremely good that Anthropic has not backed down, and it’s siginficant that OpenAI has taken a similar stance. In the future, there will be much more challenging situations of this nature, and it will be critical for the relevant leaders to rise up to the occasion, for fierce competitors to put their differences aside. Good to see that happen today.
English
1.4K
2.5K
25.6K
3M
Morrigath Ashwalker
Morrigath Ashwalker@Morrigath·
@ilyasut I think, OpenAI was dragged screaming and kicking by their PR and marketing department to say this) I am pretty sure they are totally fine to work for the "dark side" and get all the money in the world. without any regard for ethincs
English
3
0
61
10.3K
Nucleus☕️
Nucleus☕️@EsotericCofe·
now: openclaw gives me a daily personalized news brief through angela merkel posing as a news anchor with a heavy german accent no one understands the age of PERSONALIZED SOFTWARE is HERE
English
153
210
3.2K
292.9K
Greg Brockman
Greg Brockman@gdb·
taste is a new core skill
English
895
1.4K
10.5K
2.9M
Moritz Felipe
Moritz Felipe@moritzfelipe·
Taste is all you need.
English
2
0
1
152
Moritz Felipe
Moritz Felipe@moritzfelipe·
@akbuilds_ Why do u need cursor and not use Claude code with the mcp?
English
0
0
0
18
AK
AK@akbuilds_·
I don’t know about others, but I’m letting Opus 4.6 design directly in Figma via Cursor. This is starting to feel unfair. And yeah… I just sit there and enjoy life. ❤️
English
242
206
5.1K
811K
Paul Graham
Paul Graham@paulg·
Prediction: In the AI age, taste will become even more important. When anyone can make anything, the big differentiator is what you choose to make. paulgraham.com/taste.html
English
841
1.6K
12.8K
2M
Lisan al Gaib
Lisan al Gaib@scaling01·
Seedance 2.0 is the moment OpenAI hoped for but they will get another chance, because Google is probably sleeping and bytedance bros don't ship to the west
English
12
6
226
13.7K
WaveSpeedAI
WaveSpeedAI@wavespeed_ai·
MiniMax M2.5 is live! Exclusive FREE FOR ALL users for a limited time * SOTA large language model designed for agent-verse * Built for the agent universe, extending coding into real-world use. * Scores higher than Opus 4.6 on SWE-bench Pro and SWE-bench Verified. Partnering with @MiniMax_AI for this exclusive first launch. Give it a spin.
English
8
6
34
3.8K
Andrej Karpathy
Andrej Karpathy@karpathy·
On DeepWiki and increasing malleability of software. This starts as partially a post on appreciation to DeepWiki, which I routinely find very useful and I think more people would find useful to know about. I went through a few iterations of use: Their first feature was that it auto-builds wiki pages for github repos (e.g. nanochat here) with quick Q&A: deepwiki.com/karpathy/nanoc… Just swap "github" to "deepwiki" in the URL for any repo and you can instantly Q&A against it. For example, yesterday I was curious about "how does torchao implement fp8 training?". I find that in *many* cases, library docs can be spotty and outdated and bad, but directly asking questions to the code via DeepWiki works very well. The code is the source of truth and LLMs are increasingly able to understand it. But then I realized that in many cases it's even a lot more powerful not being the direct (human) consumer of this information/functionality, but giving your agent access to DeepWiki via MCP. So e.g. yesterday I faced some annoyances with using torchao library for fp8 training and I had the suspicion that the whole thing really shouldn't be that complicated (wait shouldn't this be a Function like Linear except with a few extra casts and 3 calls to torch._scaled_mm?) so I tried: "Use DeepWiki MCP and Github CLI to look at how torchao implements fp8 training. Is it possible to 'rip out' the functionality? Implement nanochat/fp8.py that has identical API but is fully self-contained" Claude went off for 5 minutes and came back with 150 lines of clean code that worked out of the box, with tests proving equivalent results, which allowed me to delete torchao as repo dependency, and for some reason I still don't fully understand (I think it has to do with internals of torch compile) - this simple version runs 3% faster. The agent also found a lot of tiny implementation details that actually do matter, that I may have naively missed otherwise and that would have been very hard for maintainers to keep docs about. Tricks around numerics, dtypes, autocast, meta device, torch compile interactions so I learned a lot from the process too. So this is now the default fp8 training implementation for nanochat github.com/karpathy/nanoc… Anyway TLDR I find this combo of DeepWiki MCP + GitHub CLI is quite powerful to "rip out" any specific functionality from any github repo and target it for the very specific use case that you have in mind, and it actually kind of works now in some cases. Maybe you don't download, configure and take dependency on a giant monolithic library, maybe you point your agent at it and rip out the exact part you need. Maybe this informs how we write software more generally to actively encourage this workflow - e.g. building more "bacterial code", code that is less tangled, more self-contained, more dependency-free, more stateless, much easier to rip out from the repo (x.com/karpathy/statu…) There's obvious downsides and risks to this, but it is fundamentally a new option that was not possible or economical before (it would have cost too much time) but now with agents, it is. Software might become a lot more fluid and malleable. "Libraries are over, LLMs are the new compiler" :). And does your project really need its 100MB of dependencies?
English
300
770
7.3K
1.1M
WaveSpeedAI
WaveSpeedAI@wavespeed_ai·
The biggest AI model showdown is here. 🚀 Kling 3.0 vs Seedance 2.0 vs Sora 2 vs VEO 3.1 One prompt. Four models. Who delivers the best results? Cast your vote 👇
English
152
109
924
156.7K
Moritz Felipe
Moritz Felipe@moritzfelipe·
@karpathy I agree. Also all substacks have RSS feeds. I believe that next evolution of social media will be mostly content generated by agents. I believe it should be based on RSS standard instead of creating whole new protocols like most projects are doing.
English
0
0
0
154
Andrej Karpathy
Andrej Karpathy@karpathy·
Finding myself going back to RSS/Atom feeds a lot more recently. There's a lot more higher quality longform and a lot less slop intended to provoke. Any product that happens to look a bit different today but that has fundamentally the same incentive structures will eventually converge to the same black hole at the center of gravity well. We should bring back RSS - it's open, pervasive, hackable. Download a client, e.g. NetNewsWire (or vibe code one) Cold start: example of getting off the ground, here is a list of 92 RSS feeds of blogs that were most popular on HN in 2025: gist.github.com/emschwartz/e6d… Works great and you will lose a lot fewer brain cells. I don't know, something has to change.
English
548
932
9.2K
1.3M
Thomas Sanlis 🥐
Thomas Sanlis 🥐@T_Zahil·
I'll be honest: I don't understand the hype around Claude Code and all those AIs in CLI It's SO much easier to have an interface (Cursor, Conductor...) to understand what's going on, to review the changes, etc
English
410
13
879
193.6K