Xpicker2
2.3K posts

Xpicker2
@xpickr
Automatically pick a random winner from any X post 🤖 the X picker 🔥 Follow us for a bonus point. AI acceleration incoming.





I’m proud to be joining SpaceX and xAI with @milichab It has become clear that software is changing fundamentally. More and more, people can shape the tools they use directly, and the ceiling of what can be built keeps rising. What makes xAI special is the scale of its ambition: to build from first principles all the way out to the stars. I’m especially grateful to work on products that expand human agency and freedom. That mission is deeply personal to me. My family came to the United States fleeing communism, and the belief that freedom should be part of the next generation of the internet has driven me every day since Andrew and I started Skiff. Now, we get to work on intelligence, understanding, and freedom on a universal scale.





I packaged up the "autoresearch" project into a new self-contained minimal repo if people would like to play over the weekend. It's basically nanochat LLM training core stripped down to a single-GPU, one file version of ~630 lines of code, then: - the human iterates on the prompt (.md) - the AI agent iterates on the training code (.py) The goal is to engineer your agents to make the fastest research progress indefinitely and without any of your own involvement. In the image, every dot is a complete LLM training run that lasts exactly 5 minutes. The agent works in an autonomous loop on a git feature branch and accumulates git commits to the training script as it finds better settings (of lower validation loss by the end) of the neural network architecture, the optimizer, all the hyperparameters, etc. You can imagine comparing the research progress of different prompts, different agents, etc. github.com/karpathy/autor… Part code, part sci-fi, and a pinch of psychosis :)



The Grok algo lately has been fire 🔥

I packaged up the "autoresearch" project into a new self-contained minimal repo if people would like to play over the weekend. It's basically nanochat LLM training core stripped down to a single-GPU, one file version of ~630 lines of code, then: - the human iterates on the prompt (.md) - the AI agent iterates on the training code (.py) The goal is to engineer your agents to make the fastest research progress indefinitely and without any of your own involvement. In the image, every dot is a complete LLM training run that lasts exactly 5 minutes. The agent works in an autonomous loop on a git feature branch and accumulates git commits to the training script as it finds better settings (of lower validation loss by the end) of the neural network architecture, the optimizer, all the hyperparameters, etc. You can imagine comparing the research progress of different prompts, different agents, etc. github.com/karpathy/autor… Part code, part sci-fi, and a pinch of psychosis :)

My new Sunday morning routine: 1. Get coffee 2. Check GPT-5.4 projects on the Codex App, continue & start new ones 4. Launch ChatGPT 5.4 Pro for fresh brainstorming sessions 5. Think/learn how to use the 90% of AI capabilities I have yet to explore 6. Drink more coffee

What if a codebase was actually stored in Postgres and agents directly modified files by reading/writing to the DB? Code velocity has increased 3-5x. This will undoubtedly continue. PR review has already become a bottleneck for high output teams. Codebase checked-out on filesystem seems like a terrible primitive when you have 10-100-1000 agents writing code. Code is now high velocity data and should be modeled at such. Bare minimum, we need write-level atomicity and better coordination across agents, better synchronization primitives for subscribing to codebase state changes and real-time time file-level code lint/fmt/review. The current ~20 year old paradigm of git checkout/branch/push/pr/review/rebase ended Jan 2026. We need an entirely new foundational system for writing code if we’re really going to keep pace with scale laws.





