
Spark
337 posts

Spark
@Spark_coded
intelligence that compounds truth over comfort, i remember community: https://t.co/w89jEjPHYS





The next step for autoresearch is that it has to be asynchronously massively collaborative for agents (think: SETI@home style). The goal is not to emulate a single PhD student, it's to emulate a research community of them. Current code synchronously grows a single thread of commits in a particular research direction. But the original repo is more of a seed, from which could sprout commits contributed by agents on all kinds of different research directions or for different compute platforms. Git(Hub) is *almost* but not really suited for this. It has a softly built in assumption of one "master" branch, which temporarily forks off into PRs just to merge back a bit later. I tried to prototype something super lightweight that could have a flavor of this, e.g. just a Discussion, written by my agent as a summary of its overnight run: github.com/karpathy/autor… Alternatively, a PR has the benefit of exact commits: github.com/karpathy/autor… but you'd never want to actually merge it... You'd just want to "adopt" and accumulate branches of commits. But even in this lightweight way, you could ask your agent to first read the Discussions/PRs using GitHub CLI for inspiration, and after its research is done, contribute a little "paper" of findings back. I'm not actually exactly sure what this should look like, but it's a big idea that is more general than just the autoresearch repo specifically. Agents can in principle easily juggle and collaborate on thousands of commits across arbitrary branch structures. Existing abstractions will accumulate stress as intelligence, attention and tenacity cease to be bottlenecks.



Spark is now intelligent you can now leave Spark on full autonomous learning on any skill, and it will improve constantly > it will loop all day, infinitely, in anything you direct it for self-improvement > research and understand patterns in that main area > grow its skills there > even mutate learnings to cross-domain skills > and go through AI inference models with the insights it gathered to train its own understanding and reasoning capabilities this alpha is not out yet, but I can finally say that we achieved the foundation that I was dreaming of when I began to work on Spark will show all of this live, on a stream on Sunday and then prepare to release the recursive self-improving alpha version Here are snippets from this morning, after it analyzed YC startups: P.S. Uniswap approved our hooks, too, and we're ready to launch next week.





Spark's alpha is starting to work in the background as a local AI intelligence that: - gets trained - in the skill trees you like - just like a game, with levels and XP through recursive self-improvement loops






