nick chen

303 posts

nick chen banner
nick chen

nick chen

@nykoras

leading frontier seed investments @goldenventures. prev: head of ventures @IEX, @Georgian_io, founder @ syzygy records. affinity for esoterica 🪡

to Katılım Nisan 2016
1K Takip Edilen420 Takipçiler
nick chen retweetledi
Zhengyao Jiang
Zhengyao Jiang@zhengyaojiang·
Is autoresearch really better than classic hyperparameter tuning? We did experiments comparing Optuna & autoresearch. Autoresearch converges faster, is more cost-efficient, and even generalizes better: 🧵(1/6)
Zhengyao Jiang tweet media
English
24
113
1.3K
132.2K
nick chen retweetledi
Zhengyao Jiang
Zhengyao Jiang@zhengyaojiang·
AutoResearch is a general purpose code optimizer, and math formulas can also be expressed as code. The emerging use case of formula discovery is really interesting, give it empirical data and let the agent search for math expressions that fit. Examples 🧵(1/5):
English
6
9
112
14.6K
nick chen retweetledi
Zhengyao Jiang
Zhengyao Jiang@zhengyaojiang·
Autoresearch has been out for 2 weeks. The community is trying to apply it to everything with a measurable metric, here are some successful attempts: 🧵 (1/6)
English
35
137
1.7K
294.2K
nick chen retweetledi
Zhengyao Jiang
Zhengyao Jiang@zhengyaojiang·
Your autoresearch needs its own Weights & Biases. We’ve turned Weco into an observability tool that lets you monitor, analyze, and share autoresearch runs. Here's what it can do: 🧵(1/4)
English
20
60
660
38.9K
nick chen retweetledi
Zhengyao Jiang
Zhengyao Jiang@zhengyaojiang·
We're excited to announce that @BingchenZhao, who built the predecessor of AutoResearch, has joined @WecoAI full-time! Bingchen is the first author of LLMSpeedrunner at Meta FAIR, which ran the automated research loop on @karpathy's NanoGPT, which later evolved into NanoChat and the speedrun community where AutoResearch operates today. Weco has been committed to ML research automation for 2.5 years, starting with AIDE. We're super pumped by how large an impact AIDE has had, topping @OpenAI's MLE-Bench and @METR_Evals' RE-Bench, and becoming a foundation for AI Scientist v2, AIRA-Dojo, and LLMSpeedrunner itself. And AutoResearch, with AIDE's simple greedy discard/keep loop reaching a mass audience, is really building consensus that the empirical research loop can and should be automated. We're excited to keep pushing this frontier, not just as a concept but seriously bringing it to the real world, and materially accelerating the knowledge generation of humanity.
Andrej Karpathy@karpathy

Love this project: nanoGPT -> recursive self-improvement benchmark. Good old nanoGPT keeps on giving and surprising :) - First I wrote it as a small little repo to teach people the basics of training GPTs. - Then it became a target and baseline for my port to direct C/CUDA re-implementation in llm.c. - Then that was modded (by @kellerjordan0 et al.) into a (small-scale) LLM research harness. People iteratively optimized the training so that e.g. reproducing GPT-2 (124M) performance takes not 45 min (original) but now only 3 min! - Now the idea is to use this process of optimizing the code as a benchmark for LLM coding agents. If humans can speed up LLM training from 45 to 3 minutes, how well do LLM Agents do, under different kinds of settings (e.g. with or without hints etc.)? (spoiler: in this paper, as a baseline and right now not that well, even with strong hints). The idea of recursive self-improvement has of course been around for a long time. My usual rant on it is that it's not going to be this thing that didn't exist and then suddenly exists. Recursive self-improvement has already begun a long time ago and is under-way today in a smooth, incremental way. First, even basic software tools (e.g. coding IDEs) fall into the category because they speed up programmers in building the N+1 version. Any of our existing software infrastructure that speeds up development (google search, git, ...) qualifies. And then if you insist on AI as a special and distinct, most programmers now already routinely use LLM code completion or code diffs in their own programming workflows, collaborating in increasingly larger chunks of functionality and experimentation. This amount of collaboration will continue to grow. It's worth also pointing out that nanoGPT is a super simple, tiny educational codebase (~750 lines of code) and for only the pretraining stage of building LLMs. Production-grade code bases are *significantly* (100-1000X?) bigger and more complex. But for the current level of AI capability, it is imo an excellent, interesting, tractable benchmark that I look forward to following.

English
1
8
83
11.2K
nick chen retweetledi
Zhengyao Jiang
Zhengyao Jiang@zhengyaojiang·
In case you want to run AutoResearch this weekend: It costs ~$300 for 85 experiments using Claude Code (opus). A quick guide to autoresearch ~60 experiments for free: 1. Use the mac/local GPU fork:github.com/miolini/autore… 2. Use weco to get some free credits: `pipx install weco` → `weco setup claude-code` Or simply give this doc to your Claude Code agent: docs.weco.ai/quickstart - You’ll get $20 in free credits 3. Tell your coding agent to run weco optimization for val_bpb on train.py. 4. Tell your coding agent to use gemini-3-flash-preview, you should get about 60 free experiments. - For better performance, use gemini-3.1-pro-preview (~15 free experiments). 5. You can watch the progress on this nice dashboard: dashboard.weco.ai/share/v5X8WV5H…
Zhengyao Jiang tweet media
English
1
32
216
21.1K
nick chen retweetledi
Weco AI
Weco AI@WecoAI·
Hard work scales linearly. Automation scales exponentially. Over 17 days, our autonomous ML agent trained 120 models and beat 90% of teams in a live $100k ML competition, with zero human intervention. Weco, now in public beta:
English
2
22
47
23.1K
nick chen retweetledi
Bread
Bread@ai_bread·
Announcing Bread Technologies. We’re building machines that learn like humans. We raised a $5 million seed round led by Menlo Ventures and have been building in stealth for 10 months. Today, we rise 🍞
English
128
129
931
761.5K
nick chen retweetledi
Tracel AI
Tracel AI@Tracel_AI·
We're excited to announce that we have raised $3M USD, led by Golden Ventures and OSS Capital, to advance our mission of democratizing and accelerating AI development. Read the full vision: burn.dev/blog/funding-a…
English
1
9
27
4.7K
nick chen retweetledi
Zhengyao Jiang
Zhengyao Jiang@zhengyaojiang·
Thrilled to announce Weco has raised an $8M seed led by @GoldenVentures to build self-evolving software! Our technology has already been used by frontier labs like OpenAI, Meta, Google and Sakana AI. We’re making every codebase a living experiment that learns to beat itself:
Zhengyao Jiang tweet media
English
12
13
110
39K
nick chen retweetledi
Zhengyao Jiang
Zhengyao Jiang@zhengyaojiang·
I always suspected people overestimate the productivity gains from AI, especially in large projects with extensive context (not just code, but deep project-specific experience). But I'm still surprised to see copilot tools actually slowing down development in practice🤯
METR@METR_Evals

We ran a randomized controlled trial to see how much AI coding tools speed up experienced open-source developers. The results surprised us: Developers thought they were 20% faster with AI tools, but they were actually 19% slower when they had access to AI than when they didn't.

English
2
4
17
2.1K
nick chen retweetledi
Commit-Boost 📻 🕶️ 🦇🔊
1/ We are excited to share that Commit-Boost is headed towards audit and production! As an open-source, public good, this could not have been possible without many teams across the Ethereum Community. Thank you for all your time and effort! See below 👇
Commit-Boost 📻 🕶️ 🦇🔊 tweet media
English
28
75
263
188.9K
nick chen retweetledi
ₕₐₘₚₜₒₙ
ₕₐₘₚₜₒₙ@hamptonism·
They literally made a Convolutional Neural Network in Minecraft before we got GPT-5.
English
106
1.8K
23.8K
2.9M
nick chen retweetledi
Anthropic
Anthropic@AnthropicAI·
New Anthropic research paper: Scaling Monosemanticity. The first ever detailed look inside a leading large language model. Read the blog post here: anthropic.com/research/mappi…
Anthropic tweet media
English
65
538
2.3K
753.9K
nick chen retweetledi
Matt Parkhurst
Matt Parkhurst@mprkhrst·
> independently discover a Zeno's paradox at age 3 > MIT at 17, grad level math in 1st year > graduate in 3 years > drive motor scooters from Boston to Bogotá with the boys > start a company in Colombia > start code breaking with the IDA for money > solve minimal varieties in riemannian manifolds > speak out against Vietnam War, get fired from IDA > take over math dept. at Stonybrook, make it a top-ranked program globally > develop Churn-Simons theory, accidentally contribute more to physics than most physicists > get bored with math, start modeling financial markets > return 60% for 4 decades straight > establish one of the most effective philanthropic organizations of all time > chain smoke cigarettes the entire time RIP Jim 🫡
Matt Parkhurst tweet media
English
212
3K
24.5K
2.5M
nick chen retweetledi
Tarun Sachdeva
Tarun Sachdeva@tarunsachdeva·
New Demos 2 last night was a special event for Toronto. Don't think I've ever seen anything quite like it. Thank you to all the builders and creatives who showcased their work....and ty @internetvin @tommytrinh, for everything
English
1
6
41
2.5K
nick chen retweetledi
Sequence
Sequence@0xsequence·
🚨NEW PARTNER: Google Cloud🚨 We’re partnering to offer our all-in-one development platform to Google Cloud customers! Now, it’s easier than ever for game makers to leverage web3 technology. 🧵 Keep reading to find out more 👇🏼 venturebeat.com/games/sequence…
English
16
71
888
1.9M
nick chen retweetledi
Matt Golden
Matt Golden@MattGoldenVP·
1/5 We’re excited to share that @GoldenVentures has closed its fifth fund, Golden Ventures V, at just over $100 Million USD. Read more here: bit.ly/3T82fTF
Matt Golden tweet media
English
50
12
259
50.7K
nick chen
nick chen@nykoras·
I'm excited to share @GoldenVentures' newest investment thesis focused on #genai, #web3 and frontier technologies at large, “Trust is all you need.”
nick chen tweet media
English
2
2
9
2.2K