TK Stohlman

1.3K posts

TK Stohlman banner
TK Stohlman

TK Stohlman

@tkstohlman

Founder https://t.co/7hwBwUpD83 Previously founded Autoscale AI (acq. 2021). Former U.S. Army officer

Encinitas, CA Katılım Ocak 2009
715 Takip Edilen796 Takipçiler
TK Stohlman retweetledi
Ken Schaefer
Ken Schaefer@kenschaefer1·
@tkstohlman Most people are still focused on better models. The real unlock is something else entirely: systems that govern how AI operates, enforce constraints during execution AND produce traceable, auditable, and repeatable outputs This is a new conversation entirely. #AI #AIGovernance
English
0
1
1
65
TK Stohlman
TK Stohlman@tkstohlman·
We’ve been building Project 777 for a while. Today marks another step forward: we’ve executed a $1M+ license with Atherion Bioresearch. More here: apnews.com/press-release/…
English
0
1
3
100
TK Stohlman
TK Stohlman@tkstohlman·
@mchadsells Thanks Chad - that means a lot. You’ve seen more of what’s behind this than most, so I really appreciate the support. 🙏
English
0
0
0
54
Chad Sells
Chad Sells@mchadsells·
@tkstohlman Really proud of what TK and the team have built. Congratulations. The idea of an execution layer sitting above frontier models to make outputs more usable, governed, and enterprise ready is exactly where the market needs to go. Excited to see this drive meaningful adoption.
English
1
0
2
76
TK Stohlman
TK Stohlman@tkstohlman·
@Drewandcrew22 Thanks Drew! I appreciate all the long talks at the football games when you listened to the vision, asked questions, and offered advice along the way. I appreciate you!
English
0
0
0
58
Drew Davis
Drew Davis@Drewandcrew22·
@tkstohlman This is incredible. TK and team continue to set the standard with a proven track record of innovation and execution in this space. The potential here—from disease treatment to broader applications—is massive. Excited to watch this unfold!
English
1
0
1
73
TK Stohlman
TK Stohlman@tkstohlman·
@kenschaefer1 Thanks Ken! You’ve seen this from the earliest days, and your guidance along the way has meant the world. Really appreciate you!
English
0
0
1
27
TK Stohlman
TK Stohlman@tkstohlman·
@JeremyBrandt Thanks Jeremy - really appreciate it! Been a long road, good to finally get it out there.
English
0
0
0
52
TK Stohlman
TK Stohlman@tkstohlman·
@chadsepulveda Thanks Chad - I really appreciate it. And I appreciate all the EO Forum advice over the years 🙏
English
0
0
0
51
TK Stohlman
TK Stohlman@tkstohlman·
@RussellAllen50 Thanks Russell - really appreciate it. Excited to finally get this out there 🙏
English
0
0
0
17
Russell Allen
Russell Allen@RussellAllen50·
Incredible! @tkstohlman and his team are changing the world. Proud to call you friend!
English
1
0
1
225
TK Stohlman
TK Stohlman@tkstohlman·
@chamath Agree. AI favorability will improve when people see it solving real problems - cures, infrastructure resilience, scientific discovery. We should focus less on benchmarks and more on the abundance AI will create.
English
0
0
0
27
TK Stohlman
TK Stohlman@tkstohlman·
@PeterDiamandis Or they already sold their last AI startup and have spent the past few years heads down building the next architectural breakthrough.
English
0
0
0
14
Peter H. Diamandis, MD
Peter H. Diamandis, MD@PeterDiamandis·
The person who will reshape the next 50 years of technology is likely writing a screenplay right now. The cell phone, the internet, the submarine, self driving cars, voice assistants, the helicopter and the space shuttle were all imagined by storytellers before they were built by engineers. Stories are blueprints.
English
139
153
1.1K
43.8K
TK Stohlman
TK Stohlman@tkstohlman·
@slow_developer Both are true. LLM capability will keep improving, but it’s also rapidly commoditizing. The real shift will come from new architectures that govern and orchestrate those models.
English
0
0
0
37
Haider.
Haider.@slow_developer·
i wonder what yann lecun's reaction is after gpt-5.4 he has been saying llms are a dead end for years, but llms have continued to improve steadily yann is highly respected in AI, and i wouldn't claim to know more than he does, but it still makes sense to keep pushing llms while others explore different architectures.
English
30
2
102
9.4K
TK Stohlman
TK Stohlman@tkstohlman·
@gdb Agree 100%. The real metric isn’t benchmarks - it’s whether AI produces solutions that materially improve human life.
English
0
0
0
49
Greg Brockman
Greg Brockman@gdb·
Benchmarks? Where we’re going, we don’t need benchmarks.
English
547
337
5.9K
629K
TK Stohlman
TK Stohlman@tkstohlman·
AI becomes the safest and most useful infrastructure humanity has ever built - helping design cures, powering trustworthy personal AI, and coordinating robots, vehicles, homes, factories, grids, etc. All of this using new architectures - not just bigger models. Abundance is achieved via new architecture breakthroughs.
English
0
0
1
30
Peter H. Diamandis, MD
Peter H. Diamandis, MD@PeterDiamandis·
Do you have a wildly optimistic vision of the future? Tell me. I want to hear it👇
English
455
36
494
47K
TK Stohlman
TK Stohlman@tkstohlman·
@karpathy good stuff. We’ve been working in this direction for ~5 years (bought autoresearch.ai in 2019). It’s evolved into a governed execution layer that turns model capability into deployable solutions across domains. Early commercial validation includes a $1M+ license.
English
1
0
0
115
Andrej Karpathy
Andrej Karpathy@karpathy·
Three days ago I left autoresearch tuning nanochat for ~2 days on depth=12 model. It found ~20 changes that improved the validation loss. I tested these changes yesterday and all of them were additive and transferred to larger (depth=24) models. Stacking up all of these changes, today I measured that the leaderboard's "Time to GPT-2" drops from 2.02 hours to 1.80 hours (~11% improvement), this will be the new leaderboard entry. So yes, these are real improvements and they make an actual difference. I am mildly surprised that my very first naive attempt already worked this well on top of what I thought was already a fairly manually well-tuned project. This is a first for me because I am very used to doing the iterative optimization of neural network training manually. You come up with ideas, you implement them, you check if they work (better validation loss), you come up with new ideas based on that, you read some papers for inspiration, etc etc. This is the bread and butter of what I do daily for 2 decades. Seeing the agent do this entire workflow end-to-end and all by itself as it worked through approx. 700 changes autonomously is wild. It really looked at the sequence of results of experiments and used that to plan the next ones. It's not novel, ground-breaking "research" (yet), but all the adjustments are "real", I didn't find them manually previously, and they stack up and actually improved nanochat. Among the bigger things e.g.: - It noticed an oversight that my parameterless QKnorm didn't have a scaler multiplier attached, so my attention was too diffuse. The agent found multipliers to sharpen it, pointing to future work. - It found that the Value Embeddings really like regularization and I wasn't applying any (oops). - It found that my banded attention was too conservative (i forgot to tune it). - It found that AdamW betas were all messed up. - It tuned the weight decay schedule. - It tuned the network initialization. This is on top of all the tuning I've already done over a good amount of time. The exact commit is here, from this "round 1" of autoresearch. I am going to kick off "round 2", and in parallel I am looking at how multiple agents can collaborate to unlock parallelism. github.com/karpathy/nanoc… All LLM frontier labs will do this. It's the final boss battle. It's a lot more complex at scale of course - you don't just have a single train. py file to tune. But doing it is "just engineering" and it's going to work. You spin up a swarm of agents, you have them collaborate to tune smaller models, you promote the most promising ideas to increasingly larger scales, and humans (optionally) contribute on the edges. And more generally, *any* metric you care about that is reasonably efficient to evaluate (or that has more efficient proxy metrics such as training a smaller network) can be autoresearched by an agent swarm. It's worth thinking about whether your problem falls into this bucket too.
Andrej Karpathy tweet media
English
970
2.1K
19.5K
3.6M
TK Stohlman retweetledi
Bambulu
Bambulu@BambuluQ·
Tattoo this to your brain: "The world is a malleable place. If you know what you want, and you go for it with maximum energy and passion, the world will often reconfigure itself around you much more quickly and easily than you would think.” – Marc Andreessen
English
53
2.8K
15K
228.9K
TK Stohlman retweetledi
U.S. Department of Energy
AI is the next Manhattan Project, and THE UNITED STATES WILL WIN. 🇺🇸
English
931
1.1K
10.2K
1.6M
Sawyer Merritt
Sawyer Merritt@SawyerMerritt·
Someone took their Cybertruck into Lake Grapevine in Texas. The Cybertruck can drive in water up to 31 inches deep using Wade Mode, which uses a technology Tesla calls the scuba pack, where it uses the truck’s built in air suspension to pressurize the battery. However, this video looks deeper than 31” lol
English
1K
2.2K
24.3K
7.1M