Joe Botsch
5.4K posts

Joe Botsch
@jbotsch
VC @Deviationcap, formerly @rtpvc, @OpenView & Head of Research at @MassChallenge. talking startups or complaining about Boston sports

New blackboard lecture w @ericjang11 He walks through how to build AlphaGo from scratch, but with modern AI tools. Sometimes you understand the future better by stepping backward. AlphaGo is still the cleanest worked example of the primitives of intelligence: search, learning from experience, and self-play. You have to go back to 2017 to get insight into how the more general AIs of the future might learn. Once he explained how AlphaGo works, it gave us the context to have a discussion about how RL works in LLMs and how it could work better – naive policy gradient RL has to figure out which of the 100k+ tokens in your trajectory actually got you the right answer, while AlphaGo’s MCTS suggests a strictly better action every single move, giving you a training target that sidesteps the credit assignment problem. The way humans learn is surely closer to the second. Eric also kickstarted an Autoresearch loop on his project. And it was very interesting to discuss which parts of AI research LLMs can already automate pretty well (implementing and running experiments, optimizing hyperparameters) and which they still struggle with (choosing the right question to investigate next, escaping research dead ends). Informative to all the recent discussion about when we should expect an intelligence explosion, and what it would look like from the inside. Timestamps: 0:00:00 – Basics of Go 0:08:06 – Monte Carlo Tree Search 0:31:53 – What the neural network does 1:00:22 – Self-play 1:25:27 – Alternative RL approaches 1:45:36 – Why doesn’t MCTS work for LLMs 2:00:58 – Off-policy training 2:11:51 – RL is even more information inefficient than you thought 2:22:05 – Automated AI researchers


Boston University's new 19 story building runs with: No gas No fossil fuels No emissions One of the greenest buildings in New England.




would people be interested in a blog series on "simulation infrastructure" where I look at popular robotics sim tools (e.g. mjlab, isaac...) and explain the pros and cons of their design decisions. The series would cover what I believe to be key pillars of good, usable, sim code




This tweet single-handedly made Anthropic and OpenAI scramble and issue full statements on secondary stock deals. $500B of paper value just got wiped out overnight.

1) I’m excited to announce @EqualVentures’s 5th Annual Emerging Manager Circle Summit 8 years ago, this was just a few of us swapping stories, trying to help each other. I couldn’t imagine what EMC would become Now, we need events like EMC more than ever newsletter.equal.vc/p/announcing-t…





Whenever something looks incredibly easy, and it becomes the conventional wisdom that everyone's going to make money, that’s usually when it blows up in your face. In 2019, The WSJ was writing that the best PE firms “never lost money.” The deals that those firms did right after are now blowing up. Overconfidence is a fatal error in investing. Something we probably should remember. wsj.com/articles/vista…










