Adam Goulburn

6.9K posts

Adam Goulburn banner
Adam Goulburn

Adam Goulburn

@AGoulburn

Founder @_dimensioncap

NYC เข้าร่วม Eylül 2012
516 กำลังติดตาม5.5K ผู้ติดตาม
ทวีตที่ปักหมุด
Adam Goulburn
Adam Goulburn@AGoulburn·
1/ Today we're proud to launch Dimension! The last 12 months have been a wild ride - the best & most exciting time of my career. Starting something from scratch, with two of my closest mates @nanli @zavaindar is a dream come true dimensioncap.com @_DimensionCap
English
7
6
118
21.3K
Adam Goulburn รีทวีตแล้ว
Simon Barnett
Simon Barnett@SimonDBarnett·
Galileo's similarity theory powered centuries of industrialization, but living cells broke it. In my newest essay, I argue that encoding this theory as an inductive bias is the only way to transform #bioprocess scaling from alchemy back into engineering.
Simon Barnett@SimonDBarnett

x.com/i/article/2038…

English
0
6
13
3K
Adam Goulburn รีทวีตแล้ว
Frank Gao
Frank Gao@ChemVagabond·
We gave AI agents varying freedom to optimize models for protein thermostability prediction. They beat a 15x larger reference model, matched Bayesian optimization. And we learned a hybrid approach may work even better. Blog: research.dimensioncap.com/p/the-melting-…
English
1
13
68
9.1K
Adam Goulburn รีทวีตแล้ว
Zavain Dar
Zavain Dar@zavaindar·
.@_DimensionCap has led, alongside DST Global and Hillhouse, $787M of new financings into Earendil Labs. The company uses cutting edge AI to redesign every joint of the discovery, design, manufacturing, and development process for new medicines, and is pioneering some of the most inventive research we’ve seen across the US and China. In terms of novel, orthogonal research: I’d describe this as the Deepseek of biotech. The company has multiple partnered and wholly owned assets entering or already in the clinic, and is building a burgeoning pipeline across I&I and onc .. pushing the boundaries of multispecific, multivalent, and next gen modalities. The team has publicly commented on pipeline, platform, and research partnerships with Sanofi, Wuxi and NVidia. I have to imagine: more to come. Finally, the company is led by famed UIUC/Tsinghua computer science professor, Jian Peng, and the original inventor of the class of PD-1/VEGF bispecifics, Zhenping Zhu. prnewswire.com/news-releases/…
English
4
11
102
14.8K
Adam Goulburn รีทวีตแล้ว
Frank Gao
Frank Gao@ChemVagabond·
We @_DimensionCap ported @karpathy's autoresearch framework to biology. We let Claude run 50 experiments over the weekend on protein thermostability prediction via @modal. It beat a recent baseline (TemBERTure) using a 20x smaller model. Code + research blog later this week!
Frank Gao tweet media
English
25
75
622
63K
Adam Goulburn รีทวีตแล้ว
Zavain Dar
Zavain Dar@zavaindar·
Memo: What If We're Right? We recently wrote a private letter to partners & friends of a common failure mode: the inability to consistently reason through the daisy chain of downstream consequence when non-consensus, low-probability, events actually occur pages: 1-3
Zavain Dar tweet mediaZavain Dar tweet mediaZavain Dar tweet media
English
4
13
96
12.3K
Adam Goulburn รีทวีตแล้ว
Simon Barnett
Simon Barnett@SimonDBarnett·
♻️ Recursive language models (RLMs) are incredibly cool and now is the time to be paying attention to them. Reasoning models are clearly the frontier. They've matured at breakneck speed. We've gone from simple chains-of-thought to sophisticated test-time scaling paradigms in a few years. Great! But how can we make reasoning more efficient at scale? -- ‼️ TL;DR Do surgery on an existing transformer. Install an internal recursion mechanism to create an RLM. The model's immune system will respond. That's okay. Conduct a 'healing phase' to reawaken the RLM to its new, hybrid reality. You don't need to do RL or SFT with valid reasoning traces to amplify higher-level thinking. Now, instead of scaling tokens/context at test-time to boost reasoning, the thinking happens within the model's latent space as it iteratively polishes its hidden state -- saving inference compute. Current reasoning models pay for every thought twice—once to generate the token, once to store it. RLMs think in place, exiting tokens only when ready. What's the new efficiency ceiling? Couldn't tell ya'. Is this a robust procedure yet? No. Do we know how the method scales or if we can reapply any other modern reasoning mechanisms? Also no. Is it obvious how far RLMs will take us or if they'll be the prevailing paradigm? Definitely not. But if we knew all these answers, it wouldn't be as interesting to read about. More detailed mini-essay below. 👇 -- ⏩ Skip this section if you don't want the background. I found out about RLMs at the inaugural workshop on efficient reasoning at @NeurIPSConf, where they were a fixture. It struck me how tech progress is rarely linear. It ebbs and flows with funding cycles, grinds to a halt if technology barriers pile up, and can explode with one eureka moment. But other times cool ideas fade into the scientific backwater if we get fixated on something that works. That's sort of what happened with the 2017 transformer unlock. We got these smooth-looking scaling laws with a simple recipe of parameters, tokens, and FLOPs. As pretraining waned, the field has moved into post / test-time training to keep the party going, and to much success. So, why not keep spamming this formula? I certainly would, especially if I'm a multi-billion $ frontier lab that can't afford to fall behind, break a narrative, etc. Engineering inertia is very real. That's why RLMs were so cool to hear about. It was like rediscovering an idea. Recursion isn't new, by any means. But every once in a while, ideas orbit back around. The immense gravity of parallel-processing via transformers seems to have pulled recurrent scaling back into the limelight. -- ▶️ RLM stuff starts here. The transformer essentially killed recurrent neural networks (RNNs) for language-modeling tasks. RNNs excel in some areas like real-time processing of sequential data, but they're notoriously hard to train as gradient updates can vanish/explode inside them. They also don't take advantage of GPU parallelism to the same extent. But was there a way to have one's cake and eat it too with some sort of hybrid model? Yes! Universal Transformers (UTs, 2018) were kind of like patient zero for this idea. You exploit the parallel attention of a transformer, but impose a recurrent inductive bias that exists in depth rather than sequence position. Basically, you're refining a hidden state representation at each token position until the model decides it's done thinking. But when's that? Ponder time (a/k/a adaptive computation time, ACT) came out a few years earlier. Here, you embed a lightweight halting classifier at each position that determines its doneness. Similar ideas were floating at the time, like neural GPUs, neural Turing machines, etc, but I like UTs because they combined global attention with recurrent depth and dynamic halting and also showed they could smoke vanilla transformers on contemporary benchmarks. arxiv.org/abs/1807.03819 -- The main RLM work I want to talk about combines several modern ideas (e.g., test-time scaling, recurrence, latent reasoning). The lead author @jonasgeiping also gave the talk! Conceptually, the idea is that humans don't vocalize our intermediate thoughts while reasoning, which is what current reasoning LLMs do -- they construct their reasoning trajectories via token-scaling. Using a prior on metastable brain waves, we talked about how neuronal activity bounces between these 'thinking modes' defined by MRI activity. So how can we bio-mimic this? Well, we can reason in the model's latent space rather than forcing thinking through the pinhole of token verbalization. Neat, but how? They made Huginn-0125, a 3.5B parameter model trained on 0.8T tokens. It's got three main parts: an encoder section (prelude), the inner recurrent block (R), and a decoder (coda), shown below. Importantly, there's a residual stream that concatenates the tokenized (but unaltered) input through each iteration to ensure training/inference stability. Huginn wasn't trained with an explicit number of iterations (k), but rather via random (Poisson distributed) k un-rollings, making sure the model stays on it's toes and doesn't expect to exit at a specific time. At test-time, the model dynamically iterates the hidden state through the R block, polishing each position until it's ready to go. The paper fully admits the model was trained sub-optimally for budget reasons, so I read this as a proof-of-principle, which makes the results even more exciting because there's a lot of work that can be stacked quickly. Without any RL or SFT on reasoning traces, Huginn is pretty competitive with models 2-3 its size on reasoning tasks (e.g., GSM8k). In effect, this is another way to scale at test-time. You simply increase iteration count in latent space instead of blowing up your token count as is currently being done. What really blew my mind was the call-back to brainwave regimes. They ran a PCA analysis of the model's latent space trajectories, finding that it sometimes converges on these orbital-like shapes during certain tasks. That's crazy to me because we saw the same thing with the MRI scans. I don't want to get heavy-handed with the bio-analogies, but this is 'pinch me' stuff. I think this is important because latent reasoning loses interpretability by default, so some semblance of a way to monitor these trajectories could be useful. arxiv.org/pdf/2502.05171 -- Many of the same authors (+ @SeanMcleish) did some important follow-up work that cements a few guardrails around RLMs and also an important concept -- You can adapt pre-trained transformers to do recurrent, latent reasoning. You don't need to start from scratch. This opens up doors for accessibility quite a lot. First, they find that you can take a many-layer, non-recurrent pre-trained transformer and cut it into the aforementioned prelude/recurrent/coda blocks. While this was originally traumatic for Llama, they noticed that additional training (healing) can adapt the network to the new inductive bias. Also, they find that initializing from the pre-trained weights is vastly more FLOPs-efficient. Compared to Huginn-0125, this retrofitted recurrence method was +12 points on MMLU and +7 points on GSM8k despite Huginn having 4-5x more parameters and being trained on 0.8T of data. Economically, this sounds like model distillation to me. At least in the sense you take a pre-trained teacher and build a smaller (in this case, depth-scalable) student without paying the full cost again. openreview.net/pdf?id=Oq3Xblt… -- I'll close with a bit of efficiency talk. We've seen how these RLMs can be created, even forked from existing open-source models, but I want to talk about economics. Another work points to the fact that Huginn-0125 was much slower (a factor of k) than the non-recurrent versions. We need some early threads about how to recover that speed. From the original paper Huginn-0125 paper, we know that some token positions mature quicker than others. Simpler iterates can exit quicker than others. So, this follow-up work addresses that by noting conceptual/mathematical similarities with diffusion. I could be wrong, but my understanding is that you can consider an RLM to be a continuous, latent diffusion model. Obviously the randomized unrolling objective is different than static denoising, but the analogy holds. So, they flip the recurrent process from batch processing to an assembly-line. Because some token positions finish early, waiting for the whole batch to be done is inefficient. They use a different sampler that fills the token position x iteration depth grid on a diagonal wavefront. At each step, you advance active positions by a step, decode draft tokens at the frontier, and freeze stable tokens. It's a bit like speculative decoding. You don't necessarily save on FLOPs, but you are exploiting GPU parallelism in a way the initial, sequential setup left on the table. This netted a 5x speed-up with minimal accuracy loss, though obviously there's still much room to run. openreview.net/pdf?id=nA5IRfA… -- Alright, here's the rub. On paper, nothing about RLMs seems economically appealing as of today. Recurrent depth imposes a speed penalty that has only been partially offset with wavefront diffusion sampling. This is substantially behind transformers and not anywhere close to cases where time-to-first-token matters. There are new failure modes (e.g., overthinking can lead to inverse scaling), new hyper-parameters, less mature tooling, etc. Current performance doesn't appear to be an OOM better than token-scaling. So, why be excited about this? Well, consider a future where frontier labs have multiple different models/architectures under the hood that they route to different requests depending on what the situation calls for (with gross margin being ubiquitous in the denominator). The fact that RLMs have fixed hidden states regardless of iteration count is important, especially if juxtaposed against KV caches that grow linearly with chain-of-thought methods. Theoretically, RLMs are Turing complete and can loop ad infinitum, maybe allowing them to address extremely difficult tasks that feed-forward networks can't, though there's zero proof of this yet? I'm not sure how many latent iterations will be needed to approximate equivalent token-scale reasoning, but if it's relatively small, I can see a place for RLMs to co-exist. There will need to be new infrastructure. Maximizing inference efficiency seems like the biggest one. We'd need routers/schedulers that assume fixed KV and dynamic iteration. Perhaps some purpose-built kernels for diagonal wavefront sampling, exit tracking, etc could be useful optimizations? I'm really hoping this stuff goes mainstream!
Simon Barnett tweet mediaSimon Barnett tweet mediaSimon Barnett tweet mediaSimon Barnett tweet media
English
7
19
154
11K
Adam Goulburn รีทวีตแล้ว
Samuel Stanton
Samuel Stanton@samuel_stanton_·
My cofounders and I have been building Coefficient Bio for the past 5 months. We're ushering biopharma into the Intelligence Age. It will change everything about how the industry learns and makes decisions. If that's a future you want to build, get in touch!
Nathan C. Frey@nc_frey

Join the early team @CoefficientBio. We have an AI team that I’m incredibly proud of, and truly think is the most effective and exceptional technical team building in AI x bio. We're looking for people to: * Build and maintain robust AI systems to an exacting standard * Design and run experiments that let us iterate fast * Work on challenging scientific and engineering problems for human flourishing NYC-based (or willing to relocate). If you’re curious about what we’re building, reach out directly. If we don’t know each other yet, a great cold DM (or email: join@coefficientbio.com), or warm intro goes a long way.

English
7
11
135
23.8K
Adam Goulburn รีทวีตแล้ว
Nathan C. Frey
Nathan C. Frey@nc_frey·
Join the early team @CoefficientBio. We have an AI team that I’m incredibly proud of, and truly think is the most effective and exceptional technical team building in AI x bio. We're looking for people to: * Build and maintain robust AI systems to an exacting standard * Design and run experiments that let us iterate fast * Work on challenging scientific and engineering problems for human flourishing NYC-based (or willing to relocate). If you’re curious about what we’re building, reach out directly. If we don’t know each other yet, a great cold DM (or email: join@coefficientbio.com), or warm intro goes a long way.
English
13
33
330
49.2K
Adam Goulburn รีทวีตแล้ว
DIMENSION
DIMENSION@_DimensionCap·
Excited to welcome @nikhilvnamburi to the investment team at Dimension!! Nikhil brings an investment background in compute / ML infrastructure to the team and will help the firm continue to expand our work in cutting-edge technology and science. In addition to his domain expertise, Nikhil brings a deep curiosity for the technologies and the businesses shaping the modern internet and a passion for the craft of investing. Welcome Nikhil! dimensioncap.com/people/nikhil-…
DIMENSION tweet media
English
2
5
37
4.2K
Derek Brand
Derek Brand@echo_nyc·
Fun fact - @AGoulburn has not changed his outfit for JPM since 2016. #JPM2026
Chris Gibson@RecursionChris

On the annual @Delta #JPM2026 party bus from SLC to SF with @_DimensionCap and about half of @BioHiveUtah. I remember my first JPM in 2016. We were frequently laughed at (literally, in multiple meetings) or disregarded by the industry… Here we are 10 years later in what is clearly the start of a golden age of TechBio. So many incredible companies, including many of the large pharma, are now making AI a central pillar in the strategy to bring new, better medicines to patients quickly. And that fills me with joy. I’m deeply committed to @RecursionPharma and I am confident we will continue to lead the TechBio field. But I want most of all for the field to progress and patients to win. So if you are a new TechBio company, a young founder, or even a large-pharma veteran and open to exploring how you can leverage technology to accelerate your work, shoot me a note - I have some time Wednesday afternoon and would love to meet up! See you on the streets of SF!

English
2
0
8
1.4K
Adam Goulburn รีทวีตแล้ว
Nan Li
Nan Li@nanli·
Congrats to the @chaidiscovery team on a massive announcement today!! This broad-sweeping asset design deal with Lilly is groundbreaking in many ways and is another signal of a broad platform shift in life science driven by AI. endpoints.news/eli-lilly-chai…
English
3
5
44
12.9K
Adam Goulburn รีทวีตแล้ว
Simon Barnett
Simon Barnett@SimonDBarnett·
2026 will be a big year for ML architecture innovation. Sure, scaling and how we spend our tokens (e.g., at test time) matters a lot. But it's fascinating to look at how our understanding of human cognition has inspired waves of ML design improvements. Check these out 👇
English
2
5
15
2.4K
Adam Goulburn รีทวีตแล้ว
Simon Barnett
Simon Barnett@SimonDBarnett·
Still going through @deepseek_ai's mHC paper? I wrote an easy-to-follow read-along guide. When I first read this paper, it sounded like word soup because so many things were new to me. I wrote this to help me understand it, so hopefully it's useful for you too!
Simon Barnett@SimonDBarnett

x.com/i/article/2006…

English
2
5
20
2.9K
Adam Goulburn รีทวีตแล้ว
Dave Sharma
Dave Sharma@DaveSharma·
Royal Commissions supported by Anthony Albanese: ✅ UN oil-for-food ✅ Institutional Child Sexual Abuse ✅ Banking and financial services ✅ Aged Care quality ✅ Disability abuse and neglect ✅ Robodebt ✅ Veteran suicide ❌ Australia’s most deadly terrorist attack.
English
345
385
1.7K
48.3K
Adam Goulburn รีทวีตแล้ว
DIMENSION
DIMENSION@_DimensionCap·
We’re thrilled to announce our partnership with the @BioRender team! As scientists, engineers, and computational biologists ourselves, we’ve been long-time users of BioRender - many of us relying on it for years to communicate complex ideas clearly and effectively. @ShizAoki + Ryan Marien + @KatyaSht have systematically built what has become an indispensable tool for builders and researchers - driving organizations globally. And the company has transformed how science is communicated - making research more visible and *legible* in a simple, highly efficient way. At a time when standardized, clear scientific communication has never been more important, BioRender is leading the way. And we’re excited to partner with a team so aligned with the future of how science is shared. biorender.com/blog/funding-a…
English
0
5
11
1.1K
Adam Goulburn รีทวีตแล้ว
StandWithUs
StandWithUs@StandWithUs·
A heartbreaking photo from the funeral of Matilda Britvan, who was brutally murdered by terrorists in the Bondi Beach Hanukkah massacre on December 14, 2025. She was just 10 years old. May her memory forever be a blessing. 🕯️
StandWithUs tweet media
English
252
684
3.8K
51.6K
Adam Goulburn รีทวีตแล้ว
StandWithUs
StandWithUs@StandWithUs·
Adam (50) was brutally murdered while on a walk with his wife at Bondi Beach as the massacre began. He leaves behind his wife and four children. Boris was murdered while attending the Hanukkah event together with his son. His son was also shot and is being treated in hospital. Boris leaves behind his wife and two sons. May their memories forever be a blessing. 🕯️
StandWithUs tweet media
English
57
200
806
8.6K