Simon Kozlov

1.3K posts

Simon Kozlov

Simon Kozlov

@sim0nsays

Researcher at MIT, generative AI for biology

New York, NY Katılım Mart 2009
317 Takip Edilen2.6K Takipçiler
Simon Kozlov
Simon Kozlov@sim0nsays·
Coding is solved by AI no more than protein folding is solved by AI
English
1
1
10
439
Simon Kozlov retweetledi
Creators on the Blox
Creators on the Blox@CreatorsOnBlox·
Former Roblox CTO Keith V Lucas joins the show to reminisce about the early days of the platform, share memories of David and Erik, and break down notable moments throughout Roblox history. New episode out now on YouTube, Spotify, and anywhere else you listen to podcasts!
English
12
42
421
51.1K
Simon Kozlov retweetledi
owl
owl@owl_posting·
What could Alphafold 4 look like? (Sergey Ovchinnikov, Ep #3) 2 hours listening time (links below) To those in the (machine-learning for protein design) space, Dr. Sergey Ovchinnikov (@sokrypton) is a very, very well-recognized name. A recent MIT professor (circa early 2024), he has played a part in a staggering number of recent great papers in the field: ColabFold, RFDiffusion, Bindcraft, automated design of soluble proxies of membrane proteins, elucidating what protein language models are learning, conformational sampling via Alphafold2, and many more. Of course, all these papers were group efforts, but Sergey's name comes up astonishingly frequently! And even beyond the research that have come from his lab in the last few years, the co-evolution work he did during his PhD/fellowship also laid some of the groundwork for the original Alphafold paper, being cited twice in it. This is a two hour conversation with him, asking every question I could think of. We talk about his own journey into biology research, an issue he has with Alphafold3, what Alphafold4-and-beyond models may look like, what research he’d want to spend a hundred million dollars on, and lots more. Topics/institutions we discuss: @arcinstitute's Evo models, @HWaymentSteele's work, @IsomorphicLabs's AF2/AF3, and @EvoscaleAI's ESM models Also, extremely grateful to Asimov Press (@asimovpress) for helping fund the travel + studio time required for this episode! They are a non-profit publisher dedicated to thoughtful writing on biology and metascience, such as articles over synthetic blood and interviews with plant geneticists. I myself have published within them twice! I highly recommend checking out their essays at asimov.press, or reaching out to editors@asimov.com if you’re interested in contributing. Timestamps: [00:00:00] Highlight clips [00:01:10] Introduction + Sergey's background and how he got into the field [00:18:14] Is conservation all you need? [00:23:26] Ambiguous vs non-ambiguous regions in proteins [00:24:59] What will AlphaFold 4/5/6 look like? [00:36:19] Diffusion vs. inversion for protein design [00:44:52] A problem with Alphafold3 [00:53:41] MSA vs. single sequence models [01:06:52] How Sergey picks research problems [01:21:06] What are DNA models like Evo learning? [01:29:11] The problem with train/test splits in biology [01:49:07] What Sergey would do with $100 million
English
8
68
337
83.4K
John Shedletsky
John Shedletsky@Shedletsky·
Hey followers who are game developers: you should play Slay the Spire because it will teach you about game design & especially UX design. Hey StS fans: you should try the Packmaster mod, it makes StS a whole new game. Hey Packmaster fans: 👍 Hey Elon: let me link to this stuff without deboosting my tweets. Thx!
English
34
12
298
52.2K
Vega Shah
Vega Shah@dr_alphalyrae·
best tea leaf salad in the bay area?
English
4
0
4
1.9K
Simon Kozlov
Simon Kozlov@sim0nsays·
@ElliotHershberg The cost of testing out ideas is formulating them cleanly :) Hmm didn’t realize connection to writing before.
English
1
0
1
134
Elliot Hershberg
Elliot Hershberg@ElliotHershberg·
I find myself more aggressively using version control when I use Cursor. That way, I can plunge down a rabbit hole and just as quickly pop back out. The cost of testing out specific ideas is basically zero.
English
2
0
10
1.7K
Andrej Karpathy
Andrej Karpathy@karpathy·
Exactly, roughly what I tried and mostly failed. I want to highlight some text in the pdf, pull out the highlight, the preceding text of the chapter, maybe the generated summaries of the other chapters, put it all together, attach nearby images if any… there’s a whole design space on how to build the context before you submit different kinds of queries to the AI book club. Queries like explain, discuss, argue in favor or opposed, take notes, create anki cards, generate quiz or exercises for thinking through the content, …
English
59
30
1.1K
147K
Andrej Karpathy
Andrej Karpathy@karpathy·
One of my favorite applications of LLMs is reading books together. I want to ask questions or hear generated discussion (NotebookLM style) while it is automatically conditioned on the surrounding content. If Amazon or so built a Kindle AI reader that “just works” imo it would be a huge hit. For now, it is possible to kind of hack it with a bunch of script. Possibly someone already tried to build a very nice AI-native reader app and I missed it.
English
351
483
7.8K
1.2M
Simon Kozlov
Simon Kozlov@sim0nsays·
@ElliotHershberg @cursor_ai @ProjectJupyter Apply doesn't work from the side pane, but works from the inline chat (Cmd-K is the shortcut). Context from prior cells is weird - as you can see, it can see it sometimes, but often can't. My theory is all the notebook output is also passed and can quickly fill all the context.
Simon Kozlov tweet media
English
0
0
0
81
Elliot Hershberg
Elliot Hershberg@ElliotHershberg·
@sim0nsays @cursor_ai @ProjectJupyter Maybe I'm missing something, but Cmd-K can't actually apply code suggestions into notebook cells, and seems to have a total lack of context for prior cells (which is surprising given the notebook file is attached as context!)
English
1
0
0
181
Simon Kozlov
Simon Kozlov@sim0nsays·
@ProfTomEllis Another theme is making all kinds of stretched arrangements for kids for an evening call and the other person not showing up
English
0
0
1
118
Tom Ellis
Tom Ellis@ProfTomEllis·
10pm zoom calls to west coast US folk are a great idea until one sleeping kid throws up in their bed and then the whole big brain plan of an uninterrupted call falls completely apart.
English
1
1
17
1.3K
Simon Kozlov
Simon Kozlov@sim0nsays·
@ferruz_noelia Appreciate the response! Ok, so predicting structure is not a relevant problem, so no good analogy. How about an analog of Protein Language Models a few years earlier in Chem ML?
English
1
0
0
78
Noelia Ferruz
Noelia Ferruz@ferruz_noelia·
@sim0nsays Predicting 3D structures of molecules was never as much of a problem as it is for proteins because they have multiple quite different conformations in solution and the torsions can in either case be solved with various levels of QM theory
English
1
0
1
98
Noelia Ferruz
Noelia Ferruz@ferruz_noelia·
The easiest way to predict what we Bio+ML people will be publishing in two years is to look at what Chem+ML people are publishing today
English
3
7
91
10.2K
Simon Kozlov
Simon Kozlov@sim0nsays·
@nickcammarata Was just playing with a prompt for Claude to reformulate comments in the private chat (say judgmental unsolicited feedback to supportive open-ended questions), kinda want to have this as a filter
English
0
0
0
53
Nick
Nick@nickcammarata·
i wish there were a way to make twitter healthier, have claude filter every tweet before i see it for the healthy ones most individual moment of twitter feel like junk food but over the long term it produces so many magical things. unclear if deleting is good
English
40
8
500
25.2K
Simon Kozlov
Simon Kozlov@sim0nsays·
@karpathy 100% agree. It feels connected to Goodhart’s Law - to measure something as a stable eval metric, you need to make it constrained, reduce the complexity of tradeoffs, get it to “right-wrong” answer, make context clear, etc. And all that makes it less and less real.
English
0
0
4
402
Andrej Karpathy
Andrej Karpathy@karpathy·
Moravec's paradox in LLM evals I was reacting to this new benchmark of frontier math where LLMs only solve 2%. It was introduced because LLMs are increasingly crushing existing math benchmarks. The interesting issue is that even though by many accounts (/evals), LLMs are inching well into top expert territory (e.g. in math and coding etc.), you wouldn't hire them over a person for the most menial jobs. They can solve complex closed problems if you serve them the problem description neatly on a platter in the prompt, but they struggle to coherently string together long, autonomous, problem-solving sequences in a way that a person would find very easy. This is Moravec's paradox in disguise, who observed 30+ years ago that what is easy/hard for humans can be non-intuitively very different to what is easy/hard for computers. E.g. humans are very impressed by computers playing chess, but chess is easy for computers as it is a closed, deterministic system with a discrete action space, full observability, etc etc. Vice versa, humans can tie a shoe or fold a shirt and don't think much of it at all but this is an extremely complex sensorimotor task that challenges the state of the art in both hardware and software. It's like that Rubik's Cube release from OpenAI a while back where most people fixated on the solving itself (which is trivial) instead of the actually incredibly difficult task of just turning one face of the cube with a robot hand. So I really like this FrontierMath benchmark and we should make more. But I also think it's an interesting challenge how we can create evals for all the "easy" stuff that is secretly hard. Very long context windows, coherence, autonomy, common sense, multimodal I/O that works, ... How do we build good "menial job" evals? The kinds of things you'd expect from any entry-level intern on your team.
Epoch AI@EpochAIResearch

1/10 Today we're launching FrontierMath, a benchmark for evaluating advanced mathematical reasoning in AI. We collaborated with 60+ leading mathematicians to create hundreds of original, exceptionally challenging math problems, of which current AI systems solve less than 2%.

English
151
505
4K
448.1K
Simon Kozlov
Simon Kozlov@sim0nsays·
@Shedletsky Business opportunity: buy used Lego in bulk, use AI to come up with sets from the pieces you got, sell full sets
English
0
0
1
146
John Shedletsky
John Shedletsky@Shedletsky·
The price of used Lego on eBay has declined 75% in real terms over the past two decades. Civilization is doomed.
English
5
1
61
7.3K
Simon Kozlov
Simon Kozlov@sim0nsays·
Salvador Dali, "Using Roblox Studio for the first time"
Simon Kozlov tweet media
English
1
0
27
1.7K