tbepler

132 posts

tbepler

tbepler

@tbepler1

Scientist and Group Leader of the Simons Machine Learning Center @SEMC_NYSBC. Co-founder and CEO of https://t.co/YcCw0bpHZo. Opinions are my own.

Inscrit le Ekim 2018
152 Abonnements645 Abonnés
Tweet épinglé
tbepler
tbepler@tbepler1·
Excited to share PoET-2, our next breakthrough in protein language modeling. It represents a fundamental shift in how AI learns from evolutionary sequences. 🧵 1/13
English
1
38
205
29K
tbepler retweeté
OpenProtein.AI
OpenProtein.AI@openprotein·
We’re excited to announce our expanded partnership with Boehringer Ingelheim. Together, we are building the future of AI‑driven antibody discovery and optimization. openprotein.ai/strategic-part…
OpenProtein.AI tweet media
English
0
1
3
173
tbepler
tbepler@tbepler1·
@PatrickSoga @openprotein Yes, we will probably take 1-2 interns next year. If you're interested feel free to send us your CV
English
0
0
1
46
tbepler
tbepler@tbepler1·
@RippaSatss @openprotein We'll consider anyone with relevant skills who's motivated. Send us your CV and tell us more about what you've worked on!
English
1
0
1
107
Rippa Sats
Rippa Sats@RippaSatss·
Quick question: Do you consider non-PhD candidates with competition-proven protein design experience? - 50 proteins designed and ranked #386/681 (Nipah Binder Competition - ProteinBase) with only laptop. - #207/337 globally (ARC Virtual Cell Challenge) - Drug discovery platforms: 1,192+ molecules validated I know I don't have the PhD, but I've built what you're building. Thoughts? You need builders who ship. I ship. Let's talk.
English
1
0
1
222
Kyle Tretina, Ph.D.
Kyle Tretina, Ph.D.@AllThingsApx·
Quietly, this paper reset the trajectory of protein language modeling earlier this month. Retrieval-at-inference (MSAs) + pair-representation beats 10–100× larger single-sequence LMs on contacts, PPIs, variant effects. Not just a benchmark win - a blueprint: query-biased MSAs, explicit pair geometry, stress-tested reliability.
Kyle Tretina, Ph.D. tweet media
English
3
17
159
10.3K
tbepler
tbepler@tbepler1·
Our preprint on sequence-to-property learning and zero-shot fitness prediction with PoET-2 is live: arxiv.org/abs/2508.04724
English
2
3
26
2K
tbepler retweeté
Biology+AI Daily
Biology+AI Daily@BiologyAIDaily·
Understanding Protein Function with a Multimodal Retrieval-Augmented Foundation Model 1. PoET-2, a new protein language model, achieves state-of-the-art performance in predicting the effects of mutations on protein function, especially for challenging cases like insertions/deletions and higher-order mutations. This model combines sequence, structure, and evolutionary information in a novel way to improve protein understanding and design capabilities. 2. The model incorporates a hierarchical transformer encoder and dual decoders with both causal and masked language modeling objectives. This dual training approach allows PoET-2 to excel in both generative tasks (like sequence generation) and bidirectional representation learning, making it versatile for various protein-related tasks. 3. PoET-2 leverages retrieval augmentation, which enables it to learn from context and incorporate new sequences not present in the original training data. This feature enhances its ability to adapt to different protein families and their specific evolutionary constraints, leading to more accurate predictions. 4. In zero-shot variant effect prediction, PoET-2 outperforms previous models significantly, especially on datasets involving multiple mutations and indels. It also shows superior performance in supervised settings with limited data, demonstrating excellent data efficiency and generalization ability. 5. The model's architecture includes a structure-based attention bias mechanism, which integrates structural information into the attention operations. This enhances the model's ability to capture 3D structural relationships, contributing to its improved performance in tasks related to protein structure and function. 6. PoET-2 is compact, with only 182 million parameters, making it efficient and scalable. Despite its smaller size, it matches or exceeds the performance of much larger models, highlighting its efficiency and practicality for real-world applications in protein engineering and design. 7. The authors demonstrate PoET-2's effectiveness across various benchmarks, including deep mutational scanning and clinical datasets. The model's ability to predict the fitness effects of mutations accurately can accelerate the development of new therapeutics and enhance our understanding of disease mechanisms. 📜Paper: arxiv.org/abs/2508.04724 #ProteinEngineering #AIinBiology #MachineLearning #ProteinFunction #MutationPrediction
Biology+AI Daily tweet media
English
0
10
52
3.8K
tbepler
tbepler@tbepler1·
If you're interested in the PLM used here, check out PoET-2 from @openprotein! It allows controllable protein generation with sequence (homologue) and optional structure conditioning and is available right now. 🦾
Boris Power@BorisMPower

At @OpenAI, we believe that AI can accelerate science and drug discovery. An exciting example is our work with @RetroBiosciences, where a custom model designed improved variants of the Nobel-prize winning Yamanaka proteins. Today we published a closer look at the breakthrough. ⬇️

English
2
1
38
3.1K
tbepler retweeté
Andrew Gordon Wilson
Andrew Gordon Wilson@andrewgwils·
Repeat after me: learning from data is only possible by making assumptions.
English
8
4
146
11.7K