Ian Quigley

6.9K posts

Ian Quigley banner
Ian Quigley

Ian Quigley

@allmeasures

Cofounder, @LeashBio. Ex-Recursion, Arima Genomics, Nanocellect, Salk, UT Austin, Baylor College of Medicine, Rice. He/him.

Salt Lake City, UT Katılım Mayıs 2010
2.6K Takip Edilen1.1K Takipçiler
Sabitlenmiş Tweet
Ian Quigley
Ian Quigley@allmeasures·
Ian Quigley tweet media
ZXX
0
1
13
0
Ian Quigley retweetledi
Michael Dempsey
Michael Dempsey@mhdempsey·
i think we're in the mid-curve era of building startups now that most of the low-hanging fruit and obvious ideas are going to get bitter lesson'd over the coming 24 months.
English
12
7
210
22K
Ian Quigley retweetledi
Anshul Kundaje
Anshul Kundaje@anshulkundaje·
"These findings highlight limitations of current pLMs for mutational effect prediction and suggest that dataset composition, rather than model architecture or training, is the primary driver of predictive success." Again and again and again. Exact same problem with DNALMs
Prof. Nikolai Slavov@slavov_n

Intrinsic dataset features drive mutational effect prediction by protein language models biorxiv.org/content/10.648…

English
4
21
146
24.2K
Becky Pferdehirt
Becky Pferdehirt@beckypferdehirt·
Big news! I’m joining @AsteraInstitute as CEO of Radial, their new life sciences division. How we fund, do, and build upon science has long needed an update. At Radial, we design, fund, and operate programs that tackle foundational scientific problems while simultaneously testing better ways to do science.
Becky Pferdehirt tweet media
English
20
22
221
28.4K
Ian Quigley retweetledi
Utah Daily Snow
Utah Daily Snow@WasatchSnow·
BAD NEWS! Usually, we don't see our statewide snowpack peak until around March 24. However, with an upcoming ridge and accompanying heatwave, I think we've already peaked. If that turns out to be the case, this season's peak statewide mean snowpack will be the lowest on record.
Utah Daily Snow tweet media
English
39
81
707
63.3K
Ian Quigley
Ian Quigley@allmeasures·
If we restrict ourselves to legacy data lacking such variation, we'll be stuck with models that can only remember what they've seen and not able to predict what they haven't. (16/16)
English
0
0
8
300
Ian Quigley
Ian Quigley@allmeasures·
Exploring mutations is rarely a goal of legacy datasets. All this work suggests that to better predict protein behaviors, we must collect *new* data, and we'll have to do it in a way that intentionally mitigates such cheating. (15/n)
English
1
1
9
387
Ian Quigley
Ian Quigley@allmeasures·
Current protein models seem to only memorize their wild-type training sets and lack physics understanding. One nice way forward is physically engineering mutations and then taking more ground-truth measurements. It works. We've already shown this! (1/n)
Biology+AI Daily@BiologyAIDaily

Adversarial Sequence Mutations in AlphaFold and ESMFold Reveal Nonphysical Structural Invariance, Confidence Failures, and Concerns for Protein Design 1. A new adversarial study systematically evaluates AlphaFold 3's robustness by introducing point mutations (up to 70%) and deletions (up to 10%) across 200 proteins, revealing striking structural invariance that raises fundamental questions about the model's biophysical reasoning capabilities. 2. The most concerning finding: AlphaFold 3 maintains virtually identical predicted structures even when 40% of residues are mutated with deliberately destabilizing substitutions, or when 10% of residues are deleted—perturbations that would catastrophically destabilize real proteins. 3. This structural invariance persists even for experimentally validated fold-switching proteins, where specific mutations are known to induce alternative conformations. AlphaFold 3 fails to capture these biologically critical transitions, suggesting limited sequence-structure coupling. 4. Confidence metrics prove unreliable: AlphaFold 3's ranking score selects the most accurate structure only ~25% of the time, and these scores correlate more strongly with template availability in the training set than with actual prediction quality. 5. Comparative analysis with ESMFold reveals that the protein language model-based approach shows significantly greater sensitivity to mutations, with structures diverging more rapidly as sequence perturbations increase—suggesting superior learned sequence-structure relationships despite lower absolute accuracy. 6. The study's template analysis provides quantitative evidence that AlphaFold 3's confidence reflects structural similarity to training-set exemplars (Pearson r=0.39) rather than genuine biophysical assessment, indicating heavy reliance on memorized patterns over learned principles. 7. These findings have profound implications for the entire AlphaFold ecosystem: protein design tools like RFdiffusion, binder design methods like BoltzGen and BindCraft, and drug discovery pipelines may inherit these fundamental limitations, potentially generating non-physical sequences or missing viable candidates. 8. The work identifies critical gaps in current structure prediction—models trained primarily on stable, wild-type proteins lack exposure to destabilized mutants and misfolded states, limiting their ability to generalize beyond the training distribution. 📜Paper: biorxiv.org/content/10.648… #AlphaFold #AlphaFold3 #ProteinStructurePrediction #StructuralBiology #ProteinDesign #MachineLearning #Bioinformatics #ComputationalBiology #AIforScience #ProteinEngineering #DeepLearning #Biophysics

English
3
16
117
15.3K