Alan DenAdel

335 posts

Alan DenAdel

Alan DenAdel

@AlanDenadel

Bioinformatics scientist @AllenInstitute. PhD in computational biology from @BrownCCMB @BrownUniversity. Previously at @illumina. he/him

Seattle, WA Beigetreten Ağustos 2018
1.3K Folgt271 Follower
Alan DenAdel retweetet
Ava Amini
Ava Amini@avapamini·
protein language models capture rich structural signals, but where that knowledge lives in the network is still unclear we show that small subnetworks inside PLMs encode structural concepts, from residues to folds journals.plos.org/ploscompbiol/a… @PLOSCompBiol work led by @riavinod_!
Ava Amini tweet media
English
0
26
170
13.8K
Alan DenAdel retweetet
Thomas Serre
Thomas Serre@tserre·
Postdoc opening in APMA @BrownUniversity — specifically looking for candidates bridging APMA + brain science or CS. Review starts April 1! Great fit for anyone interested in collaborating with faculty at the @CarneyInstitute Nancy G. Zimmerman Center for Comput. Brain Science. 🧵
English
1
8
21
2.5K
Alan DenAdel
Alan DenAdel@AlanDenadel·
@jmschreiber91 It's nice when randomize gifts you two binding motifs with the initial sequence
Alan DenAdel tweet media
English
1
0
1
28
Jacob Schreiber
Jacob Schreiber@jmschreiber91·
A cool part of designing synthetic regulatory DNA is that the results are frequently surprising. Build your intuition with a fun little design-via-editing game: programmable-genomics.github.io/dna-game.html See if you can maximize the binding of GATA, but don't use too many edits!
English
2
8
67
6.3K
Alan DenAdel retweetet
Jacob Schreiber
Jacob Schreiber@jmschreiber91·
The Programmable Genomics Lab at @UMassGCB just officially launched our site! We have a "simple" goal: to develop synthetic regulatory elements that target every cell type in every tissue, and control payload dosage/duration Check it out! programmable-genomics.github.io
English
2
40
153
22.8K
Alan DenAdel retweetet
Till Richter
Till Richter@TillRichter6·
New preprint: Beyond alignment: synergistic integration is required for multimodal cell foundation models. 🔗 biorxiv.org/content/10.648… Multimodal compositional cell foundation models are emerging as a path toward virtual cells. But when does multimodal fusion truly add info? 🧵
English
1
4
28
3.9K
Alan DenAdel retweetet
Lorin Crawford
Lorin Crawford@lorin_crawford·
Sharing our newest preprint on data synergy and multimodal cell foundation models, where we study when integration actually adds value. Main takeaway: "virtual cells will require synthesis, not just correspondence." Great work led by @TillRichter6! See more details below.
Till Richter@TillRichter6

New preprint: Beyond alignment: synergistic integration is required for multimodal cell foundation models. 🔗 biorxiv.org/content/10.648… Multimodal compositional cell foundation models are emerging as a path toward virtual cells. But when does multimodal fusion truly add info? 🧵

English
0
1
13
1.1K
Alan DenAdel retweetet
Michael 英泉 Eisen
Michael 英泉 Eisen@mbeisen·
Science would be a lot better if everyone read Franny and Zooey.
English
2
1
20
4.2K
Alan DenAdel
Alan DenAdel@AlanDenadel·
@amanpatel100 Very interesting work! Do you know if anyone has ever looked into using known motifs as convolutional filters (instead of learning them, at least for a subset of filters), similar in spirit to how expiMap uses prior knowledge of gene programs in its decoder?
English
1
0
0
24
Aman Patel
Aman Patel@amanpatel100·
ARSENAL’s masked likelihoods allow for the de novo reconstruction of a wide variety of known binding motifs, thus proving a powerful tool for analyzing individual regulatory regions in detail or studying functional syntax in aggregate.
Aman Patel tweet media
English
2
1
0
582
Alan DenAdel retweetet
Aman Patel
Aman Patel@amanpatel100·
Excited to announce our latest work! We present ARSENAL, a short-context DNA language model specifically designed to learn important sequence features in noncoding regulatory DNA. Why is such a model important? Read on to find out! biorxiv.org/content/10.648…
English
3
29
97
29.2K
Alan DenAdel retweetet
Micaela
Micaela@micaelanonsense·
Announcing our preprint: Predicting evolutionary rate as a pretraining task improves genome language model representations: biorxiv.org/content/10.648… Genome language models (gLMs) have the potential to further understanding of regulatory genomics without requiring labeled data...
English
4
23
84
26.5K
Alan DenAdel retweetet
Bo Wang
Bo Wang@BoWang87·
🚨 New preprint -- Predicting evolutionary rate as a pretraining task improves genome language model representations led by my phd student, Micaela Consens during her internship at @Microsoft 👉 paper: biorxiv.org/content/10.648… Genome language models keep getting bigger — but are we training them on the right objectives? This paper makes a strong case that biology already gives us the answer: 🧬 evolutionary rate = functional signal Instead of only doing sequence reconstruction (NTP / MLM), the team pretrains models to predict evolutionary rate — and combines it with standard objectives in a clean, controlled way. The result? 🔥 Better representations 🔥 Stronger zero-shot performance 🔥 Big gains in functional region discovery & variant effect prediction 🔥 Small models competing with much larger gLMs Key insight: Evolution isn’t just a benchmark — it’s a training signal. Even more impressive: the paper introduces biologically grounded zero-shot evaluations that avoid many pitfalls of existing gLM benchmarks. Big picture: We won’t get better genome models by scaling alone. We get them by aligning learning objectives with biology.
Micaela@micaelanonsense

Announcing our preprint: Predicting evolutionary rate as a pretraining task improves genome language model representations: biorxiv.org/content/10.648… Genome language models (gLMs) have the potential to further understanding of regulatory genomics without requiring labeled data...

English
1
10
53
11.7K
Alan DenAdel retweetet
Hugues Van Assel
Hugues Van Assel@hugues_va·
Excited to share an internship opening in Regev Lab Genentech! 🧬 We’re looking for a PhD intern to work at the frontier of Bayesian Optim and LLMs for high-throughput genomics. Co-advised by Edward de Brouwer and me in SF. RT like there is no tomorrow ! (is there ?) Link below.
English
4
18
97
10.6K
Alan DenAdel retweetet
Alan DenAdel retweetet
Jorge Bravo Abad
Jorge Bravo Abad@bravo_abad·
Hierarchical loss functions for cell-type annotation at atlas scale Automated cell-type annotation is foundational for single-cell RNA-seq analysis—and it's fundamentally a classification problem with structured labels. Cell types form a hierarchical ontology: 'leukocytes' contain 'lymphocytes', which contain 'B cells' and 'T cells'. But standard cross-entropy loss treats these labels as flat and independent, ignoring the biological relationships that define them. This mismatch becomes acute when models encounter new data. Sebastiano Cultrera di Montesano and colleagues trained linear classifiers, MLPs, and TabNet on 15.2 million cells from CELLxGENE spanning 164 cell types. In-distribution performance was strong (80–84% macro F1). But when evaluated on 2.6 million cells from 21 newly released studies—same cell types, same assays—performance dropped by 24–32%. The models had memorized study-specific patterns rather than learning generalizable cell-type signatures. The fix is elegant: a hierarchical cross-entropy (HCE) loss that propagates probability mass up the ontology DAG. For any cell type, the adjusted score equals its raw probability plus the probabilities of all its descendants. This encodes a consistency constraint: predicting 'CD4+ T cell' should increase the probability of 'T cell' and 'lymphocyte'. Implementation requires only a reachability matrix multiplication—no architecture changes, no hyperparameter tuning. The results are remarkable. HCE improves out-of-distribution macro F1 by 12–15% across all three architectures, recovering roughly half the performance drop. Gains concentrate in internal nodes embedded in densely connected regions of the ontology, where hierarchical signal can propagate across related types. The improvement is architecture-agnostic—linear models benefit as much as transformers. The broader message challenges current trends: rather than scaling model complexity, aligning training objectives with biological structure yields consistent generalization gains. For atlas curation, this suggests prioritizing studies that increase ontology connectivity in underrepresented regions over simply adding more cells. Paper: nature.com/articles/s4358…
Jorge Bravo Abad tweet media
English
6
19
101
7K
Alan DenAdel retweetet
Lorin Crawford
Lorin Crawford@lorin_crawford·
Excited to see our hierarchical cross-entropy strategy for improved cell type annotation out in @NatComputSci today! Congratulations to the team! Check out @sebacultrera's original breakdown of our approach here: x.com/sebacultrera/s…
Nature Computational Science@NatComputSci

📢Out now! @sebacultrera, @davide_dascenzo, @avapamini, @peterswinter, @lorin_crawford, and colleagues present a hierarchical cross-entropy loss that improves performance of single-cell annotation models. nature.com/articles/s4358…

English
1
2
8
623
Alan DenAdel retweetet
Alan DenAdel retweetet
Constantine Tzouanas
Constantine Tzouanas@constantine_sci·
Incredible to see our work on @CellCellPress cover! Environmental stresses like high-fat diets➡️cellular communities whose past histories & present neighbors shape future disease progression & cancer outcomes (with Akira-inspired Easter eggs too!). Appreciation to @TeamSciStories
Constantine Tzouanas tweet media
English
1
1
8
301