Ankur Parikh

108 posts

Ankur Parikh

Ankur Parikh

@ank_parikh

Staff Research Scientist at Google DeepMind. Former adjunct assistant prof at @NYU_Courant. PhD at @mldcmu. ML for Bio/Chem (Prev. NLP). All opinions my own.

New York Katılım Ağustos 2012
3.5K Takip Edilen3.2K Takipçiler
Sabitlenmiş Tweet
Ankur Parikh
Ankur Parikh@ank_parikh·
Really excited about our new work (led by @ptshaw2) on ProtEx: A Retrieval Augmented Approach to Protein Function Prediction. Gives SOTA results on several tasks (especially for rare classes and sequences far from training set) to help better characterize the protein universe!
Pete Shaw@ptshaw2

Excited to share new work from @GoogleDeepMind: “ProtEx: A Retrieval-Augmented Approach for Protein Function Prediction” biorxiv.org/content/10.110…

English
1
7
40
7.1K
Ankur Parikh retweetledi
Kevin K. Yang 楊凱筌
Kevin K. Yang 楊凱筌@KevinKaichuang·
We made FLIP2, a protein fitness benchmark spanning seven new datasets, including enzymes, protein-protein interactions, and light-sensitive proteins, as well as splits that measure generalization relevant to real-world protein engineering campaigns.
Kevin K. Yang 楊凱筌 tweet media
English
6
44
193
18.6K
James Blunt
James Blunt@JBlunt1018·
My family and I are stuck in Puerto Vallarta and was heading back home to the United States, the airport in Puerto Vallarta is also closed with the cartel violence, pray for us and our family.
English
741
535
8.8K
1.8M
Ankur Parikh retweetledi
Richard Pang
Richard Pang@yzpang_·
🚨🔔Foundational graph search task as testbed: for some distribution, transformers can learn to search (100% acc). We interpreted their algo!! But as graph size ↑, transformers struggle. Scaling up # params does not help; CoT does not help. 1.5 years of learning in 10 pages!
Richard Pang tweet media
English
2
31
117
16.7K
Ankur Parikh retweetledi
Kevin K. Yang 楊凱筌
Kevin K. Yang 楊凱筌@KevinKaichuang·
Most protein function predictors only make predictions for labels seen in training. We used LLM embeddings of text describing protein function to train ProtNote, which can generalize to new functional labels described in free text. Code is available for everybody to try!
Kevin K. Yang 楊凱筌 tweet mediaKevin K. Yang 楊凱筌 tweet mediaKevin K. Yang 楊凱筌 tweet media
English
3
46
260
43K
Ankur Parikh retweetledi
Cade Gordon
Cade Gordon@CadeGordonML·
Excited to announce our new work! 🧬 Some highlights are: - sequences likelihoods predict zero-shot fitness capabilities - a new method to calculate pLM likelihood in O(1) instead of O(L) forward passes - providing a causal between training data and outputs - suggesting a new finetuning method to improve pLM capabilities (1/9) biorxiv.org/content/10.110…
Cade Gordon tweet media
English
2
33
206
47.1K
Ankur Parikh retweetledi
Richard Pang
Richard Pang@yzpang_·
Self-rewarding LMs at #icml2024 ! Thru iterative DPO (w/ a small amount of seed data), LLM instruction following ↑ (AlpacaEval 2.0, human, MT-bench) & reward modeling ↑ (corr w human rankings). @jingxu_ml will be presenting in Vienna (Tues 7/23 11:30am); please stop by! (1/2)
Jason Weston@jaseweston

🚨New paper!🚨 Self-Rewarding LMs - LM itself provides its own rewards on own generations via LLM-as-a-Judge during Iterative DPO - Reward modeling ability improves during training rather than staying fixed ...opens the door to superhuman feedback? arxiv.org/abs/2401.10020 🧵(1/5)

English
5
11
61
7.1K
Ankur Parikh retweetledi
Kalpesh Krishna
Kalpesh Krishna@kalpeshk2011·
Check out our new @GoogleAI paper: we curate a mixture of 5M human judgments to train general-purpose foundational autoraters. Strong LLM-as-judge scores on RewardBench (87.8%), and highest perf among baselines on LLMAggreFact + 6 other benchmarks! 📰 arxiv.org/abs/2407.10817 👇
Tu Vu@tuvllms

🚨 New @GoogleDeepMind paper 🚨 We trained Foundational Large Autorater Models (FLAMe) on extensive human evaluations, achieving the best RewardBench perf. among generative models trained solely on permissive data, surpassing both GPT-4 & 4o. 📰: arxiv.org/abs/2407.10817 🧵:👇

English
6
20
124
19.8K
Ankur Parikh retweetledi
Adam Fisch
Adam Fisch@adamjfisch·
Excited to share new work from @GoogleDeepMind / @GoogleResearch on improving LLM evals using ML predictions together with a simple but effective stratified sampling approach that strategically divides the underlying data for better performance. Paper: arxiv.org/abs/2406.04291
English
4
24
120
24.5K
Ankur Parikh retweetledi
Leo Zang
Leo Zang@LeoTZ03·
ProtEx: A Retrieval-Augmented Approach for Protein Function Prediction | @GoogleDeepMind - ProtEx, combines homology-based similarity search with pre-trained model - Pretrain, Multi-sequence pretraining, span denoising and sequence similarity scoring. Finetune conditions on query and retrieved exemplar sequences to predict functional labels Preprint: biorxiv.org/content/10.110…
Leo Zang tweet mediaLeo Zang tweet media
English
0
8
29
3K
Ankur Parikh retweetledi
Anthony Gitter
Anthony Gitter@anthonygitter·
"Engineering of highly active and diverse nuclease enzymes by combining machine learning and ultra-high-throughput screening" from @countablyfinite @LucyColwell37 and team looks amazing. They compare directed evolution, hit recombination, and ML-guided design of NucB.
Anthony Gitter tweet media
English
3
28
107
12.5K