Angelica Chen

131 posts

Angelica Chen

Angelica Chen

@_angie_chen

Gemini training @GoogleDeepMind, PhD from @NYUDataScience, previously @Princeton 🐅, angie-chen at 🦋

New York, NY Katılım Şubat 2016
484 Takip Edilen1.7K Takipçiler
Kyunghyun Cho
Kyunghyun Cho@kchonyc·
please @_angie_chen fix it natively please
Kyunghyun Cho@kchonyc

@Google can't figure out how to remove "[cite_start]" issue on their own gemini web app, and thereby i wrote a browser extension to fix it myself ... 🤦 let me present you anticitestart extension; fully open sourced and implemented by antigravity. expecting a zuck-level offer from google anytime now.

English
3
1
13
6.7K
Angelica Chen retweetledi
Lavender Jiang
Lavender Jiang@lavenderjiang99·
We built Lang1: a 100M–7B family of models specialized for hospital operations. After finetuning, Lang1-1B outperforms generalist models up to 671B, transfers to unseen tasks and another hospital, and is more data-efficient. 🧵Why small specialists win in healthcare.
Lavender Jiang tweet media
English
1
8
29
4.7K
Angelica Chen retweetledi
Richard Pang
Richard Pang@yzpang_·
🚨Prompt Curriculum Learning (PCL) - Efficient LLM RL training algo! - We investigate factors that affect convergence: bsz, # prompt, # gen, prompt selection - We propose PCL: lightweight algo that *dynamically selects intermediate-difficulty prompts* using a learned value model
Richard Pang tweet media
English
2
35
170
24.2K
Angelica Chen
Angelica Chen@_angie_chen·
Working with Sadhika was one of the biggest highlights of my PhD and I couldn't be more excited to see her mentor the next generation of brilliant scientists! 🌟💫 Apply to work with her at UCSD!
Sadhika Malladi@SadhikaMalladi

Excited to share that I will be starting as an Assistant Professor in CSE at UCSD (@ucsd_cse) in Fall 2026! I am currently recruiting PhD students who want to bridge theory and practice in deep learning - see here: cs.princeton.edu/~smalladi/recr…

English
0
0
8
1.1K
Kyunghyun Cho
Kyunghyun Cho@kchonyc·
it is amazing to have @tallinzen as a colleague as he inspires so many people including myself and gemini 2.5 pro. here's the new position paper on Cardiology of Language Models written almost entirely by Gemini 2.5 Pro with my promp ... nope ... supervision. (and thanks to @gregd_nlp for motivation!)
Kyunghyun Cho tweet mediaKyunghyun Cho tweet media
Tal Linzen@tallinzen

introducing our new interpretability research paradigm, Cardiology of Language Models! it is based on a method we call the "stethoscope", where we train a linear classifier to discriminate between the LLM hidden states that represent a concept and those that do not!

English
9
12
98
23.4K
Xiao Ma
Xiao Ma@infoxiao·
I've been automating parts of my research workflow. Sometimes amazing, sometimes I'm the engineer spending 3 hours to automate a 30-minute task. I think there might be a useful framework here: Context × Action ✅ Potential Sweet Spot (Low Context + High Action): "Run 10 evals across 20 checkpoints" → Clear task, lots of repetitive work ❌ Where It Seems to Break (High Context + Low Action): "Book me a flight but I prefer aisle seats, have Delta status, need to coordinate with friends..." Takes longer to explain than to just book it yourself. Maybe this explains why so many AI features feel useless? The magic might happen when explanation is minimal but the task list is long.
Xiao Ma tweet media
English
18
55
575
70.5K
yi
yi@agihippo·
Happy birthday @vqctran
English
3
0
6
2.5K
Angelica Chen retweetledi
Nathan C. Frey
Nathan C. Frey@nc_frey·
LLMs vs Protein Design Tools: Who Wins? if you missed them @icmlconf, here's the poster for work led by @_angie_chen & @samuel_stanton_ that answers this question! Link to paper below 👇
Nathan C. Frey tweet media
English
2
9
48
5.2K
Angelica Chen retweetledi
ML Collective
ML Collective@ml_collective·
Starting in 30 min at 10am PT! @nsaphra presents at DLCT: Sudden Drops in the Loss: Syntax Acquisition, Phase Transitions, and Simplicity Bias in MLMs Angelica Chen, Ravid Shwartz-Ziv, Kyunghyun Cho, Matthew L. Leavitt, Naomi Saphra arxiv.org/abs/2309.07311 Zoom 👇
ML Collective tweet media
English
3
3
12
2.1K
Angelica Chen
Angelica Chen@_angie_chen·
We propose a new synthetic benchmark for highly-constrained biophysical optimization tasks, and propose a new semi-online RL algorithm for training an LLM to iteratively improve its own generations while using very few labels. Come chat with me and @samuel_stanton_ !
English
1
0
2
400
Angelica Chen
Angelica Chen@_angie_chen·
Have you ever wondered how your specialized biomolecule engineering model compares with a general purpose LLM on bio-optimization tasks? Come check out our work at ICML, in the East poster session (#E-2804) happening right now!. #ICML2025
Angelica Chen tweet media
English
1
3
31
6.1K
Kyunghyun Cho
Kyunghyun Cho@kchonyc·
Gemini team, can you please fix your rendering to remove "[cite_start]" everywhere? i need to copy-paste the result on a new chat to have Gemini clean up these unnecessary and wrong markers. @orf_bnw get to work, please
English
5
0
25
7K
Angelica Chen retweetledi
Jason Weston
Jason Weston@jaseweston·
🌉 Bridging Offline & Online RL for LLMs 🌉 📝: arxiv.org/abs/2506.21495 New paper shows on verifiable & non-verifiable tasks: - Online DPO & GRPO give similar performance. - Semi-online (iterative) DPO with sync every s steps (more efficient!) works very well also. - Offline DPO is way behind. - Combining verifiable + non-verifiable works! Cross-transfer gains. - Recipes for how to make this work. 🧵1/4
Jason Weston tweet media
English
1
94
452
68.2K
Angelica Chen retweetledi
Aran Komatsuzaki
Aran Komatsuzaki@arankomatsuzaki·
Bridging Offline and Online Reinforcement Learning for LLMs Investigates the effectiveness of RL for finetuning LLMs when transitioning from offline to semi-online to fully online regimes for both verifiable and nonverifiable tasks.
Aran Komatsuzaki tweet media
English
3
42
285
25.9K