Hadi Amiri

749 posts

Hadi Amiri banner
Hadi Amiri

Hadi Amiri

@amirieb

Assistant Professor of Computer Science at UMass Lowell. Language + Healthcare. Previously at MIT, Harvard, UMD College Park, NUS, and University of Tehran.

Boston, MA Katılım Haziran 2011
277 Takip Edilen393 Takipçiler
Hadi Amiri retweetledi
sijia.liu
sijia.liu@sijialiu17·
See our MSU–IBM joint work on reasoning model “unlearning” presented at @emnlpmeeting, showcasing how unlearning can be leveraged to sanitize sensitive reasoning traces and enhance model safety.
Changsheng Wang @ NeurIPS@wcsa23187

🎯 Our EMNLP 2025 Main paper “Reasoning Model Unlearning: Forgetting Traces, Not Just Answers, While Preserving Reasoning Skills” goes live soon! Catch us on Wednesday in Suzhou at #EMNLP2025 🇨🇳 🔗 Paper Link: arxiv.org/pdf/2506.12963 🏡 Project Page Link: r2mu.netlify.app 🗓 November 5, 11:00–12:30 CST (UTC+8) 📍 Hall C, Section 2, 500-Main 🧍 I won’t be there in person — but feel free to chat with my co-authors! 🧠 The Problem You’ve erased sensitive answers from your LRM. But the reasoning traces, the step-by-step “thoughts” that led there, still remain. Even after unlearning, the model can reconstruct or re-infer forgotten answers through these traces. So the question is: 👉 Can we truly forget reasoning traces, while preserving the model’s reasoning ability? 🎯 Our Solution: R²MU (Reasoning-aware Representation Misdirection for Unlearning) We go beyond answer-level forgetting and target the reasoning process itself. R²MU suppresses sensitive reasoning traces while maintaining general reasoning competence. Through representation misdirection, the model unthinks unsafe reasoning paths, while CoT supervision preserves valid reasoning skills. ⚙️ How it Works 🔄 Unthinking Loss: misaligns hidden representations of sensitive reasoning traces with randomized features. 💡 Reasoning Preservation: uses CoT datasets (like LIMO) to retain problem-solving ability. ✅ R²MU erases reasoning traces — not just answers. ✅ Preserves general reasoning and utility across diverse benchmarks. ✅ Achieves the lowest reasoning-trace leakage (RT-UA ↓) on unlearning benchmark WMDP and LRM safety benchmark STAR-1, while maintaining top reasoning accuracy on AIME, MATH-500, and GPQA. 👥 With amazing collaborators from MSU: @ChongyuFan ,@zyh2022 , @jia_jinghan , and my advisor @sijialiu17 . 🙏 Grateful to our IBM collaborators from @MITIBMLab : @NathalieBaraca1 , Dennis Wei, @p_ram_p.

English
0
1
9
593
Hadi Amiri retweetledi
uclanlp
uclanlp@uclanlp·
We’ve been running the UCLA NLP Seminar for a while now and realized it’s a waste not to share these amazing talks more broadly. So here’s our YouTube channel now! 🎥 Watch and subscribe to our channel for past and upcoming sessions: 👉 @uclanlp-plus" target="_blank" rel="nofollow noopener">youtube.com/@uclanlp-plus #AI #UCLANLP
English
3
21
105
20.8K
Hadi Amiri retweetledi
Dan Jurafsky
Dan Jurafsky@jurafsky·
Now that school is starting for lots of folks, it's time for a new release of Speech and Language Processing! Jim and I added all sorts of material for the August 2025 release! With slides to match! Check it out here: web.stanford.edu/~jurafsky/slp3/
English
9
70
400
34.7K
Hadi Amiri
Hadi Amiri@amirieb·
EqualizeIR: Mitigating Linguistic Biases in Retrieval Models with @jialicheng123. We show that IR models are biased by the ling. complexity of queries, and introduce 4 types of weak learners to build robust retrievers. shorturl.at/tsw5D #NLProc #NAACL2025
English
0
0
4
327
Hadi Amiri retweetledi
Jessy Li
Jessy Li@jessyjli·
Thank you so much @amirieb for hosting, amazing visit 🤗 New papers discussed: QUDsim (discourse similarity): arxiv.org/abs/2504.09373 Information salience in LLMs: arxiv.org/abs/2502.14613
Hadi Amiri@amirieb

This morning we had the pleasure of learning from @jessyjli about how LLMs model curiosity, discourse coherence, and information salience.👉 ow.ly/fZL250BGU4M #Linguistics #Discourse #NLP #LLMs @KCSciences_UML @UT_Linguistics @UTAustin

English
1
1
12
1.5K
Hadi Amiri
Hadi Amiri@amirieb·
We were honored to host Prof. Jiawei Han for an inspiring talk on integrating knowledge graphs with LLMs to accelerate theme-driven scientific discovery. (4/5, 11AM) @dmguiuc @KCSciences_UML 👉 ow.ly/fZL250BGU4M
English
0
0
2
118
sijia.liu
sijia.liu@sijialiu17·
Honored to receive the prestigious Withrow Rising Scholar Award. Grateful for the unwavering support from my students, advisors, collaborators, nominators, and recommenders. Excited to keep bridging foundational research and real-world impact to advance trustworthy and scalable ML! @MSU_EGR @OptML_MSU
MSU College of Engineering@MSU_EGR

Sijia Liu, Assistant Professor, Computer Science and Engineering. “Prof. Liu is one of the rare few research scholars able to span very theoretical work to very practical and empirical work, bringing forth a novel perspective that galvanizes a community.” spr.ly/60180SXYj

English
7
1
45
3.6K
Hadi Amiri retweetledi
Google AI
Google AI@GoogleAI·
Introducing an advanced Articulate Medical Intelligence Explorer (AMIE), which goes beyond diagnosis towards treating and managing disease over time, matching clinician performance on multi-visit consultations in a study with patient actors. Learn more at goo.gle/3Flpbdy
English
144
458
2K
236.1K
Hadi Amiri retweetledi
Yoo Yeon Sung@ACL2025
Yoo Yeon Sung@ACL2025@YooYeonSung1·
Thrilled to share that AdvScore paper has been accepted to NAACL Main 🚀! Looking forward to pushing forward human-centered model evaluation and benchmark creation. Huge thanks to my amazing collaborators! 🎈🌵🏜️ #NAACL2025
Yoo Yeon Sung@ACL2025@YooYeonSung1

Our new paper introduces ADVSCORE, a novel human-grounded metric that evaluates the adversarial robustness of benchmarks over time. 📜arxiv.org/abs/2406.16342 🧵👇

English
1
7
40
11.4K