Virginia Smith

46 posts

Virginia Smith

Virginia Smith

@gingsmith

ML Professor @ CMU 🦋 gingsmith

Katılım Aralık 2011
26 Takip Edilen726 Takipçiler
Virginia Smith retweetledi
Niloofar
Niloofar@niloofar_mire·
Excited to give a talk at @SimonsInstitute Trust in Decentralized Systems Workshop on Tuesday at 11am! Title: "2026 Is the New 2016" — on federated memory, contextual privacy, and personalized agents. The privacy conversation has moved way past training data memorization. With persistent memory, tool use, and agents acting on your behalf (#clawdbot👀), the real risk is what models do with the data you feed them at inference time. Context window is the new attack surface! I'll talk about our new benchmark CIMemories, where we test whether models can actually make context-dependent decisions about what to share with whom from memory. Turns out they really can't, up to 69% violation rates, and it only gets worse the more you use them. Link to slides🔻🔻🔻
Niloofar tweet mediaNiloofar tweet mediaNiloofar tweet mediaNiloofar tweet media
English
5
13
108
6.3K
Virginia Smith retweetledi
Neil Kale
Neil Kale@neilkale·
[1/n] Open Problems in AI Child Safety AI is misused to generate CSAM at alarming scale. 400% increase in AI-generated CSAM since 2024 (IWF). 1 in 17 teens are victimized by deepfake nudes. We outline 15 open problems where AI safety research can help. 🔗aichildsafety.github.io
Neil Kale tweet media
English
1
2
15
3.5K
Virginia Smith retweetledi
Chhavi Yadav
Chhavi Yadav@chhaviyadav_·
Ending the year on a high note, with 2 papers accepted at @satml_conf 🎉 check them out! 📄 One is an interview study on cross-silo Federated Learning, highlighting a misalignment between real-world challenges & current research focus. (arxiv.org/pdf/2510.12595) 📄 The other evaluates machine unlearning methods on a harder, but realistic setting involving multi-hop knowledge. (arxiv.org/pdf/2410.15153)
Chhavi Yadav tweet mediaChhavi Yadav tweet media
English
6
5
43
5.4K
Virginia Smith retweetledi
Amrith Setlur
Amrith Setlur@setlur_amrith·
🔥 New blog on scaling RL📈 by unlocking a *new* way to explore on hard problems. - ☹️Currently scaling compute on hard problems is futile. - ☠️Turns out exploration during RL is the main bottleneck! - 👎Classical exploration methods like token entropy fail. - 💡We need a new paradigm of exploration that goes beyond simply sharpening or chaining (vg-gap) capabilities in the base model. - 🗝️The key lies in the instruction-following abilities of base models to unlock a new regime of guided exploration. More here 👇:
Aviral Kumar@aviral_kumar2

🚨🚨New blog post led by CMU students: Want to know why LLM RL training plateaus on hard problems & scaling compute may not help? And how to fix this issue? Turns out it stems from a coupling of poor exploration & optimization. Classical ways to explore don't work, but ours does! 🧵⬇️

English
2
5
27
3.3K
Virginia Smith retweetledi
Chhavi Yadav
Chhavi Yadav@chhaviyadav_·
🚀 Federated Learning (FL) promises collaboration without data sharing. While Cross-Device FL is a success and deployed widely in industry, we don’t see Cross-Silo FL (collaboration between organizations) taking off despite huge demand and interest. Why could this be the case? 🤔 We conduct an interview study to dig deeper into the real-world barriers to cross-silo FL adoption! 👇 (1/3)
Chhavi Yadav tweet mediaChhavi Yadav tweet media
English
1
12
22
4.5K
Virginia Smith retweetledi
Transactions on Machine Learning Research
As Transactions on Machine Learning Research (TMLR) grows in number of submissions, we are looking for more reviewers and action editors. Please sign up! Only one paper to review at a time and <= 6 per year, reviewers report greater satisfaction than reviewing for conferences!
Transactions on Machine Learning Research tweet media
English
2
28
65
56.4K
Virginia Smith retweetledi
Steven Kolawole
Steven Kolawole@_stevenkolawole·
1/ Can model agreement replace LLM cascades routers for efficient inference? Formalizing this, we found that under broad conditions, ensemble provides provably safe, training-free routing rules. We call the method *Agreement-Based Cascading (ABC)*, and it's live now at TMLR! 🧵
English
1
15
35
2.3K
Virginia Smith retweetledi
Pratiksha Thaker
Pratiksha Thaker@prthaker_·
I'm very excited to share some new work arxiv.org/abs/2506.06488. This work started out in conversations with @thorn where we realized that shadow model MIAs couldn't be used to audit models for harmful content of children. See 🧵 for why, and our progress on solving this...
English
1
7
25
4.3K
Virginia Smith retweetledi
ICML Conference
ICML Conference@icmlconf·
ICML offers an optional poster printing service icml.myprintdesk.net Orders can be picked up the day at the Vancouver Convention Centre in West MR 104 during the following hours: Monday - Friday: 7:30 am - 5:00 pm Saturday: 8:00 am - 1:00 pm
English
0
4
13
8.1K
Virginia Smith retweetledi
Matthew Yang
Matthew Yang@_matthewyang·
🚨 NEW PAPER: What if LLMs could tackle harder problems - not by explicitly training on longer traces, but by learning how to think longer? Our recipe e3 teaches models to explore in-context, enabling LLMs to unlock longer reasoning chains without ever seeing them in training. 🤯 Website: matthewyryang.github.io/e3/ Paper: arxiv.org/abs/2506.09026 🧵[1/8]
English
1
3
10
928
Virginia Smith retweetledi
Ameet Talwalkar
Ameet Talwalkar@atalwalkar·
I’m excited to share new work from Datadog AI Research! We just released Toto, a new SOTA (by a wide margin!) time series foundation model, and BOOM, the largest benchmark of observability metrics. Both are available under the Apache 2.0 license. 🧵
Ameet Talwalkar tweet media
English
5
51
241
38.2K
Virginia Smith retweetledi
ICML Conference
ICML Conference@icmlconf·
Invited talked are announced. icml.cc/virtual/2025/e… Jon Kleinberg Pamela Samuelson Frauke Kreuter Anca Dragan Andreas Krause
Norsk
1
5
63
16.4K
Virginia Smith retweetledi
Aashiq Muhamed
Aashiq Muhamed@AashiqMuhamed·
Thrilled to share our new work on improving LLM unlearning! 🚀 Gradient-based unlearning struggle with high cost, instability & lack of precision. We introduce Dynamic SAE Guardrails (DSG): an activation-based approach using SAEs for targeted, efficient knowledge removal.
Aashiq Muhamed tweet media
English
4
8
46
6.6K
Virginia Smith retweetledi
Andrew Gordon Wilson
Andrew Gordon Wilson@andrewgwils·
The ICML 2025 workshops list is online! icml.cc/virtual/2025/e…. Many exciting topics, spanning multi-agent systems, world models, test-time adaptation, actionable interpretability, and much more.
English
4
17
115
13.9K
Virginia Smith retweetledi
Amrith Setlur
Amrith Setlur@setlur_amrith·
How to effectively unlearn finetuning data? ❌ Approx. methods leak sensitive data ✅ Exact unlearning (eg. retraining) is secure 🔒 but inefficient 🚨 New paper: *efficient* & *exact* unlearning (led by Kevin) 🗝️ Idea: model merging at scale arxiv.org/pdf/2504.04626 🧵⤵️
Amrith Setlur tweet media
English
1
10
54
4K
Virginia Smith
Virginia Smith@gingsmith·
@tin_ng_qn @icmlconf The deadline for the initial author response is Mar 31 AoE (please see here for details #discussions" target="_blank" rel="nofollow noopener">icml.cc/Conferences/20…).
English
0
0
1
84
Virginia Smith
Virginia Smith@gingsmith·
The Simons Institute Research Fellowship provides a salary & benefits for the duration of the program, along with housing and visa support. The fellowship is intended for junior researchers who received their PhD at most 5 years ago.
English
1
0
3
1.4K
Virginia Smith
Virginia Smith@gingsmith·
Are you a junior researcher interested in federated and collaborative learning? Consider applying to the Simons Institute Research Fellowship (due April 1) to participate in the Spring 2026 FL/CL Program at the Simons Institute at UC Berkeley! details below 👇
English
1
11
67
13.2K