Sophia Hager

14 posts

Sophia Hager

Sophia Hager

@SophiaNLP

PhD Student at @jhuclsp

Katılım Aralık 2025
25 Takip Edilen16 Takipçiler
Sophia Hager
Sophia Hager@SophiaNLP·
Artificial uncertainty has the potential to ensure we can keep AI reliable and interpretable without worrying that calibration data will be memorized in the next wave of models. Big thanks to my collaborators, Simon Zeng and Nick Andrews! Read the paper: arxiv.org/pdf/2605.13595
English
0
0
0
19
Sophia Hager
Sophia Hager@SophiaNLP·
(This is not the case when introducing data uncertainty through ambiguity; while it's trivial to construct questions that the model gets wrong with data uncertainty, any improvements in calibration are much more inconsistent than inducing artificial model uncertainty.)
Sophia Hager tweet media
English
1
0
0
23
Sophia Hager
Sophia Hager@SophiaNLP·
Can we learn to recognize artificial uncertainty as a proxy for real uncertainty? As LLMs memorize more of the internet, they become correctly confident on almost any existing question you can throw at them. Creating new challenging calibration data is unsustainably expensive.🧵
Sophia Hager tweet media
English
1
3
4
404
Sophia Hager retweetledi
Mark Dredze
Mark Dredze@mdredze·
Apparently, my students have brought binoculars to their office to bird watch during the day. The advantages of our new beautiful @HopkinsDSAI office space. 🦜🦆🕊️ Should I be worried about productivity?
Mark Dredze tweet media
English
5
10
66
8.1K
Sophia Hager retweetledi
Rohan Jha
Rohan Jha@Robro612·
New 📄: we replicate XTR, a multi-vector retrieval method that makes ColBERT faster by avoiding its expensive step of gathering full document embeddings XTR is not a free lunch over ColBERT, but its training objective is useful for modern efficient engines like PLAID and WARP 👇🏼
English
1
25
63
11.4K
Sophia Hager retweetledi
Drew Prinster
Drew Prinster@DrewPrinster·
Can we ensure AI agents respect our safety constraints, even as they explore & improve? - Medical LLMs that are helpful, & avoid false claims? - Bioscience agents that generate effective molecule designs, & ensure they’re safe? 📄🧵w/ @samuel_stanton_ @clara_fannjiang @jiwoncpark @kchonyc @anqi_liu33 @suchisaria Excited to share “Conformal Policy Control” ⬇️ 1/12
GIF
English
1
25
68
21.6K
Sophia Hager retweetledi
Jack Jingyu Zhang
Jack Jingyu Zhang@jackjingyuzhang·
Real-world agents juggle instructions from skill files, tools, other agents, ... each with different trust levels. When these conflict, can models reliably prioritize the most trusted one? Our ManyIH-Bench🪜 finds that even frontier models like GPT-5.4 only get ~40% accuracy! 👇
Jack Jingyu Zhang tweet media
English
1
31
120
11.8K
Sophia Hager retweetledi
arXiv Sound
arXiv Sound@ArxivSound·
``Generating Music with Structure Using Self-Similarity as Attention,'' Sophia Hager, Kathleen Hablutzel, Katherine Kinnaird, ift.tt/FNwGvzx
English
0
2
10
935
Sophia Hager retweetledi
Marc Marone
Marc Marone@ruyimarone·
I'm on the job market and at #neurips2025! Looking for research roles around data for foundation models and would love to chat with folks - resume/site in my bio. I've recently worked @AIatMeta and @databricks and publish papers with my awesome collaborators @jhuclsp!
English
4
18
49
10.7K
Sophia Hager retweetledi
Andrew Wang
Andrew Wang@andrewwnlp·
Tools break in the real world all the time, but not much attention has been given to how well LLMs deal with tool failures. We introduce HOHW, a tool-use benchmark where problems remain solvable even when tools break adversarially.
Andrew Wang tweet media
English
1
13
24
2.4K