Yining Lu

138 posts

Yining Lu banner
Yining Lu

Yining Lu

@Yining__Lu

Second year CS PhD student @NotreDame | Prev: @amazon @JHUCLSP 🦋: https://t.co/gPXvdPuesy

South Bend, IN Katılım Aralık 2019
591 Takip Edilen232 Takipçiler
Sabitlenmiş Tweet
Yining Lu
Yining Lu@Yining__Lu·
Thrilled to share that I'll start my Ph.D. at @ND_CSE this fall, working with @Meng_CS. I am so grateful for the sincere guidance from my current advisor, @DanielKhashabi, and for the unconditional support I received from my family, friends, and collaborators over the past years!
English
9
3
51
7.3K
Yining Lu retweetledi
Daniel Khashabi 🕊️
Daniel Khashabi 🕊️@DanielKhashabi·
LLMs continue to struggle with long-context tasks—such as needle-in-a-haystack problems—because of “positional bias.” What can we do if we only have 𝘣𝘭𝘢𝘤𝘬-𝘣𝘰𝘹 access to the model? (i.e., we can’t modify the model weights or attention patterns, as is often the case with API models.) We introduce ⭐𝐆𝐨𝐥𝐝-𝐏𝐚𝐧𝐧𝐢𝐧𝐠⭐, a black-box Bayesian framework that, at inference time, strategically and iteratively shuffles documents to overcome positional bias. Specifically, it searches over long contexts by (i) reordering documents to concentrate high-belief items in highly “diagnostic” positions, and (ii) updating beliefs about document relevance from model outputs. We show that GP provably identifies a target among N documents in O(log N) rounds, ensuring scalability to many-document settings. More in the paper: arxiv.org/pdf/2510.09770
Daniel Khashabi 🕊️ tweet media
English
2
19
73
6.3K
Yining Lu
Yining Lu@Yining__Lu·
Our method, CTWA, effectively mitigates cross-objective interference compared to others. Competing methods either quickly sacrifice accuracy to achieve superficially high conciseness and clarity (e.g., GradNorm in 1a, Linear and Dynamic weighting in 1b), or trying to maintain high accuracy while overlooking the improvement of others (e.g., Lagrangian in 1a and PAMA in 1b). In contrast, CTWA achieves strong, balanced performance across all three objectives.
Yining Lu tweet media
English
1
0
0
58
Yining Lu
Yining Lu@Yining__Lu·
Why does improving one objective in multi-objective RL sometimes hurt others, even when they shouldn't conflict from the MOO and MTL sense? And why only on certain models? 🤔 This isn't Pareto tradeoffs. It's 𝐜𝐫𝐨𝐬𝐬-𝐨𝐛𝐣𝐞𝐜𝐭𝐢𝐯𝐞 𝐢𝐧𝐭𝐞𝐫𝐟𝐞𝐫𝐞𝐧𝐜𝐞 in LLM alignment. We uncover this in arxiv.org/abs/2602.06869 by - answering why and when this model-dependent interference actually happens. - conducting the first systematic benchmark of classic MOO/MTL methods for LLM alignment. - proposing CTWA (Covariance-Targeted Weight Adaptation), our solution that actually works across the board. We hope this work could provide actionable insights for multi-objective alignment for LLMs. Details in the thread below 👇
Yining Lu tweet media
English
3
3
8
658
Yining Lu retweetledi
Zilong Wang
Zilong Wang@zlwang_cs·
Awesome work! 🧠 GDPO solves the signal collapse issue beautifully. Our Dynamic Reward Weighting tackles the optimization side: since objectives vary in difficulty, static weights often waste training effort or get stuck. Both findings highlight that simple linear scalarization is just not enough for complex alignment. arxiv.org/pdf/2509.11452
Shizhe Diao@shizhediao

RLVR is powerful — but how do you train with multiple rewards effectively? 🤔 🎯GDPO (not GRPO) is coming. We introduce Group reward-Decoupled Normalization Policy Optimization (GDPO), a new multi-reward RL algorithm that consistently improves per-reward convergence over GRPO across a wide range of tasks. (1/n)

English
2
12
99
9.3K
Yining Lu
Yining Lu@Yining__Lu·
Thanks for the question! In real-time queries, blockchain interactions are limited to batched reads, which are similar to querying data from a conventional service. As a result, they have minimal impact on the RAG process when generating responses. For time-consuming blockchain updates, we plan to handle them asynchronously in batches to improve the user experience.
English
0
0
0
36
Karan Jagtiani
Karan Jagtiani@karanjagtiani04·
@Yining__Lu Nice to see blockchain used for data reliability in RAG. How do you handle latency impacts in real-time queries?
English
1
0
0
12
Yining Lu
Yining Lu@Yining__Lu·
📣 My first system paper 📣 We built a decentralized RAG system that solves data reliability challenges in real-world settings. The sources provided by each data owner will be securely managed and scored on the blockchain. 🚀 one-line command to deploy: github.com/yining610/Reli…
Meng Jiang@Meng_CS

Decentralized RAG allows your database to benefit all LLM clients. On the other side, not all data sources are reliable. Managing source reliability on blockchain can avoid third-party manipulation. Introducing dRAG + Blockchain + Truth Discovery: arxiv.org/abs/2511.07577

English
2
1
4
542
Yining Lu retweetledi
Meng Jiang
Meng Jiang@Meng_CS·
Decentralized RAG allows your database to benefit all LLM clients. On the other side, not all data sources are reliable. Managing source reliability on blockchain can avoid third-party manipulation. Introducing dRAG + Blockchain + Truth Discovery: arxiv.org/abs/2511.07577
English
0
1
7
1.3K
Jack Jingyu Zhang
Jack Jingyu Zhang@jackjingyuzhang·
I’m super thrilled and honored to be named an Amazon AI PhD Fellow 💫 Huge thanks to @AmazonScience for generously supporting our research at JHU! We’ll be advancing AI alignment in collaboration with folks at Amazon.
Rohit Prasad@RohitPrasadAI

Excited to announce @amazon's new AI PhD Fellowship Program supporting 100+ students across 9 universities like Carnegie Mellon, MIT & Stanford. Fellows will be paired with senior scientists working in related fields, plus receive financial support and AWS credits for research. Learn more: amazon.science/news/amazon-la…

English
17
9
92
11.7K
Yining Lu retweetledi
Zhengping Jiang
Zhengping Jiang@zhengping_jiang·
🚀 Excited to share our COLM25 paper, “Always Tell Me The Odds”! You can train an LLM to be a generic subjective probability estimator — e.g., producing fine-grained probabilities for everyday situations that go far beyond coarse confidence scores.
Liaoyaqi Wang@LiaoyaqiW

Title: Always Tell Me The Odds: Fine-grained Conditional Probability Estimation 📄Paper: openreview.net/pdf?id=xhDcG8q… 🤗Models: huggingface.co/Zhengping/cond… 💻Code: github.com/zipJiang/decod… 🤝Coauthor: @zhengping_jiang @anqi_liu33 @ben_vandurme

English
0
1
3
494
Yining Lu retweetledi
Arda Uzunoğlu
Arda Uzunoğlu@aardauzunoglu·
🛑 What's the Flaw of Averages? 📄: arxiv.org/abs/2509.25671 We’re in an evaluation crisis. Benchmarks are saturating, creating a false sense that tasks are solved. As training/eval chase these sets, plateaued averages hide shortcutting and distributional skew. 🧵1/7
Arda Uzunoğlu tweet media
English
1
15
42
12.4K
Yining Lu
Yining Lu@Yining__Lu·
Very insightful blog! Some claims like "correct trade-off between the conflicting objectives at different states is precisely given by the distribution" were also found in our recent work (arxiv.org/pdf/2509.11452, limitation section) that different pre-trained models exhibit distinct tolerance to trade-offs between conflicting objectives.
Nan Jiang@nanjiang_cs

My 3rd blogpost on PG, the topic I am least familiar with but get asked a lot, so I thought I'd just put together the very limited stuff I know on this topic. Somehow the post gets cynical from time to time🙃 nanjiang.cs.illinois.edu/2025/09/29/pg.…

English
0
1
4
1.1K