Sun Fei

101 posts

Sun Fei

Sun Fei

@fei__sun

Associate Professor @UCAS1978, working on RecSys, NLP, AI safety. My tweets are my own.

Katılım Kasım 2009
342 Takip Edilen209 Takipçiler
Sun Fei
Sun Fei@fei__sun·
@gargighosh Hi! Is there still space available for the ICLR networking event? I’d love to register if spots are still open. Thanks!
English
0
0
0
26
Gargi Ghosh
Gargi Ghosh@gargighosh·
Few of us from Meta SuperIntelligence lab will attend ICLR this year- Happy to chat in person. If you are interested in joining the networking mixer, please register here- events.atmeta.com/iclrnetworking…
English
3
4
68
8.4K
Yi R. (May) Fung
Yi R. (May) Fung@May_F1_·
Excited to share that we have 7 papers accepted into @aclmeeting! Excited to chat more about what's next with our research findings in San Diego this summer~ 🎉🎉
Yi R. (May) Fung tweet media
English
2
1
37
1.6K
Sun Fei
Sun Fei@fei__sun·
@paria_rd Very cool work! The curvature-based perspective on safe editing is both intuitive and principled. As the author of WILD and LocFT, I’m glad to see WILD evaluation used and LocFT-BF compared. Appreciate the engagement!
English
0
0
2
60
Paria Rashidinejad
Paria Rashidinejad@paria_rd·
LLMs go stale daily: facts shift, discoveries land, hallucinations are uncovered. How do you continually keep up with knowledge drift without retraining? Our new work, CrispEdit, lets you apply 𝘁𝗵𝗼𝘂𝘀𝗮𝗻𝗱𝘀 𝗼𝗳 𝗲𝗱𝗶𝘁𝘀 to billion-parameter LLMs in 𝗷𝘂𝘀𝘁 𝗮 𝗳𝗲𝘄 𝗺𝗶𝗻𝘂𝘁𝗲𝘀 𝗼𝗻 𝗮 𝘀𝗶𝗻𝗴𝗹𝗲 𝗚𝗣𝗨, while keeping the model’s existing capabilities intact. That’s >𝟭𝟬𝟬𝘅 𝗳𝗮𝘀𝘁𝗲𝗿 than popular editors like AlphaEdit and MEMIT. 💡𝗖𝗼𝗿𝗲 𝗶𝗱𝗲𝗮: The landscape of existing capabilities is sharp in a few directions and flat in many others, so we apply edits only in the low-curvature subspace, where updates are “safe”. ✅ This avoids paying for full retraining and mitigates capability degradation and forgetting in existing editors. 𝗥𝗲𝘀𝘂𝗹𝘁𝘀: • 𝗛𝗶𝗴𝗵 𝗲𝗱𝗶𝘁 𝘀𝘂𝗰𝗰𝗲𝘀𝘀: +10% over best baselines under the real 𝘢𝘶𝘵𝘰𝘳𝘦𝘨𝘳𝘦𝘴𝘴𝘪𝘷𝘦 𝘨𝘦𝘯𝘦𝘳𝘢𝘵𝘪𝘰𝘯 (WILD), not just teacher-forced evaluation. • 𝗖𝗮𝗽𝗮𝗯𝗶𝗹𝗶𝘁𝗶𝗲𝘀 𝗶𝗻𝘁𝗮𝗰𝘁: <1% drop on average. • 𝗙𝗮𝘀𝘁: 3,000 edits on Llama-3-8B in <5 minutes on a single NVIDIA A40. • 𝗦𝗲𝗾𝘂𝗲𝗻𝘁𝗶𝗮𝗹 𝘂𝗽𝗱𝗮𝘁𝗲𝘀: Sequential CrispEdit effectively maintains both the capabilities and previous edits. 📝 arxiv.org/pdf/2602.15823
Paria Rashidinejad tweet media
English
2
5
19
1.8K
Sun Fei
Sun Fei@fei__sun·
🚀 Our paper “Fine-Tuning Done Right in Model Editing” is accepted to @iclr_conf 2026! Huge congrats to @10k_miles_yang Fine-tuning is NOT weak for model editing. The pipeline was suboptimal. Implementation matters more than we thought. We show that standard breadth-first fine-tuning + localized tuning beats SOTA, scales to 100K edits and 72B models. 📄 Paper: arxiv.org/abs/2509.22072 💻 Code: github.com/WanliYoung/FT4… #ModelEditing #KnowledgeEditing #finetuning #ICLR2026
Sun Fei tweet mediaSun Fei tweet mediaSun Fei tweet media
English
2
0
4
190
Sun Fei
Sun Fei@fei__sun·
This is @10k_miles_yang’s 4th paper so far, in his 2nd year of PhD. Really impressive progress.
English
0
0
0
54
Sun Fei
Sun Fei@fei__sun·
🎉 Our paper "AuditAgent: LLM Agent for Risks Auditing in Recommender Systems" has been accepted to the AAAI 2026 Demo Track! AuditAgent uses LLM-powered GUI agents to audit platforms for filter bubbles, unfairness, and data misuse. #AAAI2026 #LLM #guiagent #algorithm_auditing #GUIAgent
English
0
0
9
1.3K
Sun Fei
Sun Fei@fei__sun·
THX for sharing our work!😀
Rohan Paul@rohanpaul_ai

Another great @GoogleDeepMind paper. Shows how to speed up LLM agents while cutting cost and keeping answers unchanged. 30% lower total cost and 60% less wasted cost at comparable acceleration. Agents plan step by step, so each call waits for the previous one, which drags latency. Speculative planning fixes that by having a cheap draft agent guess next steps while a stronger agent checks them in parallel. Fixed guess lengths backfire, small guesses barely help, big guesses waste tokens when a check disagrees. Dynamic Speculative Planning learns how far to guess, then stops early to avoid wasted calls. A tiny online predictor learns how many steps will be right using reinforcement learning. 1 knob lets teams bias for speed or cost, either by skewing training or adding a small offset. If a guess is wrong, extra threads stop and execution resumes from the verified step. Across OpenAGI and TravelPlanner, the dynamic policy matches the fastest fixed policy while spending fewer tokens The result is clear, faster responses, lower bills, and 0 loss in task quality.

English
0
0
0
95
Sun Fei
Sun Fei@fei__sun·
If you sample an LLM 15× and get the same answer, should you trust it? 🚨Not always! ⚠️ Consistency ≠ correctness. 🚫Scaling doesn’t help: self-consistent errors stay stable—or even increase—in larger models. Our #EMNLP25 main paper (Too Consistent to Detect: A Study of Self-Consistent Errors in LLMs) formalizes self-consistent errors, shows mainstream detectors miss them, and introduces a simple cross-model probe that consistently improves AUROC across datasets and model families. preprint: arxiv.org/abs/2505.17656 #Hallucination #LLM_uncertainty #HallucinationDetection #LLM
Sun Fei tweet media
English
0
0
4
348
Sun Fei
Sun Fei@fei__sun·
🛑 Stop using teacher forcing to evaluate model editing! Our ACL 2025 poster shows why past evaluations mislead progress & how to test editing in the wild. 📍 July 30, 11:00 AM – come chat! #ModelEditing #LLM #ACL2025NLP
Sun Fei tweet media
Wanli Yang@10k_miles_yang

🎉 Excited to share that our work "The Mirage of Model Editing" has been accepted as a main conference paper at #ACL2025! Many thanks to my supervisor @fei__sun, our collaborators, and special thanks to @HuaWenyue31539 for insightful discussions!

English
0
1
3
630
Mark Dredze
Mark Dredze@mdredze·
Unclear if I should be honored or insulted.
Mark Dredze tweet media
English
3
0
57
4.7K
Sun Fei
Sun Fei@fei__sun·
Great post! Thanks for featuring the figure from our #ACL2025 paper “The Mirage of Model Editing.”arxiv.org/abs/2502.11177 Glad to see the community's attention to evaluation challenges in knowledge editing. #LLMs #ModelEditing #KnowledgeEditing
Yunzhi Yao@yyzTodd

🚨 New Blog Drop! 🚀 "Reflection on Knowledge Editing: Charting the Next Steps" is live! 💡 Ever wondered why knowledge editing in LLMs still feels more like a lab experiment than a real-world solution? In this post, we dive deep into where the research is thriving — and where it's falling short. From foundational breakthroughs to the practical roadblocks no one’s talking about, we connect the dots and propose what’s needed to move forward. Join the conversation! #KnowledgeEditing #LLMs #AI #ModelEditing 📌 If you're working on LLMs, model updates, or mechanism interpretability, you don’t want to miss this. 👉 Read the full post: yyzcowtodd.cn/rethinkedit Key insights from our analysis: 0⃣ Current evaluation metrics and benchmarks inadequately assess knowledge updates in LRMs, highlighting the need for more comprehensive evaluation frameworks. 1⃣ Scaling challenges persist, with significant memory and computational constraints limiting the practical application of editing methods for larger or quantized local models. 🎁 Resource Release: To support the research community, we release covariance matrices for Qwen2.5-32B & QwQ-32B models for the current locate-and-edit methods. 2⃣ We outline promising research directions for developing language models that can effectively learn, adapt, and evolve their knowledge base. Huge thanks to the brilliant collaborators who made this deep dive into #ModelEditing possible! @uclanlp @CanyuChen3 @Jiachen_Gu @dsmall2apple1 @ManlingLi_ @VioletNPeng

English
1
0
2
364
Sun Fei
Sun Fei@fei__sun·
@AixinSG 哈哈哈哈,我平时不用。今天nips模版有一个小问题,用了一个地方
中文
0
0
0
23
Sun Fei
Sun Fei@fei__sun·
all u need is vspace😂
Sun Fei tweet media
English
0
0
1
614
Sun Fei retweetledi
Wanli Yang
Wanli Yang@10k_miles_yang·
🎉 Excited to share that our work "The Mirage of Model Editing" has been accepted as a main conference paper at #ACL2025! Many thanks to my supervisor @fei__sun, our collaborators, and special thanks to @HuaWenyue31539 for insightful discussions!
Wanli Yang tweet media
English
4
4
10
1.7K