UIUC TIMAN Group

24 posts

UIUC TIMAN Group banner
UIUC TIMAN Group

UIUC TIMAN Group

@TIMANUIUC

Text Information Modeling and Analysis Group @UofIllinois, led by Prof. ChengXiang Zhai

Katılım Haziran 2025
38 Takip Edilen28 Takipçiler
UIUC TIMAN Group retweetledi
UIUC TIMAN Group retweetledi
UIUC TIMAN Group retweetledi
Jize Jiang
Jize Jiang@JizeJiang·
Big thanks to everyone on the team and our mentors 🌟 I’m thrilled that PlugMem has been accepted to ICML 2026. This is a big milestone for our work on memory for evolving agents. What excites me just as much is we are turning PlugMem into something people can actually build with, a truly plug-and-play memory module that works across real agent runtimes and interpretable through visualization interfaces. Making research accessible is part of pushing the frontier 🎆
Ke Yang@EmpathYang

Big PlugMem update 🧠 A plug-and-play memory module for LLM agents — turns raw trajectories into a knowledge graph your agent actually reasons over. 🎉 Accepted to ICML 2026 🔌 Drop it into OpenClaw 🦞, Claude Code, and other agent runtimes 🔍 Visualize memory · test retrieval · replay sessions 🥇 SOTA backbone on LongMemEval & HotpotQA — general enough to build on Paper: arxiv.org/abs/2603.03296 Code: github.com/TIMAN-group/Pl… #ICML2026 #LLM #Agents

English
1
3
63
335.8K
UIUC TIMAN Group
UIUC TIMAN Group@TIMANUIUC·
@EmpathYang @chenzixi23 @XuanHe21 @JizeJiang Congratulations to the team on the great work! Memory is a key component of any intelligent agent, and the idea of a plug-and-play memory module is not only theoretically interesting but also practically useful. It's so exciting that the new technology can now be used by people!
English
0
0
1
217
UIUC TIMAN Group retweetledi
Ke Yang
Ke Yang@EmpathYang·
Big PlugMem update 🧠 A plug-and-play memory module for LLM agents — turns raw trajectories into a knowledge graph your agent actually reasons over. 🎉 Accepted to ICML 2026 🔌 Drop it into OpenClaw 🦞, Claude Code, and other agent runtimes 🔍 Visualize memory · test retrieval · replay sessions 🥇 SOTA backbone on LongMemEval & HotpotQA — general enough to build on Paper: arxiv.org/abs/2603.03296 Code: github.com/TIMAN-group/Pl… #ICML2026 #LLM #Agents
Ke Yang tweet mediaKe Yang tweet mediaKe Yang tweet media
English
5
15
47
401.8K
UIUC TIMAN Group retweetledi
Yuji Zhang
Yuji Zhang@Yuji_Zhang_NLP·
🤔Hold on, I can answer better. 🔗New preprint on LLM multi-turn performance drop and recovery [arxiv.org/pdf/2604.04325]. 💡We identify a hidden tension in multi-turn reasoning: hold vs. lure. ⌛️Models can hold their intent to answer until sufficient evidence is observed, avoiding premature errors. ☔️But this ability is fragile—salient information can lure models to answer. ⬇️Even with the same information, performance drops significantly when moving from single-turn to multi-turn reasoning. ❓We ask: is this due to an overly strong intent to answer early? 🧑‍⚕️This is especially critical in medical diagnosis, a high-stakes setting with low tolerance for error, where a wrong answer at any turn can have serious consequences. 🎯To study this, we introduce MINT (Medical Incremental N-Turn Benchmark). MINT is: ✔ Information-preserving: decomposed cases can be concatenated to recover original single-turn performance, isolating the effect of interaction ✔ High-fidelity: clinically structured evidence (e.g., history, labs) with controlled turn granularity 💡Our key findings: 🏃1. Strong early-answer intent: Over 55% of answers are given within the first 2 turns, leading to a 20–50% accuracy drop from single-turn to multi-turn. ⏰2. Holding unlocks self-correction: When models are instructed to WAIT, the performance drop is greatly reduced. Incorrect→correct revisions occur up to 10.6× more often than the reverse, revealing a latent self-correction ability suppressed by early commitment. 🦴3. Strong lures override control: Clinically salient signals (e.g., lab results) trigger premature answers—even when models are explicitly told to wait. 👇4. Actionable implications: • Deferring the diagnostic question improves first-answer accuracy by up to 62.6% • Delaying salient evidence prevents up to 23.3% catastrophic accuracy drop. Thanks to all our coauthors for their amazing support! @ Jinrui Fang @ Runhan Chen @ Xu Yang @ Jian Yu @ Jiawei Xu @ Ashwin Vinod @WenqiShi0106 @TianlongChen4 @hengjinlp @ Chengxiang Zhai @TIMANUIUC @ying000
Yuji Zhang tweet mediaYuji Zhang tweet mediaYuji Zhang tweet mediaYuji Zhang tweet media
English
2
17
97
10.4K
UIUC TIMAN Group retweetledi
Microsoft Research
Microsoft Research@MSFTResearch·
PlugMem transforms AI agents’ interaction histories into structured, reusable knowledge. It integrates with any agent, supports diverse tasks and memory types, and maximizes decision quality while significantly reducing memory token use: msft.it/6017Qc9vv
Microsoft Research tweet media
English
2
34
39
9K
UIUC TIMAN Group retweetledi
Ke Yang
Ke Yang@EmpathYang·
📰New preprint: How can we build a task-agnostic plug-and-play memory module for LLM agents that supports multiple memory types? We present PlugMem🔌🧠, a plugin memory module that works across tasks by turning heterogeneous experience into knowledge. Evaluated unchanged on long-term dialogue🗣️, multi-hop QA🕵️, and web agents🕸️🤖, PlugMem improves performance while using far fewer memory tokens. 📜Paper: empathyang.github.io/files/PlugMem.… 🔨Code: github.com/TIMAN-group/Pl…
Ke Yang tweet media
English
13
64
169
12.2K
UIUC TIMAN Group retweetledi
Ke Yang
Ke Yang@EmpathYang·
We’ve been building a task-agnostic memory module for LLM agents — PlugMem. While running experiments across long-horizon QA, multi-hop retrieval, and web agents, we found several unexpected patterns about how memory actually helps (or hurts) decision-making. Code: github.com/TIMAN-group/Pl… Work with an amazing team: @chenzixi23, @XuanHe21, @JizeJiang, the deep learning group @MSFTResearch, @dmguiuc, and @TIMANUIUC. Thread ↓
Ke Yang tweet media
English
7
6
10
664
UIUC TIMAN Group retweetledi
Yuji Zhang
Yuji Zhang@Yuji_Zhang_NLP·
Looking forward to discussion!
Wenyue Hua@HuaWenyue31539

👏👏Join us for a talk by Yuji Zhang @Yuji_Zhang_NLP on Developing Robust and Trustworthy Foundation Models on this Saturday@9 pm! ⏲️Event Registration here: luma.com/eibwwv33 Yuji is a postdoctoral researcher at UIUC working on robust and trustworthy foundation models, with publications at ACL, EMNLP, ICLR, and more. We will learn about: 🔍 Making model knowledge explicit, testable, and reliable 🧠 Understanding and addressing hallucination and knowledge overshadowing 🔧 Diagnosing and repairing model failures with minimal side effects ⚛️ Representing knowledge as interpretable, composable "atomic skills" 🎯 Aligning model reasoning with real-world decision value

English
1
3
16
2.5K
UIUC TIMAN Group retweetledi
Wenyue Hua
Wenyue Hua@HuaWenyue31539·
👏👏Join us for a talk by Yuji Zhang @Yuji_Zhang_NLP on Developing Robust and Trustworthy Foundation Models on this Saturday@9 pm! ⏲️Event Registration here: luma.com/eibwwv33 Yuji is a postdoctoral researcher at UIUC working on robust and trustworthy foundation models, with publications at ACL, EMNLP, ICLR, and more. We will learn about: 🔍 Making model knowledge explicit, testable, and reliable 🧠 Understanding and addressing hallucination and knowledge overshadowing 🔧 Diagnosing and repairing model failures with minimal side effects ⚛️ Representing knowledge as interpretable, composable "atomic skills" 🎯 Aligning model reasoning with real-world decision value
Wenyue Hua tweet media
English
0
10
19
5.1K
UIUC TIMAN Group retweetledi
World Labs
World Labs@theworldlabs·
Introducing Marble by World Labs: a foundation for a spatially intelligent future. Create your world at marble.worldlabs.ai
English
358
609
3.3K
2.1M
UIUC TIMAN Group retweetledi
Neuralink
Neuralink@neuralink·
🇬🇧 We’re excited to announce our first participant in the UK! Paul, who is paralyzed due to motor neuron disease, received his Neuralink implant at @uclh earlier this month and was able to control a computer with his thoughts just hours after surgery.
Neuralink tweet media
English
365
930
5.8K
547K
UIUC TIMAN Group retweetledi
Generalist
Generalist@GeneralistAI·
Introducing GEN-0, our latest 10B+ foundation model for robots ⏱️ built on Harmonic Reasoning, new architecture that can think & act seamlessly 📈 strong scaling laws: more pretraining & model size = better 🌍 unprecedented corpus of 270,000+ hrs of dexterous data Read more 👇
English
47
281
1.5K
482.5K
UIUC TIMAN Group retweetledi
Tongyi Lab
Tongyi Lab@Ali_TongyiLab·
Qwen3-Max-Thinking-Preview is now available on Qwen Chat! Choose Qwen3-Max, enable Thinking and take a spin~
Tongyi Lab tweet media
English
25
52
625
33.6K
UIUC TIMAN Group retweetledi
OpenAI
OpenAI@OpenAI·
Meet our new browser—ChatGPT Atlas. Available today on macOS: chatgpt.com/atlas
English
2.3K
4.2K
29.8K
14M
UIUC TIMAN Group retweetledi
Anthropic
Anthropic@AnthropicAI·
We’re building tools to support research in the life sciences, from early discovery through to commercialization. With Claude for Life Sciences, we’ve added connectors to scientific tools, Skills, and new partnerships to make Claude more useful for scientific work.
English
108
345
2.3K
903.3K