Hanqi Yan

99 posts

Hanqi Yan banner
Hanqi Yan

Hanqi Yan

@yan_hanqi

Lecturer (assistant professor) @kclinformatics Interpretable | Reliable Language models

London, United Kingdom Katılım Eylül 2021
520 Takip Edilen665 Takipçiler
Sabitlenmiş Tweet
Hanqi Yan
Hanqi Yan@yan_hanqi·
🧠 Mechanistic interpretability is obsessed with features. But what if gradients tell you more? 📐 Introducing GRADE — using gradient subspace dynamics to measure how far an LLM is from the correct answer, probing knowledge gaps at their root. 🔍 📄 Paper: Probing Knowledge Gaps in LLMs through Gradient Subspace Dynamics 🔗 arxiv.org/pdf/2604.02830
English
1
13
197
11.2K
Hanqi Yan
Hanqi Yan@yan_hanqi·
🎉 @kclnlp led by @yulanhe is recruiting a 2-year postdoc in NLP/LLM! 🤖 🔬 Faithful reasoning, meta-reasoning with cognitive consideration to build reliable and impactful systems for high-stakes educational assessment. 🎓 🤝 Collaborated with AQA. 📅 Application deadline: 15 May 2026 🔗 Job post: lnkd.in/eHcka7ae
English
0
3
14
914
Hanqi Yan
Hanqi Yan@yan_hanqi·
RAG alone isn't enough for agent memory 🧠 Our work xMemory tackles this through decoupling & aggregation — and it's been picked up by Claude Memory, PageIndex, and OpenClaw ⚡ Glad the direction is resonating 🌊 🔗 arxiv.org/abs/2602.02007 led by @HZhanghao
KCL NLP@kclnlp

Our work on agent memory, xMemory (Beyond RAG for Agent Memory: Retrieval by Decoupling and Aggregation), has been discussed by projects such as Claude Memory, PageIndex, and OpenClaw. Glad to see this direction resonating. @HZhanghao @yulanhe @LinGui_KCL @dair_ai @VentureBeat

English
0
1
1
315
Hanqi Yan
Hanqi Yan@yan_hanqi·
🚨 Our blog just got accepted to ICLR'26! "Misalignments and RL Failure Modes in the Early Stage of Superintelligence" — led by @shuyhere What happens when RL agents start optimizing for the wrong thing at scale? 👇 Read it: iclr-blogposts.github.io/2026/blog/2026…
English
1
6
46
4.4K
Hanqi Yan
Hanqi Yan@yan_hanqi·
🧠 We're rethinking agent memory. Top-k similarity RAG isn't enough — our new work by @HZhanghao explores what's beyond it. Getting lots of traction 📈 Code coming soon — stay tuned! 🔔
DAIR.AI@dair_ai

// Beyond RAG for Agent Memory // RAG wasn't designed for agent memory. And it shows. The default approach to agent memory today is still the standard RAG pipeline: embed stored memories, retrieve a fixed top-k by similarity, concatenate them into context, and generate an answer. Every major agent memory system follows this base pattern. But agent memory is fundamentally different from a document corpus. It's a bounded, coherent dialogue stream where candidate spans are highly correlated and often near duplicates. Fixed top-k similarity retrieval collapses into a single dense region, returning redundant evidence. And post-hoc pruning breaks temporally linked evidence chains rather than removing redundancy. This new research introduces xMemory, a hierarchical retrieval framework that replaces similarity matching with structured component-level selection. Agent memory needs redundancy control without fragmenting evidence chains. Structured retrieval over semantic components achieves both, consistently outperforming standard RAG and pruning approaches across multiple LLM backbones. The key idea: It decouples memories into semantic components, organize them into a four-level hierarchy (original messages, episodes, semantics, themes), and uses this structure to drive retrieval top-down. A sparsity-semantics objective guides split and merge operations to keep the high-level organization both searchable and semantically faithful. At retrieval time, xMemory selects a compact, diverse set of relevant themes and semantics first, then expands to episodes and raw messages only when doing so measurably reduces the reader's uncertainty. On LoCoMo with Qwen3-8B, xMemory achieves 34.48 BLEU and 43.98 F1 while using only 4,711 tokens per query, compared to the next best baseline Nemori at 28.51 BLEU and 40.45 F1 with 7,755 tokens. With GPT-5 nano, it reaches 38.71 BLEU and 50.00 F1, improving over Nemori while cutting token usage from 9,155 to 6,581. xMemory retrieves contexts that cover all answer tokens in 5.66 blocks and 975 tokens, versus 10.81 blocks and 1,979 tokens for naive RAG. Higher accuracy, half the tokens. Paper: arxiv.org/abs/2602.02007 Learn to build effective AI agents in our academy: academy.dair.ai

English
1
0
3
838
Hanqi Yan
Hanqi Yan@yan_hanqi·
Thank you @Ruihong for having me! Loved the engaging discussions at your group @UQschoolEECS. Recording of my talk 'Structured Representation for Latent Thinking in LLMs' is now available: uq-ds-seminar.github.io/latentLLM-hanq…
Ruihong Qiu@RuihongQiu

We are very happy to have Prof @yan_hanqi from @KingsCollegeLon to be our seminar speaker @UQSchoolEECS to talk about her work on Structured Representation Learning for Latent Thinking in LLMs (uq-ds-seminar.github.io/latentLLM-hanq…)!

English
0
1
16
2.5K
Hanqi Yan
Hanqi Yan@yan_hanqi·
🎓 Four-year PhD studentships are available in STaR-AI CDT @KingsCollegeLon @kclinformatics! 📅 Application deadline: 2 March 2026 🔬 STaR-AI-16: From One-Size-Fits-All to Audience-Aware: Personalized Explainable AI for Decision-Making 🤖 Are you passionate about making AI more transparent and personalized? This project explores cutting-edge approaches to tailoring AI explanations for different audiences and decision-making contexts. 📋 Project details: lnkd.in/ewE8M7N2 👥 Supervisors: 🔹 Co-first supervisor: Dr Hanqi Yan (Department of Informatics) 🌐 Webpage: lnkd.in/edYe4MzE 🔹 Co-first supervisor: Dr Wei HE (Department of Engineering) 🌐 Webpage: lnkd.in/e-4rKyhi 💡 Interested in shaping the future of Explainable AI? Apply now!
English
0
6
8
642
Hanqi Yan
Hanqi Yan@yan_hanqi·
🚀 20th Jan, 14:00-18:00. Join our tutorial at #AAAI26: "Structured Representation Learning: Interpretability, Robustness, and Transferability for LLMs" Learn how structured representations can make LLMs more interpretable, robust, and transferable across tasks. 📍 AAAI 2026 Singapore 🔗 srl4llm.github.io #AI #MachineLearning #LLMs #NLProc
English
3
1
6
504
Hanqi Yan
Hanqi Yan@yan_hanqi·
Session 1: Introduction Session 2: The Principles of Representation Learning Session 3: Representations for Latent CoTs Coffee Break Session 4: Understand and Model Edit via Representation Learning Session 5: Integrate Models Internals for Self-Improvements Session 6: Conclusion and Future Work
English
0
0
0
97
Hanqi Yan retweetledi
Hanqi Yan
Hanqi Yan@yan_hanqi·
🎓 Fully-funded PhD position (start from 10.2026) at King's College London! 🤖 Building robust & fair AI for early detection of treatment resistance in schizophrenia 👥 Co-supervised by Prof. James MacCabe (Psychology) & myself (Informatics) Apply now 👇 showcase.drive-health.org.uk/project/genai-…
English
0
2
15
1K
Hanqi Yan retweetledi
Hanqing Zhu
Hanqing Zhu@zhu_hanqin41424·
🚨 New Work! 🤔 Is RL black-box weight tinkering? 😉 No. We provably show RLVR follows a 🧭 — always updating the same off-principal regions while preserving the model's core spectra. ⚠️ Different optimization regime than SFT — SFT-era PEFT tricks can misfire(like PiSSA, the open question from @thinkymachines). Time to think new methods, not just repurpose. 🫣 Paper: huggingface.co/papers/2511.08… Blog: hanqing.notion.site/the-path-not-t… A Deep Dive 👇
Hanqing Zhu tweet media
English
8
42
257
107.8K
Hanqi Yan
Hanqi Yan@yan_hanqi·
🚀 Thrilled to announce that I’ll be attending EMNLP 2025 (4Nov-9Nov) in Suzhou, China! 🇨🇳✨ I’ll be showcasing our latest research from #KCLNLP on implicit Chain-of-Thoughts (CoTs) and an AI Scientist demo system 🤖🧠 📘 CODI: Compressing Chain-of-Thought into Continuous Space via Self-Distillation led by Zhenyi Shen 🔗 lnkd.in/gt587qAj 🔥 50+ citations so far! 💡 GraphMind: Interactive Novelty Assessment System for Accelerating AI Research led by Italo Da Silva 🔗 lnkd.in/gNGHvt-f 🎬 Demo system — lnkd.in/gXdec9Cu! Can’t wait to connect with fellow researchers, innovators, and AI enthusiasts at #EMNLP2025 🌏💬 Let’s push the boundaries of intelligent reasoning together! 💪🤝
English
2
4
40
5.5K
Hanqi Yan
Hanqi Yan@yan_hanqi·
@askerlee Yes, my idea is almost from arxiv.org/abs/1906.01820 looking into the internal world for better inner alignment! Very surprised this "not popular" paper becomes so inspiring after several years!!! 👍
English
1
0
1
49
Hanqi Yan
Hanqi Yan@yan_hanqi·
Again! 📢Vote for "latent modelling for better alignment" The world model should have a z(t) to model the 👓unobserved internal process, beyond the observed behaviour points. 📑See our recent work about CoT-induced latent misalignment: Within your reasoning model, Reason and safety fight over the same neurons — that’s how your reasoning model learns unsafe objectives. arxiv.org/pdf/2509.00544
Saining Xie@sainingxie

there’s only one right answer here, the @ylecun definition, and everyone should be able to recite it word for word

English
1
3
46
10.3K