Liangming Pan

148 posts

Liangming Pan banner
Liangming Pan

Liangming Pan

@PanLiangming

Assistant Professor, Peking University (@PKU1898) | Former AP @UofAInfoSci | Postdoc @ucsbNLP | Ph.D. @NUSingapore | Researcher in NLP, LLMs & Reasoning

Beijing, China Katılım Aralık 2017
893 Takip Edilen2K Takipçiler
Sabitlenmiş Tweet
Liangming Pan
Liangming Pan@PanLiangming·
Life update: I've joined the School of Computer Science at Peking University @PKU1898 as an Assistant Professor! I'm looking for Ph.D./intern/visiting researchers for my new research group. If you are interested in NLP and LLM, check my research at liangmingpan.bio
Liangming Pan tweet media
English
16
32
312
24.9K
Liangming Pan
Liangming Pan@PanLiangming·
Thanks for the recognition, Joshua! We quite agree with the 'dream' you described—it’s exactly what we are aiming for. Currently, the high compute cost of Influence Functions limits our study to smaller models and Induction Heads, but we are working on optimizing efficiency. Our next goal is to apply this method to analyze the training origins of more complex emergent behaviors. Hope to connect more in the future!
English
0
0
2
114
Liangming Pan
Liangming Pan@PanLiangming·
🔍Do Latent Chain-of-Thought (latent-CoT) models think step-by-step? In our new paper, we dissect CODI in the context of polynomial-iteration tasks. We map when it rolls out real intermediates, when it compresses, and when reasoning collapses. 🧾Preprint: arxiv.org/abs/2602.00449
Liangming Pan tweet media
English
1
12
95
4.7K
Liangming Pan
Liangming Pan@PanLiangming·
By identifying the "data seeds" of intelligence, we can build more efficient and interpretable models. 📝 Preprint: arxiv.org/abs/2601.21996 👨‍💻 Code: github.com/chenjianhuii/M… Huge thanks to all co-authors for their work on uncovering the causal origins of LLM circuits! 🚀 🧵(6/6)
English
0
1
7
407
Liangming Pan
Liangming Pan@PanLiangming·
Beyond understanding, we offer a tool for action: A mechanistic data augmentation pipeline. This consistently accelerates circuit convergence across model scales, providing a principled methodology for steering the developmental trajectories of LLMs. 🧵(5/6)
Liangming Pan tweet media
English
1
1
3
466
Liangming Pan
Liangming Pan@PanLiangming·
🔥While Mechanistic Interpretability has identified interpretable circuits in LLMs, their causal origins in training data remain unknown. We introduce Mechanistic Data Attribution (MDA)—a scalable framework to bridge this gap by employing Influence Functions to trace interpretable units (like induction heads) back to specific training samples. We want to know not just what the model learned, but from where. 📝 Preprint: arxiv.org/abs/2601.21996 🧵(1/6)
Liangming Pan tweet media
English
5
37
264
19.7K
Liangming Pan
Liangming Pan@PanLiangming·
🔥 What actually happens during multi-step reasoning with LLMs? ❓What are the internal computations? ❓How is such capability acquired during training? ❓Does the latent reasoning rely on shortcuts? ❓How does CoT remodel internal computation? ❓Why does CoT enhance reasoning capability? and more... Many questions remain about the internal machinery. We wrote a paper to systematically review the existing process of revealing the mechanisms behind LLM multi-step reasoning—from implicit latent reasoning to explicit CoT reasoning. We also highlight directions for future mechanistic studies. 📄Paper: arxiv.org/pdf/2601.14270 💻Github: github.com/PKU-PILLAR-Gro…
Liangming Pan tweet media
English
1
10
36
3K
Liangming Pan
Liangming Pan@PanLiangming·
🔥 Check out our new survey on unveiling the "Black Box" of Large Reasoning Models (LRMs)! We cover: 1⃣ Training Dynamics (SFT, RL) 2⃣ Inference Behaviors 3⃣ Failures (Overthinking/Faithfulness/Hallucination/Safety) Work led by @AheadOFpotato Github: github.com/AheadOFpotato/…
Liangming Pan tweet media
English
2
7
19
708
Liangming Pan
Liangming Pan@PanLiangming·
@KeremZaman3 Very interesting study! We would include your works in our overview. Thanks for sharing.
English
0
0
1
35
Liangming Pan
Liangming Pan@PanLiangming·
🔥 One of the widely adopted assumptions is that the Chain-of-Thought explains how a model made a decision. But is this "explainability" just a mirage? We wrote a survey paper to review this emerging research field of CoT Faithfulness. 💻 Github repo: github.com/PKU-PILLAR-Gro…
Liangming Pan tweet media
English
1
1
7
365
Liangming Pan
Liangming Pan@PanLiangming·
I am seeking an international Ph.D. student🎓(non-Chinese citizen) to join my lab at Peking University in Fall 2026. Interested candidates are encouraged to apply to PKU's International Graduate Student Program and get in touch. More about my research: liangmingpan.bio🔬
Liangming Pan tweet media
English
4
6
21
2.3K
Liangming Pan
Liangming Pan@PanLiangming·
Life update: I've joined the School of Computer Science at Peking University @PKU1898 as an Assistant Professor! I'm looking for Ph.D./intern/visiting researchers for my new research group. If you are interested in NLP and LLM, check my research at liangmingpan.bio
Liangming Pan tweet media
English
16
32
312
24.9K