Qingcheng Zeng

194 posts

Qingcheng Zeng

Qingcheng Zeng

@SteveZeng7

PhD-ing with @rfpvjr and @kaize0409 / IR, search agent, LLM, social computing / Big fan of @Arsenal / Christian

Evanston, IL Katılım Haziran 2019
2.7K Takip Edilen1.2K Takipçiler
Qingcheng Zeng retweetledi
Rohan Paul
Rohan Paul@rohanpaul_ai·
Researchers found that when language models face harder questions, their internal brain activity literally shrinks into fewer paths. Language models actually compress their internal thinking when they get confused, and we can use that to help them. Standard AI models usually spread their thinking across many artificial neurons when they confidently recognize familiar information. The team discovered that if you confuse a model with tricky math or conflicting facts, this broad activation collapses into a highly concentrated signal in its final processing layer. This shrinking happens because the system drops its robust distributed memory and forces the computation into a tiny specialized space to survive the unfamiliar challenge. The big deal is that we usually have no idea when a language model is actually struggling with a weird prompt until it gives a wrong answer. This paper proves that the model actually broadcasts its confusion internally by abandoning its wide neural networks and falling back on a very tiny cluster of active neurons. Because we can measure this exact shrinking effect as a raw number, we do not have to guess if a question is too hard for the AI. We can just read that internal signal and automatically provide the system with the perfectly scaled stepping stones it needs to solve the problem. ---- Paper Link – arxiv. org/abs/2603.03415 Paper Title: "Farther the Shift, Sparser the Representation: Analyzing OOD Mechanisms in LLMs"
Rohan Paul tweet media
English
16
22
100
16.8K
Orion Weller
Orion Weller@orionweller·
Very cool results @SteveZeng7, thanks for sharing! I am a big fan of late interaction but in an instruction-following or prompting setting it’s always felt a bit of a mismatch to me. Late interaction is great because it has more fine grained token-matching while prompting is the opposite (eg avoid lexical matching). That could be because someone needs to update late interaction architectures to store more prompting-esque tokens though, so it could definitely be possible, just not obvious. Cool to see your initial start here 🤗
English
1
0
3
264
Qingcheng Zeng
Qingcheng Zeng@SteveZeng7·
In early Jan, I decided to write some blogs this year. Here comes the first one: qcznlp.github.io/blog/2026/inst…. I played with PyLate on instruction-following IR tasks and got some interesting results. Would really love to get feedbacks, especially from @antoine_chaffin if possible!
English
2
3
19
3.1K
Qingcheng Zeng
Qingcheng Zeng@SteveZeng7·
@antoine_chaffin Thanks @antoine_chaffin for the really helpful feedback! And I really expect the Qwen-Embed-based PyLate to come out. Now I will play with ColBERT-zero to see what will happen🫡
English
1
0
1
58
Antoine Chaffin
Antoine Chaffin@antoine_chaffin·
I shared my feedback here, to try raising as much awareness about your work as possible! x.com/antoine_chaffi… The Tl;Dr is: I really appreciate it! I tried to do something very similar back then but did not take it to the finish line I think the results makes sense and I wonder how it would work with ColBERT-Zero, which is an even more ColBERT-tuned model, and also has a prompting mechanism (which can be double-edged sword in this case, as it's not the same kind of prompting) Also, I think to really have outstanding results, we might need to leverage big models/LLMs that have already some kind of prompting capabilities. I'm trying to get a Qwen-embed-based PyLate model out, it could be a good base!
Antoine Chaffin@antoine_chaffin

About the results now As the biggest fan of @orionweller on earth, I've immediately thought about this when they released Promptriever (actually, directly shot him a DM about it). Anytime people release dataset on the hub, it's a free opportunity to get SOTA by plugging it into PyLate So I went ahead and tried it (there is still a branch on the main repo for it, actually), even coded a in-training evaluator running with a PLAID index and coding p-MRR! So why did I not release it? Well, I am running this kind of experiments on a daily basis and, as illustrated in the blog post, the results are good, but not insanely good. I was in a phase of beating B-scale models by a large margin, so being better on some metrics but not all was a good result, but not good enough for me to spend time digging into it Essentially, I was scared that for prompting capabilities, you might require a LLM (or at least, starting from something that have shown some prompting capabilities), so I was waiting to train some larger scale ColBERT models to iterate on this (and I still believe we should see much better performance using those kind of models, it's already pretty cool to hit those results with such small models!) The work from @SteveZeng7 is a good reminder that sometimes sharing some cool results is enough and we should aim at sharing as much as we can, not just the perfect shiny results. I should share more about all the exploration I make! Finally, about the fact that it's better to start from GTE-ModernColBERT, I would say that it's somewhat related to our ColBERT-Zero study (huggingface.co/blog/lightonai…), in the sense that it's important to be careful about training the model for late interaction and not take as granted that just taking a dense model as a basis is optimal I am a bit surprised that the scale of this training is not enough, but as raised in the BP, I suppose it's because it's more related to learning "general" retrieval before! Actually, it would be pretty cool to try this boilerplate with the ColBERT-Zero models, I wonder how would be the results! The main issue to me is that those models already leverage a "prompt" and it might conflict a bit, but it's an interesting avenue!

English
2
0
2
323
Antoine Chaffin
Antoine Chaffin@antoine_chaffin·
Scaling ColBERT is very hard (trust me) However, when a whole cracked team start to work on it, it becomes possible And when you see what's possible with late interaction models (especially multi-modal ones), you know that search will be very different in the future Congrats!
Mixedbread@mixedbreadai

We build the first production ready multi-vector and multimodal search. Now we are serving over 1 billion documents in under 50ms latency (p50). We are sharing how we build it.

English
3
7
125
15K
Chenhao Tan
Chenhao Tan@ChenhaoTan·
.@universeinanegg and I have spent quite some time debating hot takes. We often disagree, so I made a little game about it: chenhaot.com/ariorchenhao/ It is super simple. Every week we will each write three AI hot takes, and you can guess who wrote each and agree or disagree. We've found debating hot takes to be a good icebreaker maybe you will too!
English
8
1
29
3.7K
Jason Eisner
Jason Eisner@adveisner·
LOR deadline day ... so many required fields! ... luckily a few of them have a high character limit
Jason Eisner tweet media
English
1
1
33
4.2K
Qingcheng Zeng retweetledi
rob voigt
rob voigt@rfpvjr·
For everyone working on the intersections between linguistic and computational research, consider submitting to the upcoming SCiL! We're excited that it will be co-located with ACL 2026 as a workshop, and we've received NSF funding to help cover costs. sites.google.com/view/scil2026
English
1
4
15
5K
Qingcheng Zeng retweetledi
Wanru Zhao
Wanru Zhao@Renee42581826·
Excited to present our work "Learning to Solve Complex Problems via Dataset Decomposition" at #NeurIPS2025! 🕟 Thu, Dec 4, 2025, 4:30 PM – 7:30 PM PST (🚨HAPPENING TODAY) 📍 Exhibit Hall C/D/E #3410 📅 Add to calendar: tinyurl.com/datadecomp 🧠 Training LLMs on randomly shuffled data is like asking elementary students to jump straight into calculus. We introduce Dataset Decomposition (Decomp): a method that recursively decomposes complex problems into an intelligent "easy-to-hard" curriculum, making small models smarter via structured reasoning. Sadly, I cannot attend in person due to visa delays 😢, but my incredible mentors from @MSFTResearch will be there! I'll be standing by remotely, so feel free to DM me or use the online chat! Huge thanks to my co-authors and mentors @LucasPCaccia, @Zhengyan_Shi, @kim__minseon, @weijiavxu, and @murefil. Deeply grateful to my supervisor @niclane7 for the generous support, to @ericxyuan and @Cote_Marc for the invaluable help, and to Colin Raffel, @MattVMacfarlane, @veds_12, @ZhihaoZhanMila, and @xiaoyin_chen66 for the stimulating discussions! @NeurIPSConf
Wanru Zhao tweet media
English
1
8
26
1.8K
Freda Shi
Freda Shi@fredahshi·
I promise I'll submit something to the 3rd COLM (please do have it!) The papers look amazing every time, and I’ve already regretted missing out twice in a row.
English
3
3
68
6.4K
Ge Zhang
Ge Zhang@GeZhang86038849·
@eliebakouch They are like 3-4 months behind Kimi / dpsk team, only based on the text understanding perf. And like I said, I believe that the multimodality understanding performance just doesn't matter.
English
2
0
1
288
Ge Zhang
Ge Zhang@GeZhang86038849·
stepfun.com/research/zh/st… Guess that they spend too much money on Scaling Law. I'm a fan for every team's amazing work. But this one is disappointing.
English
1
0
11
1.3K
马东锡 NLP
马东锡 NLP@dongxi_nlp·
@SteveZeng7 我应该消除偏见去读读。除了OAgents,还有什么不错的工作么?
中文
2
0
2
2.7K
马东锡 NLP
马东锡 NLP@dongxi_nlp·
看到两篇Agent论文,论文title很吸引人。然后一看作者分别来自神州数码和OPPO,我的偏见让我瞬间不想看了。
中文
9
6
98
43.4K
Qingcheng Zeng retweetledi
Minghao Guo
Minghao Guo@MurphyKwok_·
📢 Just dropped: DeepSieve RAG meets real-world heterogeneity. 🧠Modular RAG across SQL, JSON, Docs. 🧩 LLM Decompose → Route → Reflect → Fuse like an info sieve 📌 arXiv: 2507.22050 🧭 Project: minghokwok.github.io/deepsieve #LLM #RAG #AI #NLP #Rutgers
Minghao Guo tweet media
English
1
3
10
1.1K
JinYeong Bak
JinYeong Bak@NoSyu·
Thank you all for attending the virtual presentation session on Dialogue and Discourse @aclmeeting 2025
JinYeong Bak tweet media
English
1
0
5
399