Keming Wu
50 posts

Keming Wu
@Keming_Charles
PhD student @Tsinghua_Uni. Focus on Generative AI and VLM. Author of EditReward, OpenMMReasoner.






For a decade, we've made models wider and deeper—but we've barely changed how layers *talk* to each other. Since ResNet's `x + F(x)` in 2015, the depth residual has been the only highway for inter-layer communication. It's time to upgrade the staircase. 🧵

Introducing TurboQuant: Our new compression algorithm that reduces LLM key-value cache memory by at least 6x and delivers up to 8x speedup, all with zero accuracy loss, redefining AI efficiency. Read the blog to learn how it achieves these results: goo.gle/4bsq2qI

Advanced Machine Intelligence (AMI) is building a new breed of AI systems that understand the world, have persistent memory, can reason and plan, and are controllable and safe. We’ve raised a $1.03B (~€890M) round from global investors who believe in our vision of universally intelligent systems centered on world models. This round is co-led by Cathay Innovation, Greycroft, Hiro Capital, HV Capital, and Bezos Expeditions, along with other investors and angels across the world. We are a growing team of researchers and builders, operating in Paris, New York, Montreal and Singapore from day one. Read more: amilabs.xyz AMI - Real world. Real intelligence.


🚀 Excited to share OpenMMReasoner: A complete open-source recipe for multimodal reasoning training! 📊 874K SFT + 74K RL data 🔬 Reproducible SFT pipeline ⚡ Advanced RL training (GSPO/GRPO/DAPO) 📈 +11.6% over Qwen2.5-VL-7B baseline 🧵 Thread below 👇

🔥 Introducing LongVT: Teaching Multimodal LLMs to "Actively Look Back" and understand long videos just like humans! We tackle the "sparse evidence" & "hallucination" issues in long-video reasoning with an end-to-end Agentic solution. Paper: arxiv.org/abs/2511.20785 More in thread

Excited to announce that TIGER-Lab has 8 papers accepted to ICLR 2026. Congrats to all the students and co-authors!












