Keming Wu

50 posts

Keming Wu

Keming Wu

@Keming_Charles

PhD student @Tsinghua_Uni. Focus on Generative AI and VLM. Author of EditReward, OpenMMReasoner.

가입일 Eylül 2025
515 팔로잉141 팔로워
고정된 트윗
Keming Wu
Keming Wu@Keming_Charles·
Why do open-source image editing models lag behind closed-source giants like GPT-Image-1, Seedream, & Google-Nano-Banana? 🤔 It’s mainly due to the quality of the training reward signal. We’re bridging the gap. Meet EditReward! 🏆
Keming Wu tweet media
English
4
15
147
44.3K
Keming Wu 리트윗함
Kaichen Zhang
Kaichen Zhang@KaichenZhang358·
Excited to share a fun project I recently collaborated on: a roadmap for thinking about where visual generation is heading next. The key question is no longer just “can it make beautiful images?”, but whether it can handle memory, interaction, and eventually world modeling.
Kaichen Zhang tweet media
English
1
7
12
1.3K
Keming Wu
Keming Wu@Keming_Charles·
Benchmarks often reward visual quality. But real progress also needs spatial reasoning, topology, symbolic structure, and code/math-grounded correctness. We stress-test physical and causal reasoning: These examples probe the boundary between image synthesis and world modeling.
Keming Wu tweet mediaKeming Wu tweet mediaKeming Wu tweet media
English
1
0
3
95
Keming Wu 리트윗함
Jianyang Gao
Jianyang Gao@gaoj0017·
The TurboQuant paper (ICLR 2026) contains serious issues in how it describes RaBitQ, including incorrect technical claims and misleading theory/experiment comparisons. We flagged these issues to the authors before submission. They acknowledged them, but chose not to fix them. The paper was later accepted and widely promoted by Google, reaching tens of millions of views. We’re speaking up now because once a misleading narrative spreads, it becomes much harder to correct. We’ve written a public comment on openreview (openreview.net/forum?id=tO3AS…). We would greatly appreciate your attention and help in sharing it.
Google Research@GoogleResearch

Introducing TurboQuant: Our new compression algorithm that reduces LLM key-value cache memory by at least 6x and delivers up to 8x speedup, all with zero accuracy loss, redefining AI efficiency. Read the blog to learn how it achieves these results: goo.gle/4bsq2qI

English
98
976
6.5K
1M
Keming Wu 리트윗함
Maryam
Maryam@Sci_Tech_Eng·
@Keming_Charles Congrats; talented! Wishing you more great milestones! 🦾
English
1
0
0
29
Keming Wu
Keming Wu@Keming_Charles·
100+ citations 🎉 In a field that moves this fast, every citation feels like a small vote of confidence. Thanks to everyone who read, used, and built on our work.
Keming Wu tweet mediaKeming Wu tweet media
English
0
1
6
165
Lei Li
Lei Li@_TobiasLee·
MMMU-Pro and Video-MME are saturated now. We need some awesome agentic multimodal benchmarks. Any suggestions?
English
8
0
20
3.2K
Keming Wu
Keming Wu@Keming_Charles·
Highlights:✨ PrismLayersPro: 200K high-res transparent layers. ✨ ART+: Our model outperforms original ART in 60% of cases & rivals FLUX.1. ✨ LayerFLUX: Expert at accurate alpha mattes.
English
1
0
1
102
Keming Wu
Keming Wu@Keming_Charles·
Inspired by Qwen-Image-Layered? Want to train multi-layer models? 🛠️ Need data? Introducing PrismLayers: foundation for multi-layer gen community. Released earlier for researchers w/ robust baseline & data source.
Keming Wu tweet media
English
1
0
3
201