Lin Gui
11 posts

Lin Gui
@LinGui_KCL
NLP lecturer (assistant professor)@King's College London, also do a bit IR, ML, and Stats for fun.

ICLR rebuttal strategy meeting: 🕐 30 min discussing the OPTIMAL time to ping reviewers 📊 Must be after Thanksgiving but before the weekend 🧠 Late enough they can't ask for more experiments ⏰ Early enough they still have time to respond Lead author right after the call: "I just sent it" Everyone: "WHAT HAVE YOU DONE. THIS IS NOT WHAT WE AGREED ON. YOU RUINED EVERYTHING. WE'RE COOKED 💀" 5 minutes later Two reviewers raise their scores by 4 points each 😱😱😱






VLM can think visually without generating pixels! VLM can think visually without generating pixels! VLM can think visually without generating pixels! 📢 We introduce Machine Mental Imagery (Mirage): a new framework that enables VLM to imagine using latent visual tokens—performing visual reasoning in latent space, no pixel rendering needed! We achieve this through a two-phase training paradigm: ✅ Stage 1: Grounding latent tokens in the visual subspace (joint supervision) ✅ Stage 2: Anchoring grounded tokens for generation (text-only supervision) Mirage demonstrates strong performance on a wide range of multimodal reasoning tasks! 📜Paper: arxiv.org/abs/2506.17218 🧑💻Code: github.com/UMass-Embodied… 📽️Project Page: vlm-mirage.github.io












