Feng Li
26 posts

Feng Li
@FengLiust
Research Scientist @GoogleDeepMind | Ex-intern @MetaAI (FAIR), @MSFTResearch, @ByteDanceTalk, @IDEACVR
Kirkland, WA Katılım Temmuz 2024
242 Takip Edilen1.4K Takipçiler

Today marks a moment I'll remember for the rest of my life.
When we started Manus, few believed that general AI agents could work. We were told it was too early, too ambitious, too hard. But we kept building. Through the doubts, the setbacks, and the countless nights wondering if we were chasing the impossible.
We weren't.
This isn't just an acquisition. It's validation that the future we've been building toward is real, and it's arriving faster than anyone expected.
But this is not the end. The era of AI that doesn't just talk, but acts, creates, and delivers, is only beginning. And now, we get to build it at a scale we never could have imagined.
To everyone who believed in us before it was obvious: thank you. The best is yet to come.
Manus@ManusAI
Manus is entering the next chapter: we’re joining forces with Meta to take general agents to the next level. Full story on our blog: manus.im/blog/manus-joi…
English

@FengLiust @GoogleDeepMind Congrats and welcome to Google!
English

Thrilled to share that I’ve joined @GoogleDeepMind as a Research Scientist.
Excited for what’s ahead and the amazing people I’ll get to work with. 🚀
I am based in Kirkland, happy to meet old and new friends here (Feel free to ping me over WeChat or email)!
English

@FengLiust @GoogleDeepMind Huge congrats! Can’t wait to see more of your pioneering work at DeepMind! 🤩
English

@FengLiust @GoogleDeepMind Wait, there is a place called Kirkland? (Owned by Costco? 😜)
English

@xyz2maureen Huge congratulations xueyan! All the best on your new adventure.
English

I will join Tsinghua University, College of AI, as an Assistant Professor in the coming month. I am actively looking for 2026 spring interns and future PhDs (ping me if you are in #NeurIPS).
It has been an incredible journey of 10 years since I attended an activity organized by Tsinghua University and decided to change my undergraduate major from Economics to Computer Science, inspired by one of the teammates. During the 10 years, I met with appreciation of many wonderful researchers/professors who led me to continued growth. 🐿️
My research focus will continue to be AI & Robotics, with a specific emphasis on Interactive Embodied Intelligence. You can check my homepage to learn more: maureenzou.github.io/lab.html.
I am currently local to San Diego and will be attending #NeurIPS. Please ping me over WeChat or Email if any old or new friends are interested in having a coffee chat! (Really looking forward to meeting as many friends as possible at #NeurIPS)
[The photo is one of the places that I will miss a lot in the US]

English
Feng Li retweetledi

Unified multimodal models can generate text and images, but can they truly reason across modalities? 🎨
Introducing ROVER, the first benchmark that evaluates reciprocal cross-modal reasoning in unified models, the next frontier of omnimodal intelligence.
🌐 Project: roverbench.github.io
📄 Paper: arxiv.org/abs/2511.01163
📂 Benchmark: huggingface.co/datasets/chery…
English

@jw2yang4ai @MSFTResearch @CVPR @AIatMeta Congrats Jianwei! So grateful for your mentorship at MSR and looking forward to your next exciting work!
English

Life Update: Now that I have finished the presentation of last @MSFTResearch project Magma at @CVPR, I am excited to share that I have joined @AIatMeta as a research scientist to further push forward the boundary of multimodal foundation models!
I have always been passionate about building multimodal AI systems that can interact with human and environments. In the past five years, I am very fortunate to lead and contribute to a number of exciting projects on (a) vision and multimodal foundation: Focal Attention, FocalNet, UniCL, RegionCLIP, GLIP, Florence; (b) generalist multimodal vision models: X-Decoder, SEEM, Semantic-SAM and Grounding-DINO; (c) multimodal large language models: LLaVA variants, Phi-3-Vision; (d) multimodal agentic model: SoM + OmniParser, TraceVLA, LAPA and Magma. I’m also very proud to have contributed to impactful projects, such as LLaVA-Med, GigaPath and BiomedParse, advancing AI for healthcare and human good. These are truly meaningful footnotes in my journey at MSR!
In the past five years while I am staying at MSR, the world has witnessed tremendous breakthroughs in AI as well as those brought by AI. Looking ahead, the opportunity to advance AI research for a better world has never been so exciting. I feel so lucky to be part of this. Now right after five years, it feels like the right time for me to "graduate" from MSR and embrace new challenges beyond! Thank you all again for the support, mentorship, and friendship!

English
Feng Li retweetledi

🚀 BAGEL — the Unified Multimodal Model with emergent capabilities and production-ready performance — is finally live!
Dive in here:
👉 bagel-ai.org
English
Feng Li retweetledi

We're happy to release BAGEL, a powerful unified model:
🔓 Open-source: weights & code
🧠 Unified MoT arch for text, image, & video
🏆 Top performance on 10+ benchmarks
✨ Emerging properties: From basic T2I/I2T to intelligent manipulation
⚡ Try it now: bagel-ai.org

English
Feng Li retweetledi

We propose GenProp✨, a generative video propagation framework, which can seamlessly propagate any first frame edit through the video. 🧵(1/n)
- Arxiv: arxiv.org/abs/2412.19761
- Project Page: genprop.github.io
- Video: youtube.com/watch?v=GC8qfW…

YouTube
English
Feng Li retweetledi

🌟 Introducing DINO-X, our groundbreaking unified vision model at IDEA Research!
Paper: arxiv.org/abs/2411.14347
Blog: deepdataspace.com/blog/7?source=x
Playground: deepdataspace.com/playground/din…
API: cloud.deepdataspace.com/docs#/api/dino…
#AI #ComputerVision #ObjectDetection #IDEAResearch
Highlights:
English







