Feng Li

26 posts

Feng Li

Feng Li

@FengLiust

Research Scientist @GoogleDeepMind | Ex-intern @MetaAI (FAIR), @MSFTResearch, @ByteDanceTalk, @IDEACVR

Kirkland, WA Katılım Temmuz 2024
242 Takip Edilen1.4K Takipçiler
Yifei Zhou
Yifei Zhou@YifeiZhou02·
Belated life update: I started my next chapter at Thinking Machines Lab this week, and it’s been an incredible experience — unmatched work culture and talent density. Extremely bullish on what the team is building 🚀
Yifei Zhou tweet media
English
51
11
759
81.5K
Red Xiao
Red Xiao@Red_Xiao_·
Today marks a moment I'll remember for the rest of my life. When we started Manus, few believed that general AI agents could work. We were told it was too early, too ambitious, too hard. But we kept building. Through the doubts, the setbacks, and the countless nights wondering if we were chasing the impossible. We weren't. This isn't just an acquisition. It's validation that the future we've been building toward is real, and it's arriving faster than anyone expected. But this is not the end. The era of AI that doesn't just talk, but acts, creates, and delivers, is only beginning. And now, we get to build it at a scale we never could have imagined. To everyone who believed in us before it was obvious: thank you. The best is yet to come.
Manus@ManusAI

Manus is entering the next chapter: we’re joining forces with Meta to take general agents to the next level. Full story on our blog: manus.im/blog/manus-joi…

English
295
153
2.3K
394.7K
Feng Li
Feng Li@FengLiust·
Thrilled to share that I’ve joined @GoogleDeepMind as a Research Scientist. Excited for what’s ahead and the amazing people I’ll get to work with. 🚀 I am based in Kirkland, happy to meet old and new friends here (Feel free to ping me over WeChat or email)!
English
55
22
1K
80.4K
Feng Li
Feng Li@FengLiust·
@xyz2maureen Huge congratulations xueyan! All the best on your new adventure.
English
1
0
2
106
Xueyan Zou
Xueyan Zou@xyz2maureen·
I will join Tsinghua University, College of AI, as an Assistant Professor in the coming month. I am actively looking for 2026 spring interns and future PhDs (ping me if you are in #NeurIPS). It has been an incredible journey of 10 years since I attended an activity organized by Tsinghua University and decided to change my undergraduate major from Economics to Computer Science, inspired by one of the teammates. During the 10 years, I met with appreciation of many wonderful researchers/professors who led me to continued growth. 🐿️ My research focus will continue to be AI & Robotics, with a specific emphasis on Interactive Embodied Intelligence. You can check my homepage to learn more: maureenzou.github.io/lab.html. I am currently local to San Diego and will be attending #NeurIPS. Please ping me over WeChat or Email if any old or new friends are interested in having a coffee chat! (Really looking forward to meeting as many friends as possible at #NeurIPS) [The photo is one of the places that I will miss a lot in the US]
Xueyan Zou tweet media
English
69
87
1.1K
111.4K
Feng Li retweetledi
Yongyuan Liang
Yongyuan Liang@cheryyun_l·
Unified multimodal models can generate text and images, but can they truly reason across modalities? 🎨 Introducing ROVER, the first benchmark that evaluates reciprocal cross-modal reasoning in unified models, the next frontier of omnimodal intelligence. 🌐 Project: roverbench.github.io 📄 Paper: arxiv.org/abs/2511.01163 📂 Benchmark: huggingface.co/datasets/chery…
English
6
29
193
74.1K
Jianwei Yang
Jianwei Yang@jw2yang4ai·
Life Update: Now that I have finished the presentation of last @MSFTResearch project Magma at @CVPR, I am excited to share that I have joined @AIatMeta as a research scientist to further push forward the boundary of multimodal foundation models! I have always been passionate about building multimodal AI systems that can interact with human and environments. In the past five years, I am very fortunate to lead and contribute to a number of exciting projects on (a) vision and multimodal foundation: Focal Attention, FocalNet, UniCL, RegionCLIP, GLIP, Florence;  (b) generalist multimodal vision models: X-Decoder, SEEM, Semantic-SAM and Grounding-DINO; (c) multimodal large language models: LLaVA variants, Phi-3-Vision; (d) multimodal agentic model: SoM + OmniParser, TraceVLA, LAPA and Magma. I’m also very proud to have contributed to impactful projects, such as LLaVA-Med, GigaPath and BiomedParse, advancing AI for healthcare and human good. These are truly meaningful footnotes in my journey at MSR! In the past five years while I am staying at MSR, the world has witnessed tremendous breakthroughs in AI as well as those brought by AI. Looking ahead, the opportunity to advance AI research for a better world has never been so exciting. I feel so lucky to be part of this. Now right after five years, it feels like the right time for me to "graduate" from MSR and embrace new challenges beyond! Thank you all again for the support, mentorship, and friendship!
Jianwei Yang tweet media
English
52
6
384
30K
Feng Li retweetledi
Haoqi Fan
Haoqi Fan@HaoqiFan·
🚀 BAGEL — the Unified Multimodal Model with emergent capabilities and production-ready performance — is finally live! Dive in here: 👉 bagel-ai.org
English
8
23
129
18.2K
Feng Li retweetledi
AK
AK@_akhaliq·
ByteDance just dropped BAGEL on Hugging Face The Open-Source Unified Multimodal Model
English
29
237
1.5K
170.9K
Feng Li
Feng Li@FengLiust·
We're happy to release BAGEL, a powerful unified model: 🔓 Open-source: weights & code 🧠 Unified MoT arch for text, image, & video 🏆 Top performance on 10+ benchmarks ✨ Emerging properties: From basic T2I/I2T to intelligent manipulation ⚡ Try it now: bagel-ai.org
Feng Li tweet media
English
1
1
14
1.7K
Feng Li
Feng Li@FengLiust·
I will be at #eccv2024 from today. Happy to chat about multimodal, classic computer vision, or anything! Our papers will be presented: - Grounding DINO - LLaVA-Plus - Semantic-SAM - TAPTR: Tracking Any Point - T-Rex2: Text-Visual Prompted Detector - LLaVA-Grounding
English
3
4
99
9.7K