Ronghang Hu

29 posts

Ronghang Hu banner
Ronghang Hu

Ronghang Hu

@RonghangHu

multimodal at @xAI | previously at Meta FAIR

Palo Alto, CA Katılım Ağustos 2014
2.1K Takip Edilen1.1K Takipçiler
Ronghang Hu retweetledi
Pengchuan Zhang
Pengchuan Zhang@PengchuanZ·
SAM 3.1 is here. SAM 3.1 is a significant improvement on video grounding: shifting from SAM2-like per-object propagation to MOT-style multi-object propagation. This change significantly reduces the computational cost, while maintaining the tracking accuracy!!!
AI at Meta@AIatMeta

We’re releasing SAM 3.1: a drop-in update to SAM 3 that introduces object multiplexing to significantly improve video processing efficiency without sacrificing accuracy. We’re sharing this update with the community to help make high-performance applications feasible on smaller, more accessible hardware. 🔗 Model Checkpoint: go.meta.me/8dd321 🔗 Codebase: go.meta.me/b0a9fb

English
3
7
40
6.2K
Ronghang Hu retweetledi
Xudong Lin
Xudong Lin@Xudong_Lin_AI·
Proud of our team that makes the huge leap happen compared to last version but this is just the start. Better models are lined up and we keep improving every week. Join us towards Superhuman Multimodal Intelligence job-boards.greenhouse.io/xai/jobs/50826… !!
Arena.ai@arena

Grok 4.20 Beta Reasoning makes @xAI a top 5 lab in Vision Arena. Scoring 1240, this model ranks #11 across all Vision models today. Congrats to the @xAI team for this milestone!

English
13
33
276
60.5K
Ronghang Hu retweetledi
xAI
xAI@xai·
Understanding requires imagining. Grok Imagine lets you bring what’s in your brain to life, and now it’s available via the world’s fastest, and most powerful video API: x.ai/news/grok-imag… Try it out and let your Imagination run wild.
English
522
672
4.4K
6.8M
Ronghang Hu retweetledi
Nikhila Ravi
Nikhila Ravi@nikhilaravi·
One of the most highly requested features since we launched SAM 1 was the ability to prompt with text! @kate_saenko_ from SAM 3 team explains how we built an efficient data engine to collect high quality mask + text label annotations at scale and our new open vocabulary benchmark Segment Anything with Concepts (SA-Co).
AI at Meta@AIatMeta

Collecting a high quality dataset with 4M unique phrases and 52M corresponding object masks helped SAM 3 achieve 2x the performance of baseline models. Kate, a researcher on SAM 3, explains how the data engine made this leap possible. 🔗 Read the SAM 3 research paper: go.meta.me/6411f7

English
0
4
38
6.2K
Ronghang Hu retweetledi
Niels Rogge
Niels Rogge@NielsRogge·
SAM-3 is out on @huggingface! A big upgrade from SAM-2, and Meta finally added support for text prompts. Here I tried it out on @hazardeden10's magical goal against @Arsenal using the text prompt "Chelsea player" Works pretty well!
English
8
34
345
32K
Ronghang Hu retweetledi
Alexandr Wang
Alexandr Wang@alexandr_wang·
Today we are releasing & open-sourcing Segment Anything 3 (SAM 3). It is a state-of-the-art model for image & video segmentation, and builds upon the work of SAM & SAM 2. SAM3 will also power features in Edits, Meta AI, & Facebook Marketplace soon. aidemos.meta.com/segment-anythi…
English
189
205
2.7K
295.5K
Ronghang Hu retweetledi
AI at Meta
AI at Meta@AIatMeta·
Meet SAM 3, a unified model that enables detection, segmentation, and tracking of objects across images and videos. SAM 3 introduces some of our most highly requested features like text and exemplar prompts to segment all objects of a target category. Learnings from SAM 3 will help power new features in Instagram Edits and Vibes, bringing advanced segmentation capabilities directly to creators. 🔗 Learn more: go.meta.me/591040
English
26
144
953
187.1K
Ronghang Hu retweetledi
Nikhila Ravi
Nikhila Ravi@nikhilaravi·
🚀 Excited to announce new SAM 2.1 model checkpoints & the SAM 2 Developer Suite: 🤖 We’re releasing full training/fine tuning code for SAM 2 so you can customize it for your use case. 💻For the first time we’re publishing the frontend & backend code for our SAM 2 web demo!
AI at Meta@AIatMeta

We’re on the ground at #ECCV2024 in Milan this week to showcase some of our latest research, new research artifacts and more. Here are 4️⃣ things you won’t want to miss from Meta FAIR, GenAI and Reality Labs Research this week whether you’re here in person or following from your feed. 1️⃣ We’re releasing SAM 2.1 an upgraded version of the Segment Anything Model 2 — and the SAM 2 Developer Suite featuring open source tools for training, inference and demos. New artifacts are live in the repo on GitHub ➡️ go.fb.me/mk6ofh 2️⃣ We’re supporting 10+ presentations and workshops in areas like computer vision for smart glasses and the metaverse, 3D vision for eCommerce, egocentric research with Project Aria and more. 3️⃣ We’re presenting seven orals at ECCV — in addition to the 50+ publications from researchers at Meta that were accepted for this year’s conference. Look out for more details on some of these papers later this week. 4️⃣ Demos and discussions with Meta researchers at our booth all week — come by our booth to discuss projects like SAM 2, Ego-Exo4D, DINOv2 and more.

English
5
23
216
24.1K
Ronghang Hu
Ronghang Hu@RonghangHu·
Below are what's contained in the SAM 2.1 Developer Suite: - A new suite of improved model checkpoints (denoted as SAM 2.1) are released. - The training (and fine-tuning) code has been released. - The frontend + backend code for the SAM 2 web demo has been released.
English
0
1
5
347
Ronghang Hu
Ronghang Hu@RonghangHu·
SAM 2.1 Developer Suite (new checkpoints, training code, web demo) is released -- check it out at github.com/facebookresear…
AI at Meta@AIatMeta

We’re on the ground at #ECCV2024 in Milan this week to showcase some of our latest research, new research artifacts and more. Here are 4️⃣ things you won’t want to miss from Meta FAIR, GenAI and Reality Labs Research this week whether you’re here in person or following from your feed. 1️⃣ We’re releasing SAM 2.1 an upgraded version of the Segment Anything Model 2 — and the SAM 2 Developer Suite featuring open source tools for training, inference and demos. New artifacts are live in the repo on GitHub ➡️ go.fb.me/mk6ofh 2️⃣ We’re supporting 10+ presentations and workshops in areas like computer vision for smart glasses and the metaverse, 3D vision for eCommerce, egocentric research with Project Aria and more. 3️⃣ We’re presenting seven orals at ECCV — in addition to the 50+ publications from researchers at Meta that were accepted for this year’s conference. Look out for more details on some of these papers later this week. 4️⃣ Demos and discussions with Meta researchers at our booth all week — come by our booth to discuss projects like SAM 2, Ego-Exo4D, DINOv2 and more.

English
1
5
11
2.1K
Ronghang Hu retweetledi
AI at Meta
AI at Meta@AIatMeta·
Today, we’re sharing a roundup of Meta AI’s recent cutting-edge multimodal research, which we believe will collectively lead to more interactive, immersive, and smarter AI systems of the future: ai.facebook.com/blog/advances-…
AI at Meta tweet media
English
6
52
234
0
Ronghang Hu retweetledi
Deepak Pathak
Deepak Pathak@pathak2206·
We are presenting Worldsheet at #ICCV2021 this week as Oral. Join QnA Wed & Fri. We updated the arXiv since v1: *Multi-layered* Worldsheets to autonomously handle sharp depth discontinuities/occlusions which a single sheet may fail to capture (Sec 3.5): worldsheet.github.io/resources/worl…
Deepak Pathak@pathak2206

Excited to share Worldsheet, a method to synthesize novel views with large camera changes from a *single* image. Turns out, simply shrink-wrapping a mesh sheet onto the image captures 3D well enough to render photorealistic far-away views. w/ @RonghangHu worldsheet.github.io

English
0
4
26
0
Ronghang Hu retweetledi
Xin Eric Wang
Xin Eric Wang@xwang_lk·
The @NAACLHLT ALVR 2021 workshop is happening tomorrow (June 11th) at 8:30am PDT!!! We have an amazing lineup of speakers to talk about recent advances in language and vision research! Talk information also provided in the program. alvr-workshop.github.io
Xin Eric Wang tweet mediaXin Eric Wang tweet media
Daniel Fried@dan_fried

Looking forward to the lineup of talks, papers, and panel discussions at tomorrow (Friday)'s ALVR workshop on language grounding (images, video, embodied control...): Program: #program" target="_blank" rel="nofollow noopener">alvr-workshop.github.io/#program Papers: #accepted-papers" target="_blank" rel="nofollow noopener">alvr-workshop.github.io/#accepted-pape… Zoom link: underline.io/events/122/ses… #NAACL2021

English
1
3
25
0
Ronghang Hu retweetledi
Wojciech Galuba
Wojciech Galuba@wgaluba·
Facebook AI Research is looking for Research Engineers to join our Embodied AI team in Menlo Park. More about what we do in the thread below. If this sounds exciting, DM me or apply here: facebook.com/careers/v2/job… 1/6
English
2
14
58
0