Mohamed El Banani

207 posts

Mohamed El Banani banner
Mohamed El Banani

Mohamed El Banani

@_mbanani

MTS @theworldlabs. Prev: @UMichCSE, @GoogleAI, @MetaAI, @GeorgiaTech. I am interested in computer vision, machine learning, and cognitive science. 🇪🇬

San Francisco, CA Katılım Haziran 2020
778 Takip Edilen921 Takipçiler
Sabitlenmiş Tweet
Mohamed El Banani
Mohamed El Banani@_mbanani·
We’re finally sharing what we’ve been up to @theworldlabs! This is the first step on our journey to build fully interactive and immersive worlds that allow you to bring your creativity to life. Check out the demos, my favorites are the Kandinsky landscape and Van Gogh terrace.
World Labs@theworldlabs

We’ve been busy building an AI system to generate 3D worlds from a single image. Check out some early results on our site, where you can interact with our scenes directly in the browser! worldlabs.ai/blog 1/n

English
1
2
29
2.8K
Mohamed El Banani retweetledi
World Labs
World Labs@theworldlabs·
We're excited to be rolling out two model updates today! Marble 1.1: Improves lighting and contrast, with a major reduction in visual artifacts. Marble 1.1-Plus: Our new model built for scale. Create larger, more complex environments than ever before.
English
40
169
1.2K
172.4K
Mohamed El Banani retweetledi
World Labs
World Labs@theworldlabs·
Introducing Marble by World Labs: a foundation for a spatially intelligent future. Create your world at marble.worldlabs.ai
English
360
607
3.3K
2.1M
Mohamed El Banani retweetledi
Keunhong Park
Keunhong Park@KeunhongP·
Today we announced RTFM (Real-Time Frame Model) — a world model that generates frames in real time from any camera viewpoint. Unlike standard video models, RTFM understands 3D geometry. You can literally move the camera through the generated world. 🎥👇
English
11
22
233
15.5K
Mohamed El Banani
Mohamed El Banani@_mbanani·
We've been exploring different ways of modeling the world at @theworldlabs. This direction combines our real-time learned renderer with posed frames as a persistent spatial memory. I am excited to see where we go next! Check out the demo here: rtfm.worldlabs.ai
World Labs@theworldlabs

Introducing RTFM (Real-Time Frame Model): a highly efficient World Model that generates video frames in real time as you interact with it, powered by a single H100 GPU. RTFM renders persistent and 3D consistent worlds, both real and imaginary. Try our demo of RTFM today!

English
2
1
23
2.1K
Mohamed El Banani retweetledi
World Labs
World Labs@theworldlabs·
Generate persistent 3D worlds from a single image, bigger and better than ever! We’re excited to share our latest results and invite you to try out our world generation model in a limited beta preview.
English
205
509
3.6K
832.5K
Mohamed El Banani retweetledi
Ian Curtis
Ian Curtis@XRarchitect·
Experimenting with image-to-world: creating consistent 3D environments from a single photo. As someone who studied architecture, this feels like a glimpse into the future of spatial creation and I’m excited to see where World Labs takes it.
English
30
97
988
61.6K
Mohamed El Banani retweetledi
Chris Rockwell
Chris Rockwell@_crockwell·
Ever wish YouTube had 3D labels? 🚀Introducing🎥DynPose-100K🎥, an Internet-scale collection of diverse videos annotated with camera pose! Applications include camera-controlled video generation🤩and learned dynamic pose estimation😯 Download: huggingface.co/datasets/nvidi…
English
2
38
177
42.9K
Mohamed El Banani retweetledi
Sarah Jabbour
Sarah Jabbour@SarahJabbour_·
I’m on the PhD internship market for Spr/Summer 2025! I have experience in multimodal AI (EHR, X-ray, text), explainability for image models w/ genAI, clinician-AI interaction (surveyed 700+ doctors), and tabular foundation models. Please reach out if you think there’s a fit!
English
1
12
64
6.5K
Mohamed El Banani retweetledi
enigmatic_e
enigmatic_e@8bit_e·
Playing around with @theworldlabs' 3D generator and combining it with @runwayml vid2vid to blend myself and the background better. Although it's only a side perspective, you can tell it doesn't have that static image feel because of how the 3D environment is moving.
English
7
21
184
19.8K
Mohamed El Banani retweetledi
Heather Cooper
Heather Cooper@HBCoop_·
The future of immersive creation is here. World Labs image to 3D worlds: • Generate 3D worlds from a single image • Navigate scenes with real-time camera control • Use different effects without changing the scene I've had access and you can see some of my early results below: @theworldlabs
English
26
48
258
22K
Mohamed El Banani retweetledi
martin_casado
martin_casado@martin_casado·
I threw a couple of Mac Baconai's images in the new WorldLabs model. The scene completion is pretty darn good.
English
8
16
162
21.1K
Mohamed El Banani retweetledi
Brittani Natali
Brittani Natali@BrittaniNatali·
Generative 3D worlds are amazing! Here is a generative AI video showing fine grained camera control. Camera movements are all hand places within a generated 3D world.
English
5
25
122
21K
Mohamed El Banani retweetledi
enigmatic_e
enigmatic_e@8bit_e·
I had the opportunity to try out @theworldlabs' incredible 3D environment-generation model. This is still in the early stages. Check out some of the use cases I demonstrated!
World Labs@theworldlabs

@8bit_e shows how our models fill a gap in his creative workflow, making it easy to stage characters within scenes and direct precise camera movements. 8/n

English
9
12
94
12.9K
Mohamed El Banani retweetledi
World Labs
World Labs@theworldlabs·
We’ve been busy building an AI system to generate 3D worlds from a single image. Check out some early results on our site, where you can interact with our scenes directly in the browser! worldlabs.ai/blog 1/n
English
197
666
3K
851.4K
Mohamed El Banani retweetledi
Ben Mildenhall
Ben Mildenhall@BenMildenhall·
we’re hiring @theworldlabs seeking insanely great engineers + designers to work alongside our world-class research team to imagine and build entirely new apps and experiences made possible at the rapidly expanding frontier of generative AI + 3D computer vision + graphics
English
8
32
357
50.4K
Mohamed El Banani
Mohamed El Banani@_mbanani·
@tylerraye @abhinav1kumar I agree with @tylerraye. Linear probes don't seem to do well on MAE; eg, compare the linear probe vs. finetune perf. on ImageNet. Larger decoder probable helps MAE. Another difference is evaluating image-level vs. dense features where you would expect CLIP to do better as well.
English
1
0
3
67
tyler bonnen
tyler bonnen@tylerraye·
@abhinav1kumar i'd guess the ordering relates to different readouts: they use a dense multi-scale decoder while we're using a lightweight classifier. CLIPs are probably better suited for our lightweight readouts then MAEs. but curious what @_mbanani thinks about this!
English
1
0
4
130
Mohamed El Banani retweetledi
Keunhong Park
Keunhong Park@KeunhongP·
Happy to announce our company has come out of stealth! We are building spatial intelligence with some of the most talented researchers and engineers -- @ZhuLuyang, @ericryanchan, @kdexd, @_mbanani, @KyleSargentAI, @chaoyuaw, and many more. We are hiring so please reach out!
World Labs@theworldlabs

Hello, world! We are World Labs, a spatial intelligence company building Large World Models (LWMs) to perceive, generate, and interact with the 3D world. Read more: worldlabs.ai/about

English
4
12
87
20.1K