Jianing “Jed” Yang

179 posts

Jianing “Jed” Yang banner
Jianing “Jed” Yang

Jianing “Jed” Yang

@jed_yang

Building robots @Figure_robot | PhD 🎓 @UMich on 3D vision. Prev. @Meta @Adobe @CarnegieMellon @GeorgiaTech.

Santa Clara, CA Katılım Ağustos 2012
1.3K Takip Edilen1K Takipçiler
Jianing “Jed” Yang retweetledi
Yinpei Dai
Yinpei Dai@YinpeiD·
🚀 The RoboMME Challenge @ CVPR 2026 is now LIVE! Timeline: • May 15 — Policy submission • June 3 — Winner announcement 🏆 Top 3 teams will be awarded $500/300/200 Let’s push the frontier for memory-augmented robotic manipulation together 💪 🔗robomme.github.io/challenge.html
Yinpei Dai tweet media
Yinpei Dai@YinpeiD

Robot memory methods are growing fast, but systematic evaluation is largely lacking. 📉 Introducing RoboMME: a new benchmark for memory-augmented robotic manipulation! 🤖🧠 Featuring 16 tasks across temporal, spatial, object, and procedural memory 🔗 robomme.github.io

English
1
5
6
458
Cheng Chi
Cheng Chi@chichengcc·
We're done demoing. Time to deploy. Robotics isn't just algorithms. It's perf traces, calibrations, vibe ingress ESD tests. It's about making the sensible decision and pragmatic tradeoffs, again and again and again. ...
Tony Zhao@tonyzzhao

We raised $165M at a $1.15B valuation to stop doing demos. 2026 is about 1) deployment and 2) research. We will start shipping Memo with our new frontier models in a few months. Our series-B is led by Coatue, with Thomas Laffont joining the board. ->🧵

English
16
20
257
24.6K
Sasha Sax
Sasha Sax@iamsashasax·
In a couple weeks I'm joining @AnthropicAI to work on pretraining after nearly 3 years at FAIR, developing post-training flywheels for physical intelligence (like SAM 3D) I'm stoked to build new capabilities for a model I personally love, with such thoughtful people
English
35
9
643
26.4K
Jianing “Jed” Yang retweetledi
Figure
Figure@Figure_robot·
Today we're showing Helix 02 that can tidy a living room fully autonomously Figure is designed so when you leave the house, your home resets exactly how you like it
English
693
1.2K
9.3K
2M
Jianing “Jed” Yang retweetledi
Corey Lynch
Corey Lynch@coreylynch·
Helix 02 tidies a living room. 100% autonomous, 1x speed, no teleop.
English
38
31
393
69K
Jianing “Jed” Yang
Jianing “Jed” Yang@jed_yang·
Memory is one of the most important topic in AI right now. But memory for robot is very under-explored. A benchmark is a first step to start measuring progress scientifically. Hope to see more work on this line!
Yinpei Dai@YinpeiD

Robot memory methods are growing fast, but systematic evaluation is largely lacking. 📉 Introducing RoboMME: a new benchmark for memory-augmented robotic manipulation! 🤖🧠 Featuring 16 tasks across temporal, spatial, object, and procedural memory 🔗 robomme.github.io

English
0
0
6
547
Jianing “Jed” Yang
Jianing “Jed” Yang@jed_yang·
If you do research on spatial intelligence, robotics, or language grounding, join us at CVPR this June! An amazing line up of speakers across vision, robotics, and generative models! Consider submitting your papers to the 2nd 3D-LLM/VLA workshop.
Yining Hong@yining_hong

LLMs are now learning space, geometry, and how to move. 🤖📐 The 2nd CVPR 3D-LLM VLA Workshop brings together language, 3D perception, and action for embodied intelligence. 📢 Call for Papers is OPEN: #tab-your-consoles" target="_blank" rel="nofollow noopener">openreview.net/group?id=thecv…
🌐 Website: 3d-llm-vla.github.io If your research lives at the intersection of words, worlds, and robots—this one’s for you. #CVPR2026 @CVPR

English
0
0
2
374
Jianing “Jed” Yang retweetledi
Figure
Figure@Figure_robot·
Introducing Helix 02 It's our most powerful model to date - it's using the whole body to do dishes end-to-end and it's fully autonomous
English
368
774
5K
1.1M
Jianing “Jed” Yang
Jianing “Jed” Yang@jed_yang·
3D pretraining from 2D images only. This would unblock the data scarcity issue in 3D vision!
Hanwen Jiang@hanwenjiang1

(1/N) Will this be the BERT/GPT moment for 3D vision? Finally, unsupervised pre-training for 3D works. Led by @qitao_zhao , we present E-RayZer — a fully self-supervised 3D reconstruction model that: 🔥Matches or surpasses supervised methods like VGGT 👀Learns transferable 3D representations, outperforming CroCo, VideoMAE, and DINO 📈Scales with more unlabeled data A new recipe for scalable 3D foundation models.

English
0
0
2
819
Jianing “Jed” Yang
Jianing “Jed” Yang@jed_yang·
SAM 3D launched 🚀! 3Dfy anything and everything in an image! What made it work? tldr: Model-in-the-loop data flywheel for the win! This was a project I contributed to during my internship at Meta this summer. Incredible team to work and grind with! ❤️ Proud to be part of it!
AI at Meta@AIatMeta

Introducing SAM 3D, the newest addition to the SAM collection, bringing common sense 3D understanding of everyday images. SAM 3D includes two models: 🛋️ SAM 3D Objects for object and scene reconstruction 🧑‍🤝‍🧑 SAM 3D Body for human pose and shape estimation Both models achieve state-of-the-art performance transforming static 2D images into vivid, accurate reconstructions. 🔗 Learn more: go.meta.me/305985

English
0
0
8
466
Jianing “Jed” Yang retweetledi
Bilawal Sidhu
Bilawal Sidhu@bilawalsidhu·
Meta just dropped SAM 3D, but more interestingly, they basically cracked the 3D data bottleneck that's been holding the field back for years. Manually creating or scanning 3D ground truth for the messy real world is basically impossible at scale. But what if you just have humans rank model outputs? Route the weird edge cases to actual 3D artists to model, loop it back in. Suddenly you can annotate like a million images. It's basically RLHF for 3D reconstruction. Synthetic data is pretraining, real world ranking is alignment. They borrowed the whole damn playbook and it actually works. Two models - one for objects/scenes, one for humans. They're already shipping it in FB Marketplace so you can see if that lamp or chair looks good in your room before buying. Also they're releasing everything - models, code, their human body rig under commercial license. And they built an eval set of actual messy real-world images to help bridge the sim-to-real gap. The data engine thing is the most interesting though. 3D has been bottlenecked by ground truth forever. If verification scales easier than creation, suddenly the whole game changes.
English
24
167
1.2K
111.8K
Jianing “Jed” Yang
Jianing “Jed” Yang@jed_yang·
I joined Figure 4 days ago. Everyday I walk into the office, it feels like walking into a sci-fi movie. Robots work, humans build, machines hum. 3D printers sculpt, CNCs carve, actuators roar—it’s Iron Man’s lab, but real.
Figure@Figure_robot

Introducing Figure 03

English
4
0
20
780
Jianing “Jed” Yang retweetledi
Figure
Figure@Figure_robot·
Today we unveiled the first humanoid robot that can fold laundry autonomously Same exact Helix architecture, only new data
English
360
800
5.2K
761.4K
Jianing “Jed” Yang retweetledi
Sukjun (June) Hwang
Sukjun (June) Hwang@sukjun_hwang·
Tokenization has been the final barrier to truly end-to-end language models. We developed the H-Net: a hierarchical network that replaces tokenization with a dynamic chunking process directly inside the model, automatically discovering and operating over meaningful units of data
GIF
GIF
English
95
732
4.7K
786.1K
Jianing “Jed” Yang retweetledi
Martin Ziqiao Ma
Martin Ziqiao Ma@ziqiao_ma·
Can we scale 4D pretraining to learn general space-time representations that reconstruct an object from a few views at any time to any view at any other time? Introducing 4D-LRM: a Large Space-Time Reconstruction Model that ... 🔹 Predicts 4D Gaussian primitives directly from multi-view tokens (no motion vectors, no HexPlane); 🔹 Uses a clean, minimal Transformer backbone; 🔹 Generalizes with fast, high-quality feedforward rendering at any view and infinite frame rate. Check out more interactive demos and scaling behaviors on our homepage/paper. 👉Website: 4dlrm.github.io 👉Paper: arxiv.org/abs/2506.18890
English
1
35
100
21.7K