Yichen Sheng

149 posts

Yichen Sheng

Yichen Sheng

@Coding_Black

Research scientist in NVIDIA. Working in graphics and vision. Opinions are my own.

West Lafayette, IN शामिल हुए Temmuz 2015
1.6K फ़ॉलोइंग340 फ़ॉलोवर्स
पिन किया गया ट्वीट
Yichen Sheng
Yichen Sheng@Coding_Black·
We’re hiring research interns at @NVIDIA 🚀 You’ll work on interactive world models and explore how GenAI will shape the next generation gaming experience. Interns are encouraged to publish papers on top-tier conferences(CVPR/Siggraph/ICCV/ICLR,etc). If you are interested, send your CV to yisheng@nvidia.com #GenAI #DLSS #DLSS45 #Internships
NVIDIA GeForce@NVIDIAGeForce

Introducing DLSS 4.5: ⚫️Second-gen Super Resolution transformer model for all RTX GPUs ⚫️Dynamic Multi Frame Gen for RTX 50 Series GPUs in Spring '26 ⚫️6X Multi Frame Gen for RTX 50 Series GPUs in Spring '26 ⚫️DLSS Overrides in NVIDIA app Learn More → nvda.ws/4aMEo5h

English
0
0
3
1.5K
Yichen Sheng रीट्वीट किया
Johan Edstedt
Johan Edstedt @Parskatt·
Introducing LoMa, the next generation of feature matcher!
Johan Edstedt  tweet media
English
8
35
293
36.1K
Yichen Sheng रीट्वीट किया
Phota Labs
Phota Labs@PhotaLabs·
📒 Phota 101: Profile Setup covers best practices to help you get the most out of Phota Studio. Profiles are at the center of Phota. Built from your personal album, your identity models learn the details of your appearance so edits and generations across different contexts preserve your identity. Your photos and models are owned by you and are not used for any other model training.
English
2
5
10
2.5K
Yichen Sheng रीट्वीट किया
Oliver Mackenzie
Oliver Mackenzie@oliemack·
Here's some of the DLSS 5 material we saw in the demos but didn't get a chance to film. Here I think you can see the strengths of DLSS 5 - reflections become much more attractive. Starfield doesn't have great lighting to begin with, so the differences can be profound.
English
190
56
960
171.7K
Yiqun Mei
Yiqun Mei@myq_1997·
This is so cool. Looks like they run a generative model doing synthetic 2real. But I am more curious about is this effect deterministic? If we see a same character twice, will it look the same?🤣
NVIDIA GeForce UK@NVIDIAGeForceUK

Announcing NVIDIA DLSS 5, an AI-powered breakthrough in visual fidelity for games, coming this fall. DLSS 5 infuses pixels with photorealistic lighting and materials, bridging the gap between rendering and reality.

English
1
0
2
230
Yichen Sheng
Yichen Sheng@Coding_Black·
@CVPR What does accept and suggest to finding mean? It means accept only to finding workshop or accept to main conference and also get suggested to finding workshop?
English
3
0
5
2.9K
#CVPR2026
#CVPR2026@CVPR·
#CVPR2026 final decisions are out! Available for now only via email. Good luck🤞
#CVPR2026 tweet media
English
16
9
146
43.7K
Yichen Sheng
Yichen Sheng@Coding_Black·
Make your agents see the world 👀 to solve 3D problems. In this path, many practical technical problems need to be solved. @LuLing26466911 solves a lot of nitty-gritty problems in the spatial optimization. Great job to @LuLing26466911 and the team!
Lu Ling@LuLing26466911

🎉 **Scenethesis** has been accepted to #ICLR2026 ! Agentic systems are everywhere right now—coding agents, robotics agents, tool-using agents #moltbook. Back around two years ago when #OpenClaw , #NanoBananaPro have not arrived, we asked: can an agentic workflow *build simulation-ready 3D worlds* from text prompt? Interactive 3D scene generation isn’t just “generate some assets”. The hard part is spatial intelligence: ✅ spatial realism ✅ support & affordances ✅ physically plausible, editable, interactive layouts Scenethesis is a *training-free* language + vision agentic framework that: - 👁Let agents see: The planner doesn’t operate blind: it gets visual feedback and can correct itself. - 🏠Go outdoor: Works beyond indoor rooms; handles more open, outdoor-style compositions. check our work at: Paper: arxiv.org/abs/2505.02836 Project page: research.nvidia.com/labs/dir/scene… #ICLR2026 #Agents #GENERATION #physicalai #spatialintelligence

English
0
0
4
270
Yichen Sheng रीट्वीट किया
Andrew Tao
Andrew Tao@drewtao·
We're on a mission to build the best open-source, open-data multi-modal LLMs. From Document understanding to Visual Agent and many more domains. With the recent release of Nemotron Nano V3 LLM, you can guess what's next. We're hiring! nvidia.wd5.myworkdayjobs.com/en-US/NVIDIAEx…
English
2
2
11
558
Yichen Sheng रीट्वीट किया
Bryan Catanzaro
Bryan Catanzaro@ctnzr·
The new 2nd gen transformer super resolution model in DLSS 4.5 is a big step forward, especially for Performance mode. nvidia.com/en-us/geforce/…
English
2
3
19
2K
Yichen Sheng
Yichen Sheng@Coding_Black·
See @LuLing26466911's thread here: x.com/LuLing26466911… It includes SAM3D's teasers and real-world images as testing cases. To clarify a little bit, I-Scene's strength is scene layout quality, while SAM3D is very good at instance quality. But it is not unsolvable problem for I-Scene. It is more related to the backbone's resolution. #TRELLIS2 has much higher resolution. And TRELLIS2-based I-Scene will highly likely no longer suffer from instance quality problem any more.
Lu Ling@LuLing26466911

Limitations & Discussion I-Scene achieves better layout than SAM-3D. 🔧 while instance quality can be futher improved: • Instance upscaling networks • Higher-resolution voxel spaces 🚀 Good news: #TRELLIS2 supports much higher resolution (1024³ vs. our 64³). 🌱 Takeaway: We identify a cheap, scalable path to training generalizable 3D scene models — a step toward immersive world-generation foundation models. 🧵 [6/6]

English
0
0
2
181
Yawar Siddiqui
Yawar Siddiqui@yawarnihal·
Cool work, but the "better than SAM3D" claim needs more backing. Would love to see a comparison on realistic clutter (vs. the clean scenes as shown in the paper) and a direct benchmark against SAM3D.
Yichen Sheng@Coding_Black

If you work in #3D generation or #worldmodel, definitely take a look at this: I-Scene is an image to 3D scene model, achieves better 3D scene-gen than #SAM3D just using limited amount of random dataset.

English
1
0
0
803
Yichen Sheng
Yichen Sheng@Coding_Black·
I gradually start to believe world model is closer than we thought. Maybe this is the correct way for unsupervised large-scale pretraining in 3D. Cannot wait to see the 3D GPT moment.
English
0
0
2
154
Yichen Sheng
Yichen Sheng@Coding_Black·
Similar to @hanwenjiang1's MegaSynth and RayZer series of work, 2D and 3D is just a camera projection relationship, non-semantic random data is under-explored in our community.
English
1
0
3
165
Yichen Sheng
Yichen Sheng@Coding_Black·
If you work in #3D generation or #worldmodel, definitely take a look at this: I-Scene is an image to 3D scene model, achieves better 3D scene-gen than #SAM3D just using limited amount of random dataset.
Yichen Sheng tweet media
Lu Ling@LuLing26466911

Do we really need massive curated 3D scene data for interactive world generation? #SAM3D, #WorldGen say yes. We say no. I-Scene learns better spatial knowlesge using only 25K randomly composed instances. 🔑 Key insight: We reprogram the instance generator to infer support, proximity, and symmetry from purely geometric cues for generating interactive scenes. 🧠 Scene-context attention 👁️ View-centric space 🧱 Random composition beats expensive curation 🌐 luling06.github.io/I-Scene-projec… 💻 github.com/LuLing06/I-Sce… 🧵 Details below [1/6]

English
1
0
5
1.4K
Yichen Sheng
Yichen Sheng@Coding_Black·
Join us in pioneering research that will revolutionize the next generation of graphics experiences. You’ll be working with a strong team that has an exceptional record of success.
Edward Liu@edliu1105

🚀 NVIDIA hiring Research Scientist (all levels). GenAI for graphics/gaming: neural rendering, world models, real-time generation, AI characters. If you like turning research into products like DLSS used by millions, great fit. Apply: nvidia.wd5.myworkdayjobs.com/en-US/NVIDIAEx… #Hiring #GenerativeAI

English
0
0
1
213
Yichen Sheng
Yichen Sheng@Coding_Black·
@YiMaTweets It's not a bad thing from my perspective. You can always do things first and learn the missing knowledge along the way. I strongly disagree traditional Chinese education mindset: you have to learn all the basics and then do research. This is one of the most terrible idea.
English
0
0
3
236
Yi Ma
Yi Ma@YiMaTweets·
I have met many students and young researchers lately who claim to be working on World Models or Embodied AI but do not even know the basics of 3D Vision or linked rigid body motions. When did we start to give students the illusion that they can *do* things right without *learning* anything right?
English
84
65
1.3K
243.4K