DeepThink Lab

25 posts

DeepThink Lab banner
DeepThink Lab

DeepThink Lab

@DeepthinkUmich

Ann Arbor Katılım Aralık 2025
162 Takip Edilen18 Takipçiler
DeepThink Lab retweetledi
Siyi Chen
Siyi Chen@ChenSiyich·
🎉 Accepted to #CVPR2026 🔎 VLMs fall short on complex spatial reasoning. They struggle with: • Precise geometric perception • Multi-step reasoning grounded in 3D • Adapting perception dynamically to task and context 🚀 We propose a solution: visual tool-augmented spatial reasoning — bridging perception and multi-step reasoning through diverse, error-aware, adaptive vision tool use. And we go one step further: 🤖 enabling robot control by treating robots themselves as tools. Our framework is powered by: ⚡ Double Interactive RL (DIRL), a new training framework combining demonstrations + real exploration 🛠 Real interaction with specialized computer vision models during RL 🤖 Toolshed, a scalable, asynchronous system for multimodal execution of vision tools and robots-as-tools 🔗 Project: spacetools.github.io Code: github.com/spacetools/Spa… Toolshed is released with frontier-model demos. Full training & evaluation release coming soon. Done during my internship @NVIDIA — big thanks to the amazing collaborators! 🙌 #CVPR2026 #EmbodiedAI #ComputerVision #Robotics #ReinforcementLearning #MultimodalAI
English
2
19
160
14.9K
DeepThink Lab
DeepThink Lab@DeepthinkUmich·
@AndrewLampinen Wow! Understanding mem/gen through the implicit bias or 'abilities' of networks is truly exciting! In our ICLR2026 paper (arxiv.org/abs/2512.20963), we prove diffusion models also generalize when learning structures from data and memorize when they store&match training samples.
English
0
1
3
449
Andrew Lampinen
Andrew Lampinen@AndrewLampinen·
What is the relationship between memorization and generalization in AI? Is there a fundamental tradeoff? In a new blog post I’ve reviewed some of the evolving perspectives on memorization & generalization in machine learning, from classic perspectives through LLMs. Link below:
Andrew Lampinen tweet media
English
12
44
426
23.5K
DeepThink Lab
DeepThink Lab@DeepthinkUmich·
Experiments indicate our results hold in nonlinear Transformers (GPT-2) and nonlinear function classes, which is beyond our simplified theoretical setting. Link to arXiv version: arxiv.org/abs/2505.14808
English
1
0
0
70
DeepThink Lab
DeepThink Lab@DeepthinkUmich·
We’re excited to share our paper, “Out-of-Distribution Generalization of In-Context Learning: A Low-Dimensional Subspace Perspective,” has been accepted to AISTATS 2026! Congratulations, Soo Min and Alec!
English
1
3
6
443
DeepThink Lab
DeepThink Lab@DeepthinkUmich·
Experiments on synthetic and real data show similar conclusions hold beyond our theoretical setting. Link to arXiv version: arxiv.org/abs/2501.02364
English
0
0
0
54
DeepThink Lab
DeepThink Lab@DeepthinkUmich·
Notably, our result states a network width that is polynomial with the intrinsic dimension suffices, bridging the gap between previous theoretical guarantees and empirical observations.
English
1
0
0
47
DeepThink Lab
DeepThink Lab@DeepthinkUmich·
We’re excited to share our paper, “Linearly Separable Features in Shallow Nonlinear Networks: Width Scales Polynomially with Intrinsic Data Dimension,” has been accepted to AISTATS 2026! Congratulations, Alec!
English
1
1
7
405
DeepThink Lab
DeepThink Lab@DeepthinkUmich·
Let’s congratulate Dr. Li on this tremendous milestone! We are incredibly proud of his achievements and wish him all the very best as he begins this new chapter at The University of Hong Kong.
English
0
0
0
28
DeepThink Lab
DeepThink Lab@DeepthinkUmich·
Dr. Li is the first PhD graduate of our lab and one of the most dedicated, charismatic, and resilient students we have had. Over the past five years, Xiao has been an integral part of our DeepThink Lab, contributing significantly to our work on representation learning.
English
1
0
1
41
DeepThink Lab
DeepThink Lab@DeepthinkUmich·
We are delighted to announce that Xiao Li has successfully defended his PhD dissertation. Congratulations, Dr. Li! 🎓
DeepThink Lab tweet mediaDeepThink Lab tweet mediaDeepThink Lab tweet mediaDeepThink Lab tweet media
English
1
0
3
51
DeepThink Lab
DeepThink Lab@DeepthinkUmich·
🔷 Learned representations can indicate whether the model is learning underlying data structures (with balanced, informative representations) or memorizing training data (with spiky representations). We elaborate below. Read the full paper on arXiv: arxiv.org/abs/2512.20963
English
1
1
3
142
DeepThink Lab
DeepThink Lab@DeepthinkUmich·
💥 We’re excited to share our DeepThink Lab's "year-end" work on the generalization of diffusion models by opening the black box of their neural network backbones.
English
1
0
3
83