CMU Center for Perceptual Computing and Learning

649 posts

CMU Center for Perceptual Computing and Learning banner
CMU Center for Perceptual Computing and Learning

CMU Center for Perceptual Computing and Learning

@roboVisionCMU

The Chronicles of Smith Hall.

Pittsburgh, PA Katılım Haziran 2019
136 Takip Edilen2K Takipçiler
CMU Center for Perceptual Computing and Learning retweetledi
CMU Center for Perceptual Computing and Learning
TWO Best Paper Awards at ICCV Generating Physically Stable and Buildable Brick Structures from Text Ava Pun*, Kangle Deng*, Ruixuan Liu*, Deva Ramanan, Changliu Liu, Jun-Yan Zhu Spatially-Varying Autofocus Yingsi Qin, Aswin C. Sankaranarayanan, Matthew O'Toole #goSmithHall
English
0
0
13
696
CMU Center for Perceptual Computing and Learning retweetledi
Yishu Li
Yishu Li@LisaYishu·
A closed door looks the same whether it pushes or pulls. Two identical-looking boxes might have different center of mass. How should robots act when a single visual observation isn't enough? Introducing HAVE 🤖, our method that reasons about past interactions online! #CORL2025
Yishu Li tweet media
English
1
19
42
7.5K
CMU Center for Perceptual Computing and Learning retweetledi
Unnat Jain
Unnat Jain@unnatjain2010·
✨New edition of our community-building workshop series!✨ Tomorrow at @CVPR, we invite speakers to share their stories, values, and approaches for navigating a crowded and evolving field, especially for early-career researchers. Cheeky title🤭: How to Stand Out in the Crowd🙋? Details & context here: sites.google.com/view/standoutcv
Anand Bhattad@anand_bhattad

In this #CVPR2025 edition of our community-building workshop series, we focus on supporting the growth of early-career researchers. Join us tomorrow (Jun 11) at 12:45 PM in Room 209 Schedule: sites.google.com/view/standoutc… We have an exciting lineup of invited talks and candid panels: @sarameghanbeery, @dimadamen, @jbhuang0604, @lealtaixe, @LerrelPinto, @lschmidt3, @shubhtuls, @gulvarol, @cvondrick, @sainingxie Co-organizing with @unnatjain2010, @ap229997, @georgiagkioxari, @akanazawa, and Lana Lazebnik @CVPR

English
3
15
66
12.5K
CMU Center for Perceptual Computing and Learning retweetledi
Guanya Shi
Guanya Shi@GuanyaShi·
ASAP learns diverse, agile, whole-body humanoid motions via learning a residual action model from the real world to align sim and real physics, enabling motions that were previously difficult to achieve. It has two stages: Stage 1 pretrains a phase-based motion tracking policy to mimic human motions in simulation. Stage 2 rolls out such a policy in real to collect data, learns a residual action model to compensate for the dynamics mismatch, and finally fine-tunes the pretrained policy with the learned residual model. ASAP is fully open-sourced: agile.human2humanoid.com ASAP is not just for sim2real. It provides a general framework to align physics in training and deployment environments. To facilitate smooth transfer between different simulators, we also released HumanoidVerse, a multi-simulator humanoid learning framework: github.com/LeCAR-Lab/Huma… A key design principle of HumanoidVerse is the separation and modularization of simulators, tasks, and algorithms, allowing switching between simulators and tasks with minimal effort. We support training & evaluation in multiple simulators (IsaacGym, IsaacSim, Genesis) by changing only ONE command line: +simulator= Led by @TairanHe99 @WinstonGu_ @_wenlixiao @Yuanhang__Zhang. Collaborations with the @nvidia GEAR lab led by @yukez and @DrJimFan.
Guanya Shi tweet media
Tairan He@TairanHe99

🚀 Can we make a humanoid move like Cristiano Ronaldo, LeBron James and Kobe Byrant? YES! 🤖 Introducing ASAP: Aligning Simulation and Real-World Physics for Learning Agile Humanoid Whole-Body Skills Website: agile.human2humanoid.com Code: github.com/LeCAR-Lab/ASAP

English
5
22
132
11.9K
CMU Center for Perceptual Computing and Learning retweetledi
Zhengyi “Zen” Luo
Zhengyi “Zen” Luo@zhengyiluo·
Should have recorded our reactions when the first successful siuuu happened! 🎉 Collecting and learning real world data will be incredibly important for humanoids moving forward, and we have just took our first step ASAP🫡
Tairan He@TairanHe99

🚀 Can we make a humanoid move like Cristiano Ronaldo, LeBron James and Kobe Byrant? YES! 🤖 Introducing ASAP: Aligning Simulation and Real-World Physics for Learning Agile Humanoid Whole-Body Skills Website: agile.human2humanoid.com Code: github.com/LeCAR-Lab/ASAP

English
0
2
13
1.3K
CMU Center for Perceptual Computing and Learning retweetledi
Mehul Agarwal
Mehul Agarwal@meh_agarwal·
🎵✨Excited to share our #NeurIPS2024 paper on personalized music video generation! We combine multimodal AI with identity protection to let listeners be co-creators, generating custom music videos that reflect both music and themselves. 🎥🔒 arxiv.org/abs/2502.02610 #CreativeAI
English
1
4
17
4.2K
CMU Center for Perceptual Computing and Learning
Spot-on tips for faculty applicants from RI postdoc @unnatjain2010. Big congrats to him and UC Irvine! 🎉
Unnat Jain@unnatjain2010

Excited to share that I'll be joining University of California at Irvine as a CS faculty in '25!🌟 Faculty apps: @_krishna_murthy, @liuzhuang1234 & I share our tips: unnat.github.io/notes/Hidden_C… PhD apps: I'm looking for students in vision, robot learning, & AI4Science. Details👇

English
1
2
2
1.3K
CMU Center for Perceptual Computing and Learning retweetledi
Unnat Jain
Unnat Jain@unnatjain2010·
Excited to share that I'll be joining University of California at Irvine as a CS faculty in '25!🌟 Faculty apps: @_krishna_murthy, @liuzhuang1234 & I share our tips: unnat.github.io/notes/Hidden_C… PhD apps: I'm looking for students in vision, robot learning, & AI4Science. Details👇
Unnat Jain tweet media
English
38
72
392
66.9K
CMU Center for Perceptual Computing and Learning retweetledi
Rohan Choudhury
Rohan Choudhury@rchoudhury997·
Excited to finally release our NeurIPS 2024 (spotlight) paper! We introduce Run-Length Tokenization (RLT), a simple way to significantly speed up your vision transformer on video with no loss in performance!
GIF
English
22
169
1.4K
155.9K
CMU Center for Perceptual Computing and Learning retweetledi
Murtaza Dalal
Murtaza Dalal@mihdalal·
Can my robot cook my food, rearrange my dresser, tidy my messy table and do so much more without ANY demos or real-world training data? Introducing ManipGen: A generalist agent for manipulation that can solve long-horizon robotics tasks entirely zero shot, from text input! 1/N
English
10
115
616
148.7K
CMU Center for Perceptual Computing and Learning retweetledi
Mihir Prabhudesai
Mihir Prabhudesai@mihirp98·
1/ Happy to share VADER: Video Diffusion Alignment via Reward Gradients. We adapt foundational video diffusion models using pre-trained reward models to generate high-quality, aligned videos for various end-applications. Below we generated a short movie using VADER 😀, we used ChatGPT to write a script and an off-the-shelf AI music generator to generate the sound. Our code & weights are open-sourced: vader-vid.github.io
English
2
15
134
13.3K