CMU Center for Perceptual Computing and Learning (@robovisioncmu) 's Twitter Profile
CMU Center for Perceptual Computing and Learning

@robovisioncmu

The Chronicles of Smith Hall.

ID: 1139796230572105730

calendar_today15-06-2019 07:26:12

640 Tweet

1,1K Takipçi

138 Takip Edilen

Sudeep Dasari (@sudeepdasari) 's Twitter Profile Photo

Robots need strong visuo-motor representations to manipulate objects, but it’s hard to learn these using demo data alone. Our #RSS2024 project vastly improves robotic representations, using human affordances mined from Ego4D! w/ Mohan Kumar Srirama Shikhar Bahl Abhinav Gupta

Mihir Prabhudesai (@mihirp98) 's Twitter Profile Photo

1/ Happy to share VADER: Video Diffusion Alignment via Reward Gradients. We adapt foundational video diffusion models using pre-trained reward models to generate high-quality, aligned videos for various end-applications. Below we generated a short movie using VADER 😀, we used

Murtaza Dalal (@mihdalal) 's Twitter Profile Photo

Can my robot cook my food, rearrange my dresser, tidy my messy table and do so much more without ANY demos or real-world training data? Introducing ManipGen: A generalist agent for manipulation that can solve long-horizon robotics tasks entirely zero shot, from text input! 1/N

Rohan Choudhury (@rchoudhury997) 's Twitter Profile Photo

Excited to finally release our NeurIPS 2024 (spotlight) paper! We introduce Run-Length Tokenization (RLT), a simple way to significantly speed up your vision transformer on video with no loss in performance!

Unnat Jain (@unnatjain2010) 's Twitter Profile Photo

Excited to share that I'll be joining University of California at Irvine as a CS faculty in '25!🌟 Faculty apps: Krishna Murthy, Zhuang Liu & I share our tips: unnat.github.io/notes/Hidden_C… PhD apps: I'm looking for students in vision, robot learning, & AI4Science. Details👇

Excited to share that I'll be joining University of California at Irvine as a CS faculty in '25!🌟

Faculty apps: <a href="/_krishna_murthy/">Krishna Murthy</a>, <a href="/liuzhuang1234/">Zhuang Liu</a> &amp; I share our tips: unnat.github.io/notes/Hidden_C…

PhD apps: I'm looking for students in vision, robot learning, &amp; AI4Science. Details👇
Tarasha Khurana (@tarashakhurana) 's Twitter Profile Photo

Excited to present new work on using diffusion priors for video amodal segmentation and content completion! with Kaihua Chen (lead author) and Deva Ramanan arXiv: arxiv.org/abs/2412.04623 project page: diffusion-vas.github.io

Guanya Shi (@guanyashi) 's Twitter Profile Photo

ASAP learns diverse, agile, whole-body humanoid motions via learning a residual action model from the real world to align sim and real physics, enabling motions that were previously difficult to achieve. It has two stages: Stage 1 pretrains a phase-based motion tracking policy

ASAP learns diverse, agile, whole-body humanoid motions via learning a residual action model from the real world to align sim and real physics, enabling motions that were previously difficult to achieve. 

It has two stages: Stage 1 pretrains a phase-based motion tracking policy
Zhengyi “Zen” Luo (@zhengyiluo) 's Twitter Profile Photo

Should have recorded our reactions when the first successful siuuu happened! 🎉 Collecting and learning real world data will be incredibly important for humanoids moving forward, and we have just took our first step ASAP🫡

Mehul Agarwal (@meh_agarwal) 's Twitter Profile Photo

🎵✨Excited to share our #NeurIPS2024 paper on personalized music video generation! We combine multimodal AI with identity protection to let listeners be co-creators, generating custom music videos that reflect both music and themselves. 🎥🔒 arxiv.org/abs/2502.02610 #CreativeAI

Unnat Jain (@unnatjain2010) 's Twitter Profile Photo

✨New edition of our community-building workshop series!✨ Tomorrow at #CVPR2025, we invite speakers to share their stories, values, and approaches for navigating a crowded and evolving field, especially for early-career researchers. Cheeky title🤭: How to Stand Out in the

Tarasha Khurana (@tarashakhurana) 's Twitter Profile Photo

Excited to share recent work with Kaihua Chen and Deva Ramanan where we learn to do novel view synthesis for dynamic scenes in a self-supervised manner, only from 2D videos! webpage: cog-nvs.github.io arxiv: arxiv.org/abs/2507.12646 code (soon): github.com/Kaihua-Chen/co…

Yishu Li (@lisayishu) 's Twitter Profile Photo

A closed door looks the same whether it pushes or pulls. Two identical-looking boxes might have different center of mass. How should robots act when a single visual observation isn't enough? Introducing HAVE 🤖, our method that reasons about past interactions online! #CORL2025

A closed door looks the same whether it pushes or pulls. Two identical-looking boxes might have different center of mass. How should robots act when a single visual observation isn't enough?

Introducing HAVE 🤖, our method that reasons about past interactions online! #CORL2025
CMU Center for Perceptual Computing and Learning (@robovisioncmu) 's Twitter Profile Photo

TWO Best Paper Awards at ICCV Generating Physically Stable and Buildable Brick Structures from Text Ava Pun*, Kangle Deng*, Ruixuan Liu*, Deva Ramanan, Changliu Liu, Jun-Yan Zhu Spatially-Varying Autofocus Yingsi Qin, Aswin C. Sankaranarayanan, Matthew O'Toole #goSmithHall

Kris Kitani (@kkitani) 's Twitter Profile Photo

Super excited to share the release of SAM 3D. It's been a year in the making. Two models for lifting object and people to 3D!

CMU Center for Perceptual Computing and Learning (@robovisioncmu) 's Twitter Profile Photo

New model from Meta, SAM 3D Body, powered by people from Smith Hall (Kris Kitani,Jinkun Cao, David Park, Jyun-Ting Song) of course! #goSmithHall Introducing SAM 3D: a New Standard for 3D Object & Human Reconstruction ... youtu.be/B7PZuM55ayc?si… via YouTube