Chris Rockwell (@_crockwell) 's Twitter Profile
Chris Rockwell

@_crockwell

PhD student in #ComputerVision at @UmichCSE
Views are my own.

ID: 1247903510

linkhttp://crockwell.github.io calendar_today07-03-2013 04:29:00

122 Tweet

534 Takipçi

627 Takip Edilen

Ang Cao (@angcao3) 's Twitter Profile Photo

Lightplane gives 1000x memory saving for differentiable rendering and feature splatting (i.e. unprojecting 2D features to 3D), which is generalizable to a variety of 3D structures. We hope it could solve memory bottleneck in current 3D pipeline and contribute to 3D research.

Tiange Luo (@tiangeluo) 's Twitter Profile Photo

We've curated a 1-million 3D-Captioning dataset for Objaverse(-XL), correcting 200k potential misalignments in the original Cap3D captions. Our method employs a pre-trained text-to-3D model to rank rendered views and utilizes GPT-4 Vision. Each caption is linked to a point

Chris Rockwell (@_crockwell) 's Twitter Profile Photo

Excited to present our #CVPR2024 *Highlight* FAR on Friday at 10:30 a.m, Arch 4A-E Poster #31. Please feel free to stop by! FAR significantly improves correspondence-based methods using end-to-end pose prediction, making it applicable to many SOTA approaches!

Sarah Jabbour (@sarahjabbour_) 's Twitter Profile Photo

📢Presenting 𝐃𝐄𝐏𝐈𝐂𝐓: Diffusion-Enabled Permutation Importance for Image Classification Tasks #ECCV2024 We use permutation importance to compute dataset-level explanations for image classifiers using diffusion models (without access to model parameters or training data!)

📢Presenting 𝐃𝐄𝐏𝐈𝐂𝐓: Diffusion-Enabled Permutation Importance for Image Classification Tasks #ECCV2024

We use permutation importance to compute dataset-level explanations for image classifiers using diffusion models (without access to model parameters or training data!)
Aayan Yadav (@ionydv) 's Twitter Profile Photo

📢 Introducing our #ECCV2024 work, COCO-ReM (COCO Refined Masks), for more reliable benchmarking of object detectors, crucial for the future of object detection research. Paper: arxiv.org/abs/2403.18819 Code: arxiv.org/abs/2403.18819 Website: cocorem.xyz

📢 Introducing our #ECCV2024 work, COCO-ReM (COCO Refined Masks), for more reliable benchmarking of object detectors, crucial for the future of object detection research.

Paper: arxiv.org/abs/2403.18819
Code: arxiv.org/abs/2403.18819
Website: cocorem.xyz
Ayush Shrivastava (@ayshrv) 's Twitter Profile Photo

We present Global Matching Random Walks, a simple self-supervised approach to the Tracking Any Point (TAP) problem, accepted to #ECCV2024. We train a global matching transformer to find cycle consistent tracks through video via contrastive random walks (CRW).

Daniel Geng (@dangengdg) 's Twitter Profile Photo

What happens when you train a video generation model to be conditioned on motion? Turns out you can perform "motion prompting," just like you might prompt an LLM! Doing so enables many different capabilities. Here’s a few examples – check out this thread 🧵 for more results!

Linyi Jin (@jin_linyi) 's Twitter Profile Photo

Introducing 👀Stereo4D👀 A method for mining 4D from internet stereo videos. It enables large-scale, high-quality, dynamic, *metric* 3D reconstructions, with camera poses and long-term 3D motion trajectories. We used Stereo4D to make a dataset of over 100k real-world 4D scenes.

Yen-Chen Lin (@yen_chen_lin) 's Twitter Profile Photo

Video generation models exploded onto the scene in 2024, sparked by the release of Sora from OpenAI. I wrote a blog post on key techniques that are used in building large video generation models: yenchenlin.me/blog/2025/01/0…

Chen-Hsuan Lin (@chenhsuanlin) 's Twitter Profile Photo

Cameras are key to modeling our dynamic 3D visual world. Can we unlock the 𝘥𝘺𝘯𝘢𝘮𝘪𝘤 3𝘋 𝘐𝘯𝘵𝘦𝘳𝘯𝘦𝘵?! 🌎 📸 𝗗𝘆𝗻𝗣𝗼𝘀𝗲-𝟭𝟬𝟬𝗞 is our answer! Chris Rockwell has curated Internet-scale videos with camera pose annotations for you 🤩 Download: huggingface.co/datasets/nvidi…

Daniel Geng (@dangengdg) 's Twitter Profile Photo

Hello! If you like pretty images and videos and want a rec for CVPR oral session, you should def go to Image/Video Gen, Friday at 9am: I'll be presenting "Motion Prompting" Ryan Burgert will be presenting "Go with the Flow" and Pascal CHANG will be presenting "LookingGlass"

Jeongsoo Park (@jespark0) 's Twitter Profile Photo

Can AI image detectors keep up with new fakes? Mostly, no. Existing detectors are trained using a handful of models. But there are thousands in the wild! Our work, Community Forensics, uses 4800+ generators to train detectors that generalize to new fakes. #CVPR2025 🧵 (1/5)

Yiming Dou (@_yimingdou) 's Twitter Profile Photo

Ever wondered how a scene sounds👂 when you interact👋 with it? Introducing our #CVPR2025 work "Hearing Hands: Generating Sounds from Physical Interactions in 3D Scenes" -- we make 3D scene reconstructions audibly interactive! yimingdou.com/hearing_hands/

Ayush Shrivastava (@ayshrv) 's Twitter Profile Photo

Excited to share our CVPR 2025 paper on cross-modal space-time correspondence! We present a method to match pixels across different modalities (RGB-Depth, RGB-Thermal, Photo-Sketch, and cross-style images) — trained entirely using unpaired data and self-supervision. Our

Excited to share our CVPR 2025 paper on cross-modal space-time correspondence!

We present a method to match pixels across different modalities (RGB-Depth, RGB-Thermal, Photo-Sketch, and cross-style images) — trained entirely using unpaired data and self-supervision.

Our
Linyi Jin (@jin_linyi) 's Twitter Profile Photo

Hello! If you are interested in dynamic 3D or 4D, don't miss the oral session 3A at 9 am on Saturday: Zhengqi Li will be presenting "MegaSaM" I'll be presenting "Stereo4D" and Qianqian Wang will be presenting "CUT3R"