Chen Geng (@gengchen01) 's Twitter Profile
Chen Geng

@gengchen01

CS Ph.D. Student @Stanford. Previously Hons. B.Eng. in CS @ZJU_China.

ID: 2257563174

linkhttps://chen-geng.com calendar_today22-12-2013 10:37:06

66 Tweet

780 Takipçi

773 Takip Edilen

Sanjana Srivastava (@sanjana__z) 's Twitter Profile Photo

🤖 Household robots are becoming physically viable. But interacting with people in the home requires handling unseen, unconstrained, dynamic preferences, not just a complex physical domain. We introduce ROSETTA: a method to generate reward for such preferences cheaply. 🧵⬇️

Zizhang Li (@zizhang_li) 's Twitter Profile Photo

In our #ICCV2025 WonderPlay, we study how to combine physical simulation and video generative prior to enable 3D action interaction with the world from a single image! Check the 🧵for more details!

Yuxi Xiao (@yuxixiaohenry) 's Twitter Profile Photo

🚀 We release SpatialTrackerV2: the first feedforward model for dynamic 3D reconstruction and 3D point tracking — all at once! Reconstruct dynamic scenes and predict pixel-wise 3D motion in seconds. 🔗 Webpage: spatialtracker.github.io 🔍 Online Demo: huggingface.co/spaces/Yuxihen…

Klemen Kotar (@klemenkotar) 's Twitter Profile Photo

📷 New Preprint: SOTA optical flow extraction from pre-trained generative video models! While it seems intuitive that video models grasp optical flow, extracting that understanding has proven surprisingly elusive.

Chen Geng (@gengchen01) 's Twitter Profile Photo

🔥 Deadline extended! The non-archival track is now open until Aug 17. Have research related to digital twins? Consider submitting it to our workshop at #ICCV2025 2025. Previously accepted or published papers are welcome as well. #ICCV2025

🔥 Deadline extended! The non-archival track is now open until Aug 17.
Have research related to digital twins? Consider submitting it to our workshop at <a href="/ICCVConference/">#ICCV2025</a> 2025.
Previously accepted or published papers are welcome as well. #ICCV2025
Hadi AlZayer (@hadizayer) 's Twitter Profile Photo

✨ Our paper Magic Fixup is accepted to ACM TOG! We show how dynamic videos can guide photo editing across many tasks — making this a solid baseline for future research. project page: magic-fixup.github.io paper: dl.acm.org/doi/10.1145/37…

✨ Our paper Magic Fixup is accepted to ACM TOG!
We show how dynamic videos can guide photo editing across many tasks — making this a solid baseline for future research.

project page: magic-fixup.github.io
paper: dl.acm.org/doi/10.1145/37…
Zhaoxi Chen (@frozen_burning) 's Twitter Profile Photo

🔥Feed-Forward 4D Generative Modeling🔥 #4DNeX is a training-efficient recipe for 4D generative *world modeling* from a single image. The 10M 4D dataset is also released! - Project: 4dnex.github.io - Data: huggingface.co/datasets/3DTop… - Code: github.com/3DTopia/4DNeX

Fei-Fei Li (@drfeifei) 's Twitter Profile Photo

(1/N) How close are we to enabling robots to solve the long-horizon, complex tasks that matter in everyday life? 🚨 We are thrilled to invite you to join the 1st BEHAVIOR Challenge @NeurIPS 2025, submission deadline: 11/15. 🏆 Prizes: 🥇 $1,000 🥈 $500 🥉 $300

Klemen Kotar (@klemenkotar) 's Twitter Profile Photo

1/ A good world model should be promptable like an LLM, offering flexible control and zero-shot answers to many questions. Language models have benefited greatly from this fact, but it's been slow to come to vision. We introduce PSI: a path to truly interactive visual world

1/ A good world model should be promptable like an LLM, offering flexible control and zero-shot answers to many questions. Language models have benefited greatly from this fact, but it's been slow to come to vision.
We introduce PSI: a path to truly interactive visual world
Fei-Fei Li (@drfeifei) 's Twitter Profile Photo

If you're curious about how the latest spatial intelligence model is doing World Labs, check out this new blog! I'm so excited by how much progress has been made in 3D world generation - bigger, more consistent, and forever persistent worlds! Moreover, everyone in the world

Elliott / Shangzhe Wu (@elliottszwu) 's Twitter Profile Photo

Looking for a PhD position? Apply to the ELLIS PhD program and get the unique opportunity to work with two different research teams across Europe! Apply by 31 Oct: ellis.eu/news/ellis-phd…

Yanjie Ze (@zeyanjie) 's Twitter Profile Photo

Our humanoid now learns loco-manip skills that generalize across space (Stanford) & time (day & night), using egocentric vision, trained only in simulation visualmimic.github.io

Weiyu Liu (@weiyu_liu_) 's Twitter Profile Photo

I’m at #CoRL2025 in Seoul this week! I’m looking for students to join my lab next year, and also for folks excited to build robotic foundation models at a startup. If you’re into generalization, planning and reasoning, or robots that use language, let's chat!

Kyle Sargent (@kylesargentai) 's Twitter Profile Photo

This is a cool idea, but this paper combines two bad practices on ImageNet (1) DinoV2 features (+142 million images) on a data-constrained benchmark (ImageNet, ~1.2 million images) (2) Much better gFID (1.13) than held-out ImageNet images (~1.8), implying serious Goodharting

Ruoshi Liu (@ruoshi_liu) 's Twitter Profile Photo

Everyone says they want general-purpose robots. We actually mean it — and we’ll make it weird, creative, and fun along the way 😎 Recruiting PhD students to work on Computer Vision and Robotics UMD Department of Computer Science for Fall 2026 in the beautiful city of Washington DC!

Everyone says they want general-purpose robots.

We actually mean it — and we’ll make it weird, creative, and fun along the way 😎

Recruiting PhD students to work on Computer Vision and Robotics <a href="/umdcs/">UMD Department of Computer Science</a> for Fall 2026 in the beautiful city of Washington DC!
Hadi AlZayer (@hadizayer) 's Twitter Profile Photo

what if you could combine diffusion models instantly? You would get exponentially better control (for free!!👀) This is exactly what we do. In ✨ coupled diffusion sampling ✨, diffusion models guide each other. The result? Diverse editing capabilities!

Chen Geng (@gengchen01) 's Twitter Profile Photo

Join us tomorrow at the #ICCV2025 workshop on generating digital twins from images and videos! Don’t miss amazing talks from Manolis Savva, Katerina Fragkiadaki, Marc Pollefeys, Jiajun Wu, Matthias Niessner, Lei Li, Yanpei Cao, and Steve Xie on cutting-edge progress! #ICCV2025

Join us tomorrow at the #ICCV2025 workshop on generating digital twins from images and videos! Don’t miss amazing talks from Manolis Savva, <a href="/KaterinaFragiad/">Katerina Fragkiadaki</a>, <a href="/mapo1/">Marc Pollefeys</a>, <a href="/jiajunwu_cs/">Jiajun Wu</a>, <a href="/MattNiessner/">Matthias Niessner</a>, <a href="/craigleili/">Lei Li</a>, <a href="/yanpei_cao/">Yanpei Cao</a>, and <a href="/bgxc/">Steve Xie</a> on cutting-edge progress!

<a href="/ICCVConference/">#ICCV2025</a>
Chengshu Li (@chengshuericli) 's Twitter Profile Photo

We are excited to release MoMaGen, a data generation method for multi-step bimanual mobile manipulation. MoMaGen turns 1 human-teleoped robot trajectory into 1000s of generated trajectories automatically.🚀 Website: momagen.github.io arXiv: arxiv.org/abs/2510.18316