An-Chieh Cheng (@anjjei) 's Twitter Profile
An-Chieh Cheng

@anjjei

PhD student @UCSanDiego; prev intern: @AdobeResearch; I love 3D vision.

ID: 203924729

linkhttp://anjiecheng.me calendar_today17-10-2010 14:20:00

226 Tweet

527 Followers

355 Following

Dorsa Sadigh (@dorsasadigh) 's Twitter Profile Photo

Generalist robot policies should be evaluated on 'generalization' metrics rather than merely reporting success rate in performance in arbitrary scenarios. ★-Gen introduces axes of generalization across inputs/outputs of policies & guides eval benchmarks for robot policies. 1/7

Roger Qiu (@rogerqiu_42) 's Twitter Profile Photo

Feature Splatting can now turn Objaverse assets into GS. With optimized kernel, 30K iters can be done in <1min on a single 4090 GPU.

Yufei Ye (@yufei_ye) 's Twitter Profile Photo

We will be hosting the workshop on "Agents in Interactions, from Humans to Robots" at CVPR2025 #CVPR2025 , welcome to join us by submitting a paper or stopping by our talks/posters! For more info please check out: agents-in-interactions.github.io

We will be hosting the workshop on "Agents in Interactions, from Humans to Robots" at CVPR2025
<a href="/CVPR/">#CVPR2025</a> 
, welcome to join us by submitting a paper or stopping by our talks/posters!

For more info please check out:
agents-in-interactions.github.io
Erik Daxberger (@edaxberger) 's Twitter Profile Photo

Check out our new work on exploring 3D Spatial Understanding with Multimodal LLMs!🚀 📀CA-VQA: A fine-tuning dataset and benchmark w/ various input signals and spatial tasks. 🤖MM-Spatial: A generalist MLLM excelling at spatial reasoning. 🔗arxiv.org/abs/2503.13111 🧵(1/n)

Check out our new work on exploring 3D Spatial Understanding with Multimodal LLMs!🚀

📀CA-VQA: A fine-tuning dataset and benchmark w/ various input signals and spatial tasks.

🤖MM-Spatial: A generalist MLLM excelling at spatial reasoning.

🔗arxiv.org/abs/2503.13111

🧵(1/n)
Roger Qiu (@rogerqiu_42) 's Twitter Profile Photo

Diverse training data leads to a more robust humanoid manipulation policy, but collecting robot demonstrations is slow. Introducing our latest work, Humanoid Policy ~ Human Policy. We advocate human data as a scalable data source for co-training egocentric manipulation policy.⬇️

Xueyan Zou (@xyz2maureen) 's Twitter Profile Photo

[1/n] We are releasing M3 (#ICLR2025): a Gaussian Splatting method that builds LMM memories for arbitrary scenes. 🔥 [Efficient] 16 degrees in each Gaussian primitive for one LMM. 🔥 [Alignment] The rendered features are directly in the source LMM embedding space.

Chan Hee (Luke) Song (@luke_ch_song) 's Twitter Profile Photo

🔥 VLMs aren’t built for spatial reasoning — yet. They hallucinate free space. Misjudge object fit. Can’t tell below from behind We built RoboSpatial to tackle that — a dataset for teaching spatial understanding to 2D/3D VLMs for robotics. 📝 Perfect review scores #CVPR2025 2025

Jiarui Xu (@jerry_xu_jiarui) 's Twitter Profile Photo

Test-Time Training (TTT) is available in video generation now! We can directly generate complete one-minute video, with great temporal and spatial coherence. We created more episodes of Tom and Jerry (my favorite cartoon in childhood) with our model. test-time-training.github.io/video-dit/

Hongxu (Danny) Yin (@yin_hongxu) 's Twitter Profile Photo

#RSS2025 NaVILA constitutes a successful attempt for VILA to drive real world robotic dogs and humanoid! Fully deployable. Money saving. Fast inference. Check out our project page: navila-bot.github.io Many more amazing things to come!

Isabella Liu (@isabella__liu) 's Twitter Profile Photo

Excited to be at #ICLR2025 in person this year! Looking forward to reconnecting and making new friends.🤩 Come chat with us about Dynamic Gaussians Mesh at poster #97 tomorrow (4/26, 3–5:30pm). See you there!🥳 Website: liuisabella.com/DG-Mesh

Excited to be at #ICLR2025 in person this year! Looking forward to reconnecting and making new friends.🤩

Come chat with us about Dynamic Gaussians Mesh at poster #97 tomorrow (4/26, 3–5:30pm). See you there!🥳

Website: liuisabella.com/DG-Mesh
Andrew Liao (@andrewliao11) 's Twitter Profile Photo

Takeaway: Structured, reflective reasoning can be taught — even in perception. We show that generating better data can unlock stronger visual reasoning. 🌐Website: andrewliao11.github.io/LongPerceptual… 🤗Dataset: huggingface.co/datasets/andre… 📜Paper: arxiv.org/abs/2504.15362

Manling Li (@manlingli_) 's Twitter Profile Photo

🚨CVPR Workshop on Foundation Models + Embodied Agents Extending non-archival submission deadline to be after NeurIPS, May 17th! 🌐Website: …models-meet-embodied-agents.github.io/cvpr2025/ 📜OpenReview: openreview.net/group?id=thecv… 👥Program committee sign up form forms.gle/bL17vmr7ZbybxE… ✉️mailing list:

🚨CVPR Workshop on Foundation Models + Embodied Agents

Extending non-archival submission deadline to be after NeurIPS, May 17th!

🌐Website: …models-meet-embodied-agents.github.io/cvpr2025/
📜OpenReview: openreview.net/group?id=thecv…
👥Program committee sign up form forms.gle/bL17vmr7ZbybxE…
✉️mailing list:
Hanwen Jiang (@hanwenjiang1) 's Twitter Profile Photo

Supervised learning has held 3D Vision back for too long. Meet RayZer — a self-supervised 3D model trained with zero 3D labels: ❌ No supervision of camera & geometry ✅ Just RGB images And the wild part? RayZer outperforms supervised methods (as 3D labels from COLMAP is noisy)

Xuxin Cheng (@xuxin_cheng) 's Twitter Profile Photo

Meet 𝐀𝐌𝐎 — our universal whole‑body controller that unleashes the 𝐟𝐮𝐥𝐥  kinematic workspace of humanoid robots to the physical world. AMO is a single policy trained with RL + Hybrid Mocap & Trajectory‑Opt. Accepted to #RSS2025. Try our open models & more 👉

yisha (@yswhynot) 's Twitter Profile Photo

For years, I’ve been tuning parameters for robot designs and controllers on specific tasks. Now we can automate this on dataset-scale. Introducing Co-Design of Soft Gripper with Neural Physics - a soft gripper trained in simulation to deform while handling load.

Xiaolong Wang (@xiaolonw) 's Twitter Profile Photo

We have been focusing on policy learning for robotics for a while. But can hardware be learned as well? Check out yisha ‘s recent co-design work that learns what a soft gripper should be if we want to do better manipulation.