Yixuan Wang (@yxwangbot) 's Twitter Profile
Yixuan Wang

@yxwangbot

CS Ph.D. student @Columbia working on robotics | Worked at Boston Dynamics AI Institute, Google X #Vision #Robotics #Learning

ID: 1184179638215397376

linkhttps://wangyixuan12.github.io/ calendar_today15-10-2019 18:50:09

152 Tweet

1,1K Followers

964 Following

Yunzhu Li (@yunzhuliyz) 's Twitter Profile Photo

**Steerability** remains one of the key issues for current vision-language-action models (VLAs). Natural language is often ambiguous and vague: "Hang a mug on a branch" vs "Hang the left mug on the right branch." Many works claim to handle language input, yet the tasks are often

Yixuan Wang (@yxwangbot) 's Twitter Profile Photo

Two releases in a row from our lab today 😆 One problem I was always pondering on is how to use structured representation while making it scalable. Super excited that Kaifeng's work pushes this direction forward and I cannot wait to see what's more in the future!!

Katherine Liu (@robo_kat) 's Twitter Profile Photo

How can we achieve both common sense understanding that can deal with varying levels of ambiguity in language and dextrous manipulation? Check out CodeDiffuser, a really neat work that bridges Code Gen with a 3D Diffusion Policy! This was a fun project with cool experiments! 🤖

Yunzhu Li (@yunzhuliyz) 's Twitter Profile Photo

We’ve been exploring 3D world models with the goal of finding the right recipe that is both: (1) structured—for sample efficiency and generalization (my personal emphasis), and (2) scalable—as we increase real-world data collection. With **Particle-Grid Neural Dynamics**

Yixuan Wang (@yxwangbot) 's Twitter Profile Photo

Just arrived at LA and excited to be at RSS! I will present CodeDiffuser at following sessions: - Presentation on June 22 (Sun.) 5:30 PM - 6:30 PM - Poster on June 22 (Sun.) 6:30 PM - 8:00 PM I will also present CuriousBot at - FM4RoboPlan Workshop on June 21 (Sat.) 9:40 - 10:10

Yunzhu Li (@yunzhuliyz) 's Twitter Profile Photo

Had a great time yesterday giving three invited talks at #RSS2025 workshops—on foundation models, structured world models, and tactile sensing for robotic manipulation. Lots of engaging conversations! One more talk coming up on Wednesday (6/25). Also excited to be presenting two

Shivansh Patel (@shivanshpatel35) 's Twitter Profile Photo

🚀 Introducing RIGVid: Robots Imitating Generated Videos! Robots can now perform complex tasks—pouring, wiping, mixing—just by imitating generated videos, purely zero-shot! No teleop. No OpenX/DROID/Ego4D. No videos of human demonstrations. Only AI generated video demos 🧵👇

Russ Tedrake (@russtedrake) 's Twitter Profile Photo

TRI's latest Large Behavior Model (LBM) paper landed on arxiv last night! Check out our project website: toyotaresearchinstitute.github.io/lbm1/ One of our main goals for this paper was to put out a very careful and thorough study on the topic to help people understand the state of the