Wenhao Yu (@stacormed) 's Twitter Profile
Wenhao Yu

@stacormed

Research Scientist @DeepMind

ID: 261236765

linkhttps://wenhaoyu.weebly.com/ calendar_today05-03-2011 14:54:44

52 Tweet

414 Takipçi

129 Takip Edilen

Ayzaan Wahid (@ayzwah) 's Twitter Profile Photo

For the past year we've been working on ALOHA Unleashed 🌋 @GoogleDeepmind - pushing the scale and dexterity of tasks on our ALOHA 2 fleet. Here is a thread with some of the coolest videos! The first task is hanging a shirt on a hanger (autonomous 1x)

Wenhao Yu (@stacormed) 's Twitter Profile Photo

At Vienna for ICML this week! Let me know if you are down to catch up and look forward to the great discussions and talks ahead! Also come check out our work on PIVOT pivot-prompt.github.io during Tuesday’s poster session!

At Vienna for ICML this week!  Let me know if you are down to catch up and look forward to the great discussions and talks ahead!

Also come check out our work on PIVOT pivot-prompt.github.io during Tuesday’s poster session!
Wenhao Yu (@stacormed) 's Twitter Profile Photo

How can we leverage the common sense knowledge from a VLM to understand the progress (and even quality!) of a robotics trajectory? Check out GVL on a surprisingly simple and elegant way to do that! Awesome work by Jason!

Wenhao Yu (@stacormed) 's Twitter Profile Photo

Wow this is really good! In some way I’m more impressed that it’s teleoperated than if it’s autonomous cuz it feels very plausible to develop a highly specialized RL-based policy to do this, but being able to tele op this opens up a wide range of data to be collected.

Wenhao Yu (@stacormed) 's Twitter Profile Photo

2nd Earth Rover Challenge is coming! Eager to see how much progress AI will make in navigating real cities against real human agents!

Sundar Pichai (@sundarpichai) 's Twitter Profile Photo

We’ve always thought of robotics as a helpful testing ground for translating AI advances into the physical world. Today we’re taking our next step in this journey with our newest Gemini 2.0 robotics models. They show state of the art performance on two important benchmarks -

Yixin Lin (@yixin_lin_) 's Twitter Profile Photo

Complementary to Gemini Robotics -- the massive vision-language-action (VLA) model released yesterday -- we also investigated how far we can push Gemini for robotics _purely from simulation data_ in Proc4Gem: 🧵

Boyuan Chen (@boyuanchen0) 's Twitter Profile Photo

Deadline extended! You now have until May 25th (10 days post-NeurIPS) to submit to our ICML World Model Workshop. Looking forward to your papers!

Wenhao Yu (@stacormed) 's Twitter Profile Photo

How do imbue robots with the ability to imagine the world and complete tasks better? Join us at CoRL 25 workshop on Robotics World Modeling and share your latest work in this area!

Wenhao Yu (@stacormed) 's Twitter Profile Photo

Excited to share our latest work on Gemini Robotics 1.5! Our model can effectively learn from experience of drastically different robots, think on its own, and act as an agent. It’s an important step towards creating a general, intelligent, and friendly robot!

Wenhao Yu (@stacormed) 's Twitter Profile Photo

Gemini Robotics 1.5 is not only general, but also fairly dexterous! Enjoy some fun videos of robot doing insertion, zipping, and more (remember this is the *same checkpoint* that also controls two other very different robots) 😆

Caden Lu (@jyluxx) 's Twitter Profile Photo

Interacting with Gemini Robotics 1.5 is so fun! Our Embodied Reasoning model planned the multi-step task and orchestrated our Vision Language Action model for precise execution!