Yevgen Chebotar (@yevgenchebotar) 's Twitter Profile
Yevgen Chebotar

@yevgenchebotar

Robot AI @Figure_robot | Former Research Scientist @GoogleDeepmind | 🤖 🦾

ID: 843188676834164737

calendar_today18-03-2017 19:53:41

42 Tweet

1,1K Followers

298 Following

Karol Hausman (@hausman_k) 's Twitter Profile Photo

Super excited to introduce SayCan (say-can.github.io): 1st publication of a large effort we've been working on for 1+ years Robots ground large language models in reality by acting as their eyes and hands while LLMs help robots execute long, abstract language instructions

Danny Driess (@dannydriess) 's Twitter Profile Photo

What happens when we train the largest vision-language model and add in robot experiences? The result is PaLM-E 🌴🤖, a 562-billion parameter, general-purpose, embodied visual-language generalist - across robotics, vision, and language. Website: palm-e.github.io

Karol Hausman (@hausman_k) 's Twitter Profile Photo

Very exited to announce our largest deep RL deployment to date: robots sorting trash end-to-end in real offices! rl-at-scale.github.io (aka RLS) This project took a long time (started before SayCan/RT-1/other newer works) but the learnings from it have been really valuable.🧵

Yevgen Chebotar (@yevgenchebotar) 's Twitter Profile Photo

Excited to present RT-2, a large unified Vision-Language-Action model! By converting robot actions to strings, we can directly train large visual-language models to output actions while retaining their web-scale knowledge and generalization capabilities! robotics-transformer2.github.io

Excited to present RT-2, a large unified Vision-Language-Action model!

By converting robot actions to strings, we can directly train large visual-language models to output actions while retaining their web-scale knowledge and generalization capabilities!

robotics-transformer2.github.io
Yevgen Chebotar (@yevgenchebotar) 's Twitter Profile Photo

Exciting times for Robot Learning! 60 datasets from 22 different robots and 21 institutions combined in a single Open-X Embodiment data repository, resulting in over 1 million episodes and improved RT-X models! Amazing and a very important collaboration across the world! 🤖🌐

Ted Xiao (@xiao_ted) 's Twitter Profile Photo

Had a great time today with Yevgen Chebotar and Quan Vuong visiting USC Viterbi School to give a talk on “Robot Learning in the Era of Foundation Models”. Slides out soon, packed with works from *just the past 5 months* 🤯 Thanks to Daniel Seita 🇺🇦 for hosting!

Had a great time today with <a href="/YevgenChebotar/">Yevgen Chebotar</a> and <a href="/QuanVng/">Quan Vuong</a> visiting <a href="/USCViterbi/">USC Viterbi School</a> to give a talk on “Robot Learning in the Era of Foundation Models”. Slides out soon, packed with works from *just the past 5 months* 🤯

Thanks to <a href="/daniel_t_seita/">Daniel Seita 🇺🇦</a> for hosting!
Yevgen Chebotar (@yevgenchebotar) 's Twitter Profile Photo

RT-H learns a hierarchy all the way from high-level tasks through low-level “language motions” to robot actions! ✅ Improved performance and generalization through better data sharing ✅ Automated grounded “bottom-up” labeling ✅ Ability to intervene and correct with language

Yevgen Chebotar (@yevgenchebotar) 's Twitter Profile Photo

The path to VLAs lies through VLMs. A very nice intro for everyone interested in working with Vision-Language Models: arxiv.org/abs/2405.17247

Yevgen Chebotar (@yevgenchebotar) 's Twitter Profile Photo

We've made great progress on Vision-Language-Action Models for humanoids in our new Helix model! Check out the technical report for more details: figure.ai/news/helix