Michael Ryoo (@ryoo_michael) 's Twitter Profile
Michael Ryoo

@ryoo_michael

prof. with Stony Brook Univ. / research scientist with Salesforce AI Research

ID: 1448359208769015811

linkhttp://michaelryoo.com/ calendar_today13-10-2021 18:45:42

35 Tweet

333 Followers

68 Following

Ted Xiao (@xiao_ted) 's Twitter Profile Photo

Looking forward to showcasing one of the first foundation models for robotics at #RSS2023 next week! Presenting "RT-1: Robotics Transformer for Real-world Control at Scale" from the Google DeepMind robotics team. Website: robotics-transformer.github.io Session: Tuesday 7/12, 3PM-5PM

Xiang Li (@xiangli54505720) 's Twitter Profile Photo

Introducing Crossway Diffusion, a diffusion-based visuomotor policy taking advantage of SSL. In short: we add state decoders to reconstruct states during training diffusion policy and it works better. More at: arxiv.org/abs/2307.01849

Introducing Crossway Diffusion, a diffusion-based visuomotor policy taking advantage of SSL. In short: we add state decoders to reconstruct states during training diffusion policy and it works better. More at: arxiv.org/abs/2307.01849
Google DeepMind (@googledeepmind) 's Twitter Profile Photo

Today, we announced 𝗥𝗧-𝟮: a first of its kind vision-language-action model to control robots. 🤖 It learns from both web and robotics data and translates this knowledge into generalised instructions. Find out more: dpmd.ai/introducing-rt2

Karol Hausman (@hausman_k) 's Twitter Profile Photo

PaLM-E or GPT-4 can speak in many languages and understand images. What if they could speak robot actions? Introducing RT-2: robotics-transformer2.github.io our new model that uses a VLM (up to 55B params) backbone and fine-tunes it to directly output robot actions!

Michael Ryoo (@ryoo_michael) 's Twitter Profile Photo

Introducing LLaRA !!! github.com/LostXine/LLaRA It's a new robot action model, dataset, and framework based on LLMs/VLMs. It's opensource and trainable at an academic scale (7B LLaVA-based), so you can finetune it for your robotics task!

AK (@_akhaliq) 's Twitter Profile Photo

Salesforce presents xGen-MM (BLIP-3) A Family of Open Large Multimodal Models discuss: huggingface.co/papers/2408.08… This report introduces xGen-MM (also known as BLIP-3), a framework for developing Large Multimodal Models (LMMs). The framework comprises meticulously curated

Salesforce presents xGen-MM (BLIP-3)

A Family of Open Large Multimodal Models

discuss: huggingface.co/papers/2408.08…

This report introduces xGen-MM (also known as BLIP-3), a framework for developing Large Multimodal Models (LMMs). The framework comprises meticulously curated
Salesforce AI Research (@sfresearch) 's Twitter Profile Photo

🚨🎥🚨🎥🚨 xGen-MM-Vid (BLIP-3-Video) is now available on Hugging Face! Our compact VLM achieves SOTA performance with just 32 tokens for video understanding. Features explicit temporal encoder + BLIP-3 architecture. Try it out! 🤗32 Token Model: bit.ly/3PBNBBz 🤗128

🚨🎥🚨🎥🚨 xGen-MM-Vid (BLIP-3-Video) is now available on <a href="/huggingface/">Hugging Face</a>!

Our compact VLM achieves SOTA performance with just 32 tokens for video understanding. Features explicit temporal encoder + BLIP-3 architecture. Try it out!

🤗32 Token Model: bit.ly/3PBNBBz
🤗128
Michael Ryoo (@ryoo_michael) 's Twitter Profile Photo

LLaRA will appear at #ICLR2025 !! It is an efficient transformation of a VLM into a robot VLA. For more details: github.com/LostXine/LLaRA

Conference on Robot Learning (@corl_conf) 's Twitter Profile Photo

#CoRL2025 Hey Robot Learning Community! CoRL 2025 will be held in Seoul, Korea, Sep 27 - 30. Submission deadline: Apr 30 AoE. It's two weeks to go! Information: corl.org We are excited to receive your great work on robot learning!