Dantong Niu (@dantong_niu) 's Twitter Profile
Dantong Niu

@dantong_niu

ID: 817027139434184704

calendar_today05-01-2017 15:17:04

5 Tweet

22 Followers

7 Following

Aran Komatsuzaki (@arankomatsuzaki) 's Twitter Profile Photo

BAIR presents LLARVA: Vision-Action Instruction Tuning Enhances Robot Learning proj: llarva24.github.io abs: arxiv.org/abs/2406.11815

Roei Herzig (@roeiherzig) 's Twitter Profile Photo

Can instruction tuning be used in robotics?🚨Excited to release LLARVA, our new Vision-Action Instruction-tuned LMM for robotics! Key aspects: • Perform vision-action instruction tuning • Align vision and action modalities • Pretrained on robotic instruction data Berkeley AI Research

Max Fu (@letian_fu) 's Twitter Profile Photo

Vision-language models perform diverse tasks via in-context learning. Time for robots to do the same! Introducing In-Context Robot Transformer (ICRT): a robot policy that learns new tasks by prompting with robot trajectories, without any fine-tuning. icrt.dev [1/N]

Roei Herzig (@roeiherzig) 's Twitter Profile Photo

Have you wondered whether LLMs could directly execute robot tasks?🚨 Excited to share our work, *RoboPrompt*, a framework that enables off-the-shelf text-only LLMs to directly predict robot actions with ICL demonstrations without training! Berkeley AI Research davidyyd.github.io/roboprompt/

Tsung-Han (Patrick) Wu @ ICLR’25 (@tsunghan_wu) 's Twitter Profile Photo

📢 Call for Papers! Last chance to hang with the CV crowd in Hawaii 🌴 We're hosting the 4th MMFM Workshop at #ICCV2025 — submit your work on vision, language, audio & more by July 1 🗓️ Also check out the CVPR edition 👉 #3 MMFM Workshop 🔗 sites.google.com/view/mmfm4thwo…

📢 Call for Papers!
Last chance to hang with the CV crowd in Hawaii 🌴

We're hosting the 4th MMFM Workshop at #ICCV2025 — submit your work on vision, language, audio & more by July 1 🗓️

Also check out the CVPR edition 👉 <a href="/MMFMWorkshop/">#3 MMFM Workshop</a> 

🔗 sites.google.com/view/mmfm4thwo…
Haoru Xue (@haoruxue) 's Twitter Profile Photo

🚀 Introducing LeVERB, the first 𝗹𝗮𝘁𝗲𝗻𝘁 𝘄𝗵𝗼𝗹𝗲-𝗯𝗼𝗱𝘆 𝗵𝘂𝗺𝗮𝗻𝗼𝗶𝗱 𝗩𝗟𝗔 (upper- & lower-body), trained on sim data and zero-shot deployed. Addressing interactive tasks: navigation, sitting, locomotion with verbal instruction. 🧵 ember-lab-berkeley.github.io/LeVERB-Website/