Max Fu (@letian_fu) 's Twitter Profile
Max Fu

@letian_fu

PhD at @UCBerkeley @berkeley_ai. Prev intern @Apple @autodesk. Robotics, Foundation Models.

ID: 769239354

linkhttps://max-fu.github.io/ calendar_today20-08-2012 10:08:45

103 Tweet

453 Takipçi

502 Takip Edilen

Arthur Allshire (@arthurallshire) 's Twitter Profile Photo

our new system trains humanoid robots using data from cell phone videos, enabling skills such as climbing stairs and sitting on chairs in a single policy (w/ Hongsuk Benjamin Choi Junyi Zhang David McAllister)

ollama (@ollama) 's Twitter Profile Photo

Multimodal model support is here in 0.7! Ollama now supports multimodal models via its new engine. Cool vision models to try👇 - Llama 4 Scout & Maverick - Gemma 3 - Qwen 2.5 VL - Mistral Small 3.1 and more 😍 Blog post 🧵👇

Multimodal model support is here in 0.7!  

Ollama now supports multimodal models via its new engine. 

Cool vision models to try👇

- Llama 4 Scout & Maverick
- Gemma 3 
- Qwen 2.5 VL 
- Mistral Small 3.1 
and more 😍

Blog post 🧵👇
Long Lian (@longtonylian) 's Twitter Profile Photo

As we all know, collecting data for robotics is very costly. This is why I’m very impressed by this work: it generates a huge amount of data for different robots without any teleoperation.

Raven Huang (@ravenhuang4) 's Twitter Profile Photo

Can we scale up robot data collection without a robot? We propose a pipeline to scale robot dataset from one human demonstration. Through a real2render2real pipeline, policies trained with the generated data can be deployed directly on a real robot.

Fangchen Liu (@fangchenliu_) 's Twitter Profile Photo

Ppl are collecting large-scale teleoperation datasets, which are often just kinematics-level trajectories. Real2Render2Real is a new framework that can generate these data w.o. teleoperation or tricky sim+rl. High data quality for BC + nice scaling effect, plz dive in for more!

Max Fu (@letian_fu) 's Twitter Profile Photo

Large language models can do new tasks from a few text prompts. What if robots could do the same—with trajectories? 🤖 ICRT enables zero-shot imitation: prompt with a few teleop demos, and it acts—no fine-tuning. Happy to chat more at ICRA! 📍 ICRA | Wed 21 May | 08:35 - 08:40

Zubair Irshad (@mzubairirshad) 's Twitter Profile Photo

Interested in collecting robot training data without robots in the loop? 🦾 Check out this cool new approach that uses a single mobile device scan and a human demo video to generate diverse data for training diffusion and VLA manipulation policies. 🚀 Great work by Max Fu

Yi Zhou (@papagina_yi) 's Twitter Profile Photo

🚀 Struggling with the lack of high-quality data for AI-driven human-object interaction research? We've got you covered! Introducing HUMOTO, a groundbreaking 4D dataset for human-object interaction, developed with a combination of wearable motion capture, SOTA 6D pose

Haonan Chen (@haonanchen_) 's Twitter Profile Photo

We hope everyone had a great time at the ICRA 2025 Workshop on Learning Meets Model-Based Methods for Contact-Rich Manipulation (contact-rich.github.io)! Big thanks to our incredible speakers, panelists, and generous sponsors — and most of all, to our amazing co-organizers

We hope everyone had a great time at the ICRA 2025 Workshop on Learning Meets Model-Based Methods for Contact-Rich Manipulation (contact-rich.github.io)!

Big thanks to our incredible speakers, panelists, and generous sponsors — and most of all, to our amazing co-organizers
Stella Li (@stellalisy) 's Twitter Profile Photo

🤯 We cracked RLVR with... Random Rewards?! Training Qwen2.5-Math-7B with our Spurious Rewards improved MATH-500 by: - Random rewards: +21% - Incorrect rewards: +25% - (FYI) Ground-truth rewards: + 28.8% How could this even work⁉️ Here's why: 🧵 Blogpost: tinyurl.com/spurious-rewar…

🤯 We cracked RLVR with... Random Rewards?!
Training Qwen2.5-Math-7B with our Spurious Rewards improved MATH-500 by:
- Random rewards: +21%
- Incorrect rewards: +25%
- (FYI) Ground-truth rewards: + 28.8%
How could this even work⁉️ Here's why: 🧵
Blogpost: tinyurl.com/spurious-rewar…
Robotic Systems Lab (@leggedrobotics) 's Twitter Profile Photo

A legged mobile manipulator trained to play badminton with humans coordinates whole-body maneuvers and onboard perception. Paper: science.org/doi/10.1126/sc……Video: youtu.be/zYuxOVQXVt8 Yuntao Ma, Andrei Cramariuc, Farbod Farshidian, Marco Hutter

Sawyer Merritt (@sawyermerritt) 's Twitter Profile Photo

Waymo in a new blog post: "We conducted a comprehensive study using Waymo’s internal dataset. Spanning 500,000 hours of driving, it is significantly larger than any dataset used in previous scaling studies in the AV domain. Our study uncovered the following: • Similar to LLMs,

Waymo in a new blog post: "We conducted a comprehensive study using Waymo’s internal dataset. Spanning 500,000 hours of driving, it is significantly larger than any dataset used in previous scaling studies in the AV domain.

Our study uncovered the following: 
• Similar to LLMs,