Wenli Xiao (@_wenlixiao) 's Twitter Profile
Wenli Xiao

@_wenlixiao

Graduate Researcher @CMU_Robotics | Intern at GEAR Lab
@NvidiaAI | I build AI brains for robots

ID: 1160530286297305093

linkhttp://wenlixiao-cs.github.io calendar_today11-08-2019 12:35:56

87 Tweet

1,1K Takipçi

437 Takip Edilen

Xiaofeng Guo (@xiaofeng2guo) 's Twitter Profile Photo

🚀Can we have a freely moving hand in the air for the manipulation policy to directly command in the real world? We introduce Flying Hand: End-Effector-Centric Framework for Versatile Aerial Manipulation Teleoperation and Policy Learning. 🎯EE-centric MPC for aerial manipulator

Physical Intelligence (@physical_int) 's Twitter Profile Photo

We got a robot to clean up homes that were never seen in its training data! Our new model, π-0.5, aims to tackle open-world generalization. We took our robot into homes that were not in the training data and asked it to clean kitchens and bedrooms. More below⤵️

Xuxin Cheng (@xuxin_cheng) 's Twitter Profile Photo

Meet 𝐀𝐌𝐎 — our universal whole‑body controller that unleashes the 𝐟𝐮𝐥𝐥  kinematic workspace of humanoid robots to the physical world. AMO is a single policy trained with RL + Hybrid Mocap & Trajectory‑Opt. Accepted to #RSS2025. Try our open models & more 👉

Yanjie Ze (@zeyanjie) 's Twitter Profile Photo

🤖Introducing TWIST: Teleoperated Whole-Body Imitation System. We develop a humanoid teleoperation system to enable coordinated, versatile, whole-body movements, using a single neural network. This is our first step toward general-purpose robots. 🌐humanoid-teleop.github.io

Arthur Allshire (@arthurallshire) 's Twitter Profile Photo

our new system trains humanoid robots using data from cell phone videos, enabling skills such as climbing stairs and sitting on chairs in a single policy (w/ Hongsuk Benjamin Choi Junyi Zhang David McAllister)

Wenli Xiao (@_wenlixiao) 's Twitter Profile Photo

A portable data collection system for general dexterous manipulation—DexWild🦾—can collect data Anywhere, enabling impressive generalization across robots/tasks/scenarios/... Huge congrats to Tony Tao and the team! 💥👏 Website: dexwild.github.io

Max Fu (@letian_fu) 's Twitter Profile Photo

Tired of teleoperating your robots? We built a way to scale robot datasets without teleop, dynamic simulation, or even robot hardware. Just one smartphone scan + one human hand demo video → thousands of diverse robot trajectories. Trainable by diffusion policy and VLA models

Jim Fan (@drjimfan) 's Twitter Profile Photo

What if robots could dream inside a video generative model? Introducing DreamGen, a new engine that scales up robot learning not with fleets of human operators, but with digital dreams in pixels. DreamGen produces massive volumes of neural trajectories - photorealistic robot

Wenli Xiao (@_wenlixiao) 's Twitter Profile Photo

Tired of watching fancy humanoid dancing? Can they just do some daily useful tasks like: "Pass me a bottle of Water🍺"? 🤔Turns out it's nontrivial to stablize whole-body manipulation and locomotion at the same time. We basically want our humanoid to be stable as a camera

Tianyuan Zhang (@tianyuanzhang99) 's Twitter Profile Photo

Bored of linear recurrent memories (e.g., linear attention) and want a scalable, nonlinear alternative? Our new paper “Test-Time Training Done Right” propose LaCT (Large Chunk Test-Time Training) — a highly efficient, massively scalable nonlinear memory with: 💡 Pure PyTorch

Dana Aubakirova (@daubakirovaa) 's Twitter Profile Photo

Today, we are introducing SmolVLA: a 450M open-source vision-language action model. Best-in-class performance and inference speed! And the best part? We trained it using all the open-source LeRobot datasets in the Hugging Face hub! But how? 🫳🏀

Wenli Xiao (@_wenlixiao) 's Twitter Profile Photo

"I believe finding such a scalable off-policy RL algorithm is the most important missing piece in machine learning today." Very insightful blog on offlineRL by Seohong Park 🫡 It's quite painful that offlineRL only works for "reduced horizon" at this stage. looking forward to

Wenli Xiao (@_wenlixiao) 's Twitter Profile Photo

💡Wow—super dynamic motion controlled by a unified general policy! 🔗 gmt-humanoid.github.io Feels like the recipe for training a general whole-body controller has almost converged: MoE oracle teacher → generalist student policy In our previous research: - HOVER

Haoyu Xiong (@haoyu_xiong_) 's Twitter Profile Photo

Your bimanual manipulators might need a Robot Neck 🤖🦒 Introducing Vision in Action: Learning Active Perception from Human Demonstrations ViA learns task-specific, active perceptual strategies—such as searching, tracking, and focusing—directly from human demos, enabling robust

Joel Jang (@jang_yoel) 's Twitter Profile Photo

🚀 GR00T Dreams code is live! NVIDIA GEAR Lab's open-source solution for robotics data via video world models. Fine-tune on any robot, generate 'dreams', extract actions with IDM, and train visuomotor policies with LeRobot datasets (GR00T N1.5, SmolVLA). github.com/NVIDIA/GR00T-D…