Jason Liu (@jasonjzliu) 's Twitter Profile
Jason Liu

@jasonjzliu

PhD Student @CMU_Robotics | Prev: Robot Learning @NvidiaAI | Engineering Science @UofT

ID: 2788592256

linkhttp://jasonjzliu.com calendar_today03-09-2014 21:12:00

176 Tweet

1,1K Followers

261 Following

Kenny Shaw (@kenny__shaw) 's Twitter Profile Photo

Exiciting to see (at 5:55) Nvidia adopting LEAP Hand in their sim2real efforts! Build your own at leaphand.com ! Lots more coming this summer, stay tuned :) Deepak Pathak Ananye Agarwal

Jason Liu (@jasonjzliu) 's Twitter Profile Photo

Robot data is expensive and hard to scale But what if we could collect rich, diverse demos—with just our hands? 🙌 Our latest work, DexWild, shows how large-scale human data 💪 + robot data 🦾 co-training enables strong generalization across tasks, scenes, and embodiments

Deepak Pathak (@pathak2206) 's Twitter Profile Photo

Introducing DexWild -- a scalable approach to diverse "in the wild" data collection for dexterous robotic hands! This data can be used to co-train policy for any downstream robotic hands on any body form factor (humanoids, AMR with arms, etc). 🚀🤖

Kenny Shaw (@kenny__shaw) 's Twitter Profile Photo

Very exciting Handy Moves workshop at ICRA 2025 this year! It's an honor to be hosting this morning session! Please join us in Room 302 😀 sites.google.com/view/dexterity…

Very exciting Handy Moves workshop at ICRA 2025 this year!  It's an honor to be hosting this morning session! Please join us in Room 302 😀 sites.google.com/view/dexterity…
Mihir Prabhudesai (@mihirp98) 's Twitter Profile Photo

Excited to share our work: Maximizing Confidence Alone Improves Reasoning Humans rely on confidence to learn when answer keys aren’t available (e.g taking an exam). Surprisingly, LLMs can also learn w/o ground-truth answers, simply by reinforcing high-confidence answers via RL!

Lili (@lchen915) 's Twitter Profile Photo

One fundamental issue with RL – whether it’s for robots or LLMs – is how hard it is to get rewards. For LLM reasoning, we need ground-truth labels to verify answers. We found that maximizing confidence alone allows LLMs to improve their reasoning with RL!

Fahim Tajwar (@fahimtajwar10) 's Twitter Profile Photo

RL with verifiable reward has shown impressive results in improving LLM reasoning, but what can we do when we do not have ground truth answers? Introducing Self-Rewarding Training (SRT): where language models provide their own reward for RL training! 🧵 1/n

RL with verifiable reward has shown impressive results in improving LLM reasoning, but what can we do when we do not have ground truth answers?

Introducing Self-Rewarding Training (SRT): where language models provide their own reward for RL training!

🧵 1/n
Deepak Pathak (@pathak2206) 's Twitter Profile Photo

Maximizing Confidence Alone Improves Reasoning Feels like the start of the "curiosity-driven learning" era for LLMs. I have spent most of my career towards building agents that can self-improve without any external rewards (e.g., curiosity work during Phd and then at CMU).

Maximizing Confidence Alone Improves Reasoning

Feels like the start of the "curiosity-driven learning" era for LLMs. I have spent most of my career towards building agents that can self-improve without any external rewards (e.g., curiosity work during Phd and then at CMU).
Robotic Systems Lab (@leggedrobotics) 's Twitter Profile Photo

A legged mobile manipulator trained to play badminton with humans coordinates whole-body maneuvers and onboard perception. Paper: science.org/doi/10.1126/sc……Video: youtu.be/zYuxOVQXVt8 Yuntao Ma, Andrei Cramariuc, Farbod Farshidian, Marco Hutter

Tyler Lum (@tylerlum23) 's Twitter Profile Photo

🧑🤖 Introducing Human2Sim2Robot!  💪🦾 Learn robust dexterous manipulation policies from just one human RGB-D video. Our Real→Sim→Real framework crosses the human-robot embodiment gap using RL in simulation. #Robotics #DexterousManipulation #Sim2Real 🧵1/7

Yitang Li (@li_yitang) 's Twitter Profile Photo

🤖Can a humanoid robot carry a full cup of beer without spilling while walking 🍺? Hold My Beer ! Introducing Hold My Beer🍺: Learning Gentle Humanoid Locomotion and End-Effector Stabilization Control Project: lecar-lab.github.io/SoFTA/ See more details below👇

Jingyun Yang (@yjy0625) 's Twitter Profile Photo

Introducing Mobi-π: Mobilizing Your Robot Learning Policy. Our method: ✈️ enables flexible mobile skill chaining 🪶 without requiring additional policy training data 🏠 while scaling to unseen scenes 🧵↓

Haoyu Xiong (@haoyu_xiong_) 's Twitter Profile Photo

Your bimanual manipulators might need a Robot Neck 🤖🦒 Introducing Vision in Action: Learning Active Perception from Human Demonstrations ViA learns task-specific, active perceptual strategies—such as searching, tracking, and focusing—directly from human demos, enabling robust