Jesse Zhang
@jesse_y_zhang
PhD Student at USC focused on Robotics/Deep RL. Advisors: Erdem Biyik, Jesse Thomason, Joseph J. Lim. Focused on scalable, sample-efficient robot adaptation.
ID: 1341621435136032770
https://jessezhang.net 23-12-2020 05:47:43
91 Tweet
729 Takipçi
374 Takip Edilen
Combinatorial complexity is often the bane of imitation learning - including VLA models! Jesse Zhang and Marius Memmel proposed a way around this, using VLMs to perform problem reduction for imitation. The insight is simple - 1) High-level VLM takes a complex scene/task and
Very excited to finally share what I’ve been up to Physical Intelligence for the past 6 months: developing advantage-conditioned VLAs! We are finally moving beyond imitating teleop data, and towards improving models with suboptimal deployment data using scalable real-world RL. 👇🧵
Some of my favorites on reward modeling, RL, and robust VLA from the community: arxiv.org/abs/2505.10911 from Jesse Zhang arxiv.org/abs/2510.14830 from Kun Lei arxiv.org/abs/2509.25358 from Qianzhong Chen and pi0.6 from Physical Intelligence today: physicalintelligence.company/blog/pistar06