Jayjun Lee (@jayjunleee) 's Twitter Profile
Jayjun Lee

@jayjunleee

Robot Learning; Incoming PhD @UMRobotics; Prev @imperialcollege

ID: 1647069002156621825

linkhttps://jayjunlee.github.io calendar_today15-04-2023 02:47:17

105 Tweet

50 Followers

269 Following

Yinpei Dai (@yinpeid) 's Twitter Profile Photo

Thrilled to introduce our new work RACER ! 🌟A great journey with Jayjun Lee @ CoRL 2025 to explore the role of rich language guidance for failure recovery in robotic manipulation. Exciting times to work on VLM+robotics🤖🚀

Martin Ziqiao Ma (@ziqiao_ma) 's Twitter Profile Photo

Do Vision-Language Models represent space, and how? Spatial terms like "left" or "right" may not be enough to match images with spatial descriptions, as we often overlook the different frames of reference (FoR) used by speakers and listeners. See Figure 1 for examples!

Brian Zheyuan Zhang (@zheyuanzhang99) 's Twitter Profile Photo

Do Vision-Language Models represent space, and how? Introducing 🛋️COnsistent Multilingual Frame Of Reference Test (COMFORT), an evaluation protocol to assess the spatial reasoning in VLMs under ambiguities. 🌐 spatial-comfort.github.io 📄 arxiv.org/abs/2410.17385 More👇

Do Vision-Language Models represent space, and how?

Introducing 🛋️COnsistent Multilingual Frame Of Reference Test (COMFORT), an evaluation protocol to assess the spatial reasoning in VLMs under ambiguities.

🌐 spatial-comfort.github.io
📄 arxiv.org/abs/2410.17385

More👇
Boston Dynamics (@bostondynamics) 's Twitter Profile Photo

Atlas is demonstrating reinforcement learning policies developed using a motion capture suit. This demonstration was developed in partnership with Boston Dynamics and RAI Institute.

Jayjun Lee (@jayjunleee) 's Twitter Profile Photo

We’re presenting RACER today 3:25 pm in Room 411 @ #ICRA2025! Check out our work if you’re interested in steering policies with language and System 1-System 2 frameworks for robot learning! Come say hi!

Nima Fazeli (@nimafazeli7) 's Twitter Profile Photo

👀🤚 Robots that see and feel at once! ViTaSCOPE fuses point-cloud vision with high-res tactile shear to nail in-hand 6-DoF pose plus contact patches—trained entirely in sim, zero-shot on hardware. Dive into the demos 👇 jayjunlee.github.io/vitascope/ #RSS2025 #robotics #tactileSensing

Youngsun Wi (@wiyoungsun) 's Twitter Profile Photo

Exciting to see the VIRDO framework extend to both (1) in-hand pose + extrinsic contact estimation and (2) high-res tactile sensing — very interesting direction!

Martin Ziqiao Ma (@ziqiao_ma) 's Twitter Profile Photo

Thanks AK for sharing our work! The core takeaway from AimBot is straightforward: explicit spatial cues (as shooting lines & scope reticles) are strong 2.5D augmentations that enhance the spatial grounding of any VLA models. These cues are interpretable and also

Jayjun Lee (@jayjunleee) 's Twitter Profile Photo

Thanks AK for posting our work! And I'm happy to share that AimBot 🎯 is accepted to CoRL 2025 Conference on Robot Learning! See you in Seoul! Project webpage: aimbot-reticle.github.io Thanks to my amazing co-lead Yinpei Dai, co-authors, and our advisors Nima Fazeli, Michigan SLED Lab

Jayjun Lee (@jayjunleee) 's Twitter Profile Photo

I’m co-organizing the H2R workshop and we’re starting soon! Come join our session in Room E3. sites.google.com/view/h2r-corl2…

I’m co-organizing the H2R workshop and we’re starting soon! Come join our session in Room E3.

sites.google.com/view/h2r-corl2…