Abrar Anwar (@_abraranwar) 's Twitter Profile
Abrar Anwar

@_abraranwar

CS PhD student at @USCViterbi | prev intern @nvidia @Cornell @SandiaLabs | undergrad @UTCompSci

ID: 1703364902

linkhttp://abraranwar.github.io calendar_today27-08-2013 00:59:26

133 Tweet

485 Followers

667 Following

Nathan Dennler (@ndennler) 's Twitter Profile Photo

User-aligned robot representations are best learned from user data, but how do we collect this data without onerous labeling processes? We can learn from data that users are ✨intrinsically motivated✨ to produce through exploratory search 🔎! 🧵1/6

Gautam Salhotra (@gautamsalhotra) 's Twitter Profile Photo

We see so many startups offering robot data collection -- how come we don't see ones focussing on real robot evals as a service? There should be real-robot benchmarks for all organisations to test their VLAs on.

Yiğit Korkmaz (@yigitkkorkmaz) 's Twitter Profile Photo

I recently wrote a post about MILE for USC Robotics blog — check it out here: rasc.usc.edu/blog/mile-mode… Feel free to reach out if you have any questions or thoughts! See you at ICRA 🤖🙂

Abrar Anwar (@_abraranwar) 's Twitter Profile Photo

Arrived at #ICRA2025 and I'll be presenting my ReMEmbR work with NVIDIA on Tuesday! Happy to chat with people on robot memory, evaluation, language+robots, and reward learning (more to come soon on this one 😉)!

Jesse Zhang (@jesse_y_zhang) 's Twitter Profile Photo

Reward models that help real robots learn new tasks—no new demos needed! ReWiND uses language-guided rewards to train bimanual arms on OOD tasks in 1 hour! Offline-to-online, lang-conditioned, visual RL on action-chunked transformers. 🧵

Erdem Bıyık (@ebiyik_) 's Twitter Profile Photo

Jesse is such a naturally talented advisor that I didn't demonstrate anything about advising, just gave some language instructions. Do you know who else is so good? The model. Check it out 👇🏻

Erdem Bıyık (@ebiyik_) 's Twitter Profile Photo

By the way, Jesse graduated last week and will move to Allen School for a postdoc. Be on the lookout for him when he enters the job market in a few years 🌟

By the way, Jesse graduated last week and will move to <a href="/uwcse/">Allen School</a> for a postdoc. Be on the lookout for him when he enters the job market in a few years 🌟
Erdem Bıyık (@ebiyik_) 's Twitter Profile Photo

Allen School Jesse Zhang Another shoutout goes to Jiahui Zhang , Yusen Luo and Abrar Anwar , who did most of the work. Yusen will apply for PhD in the next cycle, so another great candidate to be aware of. Abrar and I are both at #ICRA2025. We will be happy to chat with anyone about the work :)

Abhishek Gupta (@abhishekunique7) 's Twitter Profile Photo

Very very exciting to have Jesse Zhang join us at UW soon! He's done some incredible work - I'd recommend reading rewind-reward.github.io! Congratulations on a fantastic Ph.D. Jesse Zhang 🎉

Sumedh Sontakke (@sota_kke) 's Twitter Profile Photo

Reward learning (like DYNA) has enabled e2e policies to reach 99% SR but (1) generalization to new tasks and (2) sample efficiency are still hard! ReWiND produces better rewards for OOD tasks than SOTA like GVL & LIV from Jason Ma that inspired us! 🌐: rewind-reward.github.io

Rajat Kumar Jenamani (@rkjenamani) 's Twitter Profile Photo

Excited to share our work on continual, flexible, active, and safe robot personalization w/ Tom Silver, Ziang Liu, Ben Dodson & Tapomayukh "Tapo" Bhattacharjee. Also: Tom Silver is starting a lab at Princeton!! I HIGHLY recommend joining — thoughtful, kind, and an absolute joy to work with!

Gautam Salhotra (@gautamsalhotra) 's Twitter Profile Photo

Wondering how to get more from your robot finetuning datasets? MILE extracts more out of intervention-based demos during training, giving you more bang for your buck per demonstration. Read more in the new USC Robotics blogpost! rasc.usc.edu/blog/mile-mode… Yiğit Korkmaz Erdem Bıyık

C's Robotics Paper Notes (@roboreading) 's Twitter Profile Photo

rewind-reward.github.io ReWiND: Language-Guided Rewards Teach Robot Policies without New Demonstrations LLM generated instructions z + demo -> learning-based reward model (progress) R(o,z) -> optimize policy via RL online

rewind-reward.github.io

ReWiND: Language-Guided Rewards Teach Robot Policies without New Demonstrations

LLM generated instructions z + demo -&gt; learning-based reward model (progress) R(o,z) -&gt; optimize policy via RL online
Jesse Zhang (@jesse_y_zhang) 's Twitter Profile Photo

How can non-experts quickly teach robots a variety of tasks? Introducing HAND ✋, a simple, time-efficient method of training robots! Using just a **single hand demo**, HAND learns manipulation tasks in under **4 minutes**! 🧵

Jenny Zhang (@jennyzhangzt) 's Twitter Profile Photo

**When AIs Start Rewriting Themselves** Darwin G​​ödel Machine: Open-Ended Evolution of Self-Improving Agents The Darwin G​​ödel Machine can: 1. Read and modify its own code 2. Evaluate if the change improves performance 3. Open-endedly explore the solution space 🧵👇

Ishika Singh (@ishika_s_) 's Twitter Profile Photo

VLAs have the potential to generalize over scenes and tasks, but require a ton of data to learn robust policies. We introduce OG-VLA, a novel architecture and learning framework that combines the generalization strengths of VLAs with the robustness of 3D-aware policies. 🧵