Omar Alama عمر الأعمى (@omaralama) 's Twitter Profile
Omar Alama عمر الأعمى

@omaralama

ECE Vision and Robot Perception PhD @ Carnegie Mellon University

ID: 700919826

linkhttps://www.linkedin.com/in/omar-alama-651442169/ calendar_today17-07-2012 11:54:07

153 Tweet

179 Followers

345 Following

Nikhil Keetha (@nik__v__) 's Twitter Profile Photo

Super excited about this work pushing the boundary of online semantic mapping!! One step closer to making robots see the world 🌎🚀 Check out Omar's thread for a lot more eye candy 🤩 and impressive results!

Yuheng Qiu (@qiuyuhengqiu) 's Twitter Profile Photo

Getting tired of visual language navigation in indoor environments? Check RayFronts🏹, open-set semantic mapping for beyond depth-sensing-range observations.

Omar Alama عمر الأعمى (@omaralama) 's Twitter Profile Photo

أول ورقة بحثية لي في مسيرة الدكتوراه. فكرة المشروع ببساطة اعطاء الروبوت او الدرون المقدرة على فهم جميع الأشياء من حوله في أي بيئة. بحيث تقدر تسأل الروبوت بأي نص أو أي صورة عن أي شيء شافه. اذا كان الشيء داخل اطار احساسه بالمسافة فحيقدر يحدد مكانه بدقة. وان كان أبعد فيحدد اتجاهه.

Omar Alama عمر الأعمى (@omaralama) 's Twitter Profile Photo

SIGLIP wins over CLIP even in dense tasks like zero shot open-vocab semantic segmentation on Replica . Using the RayFronts encoder (NA attention + RADIO Pavlo Molchanov + SIGLIP Lucas Beyer (bl16)) projection to the CLS token gives you SoTA performance. No more SAM+CROP+CLIP business.

SIGLIP wins over CLIP even in dense tasks like zero shot open-vocab semantic segmentation on Replica . Using the RayFronts encoder (NA attention + RADIO <a href="/PavloMolchanov/">Pavlo Molchanov</a> + SIGLIP <a href="/giffmana/">Lucas Beyer (bl16)</a>) projection to the CLS token gives you SoTA performance. No more SAM+CROP+CLIP business.
Omar Alama عمر الأعمى (@omaralama) 's Twitter Profile Photo

نمط متكرر. الناس تكتشف أسئلة بسيطة جدا النماذج اللغوية مثل ChatGpt تفشل في الجواب عليها. الشركات القائمة على النماذج تنتبه، تعيد تدريب النموذج ليجاوب بشكل صحيح. الناس تنبهر وتقول الآن صار فاهم. وتنعاد الحلقة. الواقع هو ان النماذج اللغوية الى الآن تنطبق عليها مقولة "حافظ مش فاهم".

AirLab (@airlabcmu) 's Twitter Profile Photo

🚀 Thrilled to present ViSafe, a vision-only airborne collision avoidance system that achieved drone-to-drone avoidance at 144 km/h. In an era of congested airspace and growing autonomy, reliable self-separation is paramount 🧵👇

CMU Robotics Institute (@cmu_robotics) 's Twitter Profile Photo

#ICRA2025 alert! 🚨🥳 Congratulations to Yuheng Qui, Yutian Chen, Zihao Zhang, Wenshan Wang, and Sebastian Scherer on winning the Best Conference Paper Award for: "MAC-VO: Metrics-Aware Covariance for Leaning-Based Stereo Visual Odometry"! #CMUrobotics arxiv.org/abs/2409.09479

#ICRA2025 alert! 🚨🥳

Congratulations to Yuheng Qui, Yutian Chen, Zihao Zhang, Wenshan Wang, and Sebastian Scherer on winning the Best Conference Paper Award for:

"MAC-VO: Metrics-Aware Covariance for Leaning-Based Stereo Visual Odometry"! #CMUrobotics

arxiv.org/abs/2409.09479
Omar Alama عمر الأعمى (@omaralama) 's Twitter Profile Photo

RayFronts code has been released ! github.com/RayFronts/RayF… 🤖 Guide your robot with semantics within & beyond depth. 🖼️ Stop using slow SAM crops + CLIP pipelines. RayFronts gets dense language aligned features in one forward pass. 🚀 Test your mapping ideas in our pipeline !

Omar Alama عمر الأعمى (@omaralama) 's Twitter Profile Photo

I was waiting for the AI to make a mistake the whole time. Was shocked by the quality. It was even simplifying new concepts introduced in our paper with analogies. Really impressive tool NotebookLM Listen to the full podcast here youtu.be/_tIVlw1Wrh4?si…

Nikhil Keetha (@nik__v__) 's Twitter Profile Photo

Want to learn how to empower 🤖 with real-time scene understanding and exploration capabilities? Catch Me, Cherie Ho & Yuheng Qiu presenting RayFronts at #RSS2025 SemRob Workshop (OHE 122) & Epstein Plaza at 10:00 am PST Today!

Omar Alama عمر الأعمى (@omaralama) 's Twitter Profile Photo

"When a measure becomes a target, it ceases to be a good measure" Goodhart's law. Seems more and more true with these LLM and LVLM benchmarks. Higher numbers don't always reflect real world performance.

Luca Carlone (@lucacarlone1) 's Twitter Profile Photo

We have completed the SLAM Handbook "From Localization and Mapping to Spatial Intelligence" and released it online: asrl.utias.utoronto.ca/~tdb/slam/ . The handbook will be published by Cambridge University Press. [1/n]

Omar Alama عمر الأعمى (@omaralama) 's Twitter Profile Photo

🐦‍⬛Very excited for RAVEN which truly supports long-term and long-range aerial outdoor object-goal navigation in unseen environments. 🔥 Checkout Seungchan's post to see it in action !

Omar Alama عمر الأعمى (@omaralama) 's Twitter Profile Photo

⛔️Stop throwing away far range semantics, encode them as Rays instead ! 🔥Excited to present RayFronts at #IROS2025 in Hangzhou, China ! 🎥Catch us in the live presentation next Tuesday 16:45-16:50 Track 9.

Omar Alama عمر الأعمى (@omaralama) 's Twitter Profile Photo

➡️RayFronts can be viewed as view-conditioned semantic frontiers where the viewing angle determines the semantic feature. But I find rays more straight forward. I'll visualize the alternate interpretation some day. For now, getting ready to present RayFronts at #IROS2025 🔥!

Omar Alama عمر الأعمى (@omaralama) 's Twitter Profile Photo

First, I thought ZED-X stereo depth was unusable for flying drones. The intense vibrations can make depth estimation go crazy 👎. But with aggressive voxel grid filtering, you can get something reasonable, not for fine reconstruction, but for autonomy.🤖 See you in #IROS2025 !

Omar Alama عمر الأعمى (@omaralama) 's Twitter Profile Photo

↗️Many don't realize that RayFronts can operate at any depth range, Yes, even at 0 depth range‼️We don't show this visual anywhere in the paper or website. We probably should. Catch our RayFronts presentation live at #IROS2025 Tuesday 16:45-16:50 !