AI Habitat (@ai_habitat) 's Twitter Profile
AI Habitat

@ai_habitat

A platform for Embodied AI research.

Project by FAIR (@MetaAI) and Facebook Reality Labs, in collaboration with @gtcomputing, @SFU, @IntelAI, @UCBerkeley.

ID: 1090794734010822657

linkhttps://aihabitat.org/ calendar_today31-01-2019 02:11:45

492 Tweet

1,1K Followers

31 Following

Naoki Yokoyama (@naokiyokoyama0) 's Twitter Profile Photo

Excited to share our latest work, Vision-Language Frontier Maps – a SOTA approach for semantic navigation in robotics. VLFM enables robots to navigate and find objects in novel environments using vision-language foundation models, zero-shot! Accepted to #ICRA2024! 🧵

Freethink (@freethinkmedia) 's Twitter Profile Photo

Does AI need a “body” to become truly intelligent? Meta researchers think so. We may be on the brink of finally seeing human-level intelligence in an AI — thanks to robots. AIs that can generate videos, quickly translate languages, or write new computer code could be world

Naoki Yokoyama (@naokiyokoyama0) 's Twitter Profile Photo

Our #ICRA2024 paper was selected as a finalist for the Best Paper Award in Cognitive Robotics! I'll be giving a presentation on this paper at the conference for a shot at the award. See you there!

C Zhang (@chongzitazhang) 's Twitter Profile Photo

like this line of work so much Object-oriented navigation Habitat simulator And control the dog Solid contribution, function in real world at a scale

AI at Meta (@aiatmeta) 's Twitter Profile Photo

Today we’re releasing OpenEQA — the Open-Vocabulary Embodied Question Answering Benchmark. It measures an AI agent’s understanding of physical environments by probing it with open vocabulary questions like “Where did I leave my badge?” More details ➡️ go.fb.me/7vq6hm

Dhruv Batra (@dhruvbatradb) 's Twitter Profile Photo

I have been working on vision+language models (VLMs) for a decade. And every few years, this community re-discovers the same lesson -- that on difficult tasks, VLMs regress to being nearly blind! Visual content provides minor improvement to a VLM over an LLM, even when these

I have been working on vision+language models (VLMs) for a decade. 

And every few years, this community re-discovers the same lesson -- that on difficult tasks, VLMs regress to being nearly blind! 

Visual content provides minor improvement to a VLM over an LLM, even when these
🇺🇦Olexandr Maksymets (@o_maksymets) 's Twitter Profile Photo

Launching OpenEQA, our new benchmark for AI's understanding of physical environments. Despite AGI optimism, our tests with top VLMs reveal a significant gap to human-level comprehension. Let's bridge this gap in AI's world understanding.

The Hill (@thehill) 's Twitter Profile Photo

Washingtonians delved into the world of artificial intelligence (AI) at the Washington AI Network’s inaugural weekend TGAIFriday Lunch for White House correspondents. trib.al/FwHF9Um

Dhruv Batra (@dhruvbatradb) 's Twitter Profile Photo

FAIR researchers (AI at Meta) presented SegmentAnything and our robotics work at the White House correspondents’ weekend. Llama3 + Sim2Real skills (trained with AI Habitat) = a robot assistant

FAIR researchers (<a href="/AIatMeta/">AI at Meta</a>) presented SegmentAnything and our robotics work at the White House correspondents’ weekend. 

Llama3 + Sim2Real skills (trained with <a href="/ai_habitat/">AI Habitat</a>)  = a robot assistant
Roozbeh Mottaghi (@roozbehmottaghi) 's Twitter Profile Photo

On Tuesday at #ICLR2024, we will be presenting Habitat 3.0, our human/robot simulator, featuring a VR demo that allows you to interact firsthand with both the scenes and the robot.

Dhruv Batra (@dhruvbatradb) 's Twitter Profile Photo

Naoki Yokoyama Naoki Yokoyama presenting his best paper award finalist talk at #ICRA2024! Vision-Language Frontier Maps for Zero-Shot Semantic Navigation: show how to combine VL foundation models with a mapping+search stack. Georgia Tech School of Interactive Computing Robotics@GT Machine Learning at Georgia Tech Boston Dynamics

Naoki Yokoyama <a href="/naokiyokoyama0/">Naoki Yokoyama</a> presenting his best paper award finalist talk at #ICRA2024!

Vision-Language Frontier Maps for Zero-Shot Semantic Navigation: show how to combine VL foundation models with a mapping+search stack. 

<a href="/ICatGT/">Georgia Tech School of Interactive Computing</a> <a href="/GTrobotics/">Robotics@GT</a> <a href="/mlatgt/">Machine Learning at Georgia Tech</a> <a href="/BostonDynamics/">Boston Dynamics</a>
Georgia Tech Computing (@gtcomputing) 's Twitter Profile Photo

This is no small feat. At the robotics field's #1 research venue, our researchers are #1, earning the Best Paper Award in Cognitive Robotics, among the eligible 1,700+ papers at #ICRA2024. Partnered with Boston Dynamics. 🤖🏆🎉 Award details: ieee-ras.org/awards-recogni…