Christoph Feichtenhofer (@cfeichtenhofer) 's Twitter Profile
Christoph Feichtenhofer

@cfeichtenhofer

Research Scientist @MetaAI (FAIR)

ID: 1000007035

linkhttp://feichtenhofer.github.io calendar_today09-12-2012 19:36:31

5 Tweet

287 Takipçi

29 Takip Edilen

AK (@_akhaliq) 's Twitter Profile Photo

Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles abs: arxiv.org/abs/2306.00989 Modern hierarchical vision transformers have added several vision-specific components in the pursuit of supervised classification performance. While these components lead to

Hiera: A Hierarchical Vision Transformer without the Bells-and-Whistles

abs: arxiv.org/abs/2306.00989

Modern hierarchical vision transformers have added several vision-specific components in the pursuit of supervised classification performance. While these components lead to
Christoph Feichtenhofer (@cfeichtenhofer) 's Twitter Profile Photo

At #ICML2023, Chay (Chaitanya (Chay) Ryali) will present Hiera, a hierarchical vision transformer that is fast, powerful, and simple. Code+models at: github.com/facebookresear… If interested, please come to the oral presentation on Tue 25 Jul 5:30pm HST or poster #219 on Wed 26 Jul 2pm HST.

Christoph Feichtenhofer (@cfeichtenhofer) 's Twitter Profile Photo

We're excited to release SAM 2 to segment anything in images and videos! Data, model, and code are available under permissive open licenses. Demo: sam2.metademolab.com Code: github.com/facebookresear… Paper: ai.meta.com/research/publi…

AI at Meta (@aiatmeta) 's Twitter Profile Photo

Segment Anything Model 2 (SAM 2) is a foundation model from Meta FAIR for promptable visual segmentation in images & videos. Available now for anyone to build on for free, open source under an Apache license. Try the demo ➡️ go.fb.me/ve0y8o

merve (@mervenoyann) 's Twitter Profile Photo

Don't sleep on this! 🔥 Meta dropped swiss army knives for vision with A2.0 license ❤️ > image/video encoders for vision language and spatial understanding (object detection etc) > VLM outperforms InternVL3 and Qwen2.5VL 🔥 > Gigantic video and image datasets 👏

Don't sleep on this! 🔥

<a href="/Meta/">Meta</a> dropped swiss army knives for vision with A2.0 license ❤️
&gt; image/video encoders for vision language and spatial understanding (object detection etc)
&gt; VLM outperforms InternVL3 and Qwen2.5VL 🔥
&gt; Gigantic video and image datasets 👏