Anh-Quan Cao (@anhquancao) 's Twitter Profile
Anh-Quan Cao

@anhquancao

PhD Student in Computer Vision @inria. Previously @amazon, @TU_Muenchen, @Polytechnique and @UnivParisSaclay

ID: 1304518849572806659

linkhttp://anhquancao.github.io calendar_today11-09-2020 20:35:15

885 Tweet

387 Takipรงi

757 Takip Edilen

Quankai Gao (@uuuuusher) 's Twitter Profile Photo

๐Ÿš€ Introducing InstantSfM: Fully Sparse and Parallel Structure-from-Motion. โœ… Python + GPU-optimized implementation, no C++ anymore! โœ… 40ร— faster than COLMAP with 5K images on single GPU! โœ… Scales beyond 100 images (more than VGGT/VGGSfM can consume)! โœ… Support metric scale.

๐Ÿš€ Introducing InstantSfM: Fully Sparse and Parallel Structure-from-Motion.
โœ… Python + GPU-optimized implementation, no C++ anymore!
โœ… 40ร— faster than COLMAP with 5K images on single GPU!
โœ… Scales beyond 100 images (more than VGGT/VGGSfM can consume)!
โœ… Support metric scale.
valeo.ai (@valeoai) 's Twitter Profile Photo

Our recent research will be presented at #ICCV2025 #ICCV2025! Weโ€™ll present 5 papers about: ๐Ÿ’ก self-supervised & representation learning ๐ŸŒ 3D occupancy & multi-sensor perception ๐Ÿงฉ open-vocabulary segmentation ๐Ÿง  multimodal LLMs & explainability valeoai.github.io/posts/iccv-2025

Our recent research will be presented at #ICCV2025 <a href="/ICCVConference/">#ICCV2025</a>!
Weโ€™ll present 5 papers about:
๐Ÿ’ก self-supervised &amp; representation learning
๐ŸŒ 3D occupancy &amp; multi-sensor perception
๐Ÿงฉ open-vocabulary segmentation
๐Ÿง  multimodal LLMs &amp; explainability

valeoai.github.io/posts/iccv-2025
valeo.ai (@valeoai) 's Twitter Profile Photo

Today Corentin Dancette from Raidium stopped by to talk about Curia, a multi-modal open-weights foundation model for radiology building in the style of DINOv2 with tons of downstream applications. Do check it out!

Today <a href="/cdancette/">Corentin Dancette</a> from <a href="/raidium_med/">Raidium</a> stopped by to talk about Curia, a multi-modal open-weights foundation model for radiology building in the style of DINOv2 with tons of downstream applications.
Do check it out!
Yasser Benigmim (@yasserbenigmim) 's Twitter Profile Photo

Working with CLIP for classification, segmentation, or detection? Want to boost performance without labels and while staying training-free? Come see our poster #157 (Session 1) tomorrow!! weโ€™ll tell you more about FLOSS: Free Lunch in Open-Vocabulary Semantic Segmentation!

valeo.ai (@valeoai) 's Twitter Profile Photo

Come to the FLOSS poster #157 on Thursday #ICCV2025 to find out more about this really simple strategy for boosting open-vocabulary semantic segmentation and meet Yasser Benigmim & Tuan-Hung VU

valeo.ai (@valeoai) 's Twitter Profile Photo

Thrilled to have Jayneel Parekh visiting us over a break in the CVPR sprint to talk about his exciting works in analyzing MLLMs representations. Check out his great work jayneelparekh.github.io

Thrilled to have <a href="/TrueBeerus/">Jayneel Parekh</a> visiting us over a break in the CVPR sprint to talk about his exciting works in analyzing MLLMs representations. Check out his great work jayneelparekh.github.io
valeo.ai (@valeoai) 's Twitter Profile Photo

Honored to host a very special guest today: Huy V. Vo, PhD alumnus of valeo.ai. He presented his work on automatic data-curation strategies for self-supervised representation learning (DINOv2, DINOv3). Find out more about his research here: huyvvo.github.io

Honored to host a very special guest today: Huy V. Vo, PhD alumnus of <a href="/valeoai/">valeo.ai</a>. 
He presented his work on automatic data-curation strategies for self-supervised representation learning (DINOv2, DINOv3). Find out more about his research here: huyvvo.github.io
rsasaki0109 (@rsasaki0109) 's Twitter Profile Photo

[RA-L 2025] ActiveSplat: High-Fidelity Scene Reconstruction through Active Gaussian Splatting github.com/Li-Yuetao/Actiโ€ฆ ActiveSplat enables the agent to explore the environment autonomously to build a 3D map on the fly. The integration of a Gaussian map and a Voronoi graph assures

Loรฏck Ch (@loickch) 's Twitter Profile Photo

Need pixel-level features from your backbone (DINOv3, CLIP, RADIO, FRANCA...)? ๐Ÿš€Introducing NAF: A universal, zero-shot feature upsampler. It turns low-res ViT features into pixel-perfect maps. -โšก Model-agnostic -๐Ÿฅ‡ SoTA results -๐Ÿš€ 4ร— faster than SoTA -๐Ÿ“ˆ Scales up to 2K res.

valeo.ai (@valeoai) 's Twitter Profile Photo

Off-the-shelf feature upscaling from any vision model without retraining. ๐Ÿฅ‡ SoTA results โšก๏ธ High throughput ๐Ÿง  Zero-shot generalization ๐Ÿ˜ Code available ๐Ÿ‘‡ Check out the details!

Andrei Bursuc (@abursuc) 's Twitter Profile Photo

Check out NAF: an effective ViT feature upsampler to produce excellent (and eye-candy) pixel-level feature maps. NAF outperform both VFM-specific upsamplers (FeatUp, JAFAR) and VFM-agnostic methods (JBU, AnyUp) over multiple downstream tasks ๐Ÿ‘‡

Loรฏck Ch (@loickch) 's Twitter Profile Photo

๐ŸŽ‰ Try NAF live โ€” now on Hugging Face! ๐Ÿค— ๐Ÿ–ผ๏ธ Upload any image ๐Ÿค– Select any VFM (DINOv2, CLIP, RADIO, FRANCA...) ๐Ÿ“ Choose any resolution โšก Instant upsampling Try it now: huggingface.co/spaces/LChamboโ€ฆ

Yuan Yin (@yuanyinnn) 's Twitter Profile Photo

1/Serve your PEFT with a fresh IPA!๐Ÿบ Finetuning large models is cheaper thanks to LoRA, but is its random init optimal?๐Ÿค” Meet IPA: a feature-aware alternative to random projections #NeurIPS2025 WS #CCFM Oral+Best Paper Work w/Shashank Tuan-Hung VU Andrei Bursuc Matthieu Cord ๐Ÿงต

valeo.ai (@valeoai) 's Twitter Profile Photo

Check out our works at NeurIPS Conference #NeurIPS2025 this week! We present 5 full papers + 1 workshop about: ๐Ÿ’ก self-supervised & representation learning ๐Ÿ–ผ๏ธ generative image models ๐Ÿง  finetuning & understanding LLMs & multimodal LLMs ๐Ÿ”Ž feature upsampling valeoai.github.io/posts/neurips-โ€ฆ

Check out our works at <a href="/NeurIPSConf/">NeurIPS Conference</a> #NeurIPS2025 this week!
We present 5 full papers + 1 workshop about:
๐Ÿ’ก self-supervised &amp; representation learning
๐Ÿ–ผ๏ธ generative image models
๐Ÿง  finetuning &amp; understanding LLMs &amp; multimodal LLMs
๐Ÿ”Ž feature upsampling

valeoai.github.io/posts/neurips-โ€ฆ
Andrei Bursuc (@abursuc) 's Twitter Profile Photo

I'm speaking at #aiPULSE2025 today on Open & re-purposable foundation models for the automotive industry. The morning keynotes talked a lot about open source so my slide here might be timely.

I'm speaking at #aiPULSE2025 today on Open &amp; re-purposable foundation models for the automotive industry. 
The morning keynotes talked a lot about open source so my slide here might be timely.
Yuan Yin (@yuanyinnn) 's Twitter Profile Photo

Quick clarification: we first shared the Outstanding Paper Award thinking it was the top prize โ€” and it turns out IPA did win Best Paper! Absolutely thrilled! ๐ŸŽ‰

Matthias Niessner (@mattniessner) 's Twitter Profile Photo

Today in our TUM AI - Lecture Series we'll have the amazing Ruiqi Gao, Google DeepMind. She'll talk about "๐๐ฎ๐ข๐ฅ๐๐ข๐ง๐  ๐ ๐ž๐ง๐ž๐ซ๐š๐ญ๐ข๐ฏ๐ž ๐ฐ๐จ๐ซ๐ฅ๐ ๐ฆ๐จ๐๐ž๐ฅ๐ฌ: progress and challenges". Live stream: youtube.com/live/CkOSMqwvFโ€ฆ 7pm GMT+1 / 10am PST (Tue Dec 16th).

Today in our TUM AI - Lecture Series we'll have the amazing <a href="/RuiqiGao/">Ruiqi Gao</a>, Google DeepMind.

She'll talk about "๐๐ฎ๐ข๐ฅ๐๐ข๐ง๐  ๐ ๐ž๐ง๐ž๐ซ๐š๐ญ๐ข๐ฏ๐ž ๐ฐ๐จ๐ซ๐ฅ๐ ๐ฆ๐จ๐๐ž๐ฅ๐ฌ: progress and challenges".

Live stream: youtube.com/live/CkOSMqwvFโ€ฆ

7pm GMT+1 / 10am PST (Tue Dec 16th).