Ayush Shrivastava (@ayshrv) 's Twitter Profile
Ayush Shrivastava

@ayshrv

PhD at UMich, Georgia Tech & IIT (BHU) Varanasi Alum.

ID: 2926810536

linkhttp://ayshrv.com/ calendar_today11-12-2014 13:43:00

78 Tweet

396 Takipรงi

1,1K Takip Edilen

Sara Beery (@sarameghanbeery) 's Twitter Profile Photo

At #CVPR2023 we're hosting *Scholars and Big Models* -- a forum to discuss recent rapid changes in CV and how our academic community can adapt and thrive. Our panelists will discuss questions raised by you!! Make your voice heard ๐Ÿ‘‡ forms.gle/hVDg4EhrDXuahFโ€ฆ

At #CVPR2023 we're hosting *Scholars and Big Models* -- a forum to discuss recent rapid changes in CV and how our academic community can adapt and thrive. 

Our panelists will discuss questions raised by you!! Make your voice heard ๐Ÿ‘‡
forms.gle/hVDg4EhrDXuahFโ€ฆ
AK (@_akhaliq) 's Twitter Profile Photo

Seeing the World through Your Eyes paper page: huggingface.co/papers/2306.09โ€ฆ The reflective nature of the human eye is an underappreciated source of information about what the world around us looks like. By imaging the eyes of a moving person, we can collect multiple views of a scene

Nilesh Kulkarni (@_nileshk) 's Twitter Profile Photo

๐Ÿ“ข Learning to Predict Scene-Level Implicit 3D from Posed RGBD Data #CVPR2025. We train a model to predict a 3D implicit function from a single input image. This model is directly trained on raw RGB-D data. Website: nileshkulkarni.github.io/d2drdf Paper: arxiv.org/abs/2306.08671 (1/N)

Taranjeet (@taranjeetio) 's Twitter Profile Photo

๐Ÿš€ย We've hit some big milestones with embedchain: โ€ข 300K apps โ€ข 53K downloads โ€ข 5.7K GitHub stars Every step taught us something new. Today, we're taking those lessons & introducing a platform to manage data for LLM apps. No waitlist, link in next tweet ๐Ÿ‘‡๐Ÿป๐Ÿ‘‡๐Ÿป๐Ÿ‘‡๐Ÿป

Naihao(Neo) Deng (@naihaodeng) 's Twitter Profile Photo

Annotator disagreement is common in NLP, but is it just noise? We are introducing a new strategy for annotator representation to help models better learn from data that has inherent disagreements. ๐Ÿ™ Github code: github.com/MichiganNLP/Anโ€ฆ

Chris Rockwell (@_crockwell) 's Twitter Profile Photo

๐Ÿ“ข Presenting ๐…๐€๐‘: ๐…๐ฅ๐ž๐ฑ๐ข๐›๐ฅ๐ž, ๐€๐œ๐œ๐ฎ๐ซ๐š๐ญ๐ž ๐š๐ง๐ ๐‘๐จ๐›๐ฎ๐ฌ๐ญ ๐Ÿ”๐ƒ๐จ๐… ๐‘๐ž๐ฅ๐š๐ญ๐ข๐ฏ๐ž ๐‚๐š๐ฆ๐ž๐ซ๐š ๐๐จ๐ฌ๐ž ๐„๐ฌ๐ญ๐ข๐ฆ๐š๐ญ๐ข๐จ๐ง #CVPR2024 FAR builds upon complimentary Solver and Learning-Based works yielding accurate *and* robust pose! crockwell.github.io/far/

๐Ÿ“ข Presenting ๐…๐€๐‘: ๐…๐ฅ๐ž๐ฑ๐ข๐›๐ฅ๐ž, ๐€๐œ๐œ๐ฎ๐ซ๐š๐ญ๐ž ๐š๐ง๐ ๐‘๐จ๐›๐ฎ๐ฌ๐ญ ๐Ÿ”๐ƒ๐จ๐… ๐‘๐ž๐ฅ๐š๐ญ๐ข๐ฏ๐ž ๐‚๐š๐ฆ๐ž๐ซ๐š ๐๐จ๐ฌ๐ž ๐„๐ฌ๐ญ๐ข๐ฆ๐š๐ญ๐ข๐จ๐ง #CVPR2024

FAR builds upon complimentary Solver and Learning-Based works yielding accurate *and* robust pose!

crockwell.github.io/far/
Daniel Geng (@dangengdg) 's Twitter Profile Photo

What do you see in these images? These are called hybrid images, originally proposed by Aude Oliva et al. They change appearance depending on size or viewing distance, and are just one kind of perceptual illusion that our method, Factorized Diffusion, can make.

Ziyang Chen (@czyangchen) 's Twitter Profile Photo

These spectrograms look like images, but can also be played as a sound! We call these images that sound. How do we make them? Look and listen below to find out, and to see more examples!

Andrew Owens (@andrewhowens) 's Twitter Profile Photo

In case you were wondering whatโ€™s going on with the back of the #CVPR2024 T-shirt: itโ€™s a hybrid image made by Aaron Inbum Park and Daniel Geng! When you look at it up close, youโ€™ll just see the Seattle skyline, but when you view it from a distance, the text โ€œCVPRโ€ should appear.

Sarah Jabbour (@sarahjabbour_) 's Twitter Profile Photo

๐Ÿ“ขPresenting ๐ƒ๐„๐๐ˆ๐‚๐“: Diffusion-Enabled Permutation Importance for Image Classification Tasks #ECCV2024 We use permutation importance to compute dataset-level explanations for image classifiers using diffusion models (without access to model parameters or training data!)

๐Ÿ“ขPresenting ๐ƒ๐„๐๐ˆ๐‚๐“: Diffusion-Enabled Permutation Importance for Image Classification Tasks #ECCV2024

We use permutation importance to compute dataset-level explanations for image classifiers using diffusion models (without access to model parameters or training data!)
Zhenjun Zhao (@zhenjun_zhao) 's Twitter Profile Photo

Self-Supervised Any-Point Tracking by Contrastive Random Walks Ayush Shrivastava, Andrew Owens tl;dr: global matching transformer->self-attention->transition matrix->contrastive random walk->cycle-consistent track arxiv.org/pdf/2409.16288

Self-Supervised Any-Point Tracking by Contrastive Random Walks

<a href="/ayshrv/">Ayush Shrivastava</a>, <a href="/andrewhowens/">Andrew Owens</a>

tl;dr: global matching transformer-&gt;self-attention-&gt;transition matrix-&gt;contrastive random walk-&gt;cycle-consistent track

arxiv.org/pdf/2409.16288
Andrew Owens (@andrewhowens) 's Twitter Profile Photo

At #ECCV2024: a very simple, self-supervised tracking method! We train a transformer to perform all-pairs matching using the contrastive random walk. If you want to learn more, please come to our poster at 10:30am on Thursday (#214). w/ Ayush Shrivastava x.com/ayshrv/status/โ€ฆ

Prithvijit (@prithvijitch) 's Twitter Profile Photo

Join us at the WorldModelBench workshop at #CVPR2025 where we'll tackle systematic evaluation of World Models! Focus: benchmarks, metrics, downstream tasks, and safety. Submit papers now: worldmodelbench.github.io

Prithvijit (@prithvijitch) 's Twitter Profile Photo

Check out Cosmos-Reason1, a reasoning VLM from our team for - Physical Commonsense Reasoning (spatial, temporal, intuitive physics) - Embodied Reasoning (verifying task completion, action affordance and next plausible action prediction) Models, data curation and benchmarks

Stefan Stojanov (@sstj389) 's Twitter Profile Photo

Video prediction foundation models implicitly learn how objects move in videos. Can we learn how to extract these representations to accurately track objects in videos _without_ any supervision? Yes! ๐Ÿงต Work done with: Rahul Venkatesh, Seungwoo (Simon) Kim, Jiajun Wu and Daniel Yamins