
Ayush Shrivastava
@ayshrv
PhD at UMich, Georgia Tech & IIT (BHU) Varanasi Alum.
ID: 2926810536
http://ayshrv.com/ 11-12-2014 13:43:00
78 Tweet
396 Takipรงi
1,1K Takip Edilen


Seeing the World through Your Eyes paper page: huggingface.co/papers/2306.09โฆ The reflective nature of the human eye is an underappreciated source of information about what the world around us looks like. By imaging the eyes of a moving person, we can collect multiple views of a scene

๐ข Learning to Predict Scene-Level Implicit 3D from Posed RGBD Data #CVPR2025. We train a model to predict a 3D implicit function from a single input image. This model is directly trained on raw RGB-D data. Website: nileshkulkarni.github.io/d2drdf Paper: arxiv.org/abs/2306.08671 (1/N)

๐ย We've hit some big milestones with embedchain: โข 300K apps โข 53K downloads โข 5.7K GitHub stars Every step taught us something new. Today, we're taking those lessons & introducing a platform to manage data for LLM apps. No waitlist, link in next tweet ๐๐ป๐๐ป๐๐ป


๐ข Presenting ๐ ๐๐: ๐ ๐ฅ๐๐ฑ๐ข๐๐ฅ๐, ๐๐๐๐ฎ๐ซ๐๐ญ๐ ๐๐ง๐ ๐๐จ๐๐ฎ๐ฌ๐ญ ๐๐๐จ๐ ๐๐๐ฅ๐๐ญ๐ข๐ฏ๐ ๐๐๐ฆ๐๐ซ๐ ๐๐จ๐ฌ๐ ๐๐ฌ๐ญ๐ข๐ฆ๐๐ญ๐ข๐จ๐ง #CVPR2024 FAR builds upon complimentary Solver and Learning-Based works yielding accurate *and* robust pose! crockwell.github.io/far/




In case you were wondering whatโs going on with the back of the #CVPR2024 T-shirt: itโs a hybrid image made by Aaron Inbum Park and Daniel Geng! When you look at it up close, youโll just see the Seattle skyline, but when you view it from a distance, the text โCVPRโ should appear.

Self-Supervised Any-Point Tracking by Contrastive Random Walks Ayush Shrivastava, Andrew Owens tl;dr: global matching transformer->self-attention->transition matrix->contrastive random walk->cycle-consistent track arxiv.org/pdf/2409.16288


At #ECCV2024: a very simple, self-supervised tracking method! We train a transformer to perform all-pairs matching using the contrastive random walk. If you want to learn more, please come to our poster at 10:30am on Thursday (#214). w/ Ayush Shrivastava x.com/ayshrv/status/โฆ




Video prediction foundation models implicitly learn how objects move in videos. Can we learn how to extract these representations to accurately track objects in videos _without_ any supervision? Yes! ๐งต Work done with: Rahul Venkatesh, Seungwoo (Simon) Kim, Jiajun Wu and Daniel Yamins