vislang.ai (@vislang) 's Twitter Profile
vislang.ai

@vislang

Twitter account of the Vision, Language and Learning Lab, Computer Science @ Rice University.

ID: 1270579924169035776

linkhttps://www.vislang.ai calendar_today10-06-2020 04:54:08

86 Tweet

3,3K Followers

1,1K Following

Imagine-ENPC (@imagineenpc) 's Twitter Profile Photo

#ICCV2023 Fri Oct 6 (PM) Going Beyond Nouns With Vision & Language Models Using Synthetic Data Cascante, Shehada, James Smith, Sivan Doveh, Kim, Rameswar Panda, Gül Varol, Aude Oliva, vislang.ai, Rogerio Feris, Leonid Karlinsky pdf: arxiv.org/abs/2303.17590 web: synthetic-vic.github.io

#ICCV2023 Fri Oct 6 (PM)
Going Beyond Nouns With Vision & Language Models Using Synthetic Data
Cascante, Shehada, <a href="/jamessealesmith/">James Smith</a>, <a href="/SivanDoveh/">Sivan Doveh</a>, Kim, <a href="/rpanda89/">Rameswar Panda</a>, <a href="/gulvarol/">Gül Varol</a>, <a href="/AudeOliva/">Aude Oliva</a>, <a href="/vislang/">vislang.ai</a>, <a href="/RogerioFeris/">Rogerio Feris</a>, <a href="/leokarlin/">Leonid Karlinsky</a>
pdf: arxiv.org/abs/2303.17590
web: synthetic-vic.github.io
Paola Cascante-Bonilla (@pcascanteb) 's Twitter Profile Photo

With Code+Data available, I'm happy to share one of my latest works to be presented at #ICCV2023! Going Beyond Nouns With Vision & Language Models Using Synthetic Data Project page: synthetic-vic.github.io Data & Code: github.com/uvavision/SyViC (1/6)

Rice Computer Science (@ricecompsci) 's Twitter Profile Photo

Rice CS PhD student Paola Cascante-Bonilla introduces a 1M-scale synthetic dataset at #iccv2023. It allows users to add synthetically generated objects like furniture & humans to an image & is the result of her collaboration with her vislang.ai advisor Vicente Ordónez. bit.ly/3QIqA1p

Rice CS PhD student <a href="/pcascanteb/">Paola Cascante-Bonilla</a> introduces a 1M-scale synthetic dataset at #iccv2023. It allows users to add synthetically generated objects like furniture &amp; humans to an image &amp; is the result of her collaboration with her <a href="/vislang/">vislang.ai</a> advisor Vicente Ordónez. bit.ly/3QIqA1p
Jaspreet Ranjit (@jaspreetranjit_) 's Twitter Profile Photo

How do biases change before and after finetuning large scale visual recognition models? Our AFME 2024 @ NeurIPS paper incorporates sets of canonical images to highlight changes in biases for an array of off-the-shelf pretrained models. #NeurIPS2023 Link: arxiv.org/abs/2303.07615

How do biases change before and after finetuning large scale visual recognition models? Our <a href="/afciworkshop/">AFME 2024 @ NeurIPS</a> paper incorporates sets of canonical images to highlight changes in biases for an array of off-the-shelf pretrained models. #NeurIPS2023 

Link: arxiv.org/abs/2303.07615
MOAYED HAJi ALi (@moayedhaji) 's Twitter Profile Photo

Great news from #CVPR2024 🎉🎉🎉 Happy to share that our paper ElasticDiffusion: Training-free Arbitrary Size Image Generation was accepted #CVPR2025. Big thanks to my collaborators Vicente Ordóñez and Guha Balakrishnan. Checkout more details from here: elasticdiffusion.github.io

Great news from #CVPR2024 🎉🎉🎉

Happy to share that our paper ElasticDiffusion: Training-free Arbitrary Size Image Generation was accepted <a href="/CVPR/">#CVPR2025</a>. 
Big thanks to my collaborators <a href="/bluevincent/">Vicente Ordóñez</a> and Guha Balakrishnan. 
Checkout more details from here: elasticdiffusion.github.io
harpreet (@datascienceharp) 's Twitter Profile Photo

Chatted with Ruozhen Catherine He and Paola Cascante-Bonilla from vislang.ai at Rice University about the paper they had accepted to #CVPR2025 Their paper introduces Self-Consistency Equivalence Tuning (SelfEQ) to improve visual grounding in vision-and-language models using paraphrases. The Problem:

Ruozhen Catherine He (@cathyrzhe) 's Twitter Profile Photo

(1/4) Excited to share our latest work at #CVPR2024 #CVPR2025!🔥 Join us tomorrow, Thursday, June 20, from 10:30am to noon at Poster Session 3, # 334, to learn about "Improved Visual Grounding through Self-Consistent Explanations" with Paola Cascante-Bonilla, Ziyan, Alex Berg, vislang.ai.

(1/4) Excited to share our latest work at #CVPR2024 <a href="/CVPR/">#CVPR2025</a>!🔥

Join us tomorrow, Thursday, June 20, from 10:30am to noon at Poster Session 3, # 334, to learn about "Improved Visual Grounding through Self-Consistent Explanations" with <a href="/pcascanteb/">Paola Cascante-Bonilla</a>, Ziyan, <a href="/alexandercberg/">Alex Berg</a>, <a href="/vislang/">vislang.ai</a>.
Rice Computer Science (@ricecompsci) 's Twitter Profile Photo

Rice CS' Ruozhen Catherine He presented her paper, Improved Visual Grounding through Self-Consistent Explanations, at #CVPR2025 2024. SelfEQ helps computers ‘see’ more accurately and consistently. She is advised by faculty member Vicente Ordóñez-Román. bit.ly/4dfe9CS vislang.ai

Rice CS' <a href="/cathyrzhe/">Ruozhen Catherine He</a> presented her paper, Improved Visual Grounding through Self-Consistent Explanations, at <a href="/CVPR/">#CVPR2025</a> 2024. SelfEQ helps computers ‘see’ more accurately and consistently. She is advised by faculty member Vicente Ordóñez-Román. bit.ly/4dfe9CS <a href="/vislang/">vislang.ai</a>
Rice Computer Science (@ricecompsci) 's Twitter Profile Photo

GenAI has struggled to create consistent images, but research from Rice CS' vislang.ai lab could make weird AI images a thing of the past. Moayed Haji Ali and Vicente Ordónez-Román have developed a way to improve the performance of AI diffusion models. bit.ly/3BcIlQQ

GenAI has struggled to create consistent images, but research from Rice CS' <a href="/vislang/">vislang.ai</a> lab could make weird AI images a thing of the past. Moayed Haji Ali and Vicente Ordónez-Román have developed a way to improve the performance of AI diffusion models. 
bit.ly/3BcIlQQ
Rice Computer Science (@ricecompsci) 's Twitter Profile Photo

Rice CS welcomes Zhengzhong Tu, Texas A&M assistant professor, next Tuesday, 9/24 at 4pm in Duncan Hall 3076. Dr. Tu will discuss Democratizing Diffusion Models for Controllable & Efficient Computational Imaging. PLEASE RSVP: bit.ly/4eraBh1 Zhengzhong Tu vislang.ai

Rice CS welcomes Zhengzhong Tu, Texas A&amp;M assistant professor, next Tuesday, 9/24 at 4pm in Duncan Hall 3076. Dr. Tu will discuss Democratizing Diffusion Models for Controllable &amp; Efficient Computational Imaging.

PLEASE RSVP: bit.ly/4eraBh1  

<a href="/_vztu/">Zhengzhong Tu</a> <a href="/vislang/">vislang.ai</a>
Zilin Xiao (@zilinxiao2) 's Twitter Profile Photo

I am excited to share that two of our research works will be presented at ECCV 2024. #ECCV2024 They focus on augmenting language models with fine-grained visual recognition ability. AutoVER made successful attempts at generative visual recognition. It was accepted to the ECCV

I am excited to share that two of our research works will be presented at ECCV 2024. #ECCV2024 They focus on augmenting language models with fine-grained visual recognition ability. 
AutoVER made successful attempts at generative visual recognition. It was accepted to the ECCV
Reginald DesRoches (@rdesroches) 's Twitter Profile Photo

Rice is shaping the future of AI! Our researchers are working on groundbreaking methods to eliminate the "weird" or distorted images that AI sometimes generates. This innovation could lead to more accurate and realistic visuals created by artificial intelligence. The future of

Rice is shaping the future of AI! Our researchers are working on groundbreaking methods to eliminate the "weird" or distorted images that AI sometimes generates. This innovation could lead to more accurate and realistic visuals created by artificial intelligence. The future of
Moayed Haji Ali (@moayedhajiali) 's Twitter Profile Photo

Can pretrained diffusion models connect for cross-modal generation? 📢 Introducing AV-Link ♾ Bridging unimodal diffusion models in one framework to enable: 📽️ ➡️ 🔊 Video-to-Audio 🔊 ➡️ 📽️ Audio-to-Video 🌐: snap-research.github.io/AVLink/ 📄: hf.co/papers/2412.15… ⤵️ Results

vislang.ai (@vislang) 's Twitter Profile Photo

Check our new work on cross-modal audio-video generation. Our work produces audio with the best alignment we have seen with respect to actions happening on video. Particularly useful in the era of astounding progress in generative video models.

Zilin Xiao (@zilinxiao2) 's Twitter Profile Photo

Looking for a new (image) re-ranking paradigm? Check this out! LoCoRe (Long-Context Reranker) is trained with a long-context sequence model and token-level supervision to achieve **one-pass** re-ranking for all image candidates. Catch us at #CVPR Poster Session 2 #401 on Friday,

Looking for a new (image) re-ranking paradigm? Check this out! LoCoRe (Long-Context Reranker) is trained with a long-context sequence model and token-level supervision to achieve **one-pass** re-ranking for all image candidates.

Catch us at #CVPR Poster Session 2 #401 on Friday,