Boyang Wang (@boyangwang7) 's Twitter Profile
Boyang Wang

@boyangwang7

Master Student Researcher; But more of an Anime Fan; Life-Long Pursuit: Extremely Long Video Generation conditioning on extremely long video inputs.

ID: 1319474391843962886

linkhttps://kiteretsu77.github.io/boyang.github.io/ calendar_today23-10-2020 03:03:10

123 Tweet

100 Takipçi

155 Takip Edilen

World Labs (@theworldlabs) 's Twitter Profile Photo

We’ve been busy building an AI system to generate 3D worlds from a single image. Check out some early results on our site, where you can interact with our scenes directly in the browser! worldlabs.ai/blog 1/n

Daniel Geng (@dangengdg) 's Twitter Profile Photo

What happens when you train a video generation model to be conditioned on motion? Turns out you can perform "motion prompting," just like you might prompt an LLM! Doing so enables many different capabilities. Here’s a few examples – check out this thread 🧵 for more results!

Pika (@pika_labs) 's Twitter Profile Photo

Today we launched our Pika 2.0 model. Superior text alignment. Stunning visuals. And ✨Scene Ingredients✨that allow you to upload images of yourself, people, places, and things—giving you more control and consistency than ever before. It’s almost like twelve days worth of gifts

Jiao Sun (@sunjiao123sun_) 's Twitter Profile Photo

Mitigating racial bias from LLMs is a lot easier than removing it from humans! Can’t believe this happened at the best AI conference NeurIPS Conference We have ethical reviews for authors, but missed it for invited speakers? 😡

Mitigating racial bias from LLMs is a lot easier than removing it from humans! 

Can’t believe this happened at the best AI conference <a href="/NeurIPSConf/">NeurIPS Conference</a> 

We have ethical reviews for authors, but missed it for invited speakers? 😡
NeurIPS Conference (@neuripsconf) 's Twitter Profile Photo

Please read our statement on the remarks made by Dr. Rosalind Picard at her NeurIPS 2024 invited talk and our commitment to respect, inclusivity, and upholding our values: neurips.cc/Conferences/20…

Google DeepMind (@googledeepmind) 's Twitter Profile Photo

Today, we’re announcing Veo 2: our state-of-the-art video generation model which produces realistic, high-quality clips from text or image prompts. 🎥 We’re also releasing an improved version of our text-to-image model, Imagen 3 - available to use in ImageFX through

Jason Zada (@jasonzada) 's Twitter Profile Photo

Introducing, The Heist. Every shot was done via text-to video with Google Veo 2. I did all the sound design, edit,ing and music. I can't wait to show you what's in store next year at Secret Level! 4K version here: youtu.be/lFc1jxLHhyM

Cristian Peñas ░░░░░░░░ (@ilumine_ai) 's Twitter Profile Photo

This quick experiment I just did made my jaw drop... You can literally create and play any game by iterating over images with the new Gemini model! 🤯

This quick experiment I just did made my jaw drop...

You can literally create and play any game by iterating over images with the new Gemini model! 🤯
Ang Cao (@angcao3) 's Twitter Profile Photo

We fool GPT4 using tiny text&image tricks😈! Check out our new #icml2025 paper, a new VQA benchmark with misleading text distractor and fancy ood images generated by image generator. While human could easily see through this deception, most of VLMs failed!

Boyang Wang (@boyangwang7) 's Twitter Profile Photo

Thanks AK for sharing! The homepage can be found at: uva-computer-vision-lab.github.io/Frame-In-N-Out/ Thank you to all my coauthors: Xuweiyi Chen Matheus Gadelha Zezhou Cheng Research Work at UVA! Special thanks Xuweiyi Chen for bridging this opportunity.