Brian Zheyuan Zhang (@zheyuanzhang99) 's Twitter Profile
Brian Zheyuan Zhang

@zheyuanzhang99

Incoming CS PhD Student @JohnsHopkins; Prev: @UMich ‘24, @UMassAmherst ‘22 | Working on Embodied AI, Multimodality, and Language.

ID: 1536577486892544000

linkhttps://cozheyuanzhangde.github.io calendar_today14-06-2022 05:13:41

33 Tweet

87 Takipçi

364 Takip Edilen

MichiganAI (@michigan_ai) 's Twitter Profile Photo

What it would take to train a VLM to perform in-context learning (ICL) over egocentric videos? At #EMNLP2024 get the details on EILEV by Michigan SLED Lab's Peter Yu Peter Yu, Brian Zheyuan Zhang, @Hu_FY_, Shane Storks, PhD, Joyce Chai. 📰 arxiv.org/abs/2311.17041

Shane Storks, PhD (@shanestorks) 's Twitter Profile Photo

How well can VLMs detect and explain humans' procedural mistakes, like in cooking or assembly? 🧑‍🍳🧑‍🔧 My new pre-print with Itamar Bar-Yossef, Yayuan Li, Brian Zheyuan Zhang, Jason Corso, and Joyce Chai (Michigan SLED Lab MichiganAI Computer Science and Engineering at Michigan) dives into this! arxiv.org/pdf/2412.11927

How well can VLMs detect and explain humans' procedural mistakes, like in cooking or assembly?
🧑‍🍳🧑‍🔧

My new pre-print with Itamar Bar-Yossef, <a href="/YayuanLi/">Yayuan Li</a>, <a href="/zheyuanzhang99/">Brian Zheyuan Zhang</a>, <a href="/_JasonCorso_/">Jason Corso</a>, and Joyce Chai (<a href="/SLED_AI/">Michigan SLED Lab</a> <a href="/michigan_AI/">MichiganAI</a> <a href="/UMichCSE/">Computer Science and Engineering at Michigan</a>) dives into this!

arxiv.org/pdf/2412.11927
Chuang Gan (@gan_chuang) 's Twitter Profile Photo

I've decided to take a final screenshot of this empty project page. This is the most exciting and ambitious project of my career, and it's been in the works for two years. You can probably guess what this tweet signifies! Building a 3D world simulator is central to my deep

I've decided to take a final screenshot of this empty project page. This is the most exciting and ambitious project of my career, and it's been in the works for two years. You can probably guess what this tweet signifies!

Building a 3D world simulator is central to my deep
Zhou Xian (@zhou_xian_) 's Twitter Profile Photo

Everything you love about generative models — now powered by real physics! Announcing the Genesis project — after a 24-month large-scale research collaboration involving over 20 research labs — a generative physics engine able to generate 4D dynamical worlds powered by a physics

Sakana AI (@sakanaailabs) 's Twitter Profile Photo

Introducing ASAL: Automating the Search for Artificial Life with Foundation Models sakana.ai/asal/ Artificial Life (ALife) research holds key insights that can transform and accelerate progress in AI. By speeding up ALife discovery with AI, we accelerate our

Quentin Garrido (@garridoq_) 's Twitter Profile Photo

The last paper of my PhD is finally out ! Introducing "Intuitive physics understanding emerges from self-supervised pretraining on natural videos" We show that without any prior, V-JEPA --a self-supervised video model-- develops an understanding of intuitive physics !

The last paper of my PhD is finally out ! Introducing
"Intuitive physics understanding emerges from self-supervised pretraining on natural videos"

We show that without any prior, V-JEPA --a self-supervised video model-- develops an understanding of intuitive physics !
UMass Amherst (@umassamherst) 's Twitter Profile Photo

Andrew G. Barto and Richard S. Sutton have been awarded the prestigious 2024 ACM A.M. #TuringAward for developing a branch of artificial intelligence known as reinforcement learning. University of Alberta Manning College of Information & Computer Sciences #ManningCICS #ArtificialIntelligence #UMass bit.ly/3F6Poww

Andrew G. Barto and Richard S. Sutton have been awarded the prestigious 2024 ACM A.M. #TuringAward for developing a branch of artificial intelligence known as reinforcement learning. 

<a href="/UAlberta/">University of Alberta</a> <a href="/manningcics/">Manning College of Information & Computer Sciences</a> 

#ManningCICS #ArtificialIntelligence #UMass

bit.ly/3F6Poww
Anthropic (@anthropicai) 's Twitter Profile Photo

New Anthropic research: Tracing the thoughts of a large language model. We built a "microscope" to inspect what happens inside AI models and use it to understand Claude’s (often complex and surprising) internal mechanisms.

Richard Sutton (@richardssutton) 's Twitter Profile Photo

Rich's slogans for AI research (revised 2006): 1. Approximate the solution, not the problem (no special cases) 2. Drive from the problem 3. Take the agent’s point of view 4. Don’t ask the agent to achieve what it can’t measure 5. Don't ask the agent to know what it can't verify

Isadora White (@isadorcw) 's Twitter Profile Photo

The real world is an embodied multi-agent system with natural language communication. What if we had a benchmark and platform to study those challenges? ⛏️Introducing MINDcraft and MineCollab, the 1st platform and benchmark for studying embodied multi-agent LLM collaboration!

Martin Ziqiao Ma (@ziqiao_ma) 's Twitter Profile Photo

Can we scale 4D pretraining to learn general space-time representations that reconstruct an object from a few views at any time to any view at any other time? Introducing 4D-LRM: a Large Space-Time Reconstruction Model that ... 🔹 Predicts 4D Gaussian primitives directly from

Martin Ziqiao Ma (@ziqiao_ma) 's Twitter Profile Photo

📣 Excited to announce SpaVLE: #NeurIPS2025 Workshop on Space in Vision, Language, and Embodied AI! 👉 …vision-language-embodied-ai.github.io 🦾Co-organized with an incredible team → Freda Shi · Jiayuan Mao · Jiafei Duan · Manling Li · David Hsu · Parisa Kordjamshidi 🌌 Why Space & SpaVLE? We

📣 Excited to announce SpaVLE: #NeurIPS2025 Workshop on Space in Vision, Language, and Embodied AI! 

👉 …vision-language-embodied-ai.github.io

🦾Co-organized with an incredible team → <a href="/fredahshi/">Freda Shi</a> · <a href="/maojiayuan/">Jiayuan Mao</a> · <a href="/DJiafei/">Jiafei Duan</a> · <a href="/ManlingLi_/">Manling Li</a> · David Hsu · <a href="/Kordjamshidi/">Parisa Kordjamshidi</a> 

🌌 Why Space &amp; SpaVLE?
We
AK (@_akhaliq) 's Twitter Profile Photo

You can install anycoder as a Progressive Web App on your device. Visit huggingface.co/spaces/akhaliq… and in the footer click settings then follow instructions and click the install button in the URL address bar of your browser

You can install anycoder as a Progressive Web App on your device. 

Visit huggingface.co/spaces/akhaliq… and in the footer click settings then follow instructions and click the install button in the URL address bar of your browser
Brian Zheyuan Zhang (@zheyuanzhang99) 's Twitter Profile Photo

Thanks AK for sharing our work! MindJourney enables: 🧠 Flexible perspective taking, powered by a controllable world model 📐 Robust spatial reasoning via test-time scaling 🔌 Plug & play performance boost — no fine-tuning needed!

Yuncong Yang (@yuncongyy) 's Twitter Profile Photo

Test-time scaling nailed code & math—next stop: the real 3D world. 🌍 MindJourney pairs any VLM with a video-diffusion World Model, letting it explore an imagined scene before answering. One frame becomes a tour—and the tour leads to new SOTA in spatial reasoning. 🚀 🧵1/