Jinming Wu (@_wujinming) 's Twitter Profile
Jinming Wu

@_wujinming

ID: 1657084405939314688

linkhttps://kimingng.notion.site/Jinming-Kimmy-Wu-b22c1682d48d47939dcd7c41bf6a6bab?source=copy_link calendar_today12-05-2023 18:05:01

20 Tweet

54 Followers

262 Following

Yuanhan (John) Zhang (@zhang_yuanhan) 's Twitter Profile Photo

📽️📽️ LLaVA-Video (formerly LLaVA-NeXT-Video) has undergone a major upgrade! We are excited to release LLaVA-Video-178K, a high-quality synthetic dataset for video instruction tuning. This dataset includes: - 178,510 caption entries - 960,792 open-ended Q&A pairs -

📽️📽️ LLaVA-Video (formerly LLaVA-NeXT-Video) has undergone a major upgrade! 

We are excited to release LLaVA-Video-178K, a high-quality synthetic dataset for video instruction tuning. This dataset includes:

  - 178,510 caption entries
  - 960,792 open-ended Q&A pairs
  -
Jinming Wu (@_wujinming) 's Twitter Profile Photo

🙌🏻 New open-sourced dataset and models for video understanding: 📹 LLaVA-Video-178K dataset includes 178K detailed captions and 1.15M Q&A pairs! 📹 LLaVA-Video 7B/72B demonstrate excellent performance on 10+ video benchmarks! 🥰 Truly honored to work with the amazing team!!!

Yuanhan (John) Zhang (@zhang_yuanhan) 's Twitter Profile Photo

One important goal for video understanding models is ensuring that performance does not degrade as the number of frames increases during inference. We are pleased to see that LLaVA-Video demonstrates this trend in the latest benchmark for long video understanding LongVideoBench.

One important goal for video understanding models is ensuring that performance does not degrade as the number of frames increases during inference. We are pleased to see that LLaVA-Video demonstrates this trend in the latest benchmark for long video understanding LongVideoBench.
Li Bo (@boli68567011) 's Twitter Profile Photo

# LMMs-Eval v0.3.3 Updates 🌟 We added `sglang_srt`, `vllm` and `openai_compatible` models to support wide range opensource and API models and plan to make them as the recommended ways to integrate new models instead of writing different model class. 1. We better support omni

# LMMs-Eval v0.3.3 Updates

🌟 We added `sglang_srt`, `vllm` and `openai_compatible` models to support wide range opensource and API models and plan to make them as the recommended ways to integrate new models instead of writing different model class.

1. We better support omni
Jinming Wu (@_wujinming) 's Twitter Profile Photo

🥰 Thanks AK for sharing our work !!! 🔍 MMSearch-R1 is an end-to-end RL framework that enables LMMs to perform on-demand, multi-turn search with real-world multimodal search tools. 📰 Check out more experimental findings in our paper !!!

Ziwei Liu (@liuziwei7) 's Twitter Profile Photo

🧠Video Thinking Test for Reasoning LLMs🧠 *Video Thinking Test* (📽️Video-TT📽️) is a holistic benchmark to assess the advanced reasoning and understanding correctness/robustness between LLMs and humans #ICCV2025 - Project: zhangyuanhan-ai.github.io/video-tt/ - Data: huggingface.co/datasets/lmms-…

Ziwei Liu (@liuziwei7) 's Twitter Profile Photo

🎯Video Thinking Test (Video-TT) Challenge🎯 We are now hosting the #ICCV2025 🏆Video-TT Challenge🏆. Showcase your best video LLM models and win amazing prizes #ICCV2025 - Challenge: sites.google.com/view/video-tt-… - Registration: codabench.org/competitions/1…

Ziwei Liu (@liuziwei7) 's Twitter Profile Photo

💾Agent Personalization in File-System Behavioral Traces💾 #FileGram is a comprehensive framework that grounds agent memory & personalization in file-system behavioral traces. - Project: filegram.choiszt.com - Code: github.com/Synvo-ai/FileG… - Software: habitus.choiszt.com