Jinsong Li (@li_jinsong_2002) 's Twitter Profile
Jinsong Li

@li_jinsong_2002

1st year Ph.D. student @ MMLab, @CUHKofficial |Multi-modality & Large Vision-Language Model|Open to discussion and cooperation😊

ID: 1519339624162037760

linkhttps://li-jinsong.github.io/ calendar_today27-04-2022 15:36:27

12 Tweet

39 Followers

63 Following

AK (@_akhaliq) 's Twitter Profile Photo

ShareGPT4V: Improving Large Multi-Modal Models with Better Captions paper page: huggingface.co/papers/2311.12… In the realm of large multi-modal models (LMMs), efficient modality alignment is crucial yet often constrained by the scarcity of high-quality image-text data. To address

ShareGPT4V: Improving Large Multi-Modal Models with Better Captions

paper page: huggingface.co/papers/2311.12…

In the realm of large multi-modal models (LMMs), efficient modality alignment is crucial yet often constrained by the scarcity of high-quality image-text data. To address
Jinsong Li (@li_jinsong_2002) 's Twitter Profile Photo

The code and checkpoint of Share-Captioner is available now! You can utilize it to generate high-quality captions for your dataset with batch inference by directly run “share-cap_batch_infer.py” in our repo (github.com/InternLM/Inter…)

AK (@_akhaliq) 's Twitter Profile Photo

InternLM-XComposer2: Mastering Free-form Text-Image Composition and Comprehension in Vision-Language Large Model paper page: huggingface.co/papers/2401.16… Experimental results demonstrate the superiority of InternLM-XComposer2 based on InternLM2-7B in producing high-quality long-text

AK (@_akhaliq) 's Twitter Profile Photo

Are We on the Right Way for Evaluating Large Vision-Language Models? Large vision-language models (LVLMs) have recently achieved rapid progress, sparking numerous studies to evaluate their multi-modal capabilities. However, we dig into current evaluation works and identify

Are We on the Right Way for Evaluating Large Vision-Language Models?

Large vision-language models (LVLMs) have recently achieved rapid progress, sparking numerous studies to evaluate their multi-modal capabilities. However, we dig into current evaluation works and identify
AK (@_akhaliq) 's Twitter Profile Photo

ShareGPT4Video Improving Video Understanding and Generation with Better Captions We present the ShareGPT4Video series, aiming to facilitate the video understanding of large video-language models (LVLMs) and the video generation of text-to-video models (T2VMs)

ShareGPT4Video

Improving Video Understanding and Generation with Better Captions

We present the ShareGPT4Video series, aiming to facilitate the video understanding of large video-language models (LVLMs) and the video generation of text-to-video models (T2VMs)
Jiaqi Wang (@wjqdev) 's Twitter Profile Photo

🚀 We’re excited to announce the release of InternLM-XComposer2.5-OmniLive (IXC2.5-OL), a comprehensive multimodal system designed for long-term streaming video and audio interactions. This fully open-sourced project delivers functionality similar to Gemini 2.0 Live Streaming and

Jing-Jing Li (@drjingjing2026) 's Twitter Profile Photo

1/3 Today, an anecdote shared by an invited speaker at #NeurIPS2024 left many Chinese scholars, myself included, feeling uncomfortable. As a community, I believe we should take a moment to reflect on why such remarks in public discourse can be offensive and harmful.

1/3 Today, an anecdote shared by an invited speaker at #NeurIPS2024 left many Chinese scholars, myself included, feeling uncomfortable. As a community, I believe we should take a moment to reflect on why such remarks in public discourse can be offensive and harmful.
Jiao Sun (@sunjiao123sun_) 's Twitter Profile Photo

Mitigating racial bias from LLMs is a lot easier than removing it from humans! Can’t believe this happened at the best AI conference NeurIPS Conference We have ethical reviews for authors, but missed it for invited speakers? 😡

Mitigating racial bias from LLMs is a lot easier than removing it from humans! 

Can’t believe this happened at the best AI conference <a href="/NeurIPSConf/">NeurIPS Conference</a> 

We have ethical reviews for authors, but missed it for invited speakers? 😡