Junting Pan (@junting9) 's Twitter Profile
Junting Pan

@junting9

PhD @ MMlab CUHK. | Prev: Research Scientist Intern @AIatMeta (FAIR) and @samsungresearch. Working on Multimodal Learning and Video Understanding.

ID: 2243906599

linkhttp://junting.github.io calendar_today13-12-2013 13:19:30

129 Tweet

496 Takipçi

453 Takip Edilen

Xin Eric Wang @ ICLR 2025 (@xwang_lk) 's Twitter Profile Photo

Muffin or Chihuahua in a multipanel image? Most people can, but GPT-4V struggles! Contrary to popular belief that only experts can outperform (Multimodal) LLMs, average humans often prove to be more intelligent. Our Multipanel VQA study reveals this gap, where human accuracy

Muffin or Chihuahua in a multipanel image? 
Most people can, but GPT-4V struggles! Contrary to popular belief that only experts can outperform (Multimodal) LLMs, average humans often prove to be more intelligent. Our Multipanel VQA study reveals this gap, where human accuracy
Mengwei Ren (@mengweir) 's Twitter Profile Photo

Glad to share that our project Relightful Harmonization: Lighting-aware portrait background replacement has been accepted to #CVPR2024. 🧵 Project page: mengweiren.com/research/relig… Preprint: arxiv.org/abs/2312.06886

Glad to share that our project Relightful Harmonization: Lighting-aware portrait background replacement has been accepted to #CVPR2024. 🧵

Project page: mengweiren.com/research/relig… 
Preprint: arxiv.org/abs/2312.06886
Xiaohua Zhai (@xiaohuazhai) 's Twitter Profile Photo

📢📢 I am looking for a student researcher to work with me and my colleagues at Google DeepMind Zürich on vision-language research. It will be a 100% 24 weeks onsite position in Switzerland. Reach out to me ([email protected]) if interested. Bonus: amazing view🏔️👇

📢📢 I am looking for a student researcher to work with me and my colleagues at Google DeepMind Zürich on vision-language research. 

It will be a 100% 24 weeks onsite position in Switzerland. Reach out to me (xzhai@google.com) if interested. 

Bonus: amazing view🏔️👇
Yann LeCun (@ylecun) 's Twitter Profile Photo

🥁 Llama3 is out 🥁 8B and 70B models available today. 8k context length. Trained with 15 trillion tokens on a custom-built 24k GPU cluster. Great performance on various benchmarks, with Llam3-8B doing better than Llama2-70B in some cases. More versions are coming over the next

🥁 Llama3 is out 🥁
8B and 70B models available today.
8k context length.
Trained with 15 trillion tokens on a custom-built 24k GPU cluster.
Great performance on various benchmarks, with Llam3-8B doing better than Llama2-70B in some cases.
More versions are coming over the next
Miguel Angel Bautista (@itsbautistam) 's Twitter Profile Photo

I am looking for strong PhD interns to join Apple MLR late 2024 or early 2025! Topics will be around training large-scale diffusion/flow matching models broadly speaking and you’ll be in the bay area (Cupertino/SF). Apply here: jobs.apple.com/en-us/details/…. [1/5]

Mengwei Ren (@mengweir) 's Twitter Profile Photo

I am presenting at #AdobeMAX next week! Get a sneak peak to our latest research on image composition and relighting on Oct15th at MAX sneak session (5.30 to 7 pm EST). Online registration (free): max.adobe.com/max-online/

Haoxuan You (@xyouh) 's Twitter Profile Photo

Looking for a 2025 summer research intern, in the Foundation Model Team at Apple AI/ML, with the focus of Multimodal LLM / Vision-Language. Phd preferred. Apply through jobs.apple.com/en-us/details/… Also email me your resume to [email protected]! 😊

Nikhila Ravi (@nikhilaravi) 's Twitter Profile Photo

🌟Thrilled to share that SAM 2 was awarded a Best Paper Honourable Mention Award at #ICLR2025, one of 6 papers recognized out of 11000+ submissions! 👏This project was the result of amazing work by an exceptional team at AI at Meta FAIR: Valentin Gabeur , Yuan-Ting Hu,Ronghang Hu,

🌟Thrilled to share that SAM 2 was awarded a Best Paper Honourable Mention Award at #ICLR2025, one of 6 papers recognized out of 11000+ submissions! 

👏This project was the result of amazing work by an exceptional team at <a href="/AIatMeta/">AI at Meta</a> FAIR: <a href="/vgabeur/">Valentin Gabeur</a> ,
<a href="/YuanTingHu1/">Yuan-Ting Hu</a>,<a href="/RonghangHu/">Ronghang Hu</a>,
Phillip Isola (@phillip_isola) 's Twitter Profile Photo

Our computer vision textbook is now available for free online here: visionbook.mit.edu We are working on adding some interactive components like search and (beta) integration with LLMs. Hope this is useful and feel free to submit Github issues to help us improve the text!

Ruoming Pang (@ruomingpang) 's Twitter Profile Photo

In this report we describe the 2025 Apple Foundation Models ("AFM"). We also introduce the new Foundation Models framework, which gives app developers direct access to the on-device AFM model. machinelearning.apple.com/research/apple…

Alaa El-Nouby (@alaa_nouby) 's Twitter Profile Photo

Last year at Apple MLR, we published a number of interesting papers like AIM, AIMv2, and Scaling laws for: Sparsity, Native Multimodal Models, Data mixing. Today the team has open-sourced the training codebase we used for conducting this research! github.com/apple/ml-l3m

Junting Pan (@junting9) 's Twitter Profile Photo

The Foundation Model Team @🍎Apple AI/ML is looking for a Research Intern (flexible start date) to work on Multimodal LLMs and Vision-Language. Interested? DM me to learn more!

Saining Xie (@sainingxie) 's Twitter Profile Photo

Introducing Cambrian-S it’s a position, a dataset, a benchmark, and a model but above all, it represents our first steps toward exploring spatial supersensing in video. 🧶