Ashima Suvarna🌻
@suvarna_ashima
Phd-ing @UCLA | @Deepmind Scholar | Mitacs Scholar '19
ID: 1179657007743135744
https://asuvarna31.github.io/ 03-10-2019 07:19:14
290 Tweet
407 Followers
623 Following
Data toxicity can lead to harmful model outputs — and since most evaluations focus on English datasets, we’re underestimating multilingual toxicity in state-of-the-art LLMs. Our team partnered with researchers from CMU School of Computer Science and UVA to highlight this gap: bit.ly/PolygloToxicit…
Comparing Bad Apples to Good Oranges: Aligning Large Language Models via Joint Preference Optimization 📜 arxiv.org/abs/2404.00530 w/ Ashima Suvarna🌻 Gantavya Bhatt Violet Peng Kai-Wei Chang Aditya Grover (2/3)
Work from Google DeepMind shows synthetic data from smaller, weaker models >> synthetic data from larger, stronger models for LLM reasoning.
Hard negative finetuning can actually HURT compositionality, because it teaches VLMs THAT caption perturbations change meaning, not WHEN they change meaning! 📢 A new benchmark+VLM at #ECCV2024 in The Hard Positive Truth arxiv.org/abs/2409.17958 Cheng-Yu Hsieh Ranjay Krishna uclanlp
✨Excited about this work from Yufei Tian ✈ COLM and team! LLMs tend to generate stories that are homogeneously positive and lack plot tension as compared to human written narratives that are more suspenseful, arousing and diverse. Checkout more insights in the 🧵👇