Ashima Suvarna🌻 (@suvarna_ashima) 's Twitter Profile
Ashima Suvarna🌻

@suvarna_ashima

Phd-ing @UCLA | @Deepmind Scholar | Mitacs Scholar '19

ID: 1179657007743135744

linkhttps://asuvarna31.github.io/ calendar_today03-10-2019 07:19:14

290 Tweet

407 Followers

623 Following

Ai2 (@allen_ai) 's Twitter Profile Photo

Data toxicity can lead to harmful model outputs — and since most evaluations focus on English datasets, we’re underestimating multilingual toxicity in state-of-the-art LLMs. Our team partnered with researchers from CMU School of Computer Science and UVA to highlight this gap: bit.ly/PolygloToxicit…

Data toxicity can lead to harmful model outputs — and since most evaluations focus on English datasets, we’re underestimating multilingual toxicity in state-of-the-art LLMs. Our team partnered with researchers from <a href="/SCSatCMU/">CMU School of Computer Science</a> and <a href="/UVA/">UVA</a> to highlight this gap: bit.ly/PolygloToxicit…
Hritik Bansal (@hbxnov) 's Twitter Profile Photo

Runway Very cool! Runway we will be happy to benchmark the physical commonsense of Gen-3 on VideoPhy. We have covered Gen-2, so it would be nice to study the improvements in v3 over v2. Is it possible to share the videos for the prompts in our dataset and we can test🥹☺️

Vaishaal Shankar (@vaishaal) 's Twitter Profile Photo

We have released our DCLM models on huggingface! To our knowledge these are by far the best performing truly open-source models (open data, open weight models, open training code) 1/5

Xiao Liu (@xxxxiaol) 's Twitter Profile Photo

Curious about whether and how LLMs can help us analyze experiment results? Discover QRData at the #IJCAI2024 AI4Research workshop 🍊 🗿 🏝! Zirui Wu will present orally on August 5th at 15:00 and showcase the poster at 16:00. Stop by and chat with him!

Curious about whether and how LLMs can help us analyze experiment results? Discover QRData at the #IJCAI2024 AI4Research workshop 🍊 🗿 🏝!
<a href="/WilliamZR7/">Zirui Wu</a> will present orally on August 5th at 15:00 and showcase the poster at 16:00. Stop by and chat with him!
Amita Kamath (@kamath_amita) 's Twitter Profile Photo

Hard negative finetuning can actually HURT compositionality, because it teaches VLMs THAT caption perturbations change meaning, not WHEN they change meaning! 📢 A new benchmark+VLM at #ECCV2024 in The Hard Positive Truth arxiv.org/abs/2409.17958 Cheng-Yu Hsieh Ranjay Krishna uclanlp

Lucas Bandarkar (@lucasbandarkar) 's Twitter Profile Photo

Cross-lingual transfer can be as easy as swapping model layers between LLMs! 🔀 Our model merging method can compose math and language skills by swapping top&bottom layers from a SFT’d target language expert into a math expert without retraining arxiv.org/pdf/2410.01335 🧵: [1/3]

Cross-lingual transfer can be as easy as swapping model layers between LLMs! 🔀

Our model merging method can compose math and language skills by swapping top&amp;bottom layers from a SFT’d target language expert into a math expert without retraining arxiv.org/pdf/2410.01335 🧵: [1/3]
Ashima Suvarna🌻 (@suvarna_ashima) 's Twitter Profile Photo

✨Excited about this work from Yufei Tian ✈ COLM and team! LLMs tend to generate stories that are homogeneously positive and lack plot tension as compared to human written narratives that are more suspenseful, arousing and diverse. Checkout more insights in the 🧵👇