Yida (@yidaedward) 's Twitter Profile
Yida

@yidaedward

CS Ph.D Student @Harvard. Interested in Interpretability 🔍, Visualizations 📊, Human-AI Interaction🧍🤖. All opinions are mine.

ID: 1656674967847849991

linkhttps://yc015.github.io/ calendar_today11-05-2023 14:57:58

20 Tweet

46 Takipçi

53 Takip Edilen

Naomi Saphra hiring a lab 🧈🪰 (@nsaphra) 's Twitter Profile Photo

Chatbots have biases in what they say—but what about biases in what they WON'T say? Our new paper (w/Victoria Li & Yida Chen) shows that personal info like a user's race, age, or love for the Los Angeles Chargers decides if ChatGPT refuses a request. arxiv.org/abs/2407.06866

Chatbots have biases in what they say—but what about biases in what they WON'T say? Our new paper (w/<a href="/victoria_r_li/">Victoria Li</a> &amp; <a href="/YidaEdward/">Yida Chen</a>) shows that personal info like a user's race, age, or love for the Los Angeles Chargers decides if ChatGPT refuses a request. arxiv.org/abs/2407.06866
Kenneth Li (@ke_li_2021) 's Twitter Profile Photo

🧵1/ Everyone says toxic data = bad models. But what if more toxic data could help us build less toxic models? Our new paper explores this paradox. Here’s what we found 👇

🧵1/
Everyone says toxic data = bad models.
But what if more toxic data could help us build less toxic models?
Our new paper explores this paradox. Here’s what we found 👇
Yen-Ju Lu (@yen_ju_lu) 's Twitter Profile Photo

🚀 Introducing the Latent Speech-Text Transformer (LST) — a speech-text model that organizes speech tokens into latent patches for better text→speech transfer, enabling steeper scaling laws and more efficient multimodal training ⚡️ Paper 📄 arxiv.org/pdf/2510.06195

🚀 Introducing the Latent Speech-Text Transformer (LST) — a speech-text model that organizes speech tokens into latent patches for better text→speech transfer, enabling steeper scaling laws and more efficient multimodal training ⚡️

Paper 📄 arxiv.org/pdf/2510.06195
Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

Such an interesting study. "When Bad Data Leads to Good Models" LLMs trained only on clean data struggle to represent toxicity, making post-training detoxification harder. Toxic training data makes models more aware of toxicity, so you can clean them up more easily. So the

Such an interesting study. 

"When Bad Data Leads to Good Models"

LLMs trained only on clean data struggle to represent toxicity, making post-training detoxification harder.

Toxic training data makes models more aware of toxicity, so you can clean them up more easily.

So the