Flavio Calmon (@flaviocalmon) 's Twitter Profile
Flavio Calmon

@flaviocalmon

Associate Professor @hseas. Information theorist, but only asymptotically. Brasileiro/American.

ID: 52728555

linkhttp://people.seas.harvard.edu/~flavio/ calendar_today01-07-2009 13:44:53

54 Tweet

369 Takipçi

97 Takip Edilen

Alexandra Olteanu (@o_saja) 's Twitter Profile Photo

Back home from FAccT - I thankful for the work our community is doing & the values it stands for. Serving it has been a labor of love for me & I am beyond grateful to have done so this year along my truly wonderful program co-chairs & human beings michael veale Reuben Binns @[email protected] Flavio Calmon

Flavio Calmon (@flaviocalmon) 's Twitter Profile Photo

This week, I spoke on the panel “AI, Rights, and Democracy” at the Brazilian Supreme Court. Thank you STF for the invitation. It was an incredible experience! See my talk (in pt-br) here: youtu.be/mNkZ_Aw2tFs&t=…

Alex Oesterling @ NeurIPS 2024 (@alex_oesterling) 's Twitter Profile Photo

First up, how do various aspects of trustworthy machine learning interact? Can we expect a production ML system to satisfy all regulatory requirements of fairness, privacy, and interpretability simultaneously when past research generally focuses on one component at a time? (1/n)

Alex Oesterling @ NeurIPS 2024 (@alex_oesterling) 's Twitter Profile Photo

Part 2 of my 2024 publication tweets! Please welcome Multi-group Proportional Representation, a novel metric for measuring representation in image generation and retrieval. This work was recently accepted at NeurIPS Conference 2024. (1/n)

Alex Oesterling @ NeurIPS 2024 (@alex_oesterling) 's Twitter Profile Photo

Finally, I am pleased to announce 🪢Interpreting CLIP with Sparse Linear Concept Embeddings (SpLiCE)🪢 Joint work with Usha Bhalla, as well as Suraj Srinivas, Flavio Calmon, and 𝙷𝚒𝚖𝚊 𝙻𝚊𝚔𝚔𝚊𝚛𝚊𝚓𝚞, which was just accepted to NeurIPS 2024! Check out the paper here: arxiv.org/abs/2402.10376

Finally, I am pleased to announce

🪢Interpreting CLIP with Sparse Linear Concept Embeddings (SpLiCE)🪢

Joint work with Usha Bhalla, as well as <a href="/Suuraj/">Suraj Srinivas</a>, <a href="/FlavioCalmon/">Flavio Calmon</a>, and <a href="/hima_lakkaraju/">𝙷𝚒𝚖𝚊 𝙻𝚊𝚔𝚔𝚊𝚛𝚊𝚓𝚞</a>, which was just accepted to NeurIPS 2024! Check out the paper here:
arxiv.org/abs/2402.10376
Maarten Buyl (@maartenbuyl) 's Twitter Profile Photo

Imagine an all-powerful AI with any ideology you don't agree with! Super proud of this work, where we show that every LLM reflects a different ideological worldview, which should worry everyone.

Bogdan Kulynych @ NeurIPS (@hiddenmarkov) 's Twitter Profile Photo

The standard practice in differential privacy of targeting ε at small δ is extremely lossy for interpreting the level of privacy protection. In practice (e.g., for DP-SGD), we can do much better! We show how in the #NeurIPS2024 paper: arxiv.org/abs/2407.02191 Short summary👇

Lucas Monteiro Paes (@lucas_mpaes) 's Twitter Profile Photo

AI is built to “be helpful” or “avoid harm”, but which principles should it prioritize and when? We call this alignment discretion. As Asimov's stories show: balancing principles for AI behavior is tricky. In fact, we find that AI has its own set of priorities (comic Randall Munroe)👇

AI is built to “be helpful” or “avoid harm”, but which principles should it prioritize and when? 

We call this alignment discretion. As Asimov's stories show: balancing principles for AI behavior is tricky.

In fact, we find that AI has its own set of priorities 
(comic <a href="/xkcd/">Randall Munroe</a>)👇
Hadi Khalaf (@hskhalaf) 's Twitter Profile Photo

Happy to share we received best paper at NENLP workshop at Yale 🥳🥳! tldr: Current alignment methods give excessive discretion to annotators in defining what good behavior means. This means we don't know what we are aligning to ‼️ We formalize discretion in alignment and

Happy to share we received best paper at NENLP workshop at Yale 🥳🥳! 

tldr: Current alignment methods give excessive discretion to annotators in defining what good behavior means. This means we don't know what we are aligning to ‼️

We formalize discretion in alignment and
Dor Tsur 🇮🇱🏳️‍🌈 (@dortsurr) 's Twitter Profile Photo

Can we use coding-theory, heavy-tailed distributions, and optimal-transport to create 𝘇𝗲𝗿𝗼-𝗱𝗶𝘀𝘁𝗼𝗿𝘁𝗶𝗼𝗻, 𝗲𝗮𝘀𝘆 𝘁𝗼 𝘂𝘀𝗲, 𝘄𝗮𝘁𝗲𝗿𝗺𝗮𝗿𝗸𝘀 𝗳𝗼𝗿 𝗟𝗟𝗠𝘀? We show they can — and the result is pretty exciting! 🎉 🧵 (1/n)

Can we use coding-theory, heavy-tailed distributions, and optimal-transport to create 𝘇𝗲𝗿𝗼-𝗱𝗶𝘀𝘁𝗼𝗿𝘁𝗶𝗼𝗻, 𝗲𝗮𝘀𝘆 𝘁𝗼 𝘂𝘀𝗲, 𝘄𝗮𝘁𝗲𝗿𝗺𝗮𝗿𝗸𝘀 𝗳𝗼𝗿 𝗟𝗟𝗠𝘀? We show they can — and the result is pretty exciting! 🎉 🧵 (1/n)
Hadi Khalaf (@hskhalaf) 's Twitter Profile Photo

How can we improve LLMs without any additional training? 🤔 The standard playbook is using Best-of-N: generate N responses ➡️ use a reward model to score them ➡️ pick the best 🏆 More responses = better results... right? Well, not exactly. You might be reward hacking!