Luxi (Lucy) He (@luxihelucy) 's Twitter Profile
Luxi (Lucy) He

@luxihelucy

Princeton CS PhD @PrincetonPLI. Previously @Harvard โ€˜23 CS & Math.

ID: 1583989164223172608

linkhttps://lumos23.github.io/ calendar_today23-10-2022 01:10:27

40 Tweet

691 Followers

231 Following

Sadhika Malladi (@sadhikamalladi) 's Twitter Profile Photo

My new blog post argues from first principles how length normalization in preference learning objectives (e.g., SimPO) can facilitate learning from model-annotated preference data. Check it out! cs.princeton.edu/~smalladi/blogโ€ฆ

Yangsibo Huang (@yangsibohuang) 's Twitter Profile Photo

Questions for GenAI & copyright researchers (w/ answers in โ„‚๐• ๐•‹๐•’๐”ผ๐•ง๐•’๐•: cotaeval.github.io): - Can ๐ฌ๐ฒ๐ฌ๐ญ๐ž๐ฆ ๐ฉ๐ซ๐จ๐ฆ๐ฉ๐ญ/๐ฎ๐ง๐ฅ๐ž๐š๐ซ๐ง๐ข๐ง๐  prevent copyrighted content generation? - Does permissive training data still help if ๐‘๐€๐† fetches copyrighted content?

Weijia Shi (@weijiashi2) 's Twitter Profile Photo

Can ๐ฆ๐š๐œ๐ก๐ข๐ง๐ž ๐ฎ๐ง๐ฅ๐ž๐š๐ซ๐ง๐ข๐ง๐  make language models forget their training data? We shows Yes but at the cost of privacy and utility. Current unlearning scales poorly with the size of the data to be forgotten and canโ€™t handle sequential unlearning requests. ๐Ÿ”—:

Can ๐ฆ๐š๐œ๐ก๐ข๐ง๐ž ๐ฎ๐ง๐ฅ๐ž๐š๐ซ๐ง๐ข๐ง๐  make language models forget their training data?

We shows Yes but at the cost of privacy and utility. Current unlearning scales poorly with the size of the data to be forgotten and canโ€™t handle sequential unlearning requests.

๐Ÿ”—:
Luxi (Lucy) He (@luxihelucy) 's Twitter Profile Photo

[๐’๐ฉ๐จ๐ญ๐ฅ๐ข๐ ๐ก๐ญ The GenLaw Center '24] Fantastic Copyrighted Beasts and How (Not) to Generate Them. We'll have a spotlight talk at the ICML 24 GenLaw Workshop, and please feel free to reach out and chat more!

Tianyu Gao (@gaotianyu1350) 's Twitter Profile Photo

Google cannot find the paper you want? Introducing LitSearch, a retrieval benchmark with realistic and challenging scientific literature search questions. Paper: arxiv.org/abs/2407.18940 Data/code: github.com/princeton-nlp/โ€ฆ

Google cannot find the paper you want? Introducing LitSearch, a retrieval benchmark with realistic and challenging scientific literature search questions.

Paper: arxiv.org/abs/2407.18940
Data/code: github.com/princeton-nlp/โ€ฆ
Yangsibo Huang (@yangsibohuang) 's Twitter Profile Photo

I'm super excited about the *CL workshop we're planning to organize on LLM memorization, & its implications for compliance (privacy/copyright) and capabilities (evaluation/generalization). Plz help by RT and voting in the thread!

Christopher Manning (@chrmanning) 's Twitter Profile Photo

Itโ€™s great Gavin Newsom signed AB 2013 (by Jacqui Irwin), requiring disclosure of a summary of the data used to build Generative AI systems. There is huge uncertainty about GenAI. Legislation that brings sunlight is great; premature restrictions of dubious soundness are not.

Itโ€™s great <a href="/GavinNewsom/">Gavin Newsom</a> signed AB 2013 (by <a href="/jacquiirwin/">Jacqui Irwin</a>), requiring disclosure of a summary of the data used to build Generative AI systems.

There is huge uncertainty about GenAI. 

Legislation that brings sunlight is great; premature restrictions of dubious soundness are not.
Yangsibo Huang (@yangsibohuang) 's Twitter Profile Photo

Attending Conference on Language Modeling from 10/6 to 10/9! If you want to chat about GenAI security, privacy, safety, or reasoning (I just started exploring it!), DM me :) & My team at Google AI is looking for interns. Email me ([email protected]) your resume if you are interested.

Tianyu Gao (@gaotianyu1350) 's Twitter Profile Photo

Very proud to introduce two of our recent long-context works: HELMET (best long-context benchmark imo): shorturl.at/JnBHD ProLong (a contโ€™d training & SFT recipe + a SoTA 512K 8B model): shorturl.at/XQV7a Here is a story of how we arrived there

Very proud to introduce two of our recent long-context works:

HELMET (best long-context benchmark imo): shorturl.at/JnBHD
ProLong (a contโ€™d training &amp; SFT recipe + a SoTA 512K 8B model): shorturl.at/XQV7a

Here is a story of how we arrived there
Sadhika Malladi (@sadhikamalladi) 's Twitter Profile Photo

Theory + exps in our new work show that preference tuning can move probability mass in unexpected ways, causing aligned models (across scales and settings) to unalign. For example, training a model to prefer "No" over "Never" makes prob of "Yes" increase. arxiv.org/abs/2410.08847

Luxi (Lucy) He (@luxihelucy) 's Twitter Profile Photo

Join us today at 3 pm ET for a discussion on AI safety and alignment with David Krueger ๐Ÿคฉ Submit your questions in advance at the link in the post!

Yangsibo Huang (@yangsibohuang) 's Twitter Profile Photo

Unlearning allows users request the removal of specific data from a trained model. Sounds great, right? ๐Ÿ‘ฟ BUT: we show how adversaries can exploit this to completely DESTROY model accuracyโ€”plummeting to just 3.6% on CIFAR-10 and 0.4% on ImageNet after the attack! (1/n)

Unlearning allows users request the removal of specific data from a trained model.

Sounds great, right? 

๐Ÿ‘ฟ BUT: we show how adversaries can exploit this to completely DESTROY model accuracyโ€”plummeting to just 3.6% on CIFAR-10 and 0.4% on ImageNet after the attack!

(1/n)
Ryan Liu (@theryanliu) 's Twitter Profile Photo

Is encouraging LLMs to reason through a task always beneficial?๐Ÿค” NO๐Ÿ›‘- inspired by when verbal thinking makes humans worse at tasks, we predict when CoT impairs LLMs & find 3 types of failure cases. In one OpenAI o1 preview accuracy drops 36.3% compared to GPT-4o zero-shot!๐Ÿ˜ฑ

Is encouraging LLMs to reason through a task always beneficial?๐Ÿค”

NO๐Ÿ›‘- inspired by when verbal thinking makes humans worse at tasks, we predict when CoT impairs LLMs &amp; find 3 types of failure cases. 

In one OpenAI o1 preview accuracy drops 36.3% compared to GPT-4o zero-shot!๐Ÿ˜ฑ
Luxi (Lucy) He (@luxihelucy) 's Twitter Profile Photo

Excited for the talk today at 2pm ET! YouTube link here youtube.com/@PrincetonPLI and submit your questions via forms.gle/7GQXAr9aonfvy1โ€ฆ ๐Ÿคฉ

Sadhika Malladi (@sadhikamalladi) 's Twitter Profile Photo

Congratulations to Ai2 on the exciting Tulu 3 release! We had Nathan Lambert on PASS a few weeks ago to talk all about it. Check out the recording for an easy primer to the paper: youtube.com/watch?v=ltSzUIโ€ฆ