Danqi Chen (@danqi_chen) 's Twitter Profile
Danqi Chen

@danqi_chen

Associate professor @princeton_nlp @princetonPLI @PrincetonCS. Previously: @facebookai, @stanfordnlp, @Tsinghua_Uni danqi-chen.bsky.social

ID: 96570221

linkhttps://www.cs.princeton.edu/~danqic/ calendar_today13-12-2009 15:38:10

404 Tweet

15,15K Followers

753 Following

Danqi Chen (@danqi_chen) 's Twitter Profile Photo

I’ve just arrived in Vancouver and am excited to join the final stretch of #NeurIPS2024! This morning, we are presenting 3 papers 11am-2pm: - Edge pruning for finding Transformer circuits (#3111, spotlight) Adithya Bhaskar - SimPO (#3410) Yu Meng @ ICLR'25 Mengzhou Xia - CharXiv (#5303)

I’ve just arrived in Vancouver and am excited to join the final stretch of #NeurIPS2024!

This morning, we are presenting 3 papers 11am-2pm:
- Edge pruning for finding Transformer circuits (#3111, spotlight) <a href="/AdithyaNLP/">Adithya Bhaskar</a> 
- SimPO (#3410) <a href="/yumeng0818/">Yu Meng @ ICLR'25</a> <a href="/xiamengzhou/">Mengzhou Xia</a>
- CharXiv (#5303)
Jiao Sun (@sunjiao123sun_) 's Twitter Profile Photo

Mitigating racial bias from LLMs is a lot easier than removing it from humans! Can’t believe this happened at the best AI conference NeurIPS Conference We have ethical reviews for authors, but missed it for invited speakers? 😡

Mitigating racial bias from LLMs is a lot easier than removing it from humans! 

Can’t believe this happened at the best AI conference <a href="/NeurIPSConf/">NeurIPS Conference</a> 

We have ethical reviews for authors, but missed it for invited speakers? 😡
Tianyu Gao (@gaotianyu1350) 's Twitter Profile Photo

Introducing MeCo (metadata conditioning then cooldown), a remarkably simple method that accelerates LM pre-training by simply prepending source URLs to training documents. arxiv.org/abs/2501.01956

Introducing MeCo (metadata conditioning then cooldown), a remarkably simple method that accelerates LM pre-training by simply prepending source URLs to training documents.

arxiv.org/abs/2501.01956
Xi Ye (@xiye_nlp) 's Twitter Profile Photo

🤔Now most LLMs have >= 128K context sizes, but are they good at generating long outputs, such as writing 8K token chain-of-thought for a planning problem? 🔔Introducing LongProc (Long Procedural Generation), a new benchmark with 6 diverse tasks that challenge LLMs to synthesize

🤔Now most LLMs have &gt;= 128K context sizes, but are they good at generating long outputs, such as writing 8K token chain-of-thought for a planning problem?
🔔Introducing LongProc (Long Procedural Generation), a new benchmark with 6 diverse tasks that challenge LLMs to synthesize
Manos Koukoumidis (@koukoumidis) 's Twitter Profile Photo

If AI isn’t truly open, it will fail us. We can’t close in a black box our greatest invention yet just so that a few can freely monetize. AI needs its Linux moment, and so we started working towards it. This can only succeed if we all work together! #oumi #opensource

Yong Lin (@yong18850571) 's Twitter Profile Photo

🚀 Introducing Goedel-Prover: A 7B LLM achieving SOTA open-source performance in automated theorem proving! 🔥 ✅ Improving +7% over previous open source SOTA on miniF2F 🏆 Ranking 1st on the PutnamBench Leaderboard 🤖 Solving 1.9X total problems compared to prior works on Lean

🚀 Introducing Goedel-Prover: A 7B LLM achieving SOTA open-source performance in automated theorem proving! 🔥

✅ Improving +7% over previous open source SOTA on miniF2F
🏆 Ranking 1st on the PutnamBench Leaderboard
🤖 Solving 1.9X total problems compared to prior works on Lean
Yong Lin (@yong18850571) 's Twitter Profile Photo

🚀 Exciting news! Our Goedel-Prover paper is now live on arXiv: arxiv.org/pdf/2502.07640 🎉 We're currently developing the RL version and have a stronger checkpoint than before (currently not included in the report)!🚀🚀🚀 Plus, we’ll be open-sourcing 1.64M formalized

🚀 Exciting news! Our Goedel-Prover paper is now live on arXiv: arxiv.org/pdf/2502.07640 🎉 

We're currently developing the RL version and have  a stronger checkpoint than before (currently not included in the report)!🚀🚀🚀

Plus, we’ll be open-sourcing 1.64M formalized
Stanford NLP Group (@stanfordnlp) 's Twitter Profile Photo

Congratulations to Stanford NLP Group founder Christopher Manning for being elected to The National Academy of Engineering (NAE, National Academies) Class of 2025 for the development and dissemination of natural language processing methods.

Congratulations to <a href="/stanfordnlp/">Stanford NLP Group</a> founder <a href="/chrmanning/">Christopher Manning</a> for being elected to The National Academy of Engineering (NAE,  <a href="/theNASEM/">National Academies</a>) Class of 2025 for the development and dissemination of natural language processing methods.
Alex Wettig (@_awettig) 's Twitter Profile Photo

🤔 Ever wondered how prevalent some type of web content is during LM pre-training? In our new paper, we propose WebOrganizer which *constructs domains* based on the topic and format of CommonCrawl web pages 🌐 Key takeaway: domains help us curate better pre-training data! 🧵/N

🤔 Ever wondered how prevalent some type of web content is during LM pre-training?

In our new paper, we propose WebOrganizer which *constructs domains* based on the topic and format of CommonCrawl web pages 🌐

Key takeaway: domains help us curate better pre-training data! 🧵/N
Danqi Chen (@danqi_chen) 's Twitter Profile Photo

V. happy with this work! We’ve explored domain mixtures and quality filtering (including Alex’s previous work!), but what is even a “domain” in Common Crawl? Can we use these domains to better understand quality filters, and combine them for data curation? Cool visuals too!

Noam Razin (@noamrazin) 's Twitter Profile Photo

The success of RLHF depends heavily on the quality of the reward model (RM), but how should we measure this quality? 📰 We study what makes a good RM from an optimization perspective. Among other results, we formalize why more accurate RMs are not necessarily better teachers! 🧵

The success of RLHF depends heavily on the quality of the reward model (RM), but how should we measure this quality?

📰 We study what makes a good RM from an optimization perspective. Among other results, we formalize why more accurate RMs are not necessarily better teachers!
🧵
Princeton Laboratory for Artificial Intelligence (@princetonainews) 's Twitter Profile Photo

Welcome to the official X for the Princeton Laboratory for Artificial Intelligence (“AI Lab” for short). Our mission is to support and expand the scope of AI research Princeton University Follow our page for the latest updates on events, news, research, and more at the AI Lab

Welcome to the official X for the Princeton Laboratory for Artificial Intelligence (“AI Lab” for short). Our mission is to support and expand the scope of AI research <a href="/Princeton/">Princeton University</a> 

Follow our page for the latest updates on events, news, research, and more at the AI Lab
Howard Yen (@howardyen1) 's Twitter Profile Photo

Llama 4 Scout claims to support a context window of 10M tokens; the needle-in-a-haystack results are perfect, but can it handle real long-context tasks? We evaluate them on HELMET, our diverse and application-centric long-context benchmark, to be presented at #ICLR2025!

Princeton PLI (@princetonpli) 's Twitter Profile Photo

We are proud to highlight the work of the PLI students, post-docs, and faculty which is being showcased at this year's ICLR 2025: pli.princeton.edu/blog/2025/prin…

We are proud to highlight the work of the PLI students, post-docs, and faculty which is being showcased at this year's <a href="/iclr_conf/">ICLR 2025</a>: pli.princeton.edu/blog/2025/prin…