Yinghui He (@gracie_huihui) 's Twitter Profile
Yinghui He

@gracie_huihui

First-Year Computer Science PhD @PrincetonPLI

ID: 1577860052811218944

linkhttp://ying-hui-he.github.io calendar_today06-10-2022 03:15:38

13 Tweet

109 Followers

179 Following

Andrea May Sahouri (@andreamsahouri) 's Twitter Profile Photo

At least 1k here at UofM’s Diag, mourning those lost in the mass shooting at MSU. Students laid flowers on the Block M. A banner placed at the bottom of Hatcher’s stairs reads “Wolverines for Spartans”

At least 1k here at UofM’s Diag, mourning those lost in the mass shooting at MSU. 

Students laid flowers on the Block M. A banner placed at the bottom of Hatcher’s stairs reads “Wolverines for Spartans”
Yinghui He (@gracie_huihui) 's Twitter Profile Photo

I just did a poster presentation on my paper “Hi-ToM: A Benchmark for Evaluating Higher-Order Theory of Mind Reasoning in Large Language Models” at the ICML 2023 Theory of Mind Workshop. Nice experience!

I just did a poster presentation on my paper “Hi-ToM: A Benchmark for Evaluating Higher-Order Theory of Mind Reasoning in Large Language Models” at the ICML 2023 Theory of Mind Workshop. Nice experience!
Sanjeev Arora (@prfsanjeevarora) 's Twitter Profile Photo

1/ New instruction-following dataset INSTRUCT-SKILLMIX! Supervised fine-tuning (SFT) with just 2K-4K (query, answer) pairs gives small “base LLMs” Mistral v0.2 7B and LLaMA3 8B performance rivalling some frontier models (AlpacaEval 2.0 score). No RL, no expensive human data.

1/ New instruction-following dataset INSTRUCT-SKILLMIX! Supervised fine-tuning (SFT) with  just 2K-4K (query, answer) pairs gives  small “base LLMs”  Mistral v0.2 7B and LLaMA3 8B performance rivalling some frontier models (AlpacaEval 2.0 score). No RL, no expensive human data.
Jiao Sun (@sunjiao123sun_) 's Twitter Profile Photo

Mitigating racial bias from LLMs is a lot easier than removing it from humans! Can’t believe this happened at the best AI conference NeurIPS Conference We have ethical reviews for authors, but missed it for invited speakers? 😡

Mitigating racial bias from LLMs is a lot easier than removing it from humans! 

Can’t believe this happened at the best AI conference <a href="/NeurIPSConf/">NeurIPS Conference</a> 

We have ethical reviews for authors, but missed it for invited speakers? 😡
Xi Ye (@xiye_nlp) 's Twitter Profile Photo

🤔Now most LLMs have >= 128K context sizes, but are they good at generating long outputs, such as writing 8K token chain-of-thought for a planning problem? 🔔Introducing LongProc (Long Procedural Generation), a new benchmark with 6 diverse tasks that challenge LLMs to synthesize

🤔Now most LLMs have &gt;= 128K context sizes, but are they good at generating long outputs, such as writing 8K token chain-of-thought for a planning problem?
🔔Introducing LongProc (Long Procedural Generation), a new benchmark with 6 diverse tasks that challenge LLMs to synthesize
Yinghui He (@gracie_huihui) 's Twitter Profile Photo

Excited to introduce AdaptMI, a test-time adaptation method to provide **minimal**, **targeted** in-context feedback on math reasoning. Shout out to my collaborators Sanjeev Arora Abhishek Panigrahi Yong Lin for pushing my first first-authored project in Princeton!