Yunzhen Feng (@feeelix_feng) 's Twitter Profile
Yunzhen Feng

@feeelix_feng

PhD at CDS, NYU. Ex-Intern at FAIR @AIatMeta. Previously undergrad at @PKU1898

ID: 1523345547565879298

calendar_today08-05-2022 16:54:28

87 Tweet

326 Takipรงi

588 Takip Edilen

Xiaosen Zheng @ NeurIPS 2024 (@xszheng2020) 's Twitter Profile Photo

๐ŸŽ‰ Excited to share that "NullModel" has been accepted to #ICLR2025 as an oral presentation! I am on the Job Market! I am seeking a Research Scientist position aligned with, but not limited to Data-Centric AI and AI Safety. Feel free to reach out if interested. RT appreciated!

Vivek Myers (@vivek_myers) 's Twitter Profile Photo

Current robot learning methods are good at imitating tasks seen during training, but struggle to compose behaviors in new ways. When training imitation policies, we found something surprisingโ€”using temporally-aligned task representations enabled compositional generalization. 1/

Zhuang Liu (@liuzhuang1234) 's Twitter Profile Photo

How different are the outputs of various LLMs, and in what ways do they differ? Turns out, very very different, up to the point that a text encoding classifier could tell the source LLM with 97% accuracy. This is classifying text generated by LLMs, between ChatGPT, Claude,

How different are the outputs of various LLMs, and in what ways do they differ?

Turns out, very very different, up to the point that a text encoding classifier could tell the source LLM with 97% accuracy.

This is classifying text generated by LLMs, between ChatGPT, Claude,
Aviral Kumar (@aviral_kumar2) 's Twitter Profile Photo

A lot of work focuses on test-time scaling. But we aren't scaling it optimally, simply training a long CoT doesn't mean we use it well. My students developed "v0" of a paradigm to do this optimally by running RL with dense rewards = minimizing regret over long CoT episodes. ๐Ÿงตโฌ‡๏ธ

A lot of work focuses on test-time scaling. But we aren't scaling it optimally, simply training a long CoT doesn't mean we use it well.

My students developed "v0" of a paradigm to do this optimally by running RL with dense rewards = minimizing regret over long CoT episodes. ๐Ÿงตโฌ‡๏ธ
Elvis Dohmatob (@dohmatobelvis) 's Twitter Profile Photo

Papers accepted at ICLR 2026 2025: - An Effective Theory of Bias Amplification arxiv.org/abs/2410.17263 - Pitfalls of Memorization arxiv.org/abs/2412.07684 - Strong Model Collapse arxiv.org/abs/2410.04840 - Beyond Model Collapse arxiv.org/abs/2406.07515 With Julia Kempe,

Elvis Dohmatob (@dohmatobelvis) 's Twitter Profile Photo

We refused to cite the paper due to severe misconduct of the authors of that paper: plagiarism of our own prior work, predominantly AI-generated content (ya, the authors plugged our paper into an LLM and generated another paper), IRB violations, etc. Revealed during a long

Dan Roy (@roydanroy) 's Twitter Profile Photo

I need to offer some clarification for this post because it would be wrong for people to lump this situation in with ones where work that is considered low quality (or inconvenient for priority arguments) is not cited. (This is my gut reaction whenever citations are excluded.)

Yunzhen Feng (@feeelix_feng) 's Twitter Profile Photo

Check out our poster tmr at 10am at the ICLR Bidirectional Human-AI Alignment workshop! We cover how on-policy preference sampling can be biased and our optimal response sampling for human labeling. NYU Center for Data Science AI at Meta Julia Kempe Yaqi Duan x.com/feeelix_feng/sโ€ฆ

Check out our poster tmr at 10am at the ICLR Bidirectional Human-AI Alignment workshop! We cover how on-policy preference sampling can be biased and our optimal response sampling for human labeling.
<a href="/NYUDataScience/">NYU Center for Data Science</a>
<a href="/AIatMeta/">AI at Meta</a>
<a href="/KempeLab/">Julia Kempe</a>
<a href="/YaqiDuanPKU/">Yaqi Duan</a>
x.com/feeelix_feng/sโ€ฆ
Kunhao Zheng @ ICLR 2025 (@kunhaoz) 's Twitter Profile Photo

๐Ÿšจ Your RL only improves ๐—ฝ๐—ฎ๐˜€๐˜€@๐Ÿญ, not ๐—ฝ๐—ฎ๐˜€๐˜€@๐—ธ? ๐Ÿšจ Thatโ€™s not a bug โ€” itโ€™s a ๐—ณ๐—ฒ๐—ฎ๐˜๐˜‚๐—ฟ๐—ฒ ๐—ผ๐—ณ ๐˜๐—ต๐—ฒ ๐—ผ๐—ฏ๐—ท๐—ฒ๐—ฐ๐˜๐—ถ๐˜ƒ๐—ฒ youโ€™re optimizing. You get what you optimize for. If you want better pass@k, you need to optimize for pass@k at training time. ๐Ÿงต How?

๐Ÿšจ Your RL only improves ๐—ฝ๐—ฎ๐˜€๐˜€@๐Ÿญ, not ๐—ฝ๐—ฎ๐˜€๐˜€@๐—ธ? ๐Ÿšจ

Thatโ€™s not a bug โ€” itโ€™s a ๐—ณ๐—ฒ๐—ฎ๐˜๐˜‚๐—ฟ๐—ฒ ๐—ผ๐—ณ ๐˜๐—ต๐—ฒ ๐—ผ๐—ฏ๐—ท๐—ฒ๐—ฐ๐˜๐—ถ๐˜ƒ๐—ฒ youโ€™re optimizing.

You get what you optimize for. If you want better pass@k, you need to optimize for pass@k at training time.

๐Ÿงต How?