Bingbin Liu (@bingbinl) 's Twitter Profile
Bingbin Liu

@bingbinl

PhD student at CMU machine learning department.

ID: 1039263012123758595

linkhttps://clarabing.github.io/ calendar_today10-09-2018 21:23:05

102 Tweet

841 Followers

246 Following

Kempner Institute at Harvard University (@kempnerinst) 's Twitter Profile Photo

The application is now open for our #KempnerInstitute Research #Fellowship! Postdocs studying the foundations of #intelligence or applications of #AI are encouraged to apply. Learn more and apply by Oct. 1: bit.ly/4djPSf8 #LLMs #NeuroAI #ML Sham Kakade Bernardo Sabatini

The application is now open for our #KempnerInstitute Research #Fellowship! Postdocs studying the foundations of #intelligence or applications of #AI are encouraged to apply. Learn more and apply by Oct. 1: bit.ly/4djPSf8 #LLMs #NeuroAI #ML <a href="/ShamKakade6/">Sham Kakade</a> <a href="/blsabatini/">Bernardo Sabatini</a>
Bingbin Liu (@bingbinl) 's Twitter Profile Photo

CFP & join us at M3L Workshop @ NeurIPS 2024 at #NeurIPS2024! We look forward to learning about your insights & findings on the theoretical and scientific understanding of ML phenomena💡

Christina Baek (@_christinabaek) 's Twitter Profile Photo

Chatbots are often augmented w/ new facts by context from the user or retriever. Models must adapt instead of hallucinating outdated facts. In this work w/Sachin Goyal, Zico Kolter, Aditi Raghunathan, we show that instruction tuning fails to reliably improve this behavior! [1/n]

Sadhika Malladi (@sadhikamalladi) 's Twitter Profile Photo

Our work has been selected as an Oral at ICLR 25! We find theoretical and empirical explanations for the benefits of progressive distillation. Amazing work led by Abhishek Panigrahi and Bingbin Liu, done in collaboration with Andrej Risteski and Surbhi Goel :)

Gokul Swamy (@g_k_swamy) 's Twitter Profile Photo

1.5 yrs ago, we set out to answer a seemingly simple question: what are we *actually* getting out of RL in fine-tuning? I'm thrilled to share a pearl we found on the deepest dive of my PhD: the value of RL in RLHF seems to come from *generation-verification gaps*. Get ready to🤿!

1.5 yrs ago, we set out to answer a seemingly simple question: what are we *actually* getting out of RL in fine-tuning? I'm thrilled to share a pearl we found on the deepest dive of my PhD: the value of RL in RLHF seems to come from *generation-verification gaps*. Get ready to🤿!
Sadhika Malladi (@sadhikamalladi) 's Twitter Profile Photo

1⃣ Distillation (Oral; led by Abhishek Panigrahi and Bingbin Liu): theory + exps on when and how progressive distillation (training w a few intermediate teachers) is beneficial for the student. Prog distillation induces a provably beneficial implicit curriculum. arxiv.org/abs/2410.05464

Tanya Marwah (@__tm__157) 's Twitter Profile Photo

What is the role of memory for modeling time dependent PDEs? I will be at ICLR presenting our paper (Oral) where we study when it is beneficial for modeling time-dependent PDEs! 🔗openreview.net/forum?id=o9kqa… [Oral]: Thu 24 Apr 10:30 am @ Session 1E [Poster]: Thu 24 Apr 3 pm #617

Bingbin Liu (@bingbinl) 's Twitter Profile Photo

If you're at #ICLR2025, come to our oral talk & poster on progressive distillation presented by the amazing Abhishek Panigrahi! ✨🌴 Joint work with (the equally amazing) Sadhika Malladi, Andrej Risteski, Surbhi Goel More details at our blog: unprovenalgos.github.io/progressive-di…

Kempner Institute at Harvard University (@kempnerinst) 's Twitter Profile Photo

New in the Deeper Learning blog: Kempner researchers show how VLMs speak the same semantic language across images and text. bit.ly/KempnerVLM by Isabel Papadimitriou, Chloe H. Su, Thomas Fel, Stephanie Gil, and Sham Kakade #AI #ML #VLMs #SAEs

Andrej Risteski (@risteski_a) 's Twitter Profile Photo

Misha Khodak, Tanya Marwah, along with myself, Nicholas Boffi and Jianfeng Lu are organizing a COLT 2025 workshop on the Theory of AI for Scientific Computing, to be held on the first day of the conference (June 30).

<a href="/khodakmoments/">Misha Khodak</a>, <a href="/__tm__157/">Tanya Marwah</a>, along with myself, <a href="/nmboffi/">Nicholas Boffi</a> and Jianfeng Lu are organizing a COLT 2025 workshop on the Theory of AI for Scientific Computing, to be held on the first day of the conference (June 30).
Bingbin Liu (@bingbinl) 's Twitter Profile Photo

Excited to announce MOSS, our ICML workshop focused on discoveries at small scale! We believe there's tremendous potential & creativity in research done with limited resources and would love to hear your ideas. The submission (due May 22nd) can literally be a Jupyter notebook! :)

Songlin Yang (@songlinyang4) 's Twitter Profile Photo

📢 (1/16) Introducing PaTH 🛣️ — a RoPE-free contextualized position encoding scheme, built for stronger state tracking, better extrapolation, and hardware-efficient training. PaTH outperforms RoPE across short and long language modeling benchmarks arxiv.org/abs/2505.16381

Hanlin Zhang (@_hanlin_zhang_) 's Twitter Profile Photo

[1/n] New work [JSKZ25] w/ Jikai Jin, Vasilis Syrgkanis, Sham Kakade. We introduce new formulations and tools for evaluating language model capabilities, which help explain recent observations of post-training behaviors of Qwen-series models — there is a sensitive causal link

[1/n] New work [JSKZ25] w/ <a href="/JikaiJin2002/">Jikai Jin</a>, <a href="/syrgkanis/">Vasilis Syrgkanis</a>, <a href="/ShamKakade6/">Sham Kakade</a>. 

We introduce new formulations and tools for evaluating language model capabilities, which help explain recent observations of post-training behaviors of Qwen-series models — there is a sensitive causal link