
Alex Meterez
@alexmeterez
cs phd student @harvard
deep learning theory, optimization, scale pilled
ID: 1243303287752925190
https://alexandrumeterez.github.io/ 26-03-2020 22:26:27
197 Tweet
173 Takipçi
788 Takip Edilen








Great to see this one finally out in PNAS! Asymptotic theory of in-context learning by linear attention pnas.org/doi/10.1073/pn… Many thanks to my amazing co-authors Yue Lu, Mary Letey, Jacob Zavatone-Veth and Anindita Maiti



First #ICML2025 conference proceeding (icml.cc/virtual/2025/p…)! We (Blake Bordelon, Jacob Zavatone-Veth, Cengiz Pehlevan) developed a simple model to better understand state representation learning dynamics in both artificial and biological intelligent systems! Comments appreciated!

A team from #KempnerInstitute, Harvard SEAS & Computer Science at UT Austin has won a best paper award at #ICML2025 for work unlocking the potential of masked diffusion models. Congrats to Jaeyeon (Jay) Kim @ICML, Kulin Shah, Vasilis Kontonis, Sham Kakade and Sitan Chen. kempnerinstitute.harvard.edu/news/kempner-i… #AI

ICML this week! Come by T PM Clarissa Lauditi's work on muP BNNs arxiv.org/abs/2502.07998 W AM, model of place field adaptation@mgkumar138, Jacob ZV biorxiv.org/content/10.110… W PM a model of LR transfer in linear NNs arxiv.org/abs/2502.02531 all from senior author Cengiz Pehlevan!

As Micah Goldblum and coauthors, we also found that small batches make SGD effective in LM training. It's cool that our papers came out around the same time, and each has a different perspective! Below, our take on why this happens. Our awesome team: Teodora Jonas Geiping


Come to HilD tomorrow International Conference on Minority Languages ! We have 4 posters on optimization: - In Search of Adam’s Secret Sauce - Is your batch size the problem? Revisiting the Adam-SGD gap in language modeling - On the Interaction of Noise, Compression Role, and Adaptivity under (L0,L1)-Smoothness



Very excited to be a part of this amazing collaboration! Thanks Simons Foundation for the support!

