Yizhe Zhu (@yizhezhu_) 's Twitter Profile
Yizhe Zhu

@yizhezhu_

Assistant Professor of Mathematics at University of Southern California. Probability, combinatorics, data science

ID: 1551301806763847680

linkhttps://sites.google.com/usc.edu/yizhezhu calendar_today24-07-2022 20:22:35

94 Tweet

304 Takipรงi

395 Takip Edilen

Samuel Vaiter (@vaiter) 's Twitter Profile Photo

The Davis-Kahan sin(ฮธ) theorem bounds the difference between the subspaces spanned by eigenvectors of two symmetric matrices, based on the difference between the matrices. It quantifies how small perturbations in a matrix affect its eigenvectors. jstor.org/stable/2949580โ€ฆ

The Davis-Kahan sin(ฮธ) theorem bounds the difference between the subspaces spanned by eigenvectors of two symmetric matrices, based on the difference between the matrices. It quantifies how small perturbations in a matrix affect its eigenvectors. jstor.org/stable/2949580โ€ฆ
Xiaohui Chen (@xiaohuichen18) 's Twitter Profile Photo

Introducing M.S. in Mathematical Data Science at USC Dornsife. Now accepting applications!! youtu.be/ELOFHqe2BFs?siโ€ฆ via YouTube

Subhabrata Sen (@subhabratasen90) 's Twitter Profile Photo

I have an opening for a two year postdoc position at Harvard. The application link: academicpositions.harvard.edu/postings/14420

Lenka Zdeborova (@zdeborova) 's Twitter Profile Photo

Excited about our progress in characterizing The Computational Advantage of Depth in Learning with Neural Networks. Check out the number of samples that can be saved when GD runs on a multi-layer rather than on a two-layer neural network. arxiv.org/pdf/2502.13961

Excited about our progress in characterizing The Computational Advantage of Depth in Learning with Neural Networks. Check out the number of samples that can be saved when GD runs on a multi-layer rather than on a two-layer neural network.    arxiv.org/pdf/2502.13961
Yiqiao Zhong (@yiqiao_zhong) 's Twitter Profile Photo

I'm glad that my paper on LLMs is published in the prestigious journal PNAS. The main finding is that LLMs represent compositions and achieve OOD generalization using latent subspaces so that multiple self-attention layers are aligned. See the blog post: yiqiao-zhong.github.io/jekyll/update/โ€ฆ

I'm glad that my paper on LLMs is published in the prestigious journal PNAS. The main finding is that LLMs represent compositions and achieve OOD generalization using latent subspaces so that multiple self-attention layers are aligned. See the blog post: yiqiao-zhong.github.io/jekyll/update/โ€ฆ
Molei Tao (@moleitaomath) 's Twitter Profile Photo

I'm trying to compile a reading list for math graduate students, on existing/popular theories of generalization in deep learning. I can think of tens of beautiful research articles, but any book / survey you would recommend? Any suggestion would be appreciated!

Petar Veliฤkoviฤ‡ (@petarv_93) 's Twitter Profile Photo

"This is one of the first results where a neural model, when trained to sufficiently low loss, can ๐ ๐ฎ๐š๐ซ๐š๐ง๐ญ๐ž๐ž ๐Ž๐Ž๐ƒ ๐ฌ๐ข๐ณ๐ž ๐ ๐ž๐ง๐ž๐ซ๐š๐ฅ๐ข๐ณ๐š๐ญ๐ข๐จ๐ง for a non-linear algorithm" ... ๐Ÿคฏ

"This is one of the first results where a neural model, when trained to sufficiently low loss, can ๐ ๐ฎ๐š๐ซ๐š๐ง๐ญ๐ž๐ž ๐Ž๐Ž๐ƒ ๐ฌ๐ข๐ณ๐ž ๐ ๐ž๐ง๐ž๐ซ๐š๐ฅ๐ข๐ณ๐š๐ญ๐ข๐จ๐ง for a non-linear algorithm"

...

๐Ÿคฏ
Max Hopkins (@mhop_theory) 's Twitter Profile Photo

For those that couldn't make it, we've uploaded our full STOC workshop on High Dimensional Expanders to Youtube! Hopefully a useful resource for learning the basics of HDX and how they're applied in TCS. Talk 1: An introduction to HDX youtube.com/watch?v=r6mrG3โ€ฆ