
Roy Frostig
@froystig
research scientist at @googledeepmind. co-author of JAX (github.com/jax-ml/jax)
ID: 14295070
https://cs.stanford.edu/~rfrostig/ 03-04-2008 17:26:40
123 Tweet
1,1K Followers
603 Following


Today, Iām excited to announce the release of Levanter 1.0, our new JAX-based framework for training foundation models, which weāve been working on Center for Research on Foundation Models. Levanter is designed to be legible, scalable and reproducible. crfm.stanford.edu/2023/06/16/levā¦


Excited to share Penzai, a JAX research toolkit from Google DeepMind for building, editing, and visualizing neural networks! Penzai makes it easy to see model internals and lets you inject custom logic anywhere. Check it out on GitHub: github.com/google-deepminā¦

Many of you are excited about H100 attention, so itās a good time to show you Mosaic GPU: a Python DSL for H100s. The attention example matches FA3 performance, while being only ~200 lines of Python: github.com/google/jax/blo⦠It's easy to install too! Latest JAX packages have it.



Modula x JAX = Modulax theseriousadult is cracked and ported Modula into JAX in a few days. I haven't had a chance to test yet, but I'm really excited about this project. Tagging Roy Frostig and Matthew Johnson github.com/GallagherComma⦠(1/3)



Training our most capable Gemini models relies heavily on our JAX software stack + Google's TPU hardware platforms. If you want to learn more, see this awesome book "How to Scale Your Model": jax-ml.github.io/scaling-book/ It was put together by my Google DeepMind colleagues

