Alberto Bietti (@albertobietti) 's Twitter Profile
Alberto Bietti

@albertobietti

Machine learning research. Research scientist @FlatironCCM, previously @MetaAI, @NYUDataScience, @Inria, @Quora.

ID: 11056912

linkhttp://alberto.bietti.me calendar_today11-12-2007 18:03:01

1,1K Tweet

1,1K Followers

1,1K Following

Alberto Bietti (@albertobietti) 's Twitter Profile Photo

Want to do research in ML for science in NYC and develop the next generation of foundation models for scientific data? Apply here to be a postdoc at PolymathicAI Simons Foundation : forms.gle/JbdWA3VczH3Xfg…

Want to do research in ML for science in NYC and develop the next generation of foundation models for scientific data? Apply here to be a postdoc at <a href="/PolymathicAI/">PolymathicAI</a> <a href="/SimonsFdn/">Simons Foundation</a> : forms.gle/JbdWA3VczH3Xfg…
Harmonic (@harmonicmath) 's Twitter Profile Photo

We are excited to share three major updates today on our path to mathematical superintelligence 🦾 1. A new state-of-the-art of 90% on the MiniF2F benchmark. This beats our previously announced 83% from under a month ago.

Eran Malach (@eranmalach) 's Twitter Profile Photo

Can Transformers extrapolate from short training sequences to long ones? Our new work shows that they display surprising “length generalization” capabilities on many algorithmic tasks: addition, multiplication, and even in-context simulation of SGD! arxiv.org/abs/2407.03310

Can Transformers extrapolate from short training sequences to long ones? Our new work shows that they display surprising “length generalization” capabilities on many algorithmic tasks: addition, multiplication, and even in-context simulation of SGD! arxiv.org/abs/2407.03310
Vivien Cabannes (@cabannesvivien) 's Twitter Profile Photo

Actively learning from weak information is a key problem in machine learning. Presented at COLT last week, our paper captures core aspects of active weakly supervised learning, showing how to combine information-theoretical and bandit likes methods. arxiv.org/abs/2402.13079

Actively learning from weak information is a key problem in machine learning. Presented at COLT last week, our paper captures core aspects of active weakly supervised learning, showing how to combine information-theoretical and bandit likes methods.
arxiv.org/abs/2402.13079
Sam Rodriques (@sgrodriques) 's Twitter Profile Photo

Today, we're releasing LAB-Bench, a set of >2000 evaluations for language models and agents on scientific research tasks in biology. Public models underperform PhD/postdoc-level humans on nearly all tasks. Claude 3.5 Sonnet is the clear frontrunner atm, but long way to go. 1/

Today, we're releasing LAB-Bench, a set of &gt;2000 evaluations for language models and agents on scientific research tasks in biology. Public models underperform PhD/postdoc-level humans on nearly all tasks. Claude 3.5 Sonnet is the clear frontrunner atm, but long way to go. 1/
Eshaan Nichani (@eshaannichani) 's Twitter Profile Photo

In Vienna for ICML 2024! I'll be presenting our paper on how transformers learn causal structure (like the induction head) via GD at the Wednesday 1:30pm poster session (Hall C, poster #1600). Excited to chat with all who are around! x.com/EshaanNichani/…

Diana Cai (@dianarycai) 's Twitter Profile Photo

We have a spotlight paper at #ICML2024 on batch and match: black box variational inference with a score-based divergence. Our poster is Thursday 11:30a Hall C 4–9 #1307. 1/n

We have a spotlight paper at #ICML2024 on batch and match: black box variational inference with a score-based divergence. Our poster is Thursday 11:30a Hall C 4–9 #1307.
1/n
Robert M. Gower 🇺🇦 (@gowerrobert) 's Twitter Profile Photo

New method for VI (Variational Inference) and a new optimization trick! Let me tell you about our new (spotlight) paper at ICML (Thu 11:30am Hall C 4-9 #1307). arxiv.org/pdf/2402.14758 The story involves a new divergence and a (potentially) new proximal point update. 1/n

Alberto Bietti (@albertobietti) 's Twitter Profile Photo

Training LLMs involves learning associations. We study training dynamics in a simple model that yields useful insight on the role of token interference and imbalance. Go talk to Vivien Cabannes and Berfin Simsek @ ICML at our #ICML poster #1114 tmrw 11:30! Paper: arxiv.org/abs/2402.18724

Dylan Foster (@canondetortugas) 's Twitter Profile Photo

Now that I have started using twitter somewhat regularly, let me take a minute to advertise the RL theory lecture notes I have been developing with Sasha Rakhlin: arxiv.org/abs/2312.16730

Now that I have started using twitter somewhat regularly, let me take a minute to advertise the RL theory lecture notes I have been developing with Sasha Rakhlin: arxiv.org/abs/2312.16730
Miles Cranmer (@milescranmer) 's Twitter Profile Photo

Some exciting PolymathicAI news... We're expanding!! New Research Software Engineer positions opening in Cambridge UK, NYC, and remote. Come build generalist foundation models for science with us! Please indicate your interest on the form here: docs.google.com/forms/d/e/1FAI…

Siavash Golkar (@siavashgolkar) 's Twitter Profile Photo

SOTA models often use bidirectional transformers for non-NLP tasks but did you know causal transformers can outperform them even on tasks without a causal structure? Our recent work shows causal transformers learn circuits bidirectional ones can't, leading to better performance!

SOTA models often use bidirectional transformers for non-NLP tasks but did you know causal transformers can outperform them even on tasks without a causal structure?

Our recent work shows causal transformers learn circuits bidirectional ones can't, leading to better performance!
Konstantin Mishchenko (@konstmish) 's Twitter Profile Photo

I'm looking for more reviewers (especially in optimization theory) for Transactions on Machine Learning Research, please reach out with your email (you can send it to me in DM) if you're interested in becoming a reviewer! TMLR has the best system I know: 1. You don't get assigned more than 1 submission at a

Mistral AI (@mistralai) 's Twitter Profile Photo

magnet:?xt=urn:btih:7278e625de2b1da598b23954c13933047126238a&dn=pixtral-12b-240910&tr=udp%3A%2F%2Ftracker.opentrackr.org%3A1337%2Fannounce&tr=udp%3A%2F%2Fopen.demonii.com%3A1337%2Fannounce&tr=http%3A%2F%2Ftracker.ipv6tracker.org%3A80%2Fannounce

ML-NYC Speaker Series (@mlnycseries) 's Twitter Profile Photo

Excited to announce that ML-NYC is back this semester. Our first speaker is Bin Yu from UC Berkeley on Monday Sept 23rd 4pm. Register here: eventbrite.com/e/ml-nyc-speak…