Damien Ferbach (@damien_ferbach) 's Twitter Profile
Damien Ferbach

@damien_ferbach

PhD at @Mila_Quebec/ Previously maths at @ENS_ULM

ID: 1497811031703408642

linkhttps://damienferbach.github.io calendar_today27-02-2022 05:49:27

31 Tweet

169 Followers

145 Following

Joey Bose (@bose_joey) 's Twitter Profile Photo

Some exciting news coverage of iterative retraining of generative models by the NYT! nytimes.com/interactive/20… The article includes several papers including some of my co-authors from Mila - Institut québécois d'IA Quentin Bertrand Damien Ferbach Gauthier Gidel were excited to write: Broad strokes

Damien Ferbach (@damien_ferbach) 's Twitter Profile Photo

I am delighted to share that our paper has been accepted at #NeurIPS as a spotlight!🚀 A huge thanks to my amazing collaborators Quentin Bertrand, Joey Bose and my supervisor Gauthier Gidel !!

Damien Ferbach (@damien_ferbach) 's Twitter Profile Photo

Dreaming of pushing the boundaries in AI? 🌟 Apply for MSc or PhD at Mila, a world-leading research hub, for Fall 2025! 🚀 #AI #Research #Mila #PhD #MSc

Gauthier Gidel (@gauthier_gidel) 's Twitter Profile Photo

I am hiring Ph.D. and/or M.Sc. students to work at the intersection of game theory, optimization, and Machine Learning in Fall 2025. If you are interested in working in my group, apply via mila.quebec/en/prospective… More details on the topics in the 🧵👇 1/n

Damien Ferbach (@damien_ferbach) 's Twitter Profile Photo

I will be at NeurIPS in Vancouver this week to present our work on self-consuming generative models. arxiv.org/abs/2407.09499 Please reach out to talk about high-dimensional optimization and generative models !

Joey Bose (@bose_joey) 's Twitter Profile Photo

Come checkout our spotlight poster in the East ballroom #2510 happening now with Damien Ferbach Quentin Bertrand Gauthier Gidel Especially come if you're interested in self Consuming generative models and model collapse!!

Come checkout our spotlight poster in the East ballroom #2510 happening now with <a href="/damien_ferbach/">Damien Ferbach</a> <a href="/Qu3ntinB/">Quentin Bertrand</a> <a href="/gauthier_gidel/">Gauthier Gidel</a> 

Especially come if you're interested in self Consuming generative models and model collapse!!
Reza Bayat (@reza_byt) 's Twitter Profile Photo

New Paper! 📄 Once a model memorizes an example, it stops learning from it! Our latest work explores this phenomenon and the nuanced interplay between memorization and generalization. Let’s dive in! 🚀🧵

New Paper! 📄 Once a model memorizes an example, it stops learning from it!

Our latest work explores this phenomenon and the nuanced interplay between memorization and generalization.

Let’s dive in! 🚀🧵
Katie Everett (@_katieeverett) 's Twitter Profile Photo

1. We often observe power laws between loss and compute: loss = a * flops ^ b + c 2. Models are rapidly becoming more efficient, i.e. use less compute to reach the same loss But: which innovations actually change the exponent in the power law (b) vs change only the constant (a)?

Katie Everett (@_katieeverett) 's Twitter Profile Photo

Hestness et al 2017 asked this question and concluded "model improvements only shift the error but do not appear to affect the power-law exponent". Are we doomed to chase constant factor improvements? Is there anything we can do to improve the exponent?

Hestness et al 2017 asked this question and concluded "model improvements only shift the error but do not appear to affect the power-law exponent".

Are we doomed to chase constant factor improvements? Is there anything we can do to improve the exponent?
Katie Everett (@_katieeverett) 's Twitter Profile Photo

In summary, while data can change the power law exponent significantly, it's quite challenging to find architectures and optimizers that change the exponent! If you want to see something that *does* change the power law exponent, stay tuned and follow Damien Ferbach for more 😁