
Andrew Saxe
@saxelab
Prof at @GatsbyUCL and @SWC_Neuro, trying to figure out how we learn.
Bluesky: @SaxeLab
Mastodon: @[email protected]
ID: 1193222240202035200
http://www.saxelab.org 09-11-2019 17:42:48
684 Tweet
5,5K Takipçi
380 Takip Edilen

New paper with Leon and Erin Grant! Why do we see localized receptive fields so often, even in models without sparisity regularization? We present a theory in the minimal setting from Alessandro Ingrosso and Sebastian Goldt





Our paper, “Make Haste Slowly: A Theory of Emergent Structured Mixed Selectivity in Feature Learning ReLU Networks” will be presented at ICLR 2025 this week (openreview.net/forum?id=27SSn…)! We derive closed-form dynamics for some, remarkably linear, feature learning ReLU networks (1/9)


Our paper just came out in PRX! Congrats to Nishil Patel and the rest of the team. TL;DR : We analyse neural network learning through the lens of statistical physics, revealing distinct scaling regimes with sharp transitions. 🔗 journals.aps.org/prx/abstract/1…

Really happy to see this paper out, led by Nishil Patel in collaboration with Stefano Sarao Mannelli and Andrew Saxe: we apply the statistical physics toolbox to analyse a simple model of reinforcement learning, and find some cool effects, like a speed-accuracy trade-off for generalisation 🚀



How do task dynamics impact learning in networks with internal dynamics? Excited to share our ICML Oral paper on learning dynamics in linear RNNs! with Clémentine Dominé 🍊 Murray Shanahan Pedro Mediano openreview.net/forum?id=KGOcr…