Elan Rosenfeld(@ElanRosenfeld) 's Twitter Profileg
Elan Rosenfeld

@ElanRosenfeld

Final year ML PhD Student at CMU working on principled approaches to robustness, security, generalization, and representation learning.

ID:1045447599090794497

linkhttp://cs.cmu.edu/~elan calendar_today27-09-2018 22:58:25

323 Tweets

1,1K Followers

184 Following

ML@CMU(@mlcmublog) 's Twitter Profile Photo

Imagine you're a data scientist who solves several related linear regression problems from the same application domain.

Can you learn how to best use a combination of L1 and L2 regularization penalties? We show that you can! How much data is needed?

blog.ml.cmu.edu/2024/04/12/how…

account_circle
Elan Rosenfeld(@ElanRosenfeld) 's Twitter Profile Photo

I can't speak for other schools, but I can tell you this is definitely not the case at CMU. This seems like a surefire way to lose your school's status as a top program.

account_circle
Samuel Sokota(@ssokota) 's Twitter Profile Photo

SOTA AI for games like poker & Hanabi rely on search methods that don’t scale to games w/ large amounts of hidden information.

In our ICLR paper, we introduce simple search methods that scale to large games & get SOTA for Hanabi w/ 100x less compute. 1/N

arxiv.org/abs/2304.13138

SOTA AI for games like poker & Hanabi rely on search methods that don’t scale to games w/ large amounts of hidden information. In our ICLR paper, we introduce simple search methods that scale to large games & get SOTA for Hanabi w/ 100x less compute. 1/N arxiv.org/abs/2304.13138
account_circle
Vaishnavh Nagarajan(@_vaishnavh) 's Twitter Profile Photo

🗣️ “Next-token predictors can’t plan!” ⚔️ ​​“False! Every distribution is expressible as product of next-token probabilities!” 🗣️

In work w/ Gregor Bachmann , we carefully flesh out this emerging, fragmented debate & articulate a key new failure. 🔴 arxiv.org/abs/2403.06963

account_circle
Francesco Orabona(@bremen79) 's Twitter Profile Photo

How you ever wondered why the KL divergence is in all the PAC-Bayes bounds? Are we sure is it the optimal choice?

We now know: for sure KL is *not* the optimal one!

New work with Ilja Kuzborskij, Kwang-Sung Jun , yulian wu, and Kyoungseok Jang

twitter.com/StatMLPapers/s…

account_circle
Elan Rosenfeld(@ElanRosenfeld) 's Twitter Profile Photo

A sequence of videos of Will Smith eating spaghetti, overlaid with the shutterstock logo. In some clips he uses a fork and in others his hands overflow with spaghetti as he shovels it into his mouth. In each clip he is wearing a different outfit. One clip has two Will Smiths.

account_circle
Elan Rosenfeld(@ElanRosenfeld) 's Twitter Profile Photo

Wow, the two points in the abstract of number (1) are clever and I'm surprised I haven't seen this interpretation before. Is this widely known?

account_circle
Elan Rosenfeld(@ElanRosenfeld) 's Twitter Profile Photo

Related, possibly controversial:

As a rule of thumb, if the abstract uses the phrase 'We present/propose', I regard it with greater skepticism. It sounds to me like the authors are marketing a product, and you should always be skeptical of marketing.

(skeptical ≠ disbelieving)

account_circle
Jason Hartford(@jasonhartford) 's Twitter Profile Photo

CARE talks kick off again this week, with
Goutham Rajendran talking about learning disentangled representations (portal.valencelabs.com/events/post/le…). It's a really nice paper showing with linear + Gaussian latents we don't need many interventions to disentangle latents. Thursday, 11am EST

account_circle
Zico Kolter(@zicokolter) 's Twitter Profile Photo

I feel like a lot of people leverage LLMs suboptimally, especially for long-form interactions that span a whole project. So I wrote a VSCode extension that supports what I think is a better use paradigm. 🧵 1/N

Extension: marketplace.visualstudio.com/items?itemName…
Code: github.com/locuslab/chatl…

account_circle
Gokul Swamy(@g_k_swamy) 's Twitter Profile Photo

I'm excited to share a preview of what I've spent the last few months working on at Google AI: SPO, a new RLHF algorithm with strikingly simple implementation (no reward models) and shockingly strong guarantees (handles messy, intransitive prefs.): arxiv.org/abs/2401.04056

I'm excited to share a preview of what I've spent the last few months working on at @GoogleAI: SPO, a new RLHF algorithm with strikingly simple implementation (no reward models) and shockingly strong guarantees (handles messy, intransitive prefs.): arxiv.org/abs/2401.04056
account_circle