Amil Dravid (@_amildravid) 's Twitter Profile
Amil Dravid

@_amildravid

PhD @Berkeley_AI

ID: 1639674617378795520

linkhttp://avdravid.github.io calendar_today25-03-2023 17:04:54

170 Tweet

601 Followers

459 Following

Nick Jiang @ ICLR (@nickhjiang) 's Twitter Profile Photo

Vision transformers have high-norm outliers that hurt performance and distort attention. While prior work removed them by retraining with “register” tokens, we find the mechanism behind outliers and make registers at ✨test-time✨—giving clean features and better performance! 🧵

Vision transformers have high-norm outliers that hurt performance and distort attention. While prior work removed them by retraining with “register” tokens, we find the mechanism behind outliers and make registers at ✨test-time✨—giving clean features and better performance! 🧵
Tamar Rott Shaham (@tamarrottshaham) 's Twitter Profile Photo

How do LMs track what humans believe? In our new work, we show they use a pointer-like mechanism we call lookback. Super proud of this work by Nikhil Prakash and team! This is the most intricate piece of LM reverse engineering I’ve seen!

Andrew Ilyas (@andrew_ilyas) 's Twitter Profile Photo

“How will my model behave if I change the training data?” Recent(-ish) work w/ Logan Engstrom: we nearly *perfectly* predict ML model behavior as a function of training data, saturating benchmarks for this problem (called “data attribution”).

“How will my model behave if I change the training data?”

Recent(-ish) work w/ <a href="/logan_engstrom/">Logan Engstrom</a>: we nearly *perfectly* predict ML model behavior as a function of training data, saturating benchmarks for this problem (called “data attribution”).
Zhuang Liu (@liuzhuang1234) 's Twitter Profile Photo

Can diffusion models appear to be learning, when they’re actually just memorizing the training data? We show and investigate this phenomenon in the context of neural network weight generation, in our recent paper “Generative Modeling of Weights: Generalization or Memorization?"

Can diffusion models appear to be learning, when they’re actually just memorizing the training data?

We show and investigate this phenomenon in the context of neural network weight generation, in our recent paper “Generative Modeling of Weights: Generalization or Memorization?"
TTIC (@ttic_connect) 's Twitter Profile Photo

We’re proud to announce three new tenure-track assistant professors joining TTIC in Fall 2026: Yossi Gandelsman (Yossi Gandelsman), Will Merrill (William Merrill), and Nick Tomlin (Nicholas Tomlin). Meet them here: buff.ly/JH1DFtT

We’re proud to announce three new tenure-track assistant professors joining TTIC in Fall 2026: Yossi Gandelsman (<a href="/YGandelsman/">Yossi Gandelsman</a>), Will Merrill (<a href="/lambdaviking/">William Merrill</a>), and Nick Tomlin (<a href="/NickATomlin/">Nicholas Tomlin</a>). Meet them here: buff.ly/JH1DFtT
Sonia (@soniajoseph_) 's Twitter Profile Photo

The vision mechanistic interpretability workshop Mechanistic Interpretability for Vision @ CVPR2025 earlier this month at CVPR was very informative and fun! Looking forward to seeing this community grow. Thank you to the speakers and organizers trevordarrell David Bau Tamar Rott Shaham Yossi Gandelsman Joanna

The vision mechanistic interpretability workshop <a href="/miv_cvpr2025/">Mechanistic Interpretability for Vision @ CVPR2025</a> earlier this month at CVPR was very informative and fun! Looking forward to seeing this community grow.

Thank you to the speakers and organizers <a href="/trevordarrell/">trevordarrell</a> <a href="/davidbau/">David Bau</a> <a href="/TamarRottShaham/">Tamar Rott Shaham</a> <a href="/YGandelsman/">Yossi Gandelsman</a> <a href="/materzynska/">Joanna</a>
Mechanistic Interpretability for Vision @ CVPR2025 (@miv_cvpr2025) 's Twitter Profile Photo

Thank you very much to our wonderful speakers and attendees Mechanistic Interpretability for Vision @ CVPR2025 who made the workshop a huge success. We hope to see you again next year! The workshop recording link can be accessed at: youtu.be/LTh86RMAWsI?si….

Quanta Magazine (@quantamagazine) 's Twitter Profile Photo

In a recent paper, physicists used two predictable factors to reproduce the “creativity” seen from image-generating AI. Webb Wright reports: quantamagazine.org/researchers-un…

Amil Dravid (@_amildravid) 's Twitter Profile Photo

Check out the updates to our paper. There are other ways to play around with these intriguing "register neurons." Also, check out our updated set of models with test-time registers. We have VLMs too! huggingface.co/collections/am…