
Michal Golovanevsky
@michalgolov
CS PhD student @BrownCSDept | Multimodal Learning | Mechanistic Interpretability | Clinical Deep Learning.
ID: 1573399875278049280
https://github.com/michalg04 23-09-2022 19:52:33
22 Tweet
32 Followers
42 Following







The finding that important attention heads implement one of a small set of interpretable functions boosts transparency and trust in VLMs. Michal Golovanevsky Vedant Palit #nlp #mechinterp Paper: export.arxiv.org/pdf/2406.16320 GitHub: github.com/wrudman/NOTICE… [5/5]






How do VLMs balance visual information presented in-context with linguistic priors encoded in-weights? In this project, Michal Golovanevsky and William Rudman find out! My favorite result: you can find a vector that shifts attention to image tokens and changes the VLM's response!