Neil Rathi (@neil_rathi) 's Twitter Profile
Neil Rathi

@neil_rathi

humans & machines @stanfordnlp he/him

ID: 1403759668510351364

linkhttps://www.stanford.edu/~rathi calendar_today12-06-2021 17:02:45

26 Tweet

210 Followers

131 Following

Neil Rathi (@neil_rathi) 's Twitter Profile Photo

really basic naive q but how does distributional semantics handle function words ?? like i cannot imagine them having meaningful vector representations

Neil Rathi (@neil_rathi) 's Twitter Profile Photo

postering at #blackboxnlp this thursday in-person!! talking about memory effects in production and abstract "soft" ordering preferences, i'd love to chat about this or anything nlp/cognition/psycholinguistics if you'll also be there!

Badr AlKhamissi (@bkhmsi) 's Twitter Profile Photo

I am at #CCN2024 CogCompNeuro, will be presenting our work "Brain-Like Language Processing via a Shallow Untrained Multihead Attention Network" Poster C57 on Friday! Joint work w/ Greta Tuckute, Antoine Bosselut and Martin Schrimpf Looking forward to all the discussions! :D

I am at #CCN2024 <a href="/CogCompNeuro/">CogCompNeuro</a>, will be presenting our work "Brain-Like Language Processing via a Shallow Untrained Multihead Attention Network" 

Poster C57 on Friday!

Joint work w/ <a href="/GretaTuckute/">Greta Tuckute</a>, <a href="/ABosselut/">Antoine Bosselut</a> and <a href="/martin_schrimpf/">Martin Schrimpf</a> 

Looking forward to all the discussions! :D
Richard Futrell (@rljfutrell) 's Twitter Profile Photo

Language Models learn a lot about language, much more than we expected, without much built-in structure. This matters for linguistics and opens up enormous opportunities. So should we just throw out linguistics? No! Quite the opposite: we need theory and structure.

Language Models learn a lot about language, much more than we expected, without much built-in structure. This matters for linguistics and opens up enormous opportunities. So should we just throw out linguistics? No! Quite the opposite: we need theory and structure.
Aryaman Arora (@aryaman2020) 's Twitter Profile Photo

new paper! 🫡 we introduce 🪓AxBench, a scalable benchmark that evaluates interpretability techniques on two axes: concept detection and model steering. we find that: 🥇prompting and finetuning are still best 🥈supervised interp methods are effective 😮SAEs lag behind

new paper! 🫡

we introduce  🪓AxBench, a scalable benchmark that evaluates interpretability techniques on two axes: concept detection and model steering.

we find that:
🥇prompting and finetuning are still best
🥈supervised interp methods are effective
😮SAEs lag behind
Neil Rathi (@neil_rathi) 's Twitter Profile Photo

i'm at ICLR this week presenting TopoLM as an oral(!) reach out if you want to chat about anything humans + machines (cogsci, hci, fairness, safety, interp, etc.)

Neil Rathi (@neil_rathi) 's Twitter Profile Photo

presenting our work on inducing brain-like topography in Transformers today 2x w/ Hannes Mehrer! come by to talk neuro, interp, etc. talk — Session 4C, 4:15pm [Garnet 216-218] poster — Session 3, 10am [Hall 2B #599]