Taylor Webb (@taylorwwebb) 's Twitter Profile
Taylor Webb

@taylorwwebb

Studying cognition in humans and machines.

ID: 921368137668362240

linkhttps://scholar.google.com/citations?user=WCmrJoQAAAAJ&hl=en calendar_today20-10-2017 13:30:57

425 Tweet

889 Followers

547 Following

Dongyu Gong (@dongyu_gong) 's Twitter Profile Photo

Introducing our new work on mechanistic intepretability of LLM cognition🤖🧠: why do Transformer-based LLMs have limited working memory capacity, as measured by N-back tasks? (1/7) openreview.net/pdf?id=dXjQgm9…

Introducing our new work on mechanistic intepretability of LLM cognition🤖🧠: why do Transformer-based LLMs have limited working memory capacity, as measured by N-back tasks? (1/7)

openreview.net/pdf?id=dXjQgm9…
Earl K. Miller (@millerlabmit) 's Twitter Profile Photo

More evidence that working memory is not persistent activity. Instead, it is dynamic on/off states with short-term synaptic plasticity. Intermittent rate coding and cue-specific ensembles support working memory nature.com/articles/s4158… #neuroscience

Laura Ruis (@lauraruis) 's Twitter Profile Photo

How do LLMs learn to reason from data? Are they ~retrieving the answers from parametric knowledge🦜? In our new preprint, we look at the pretraining data and find evidence against this: Procedural knowledge in pretraining drives LLM reasoning ⚙️🔢 🧵⬇️

How do LLMs learn to reason from data? Are they ~retrieving the answers from parametric knowledge🦜? In our new preprint, we look at the pretraining data and find evidence against this:

Procedural knowledge in pretraining drives LLM reasoning ⚙️🔢

🧵⬇️
Taylor Webb (@taylorwwebb) 's Twitter Profile Photo

It would be great to have a precise enough formulation of ‘approximate retrieval’ for this hypothesis to be rigorously tested. There is a concern that virtually any task can be characterized in this way, by appealing to a vague notion of similarity with other tasks.

Valentina Pyatkin (@valentina__py) 's Twitter Profile Photo

Open Post-Training recipes! Some of my personal highlights: 💡 We significantly scaled up our preference data! (using more than 330k preference pairs for our 70b model!) 💡 We used RL with Verifiable Rewards to improve targeted skills like math and precise instruction following

Open Post-Training recipes! 

Some of my personal highlights:
💡 We significantly scaled up our preference data! (using more than 330k preference pairs for our 70b model!)
💡 We used RL with Verifiable Rewards to improve targeted skills like math and precise instruction following
Matthias Michel (@matthiasmichel_) 's Twitter Profile Photo

In this new preprint @smfleming and I present a theory of the functions and evolution of conscious vision. This is a big project: osf.io/preprints/psya…. We'd love to get your comments!

In this new preprint @smfleming and I present a theory of the functions and evolution of conscious vision. This is a big project: osf.io/preprints/psya…. We'd love to get your comments!
Alexa R. Tartaglini (@artartaglini) 's Twitter Profile Photo

🚨 New paper at NeurIPS Conference w/ Michael Lepori! Most work on interpreting vision models focuses on concrete visual features (edges, objects). But how do models represent abstract visual relations between objects? We adapt NLP interpretability techniques for ViTs to find out! 🔍

🚨 New paper at <a href="/NeurIPSConf/">NeurIPS Conference</a>  w/ <a href="/Michael_Lepori/">Michael Lepori</a>! Most work on interpreting vision models focuses on concrete visual features (edges, objects). But how do models represent abstract visual relations between objects? We adapt NLP interpretability techniques for ViTs to find out! 🔍
Michael Lepori (@michael_lepori) 's Twitter Profile Photo

Even ducklings🐣can represent abstract visual relations. Can your favorite ViT? In our new NeurIPS Conference paper, we use mechanistic interpretability to find out!

Dylan Foster 🐢 (@canondetortugas) 's Twitter Profile Photo

Given a high-quality verifier, language model accuracy can be improved by scaling inference-time compute (e.g., w/ repeated sampling). When can we expect similar gains without an external verifier? New paper: Self-Improvement in Language Models: The Sharpening Mechanism

Given a high-quality verifier, language model accuracy can be improved by scaling inference-time compute (e.g., w/ repeated sampling). When can we expect similar gains without an external verifier? 

New paper: Self-Improvement in Language Models: The Sharpening Mechanism
François Chollet (@fchollet) 's Twitter Profile Photo

Today OpenAI announced o3, its next-gen reasoning model. We've worked with OpenAI to test it on ARC-AGI, and we believe it represents a significant breakthrough in getting AI to adapt to novel tasks. It scores 75.7% on the semi-private eval in low-compute mode (for $20 per task

Today OpenAI announced o3, its next-gen reasoning model. We've worked with OpenAI to test it on ARC-AGI, and we believe it represents a significant breakthrough in getting AI to adapt to novel tasks.

It scores 75.7% on the semi-private eval in low-compute mode (for $20 per task
Mikel Bober-Irizar (@mikb0b) 's Twitter Profile Photo

Why do pre-o3 LLMs struggle with generalization tasks like ARC Prize? It's not what you might think. OpenAI o3 shattered the ARC-AGI benchmark. But the hardest puzzles didn’t stump it because of reasoning, and this has implications for the benchmark as a whole. Analysis below🧵

Why do pre-o3 LLMs struggle with generalization tasks like <a href="/arcprize/">ARC Prize</a>? It's not what you might think.

OpenAI o3 shattered the ARC-AGI benchmark. But the hardest puzzles didn’t stump it because of reasoning, and this has implications for the benchmark as a whole.

Analysis below🧵
Stephanie Chan (@scychan_brains) 's Twitter Profile Photo

New work led by Aaditya Singh: "Strategy coopetition explains the emergence and transience of in-context learning in transformers." We find some surprising things!! E.g. that circuits can simultaneously compete AND cooperate ("coopetition") 😯 🧵👇

Mengdi Wang (@mengdiwang10) 's Twitter Profile Photo

🚨 Discover the Science of LLM! We uncover how LLMs (Llama3-70B) achieve abstract reasoning through emergent symbolic mechanisms: 1️⃣ Symbol Abstraction Heads: Early layers convert input tokens into abstract variables based on their relationships. 2️⃣ Symbolic Induction Heads:

🚨 Discover the Science of LLM! We uncover how LLMs (Llama3-70B) achieve abstract reasoning through emergent symbolic mechanisms: 

1️⃣ Symbol Abstraction Heads: Early layers convert input tokens into abstract variables based on their relationships. 
2️⃣ Symbolic Induction Heads:
Brian Odegaard (@brianodegaard2) 's Twitter Profile Photo

Led by postdoc Doyeon Lee and grad student Joseph Pruitt, our lab has a new Perspectives piece in PNAS Nexus: "Metacognitive sensitivity: The key to calibrating trust and optimal decision-making with AI" academic.oup.com/pnasnexus/arti… With co-authors Tianyu Zhou and Eric Du 1/

Raphaël Millière (@raphaelmilliere) 's Twitter Profile Photo

The final version of this paper has now been published in open access in the Journal of Memory and Language (link below). This was a long-running but very rewarding project. Here are a few thoughts on our methodology and main findings. 1/9

The final version of this paper has now been published in open access in the Journal of Memory and Language (link below). This was a long-running but very rewarding project. Here are a few thoughts on our methodology and main findings. 1/9
Tom McCoy (@rtommccoy) 's Twitter Profile Photo

🤖🧠 NEW PAPER ON COGSCI & AI 🧠🤖 Recent neural networks capture properties long thought to require symbols: compositionality, productivity, rapid learning So what role should symbols play in theories of the mind? For our answer...read on! Paper: arxiv.org/abs/2508.05776 1/n

🤖🧠 NEW PAPER ON COGSCI &amp; AI 🧠🤖

Recent neural networks capture properties long thought to require symbols: compositionality, productivity, rapid learning

So what role should symbols play in theories of the mind? For our answer...read on!

Paper: arxiv.org/abs/2508.05776

1/n