Thomas Fel (@napoolar) 's Twitter Profile
Thomas Fel

@napoolar

Explainability, Computer Vision, Neuro-AI. Research Fellow @KempnerInst, @Harvard. Prev. @tserre lab, @Google, @GoPro. Crêpe lover.

ID: 831571293392732162

linkhttps://thomasfel.me calendar_today14-02-2017 18:30:21

1,1K Tweet

1,1K Takipçi

650 Takip Edilen

Subbarao Kambhampati (కంభంపాటి సుబ్బారావు) (@rao2z) 's Twitter Profile Photo

Do Intermediate Tokens Produced by LRMs (need to) have any semantics? Our new study "Beyond Semantics: The Unreasonable Effectiveness of Reasonless Intermediate Tokens" lead by kstechly, Karthik Valmeekam Atharva & Vardhan Palod dives into this question 🧵 1/

Do Intermediate Tokens Produced by LRMs (need to) have any semantics? Our new study "Beyond Semantics: The Unreasonable Effectiveness of Reasonless Intermediate Tokens"  lead by <a href="/kayastechly/">kstechly</a>, <a href="/karthikv792/">Karthik Valmeekam</a> <a href="/_gundawar/">Atharva</a> &amp; <a href="/PalodVardh12428/">Vardhan Palod</a> dives into this question 🧵 1/
Thomas Fel (@napoolar) 's Twitter Profile Photo

Keep stumbling on older gems doing sparse dictionary learning for concept extraction. This one’s from 2018, tldr: transform deep features into sparse convex combos of archetypes. arxiv.org/pdf/1805.11155

Keep stumbling on older gems doing sparse dictionary learning for concept extraction.
This one’s from 2018, 
tldr: transform deep features into sparse convex combos of archetypes.
arxiv.org/pdf/1805.11155
Greta Tuckute (@gretatuckute) 's Twitter Profile Photo

What are the organizing dimensions of language processing? We show that voxel responses are organized along 2 main axes: processing difficulty & meaning abstractness—revealing an interpretable, topographic representational basis for language processing shared across individuals.

Adil D. Ztn 👒 (@adilztn) 's Twitter Profile Photo

Shreyas Gite Ville 🤖 the robotics field needs an observability layer to scan models and datasets. the most convenient way would be an extension of CRAFT for multimodal inputs. huggingface.co/papers/2211.10… Thomas Fel

Bahareh Tolooshams (@btolooshams) 's Twitter Profile Photo

My research group is recruiting MSc and PhD students at the University of Alberta in Canada. Research topics include generative modeling, representation learning, interpretability, inverse problems, and neuroAI. If interested, check my website and consider applying.

Mustafa Shukor (@mustafashukor1) 's Twitter Profile Photo

The Worldwide LeRobot hackathon is in 2 weeks, and we have been cooking something for you… Introducing SmolVLA, a Vision-Language-Action model with light-weight architecture, pretrained on community datasets, with an asynchronous inference stack, to control robots🧵

The Worldwide <a href="/LeRobotHF/">LeRobot</a>  hackathon is in 2 weeks, and we have been cooking something for you… 
Introducing SmolVLA, a Vision-Language-Action model with light-weight architecture, pretrained on community datasets, with an asynchronous inference stack, to control robots🧵
Raphaël Sourty (@raphaelsrty) 's Twitter Profile Photo

I'm thrilled to announce the release of FastPlaid ! 🚀🚀 FastPlaid is a high-performance engine for multi-vector search, built from the ground up in Rust (with the help of Torch C++)⚡️ You can view FastPlaid as the counterpart of Faiss for multi vectors.

Surya Ganguli (@suryaganguli) 's Twitter Profile Photo

Looking forward to speaking today at Harvard’s Kempner Institute at Harvard University NeuroAI conference on theories of learning, creativity and reasoning. Looks like a great set of speakers. Virtual attendance is possible: kempnerinstitute.harvard.edu/frontiers-in-n…

Ekdeep Singh Lubana (@ekdeepl) 's Twitter Profile Photo

🚨 New paper alert! Linear representation hypothesis (LRH) argues concepts are encoded as **sparse sum of orthogonal directions**, motivating interpretability tools like SAEs. But what if some concepts don’t fit that mold? Would SAEs capture them? 🤔 1/11

Ekdeep Singh Lubana (@ekdeepl) 's Twitter Profile Photo

We then analyze what the hierarchical concepts look like, finding an intriguing set of **shared** concepts across modality-specific backbones to enable hierarchical structures. This stands in contrast to the “usual understanding” of disparate representations spaces in VLMs! 8/11

Kempner Institute at Harvard University (@kempnerinst) 's Twitter Profile Photo

NEW: Asma Ghandeharioun (Asma Ghandeharioun) of Google DeepMind demonstrates techniques to enhance interpretability of large language models, including the Patchscopes framework. Watch the video: youtu.be/Og8FTqUrvtA #NeuroAI2025 #AI #ML #LLMs #NeuroAI

NEW: Asma Ghandeharioun (<a href="/ghandeharioun/">Asma Ghandeharioun</a>) of <a href="/GoogleDeepMind/">Google DeepMind</a> demonstrates techniques to enhance interpretability of large language models, including the Patchscopes framework. 

Watch the video: youtu.be/Og8FTqUrvtA

#NeuroAI2025 #AI #ML #LLMs #NeuroAI
Simone Scardapane (@s_scardapane) 's Twitter Profile Photo

*Universal Sparse Autoencoders* by Harry Thasarathan Thomas Fel Matthew Kowal Kosta Derpanis They train a shared SAE latent space on several vision encoders at once, showing, e.g., how the same concept activates in different models. arxiv.org/abs/2502.03714

*Universal Sparse Autoencoders*
by <a href="/HThasarathan/">Harry Thasarathan</a> <a href="/Napoolar/">Thomas Fel</a> <a href="/MatthewKowal9/">Matthew Kowal</a> <a href="/CSProfKGD/">Kosta Derpanis</a> 

They train a shared SAE latent space on several vision encoders at once, showing, e.g., how the same concept activates in different models.

arxiv.org/abs/2502.03714
Emanuele Marconato (@ema_marconato) 's Twitter Profile Photo

🧵Why are linear properties so ubiquitous in LLM representations? We explore this question through the lens of 𝗶𝗱𝗲𝗻𝘁𝗶𝗳𝗶𝗮𝗯𝗶𝗹𝗶𝘁𝘆: “All or None: Identifiable Linear Properties of Next-token Predictors in Language Modeling” Published at #AISTATS2025🌴 1/9

🧵Why are linear properties so ubiquitous in LLM representations?

We explore this question through the lens of 𝗶𝗱𝗲𝗻𝘁𝗶𝗳𝗶𝗮𝗯𝗶𝗹𝗶𝘁𝘆:

“All or None: Identifiable Linear Properties of Next-token Predictors in Language Modeling”

Published at #AISTATS2025🌴

1/9
Kempner Institute at Harvard University (@kempnerinst) 's Twitter Profile Photo

NEW: George Alvarez (George Alvarez) of Harvard Psychology and the #KempnerInstitute shows how long-range feedback projections can enhance the alignment between #ML models and human visual processing. Watch the video: youtu.be/Ju_eD0Jwa8Q #NeuroAI20205 #AI #neuroscience #NeuroAI

NEW: George Alvarez (<a href="/grez72/">George Alvarez</a>) of <a href="/PsychHarvard/">Harvard Psychology</a> and the #KempnerInstitute shows how long-range feedback projections can enhance the alignment between #ML models and human visual processing. 

Watch the video: youtu.be/Ju_eD0Jwa8Q

#NeuroAI20205 #AI #neuroscience #NeuroAI
Isabel Papadimitriou (@isabelpapad) 's Twitter Profile Photo

Check out our ACL paper! We use shapley interactions to see which words (and phones) interact non-linearly -- what we lose when we assume linear relationships between features. Chat to Diganta in Vienna!

Kempner Institute at Harvard University (@kempnerinst) 's Twitter Profile Photo

The Kempner Institute congratulates its research fellows Isabel Papadimitriou (Isabel Papadimitriou) and Jenn Hu (Jennifer Hu) for their faculty appointments (UBC Linguistics & JHU Cognitive Science) and celebrates their innovative research. Read more here: bit.ly/448heBy #AI #LLMs