Stella Biderman (@blancheminerva) 's Twitter Profile
Stella Biderman

@blancheminerva

Open source LLMs and interpretability research at @BoozAllen and @AiEleuther. My employers disown my tweets. She/her

ID: 1125849026308575239

linkhttp://www.stellabiderman.com calendar_today07-05-2019 19:44:59

11,11K Tweet

15,15K Followers

743 Following

Alisa Liu (@alisawuffles) 's Twitter Profile Photo

What do BPE tokenizers reveal about their training data?🧐 We develop an attack🗡️ that uncovers the training data mixtures📊 of commercial LLM tokenizers (incl. GPT-4o), using their ordered merge lists! Co-1⃣st Jonathan Hayase arxiv.org/abs/2407.16607 🧵⬇️

What do BPE tokenizers reveal about their training data?🧐

We develop an attack🗡️ that uncovers the training data mixtures📊 of commercial LLM tokenizers (incl. GPT-4o), using their ordered merge lists!

Co-1⃣st <a href="/JonathanHayase/">Jonathan Hayase</a>
arxiv.org/abs/2407.16607 🧵⬇️
caden (@kh4dien) 's Twitter Profile Photo

Sparse autoencoders recover a diversity of interpretable features but present an intractable problem of scale to human labelers. We build new automated pipelines to close the gap, scaling our understanding to GPT-2 and LLama-3 8b features. @goncaloSpaulo Jacob Drori Nora Belrose

Sparse autoencoders recover a diversity of interpretable features but present an intractable problem of scale to human labelers. We build new automated pipelines to close the gap, scaling our understanding to GPT-2 and LLama-3 8b features.

@goncaloSpaulo <a href="/jacobcd52/">Jacob Drori</a> <a href="/norabelrose/">Nora Belrose</a>
EleutherAI (@aieleuther) 's Twitter Profile Photo

As models become larger and more unwieldy, auto-interp methods have becoming increasingly important. We are excited to be releasing the most comprehensive auto interp library to enable wider research on this topic. github.com/EleutherAI/sae…

Stella Biderman (@blancheminerva) 's Twitter Profile Photo

Very cool paper that shows impressive performance with ternary LLMs. Discovering new papers that use EleutherAI's GPT-NeoX library in the wild is always a treat as well :D

Stella Biderman (@blancheminerva) 's Twitter Profile Photo

One of the best and least-acknowledged use cases for LLMs is in data processing. This is already making waves behind the scenes at companies and it's great to see pleias and Alexander Doria making it happen.

Stella Biderman (@blancheminerva) 's Twitter Profile Photo

If you're looking to learn about training large language models, this cookbook lead by Quentin Anthony details essential information often glossed over in papers and resources for learning.

RWKV (@rwkv_ai) 's Twitter Profile Photo

The RWKV v6 Finch lines of models are here Scaling from 1.6B all the way to 14B Pushing the boundary for an Attention-free transformer, and Multi-lingual models. Cleanly licensedm Apache 2, under The Linux Foundation Find out more from the writeup here: blog.rwkv.com/p/rwkv-v6-finc…

Nathan (@nathanhabib1011) 's Twitter Profile Photo

The Open LLM Leaderboard is now the most liked repo on all of HuggingFace 👀 - open-llm-leaderboard/open_llm_leaderboard: 11,236 - stabilityai/stable-diffusion: 10,618 - jbilcke-hf/ai-comic-factory: 7,911 - CompVis/stable-diffusion-v1-4: 6,431

Stella Biderman (@blancheminerva) 's Twitter Profile Photo

This is important: only 25% of respondents that chose an answer got it right. I suspect the rate would be lower among a random sample of AI audiences too. If ppl don't know what a tool does they won't use it correctly. And if they wrongly think it's a watermark, that's worse.

Stella Biderman (@blancheminerva) 's Twitter Profile Photo

GPT-4 can't draw a basic diagram, but by telling the model to draw it in ascii you (I suspect) bypass the diffusion model call and use something that's mostly the language model which does know what I want.

GPT-4 can't draw a basic diagram, but by telling the model to draw it in ascii you (I suspect) bypass the diffusion model call and use something that's mostly the language model which does know what I want.