Charlie Snell (@sea_snell) 's Twitter Profile
Charlie Snell

@sea_snell

PhD @berkeley_ai; prev SR @GoogleDeepMind. My friend told me to tweet more. I stare at my computer a lot and make things

ID: 1388361811

linkhttps://sea-snell.github.io calendar_today29-04-2013 01:24:30

4,4K Tweet

6,6K Takipçi

5,5K Takip Edilen

kalomaze (@kalomaze) 's Twitter Profile Photo

the most underused, underappreciated, and underrated trick that is basically only used by Anthropic; context distillation

Letta (@letta_ai) 's Twitter Profile Photo

We're excited to release our latest paper, “Sleep-time Compute: Beyond Inference Scaling at Test-Time”, a collaboration with Charlie Snell from UC Berkeley and Letta advisors / UC Berkeley faculty Ion Stoica and Joey Gonzalez letta.com/blog/sleep-tim…

Jiayi Pan (@jiayi_pirate) 's Twitter Profile Photo

We explore a new dimension in scaling reasoning models in Adaptive Parallel Reasoning APR lets LMs learn to orchestrate both serial & parallel compute E2E via supervised training + RL — w/ better efficiency and scalability than long CoT on Countdown 🧵 arxiv.org/abs/2504.15466

We explore a new dimension in scaling reasoning models in Adaptive Parallel Reasoning

APR lets LMs learn to orchestrate both serial & parallel compute E2E via supervised training + RL — w/ better efficiency and scalability than long CoT on Countdown

🧵 arxiv.org/abs/2504.15466
Charlie Snell (@sea_snell) 's Twitter Profile Photo

If you’re at ICLR and wondering why I just gave my talk in sweatpants this morning, it’s a long story, but it has something to do with flying in at 1AM last night and my luggage not making it to Singapore

Ruiqi Zhong (@zhongruiqi) 's Twitter Profile Photo

Last day of PhD! I pioneered using LLMs to explain dataset&model. It's used by interp at OpenAI and societal impact Anthropic Tutorial here. It's a great direction & someone should carry the torch :) Thesis available, if you wanna read my acknowledgement section=P

Last day of PhD! 

I pioneered using LLMs to explain dataset&amp;model. It's used by interp at <a href="/OpenAI/">OpenAI</a>  and societal impact <a href="/AnthropicAI/">Anthropic</a> 

Tutorial here. It's a great direction &amp; someone should carry the torch :)

Thesis available, if you wanna read my acknowledgement section=P
Cursor (@cursor_ai) 's Twitter Profile Photo

Cursor 1.0 is out now! Cursor can now review your code, remember its mistakes, and work on dozens of tasks in the background.

Dylan Patel ✈️ ICLR (@dylan522p) 's Twitter Profile Photo

RL is very inference heavy and shifts infrastructure build outs heavily Scaling well engineered environments is difficult Reward hacking and non verifiable rewards are key areas of research Recursive self improvement already playing out Major shift in o4 and o5 RL training