
Samuel Marks
@saprmarks
AI safety research @AnthropicAI. Prev postdoc in LLM interpretability with @davidbau, math PhD at @Harvard, director of technical programs at haist.ai
ID: 1712234210109587456
11-10-2023 22:30:42
382 Tweet
1,1K Followers
110 Following

𝔊𝔴𝔢𝔯𝔫
@gwern
Internet besserwisser; pedantic, mean reply guy. 𝘞𝘢𝘵𝘢𝘴𝘩𝘪 𝘬𝘪𝘯𝘪𝘯𝘢𝘳𝘪𝘮𝘢𝘴𝘶! (Follow requests ignored due to terrible UI.)
Chana
@chanamessinger
the stakes and the world and the stars Head of Video for 80,000 Hours - send me content ideas! Opinions are my own, though.
ringworm
@prawn_meat
kick damn ass and rock on. check out my shit infinitegossip.ghost.io
Dean W. Ball
@deanwball
Senior Policy Advisor for AI and Emerging Technology, White House Office of Science and Technology Policy | Strategic Advisor for AI, @NSF
Dan Hendrycks
@danhendrycks
• Center for AI Safety Director • xAI and Scale AI advisor • GELU/MMLU/MATH/HLE • PhD in AI from UC Berkeley newsletter.safe.ai aisafetybook.com
Dylan HadfieldMenell
@dhadfieldmenell
Assistant Prof @MITEECS working on value (mis)alignment in AI systems; @[email protected]; he/him
akbir.
@akbirkhan

Yonatan Belinkov
@boknilev
Assistant professor of computer science @TechnionLive. #NLProc
James Bradbury
@jekbradbury
Compute at @AnthropicAI! Previously JAX, TPUs, and LLMs at Google, MetaMind/@SFResearch, @Stanford Linguistics, @Caixin.
Michele Wang
@michelelwang
preparedness @openai
David Duvenaud
@davidduvenaud
Machine learning prof @UofT. Former team lead at Anthropic. Working on generative models, inference, & latent structure.
Séb Krier
@sebkrier
🪼 policy dev & strategy @GoogleDeepMind | rekkid junkie, dimensional glider, deep ArXiv dweller, interstellar fugitive, uncertain | 🛸
Trenton Bricken
@trentonbricken
Trying to figure out what makes minds and machines go "Beep Bop!" @AnthropicAI
Ben Edelman
@edelmanben
Thinking about how/why AI works/doesn't, and how to make it go well for us. Currently: Technical research × policy @ U.S. AI Safety Institute
Buck Shlegeris
@bshlgrs
CEO@Redwood Research (@redwood_ai), working on technical research to reduce catastrophic risk from AI misalignment. [email protected]
Jaime Sevilla
@jsevillamol
Director of @EpochAIResearch. Trying to glimpse the future of AI.
Kshitij Sachan
@sachankshitij
beep boop at @AnthropicAI
Roger Grosse
@rogergrosse

Sarah Schwettmann
@cogconfluence
Co-founder and CSO, @TransluceAI // Research Scientist, @MIT_CSAIL
Kelsey Piper
@kelseytuoc
Senior writer at Vox's Future Perfect. We're not doomed, we just have a big to-do list.
Tamar Rott Shaham
@tamarrottshaham
Postdoctoral fellow at @MIT_csail
Jack Lindsey
@jack_w_lindsey
Neuroscience of AI brains @AnthropicAI. Previously neuroscience of real brains @cu_neurotheory. Alignment McGee.
Peter Hase
@peterbhase
AI safety researcher. PhD from UNC Chapel Hill (Google PhD Fellow). Previously: Anthropic, AI2, Google, Meta
Jacob Pfau
@jacob_pfau
Alignment at UKAISI and PhD student at NYU
Leo Gao
@nabla_theta
working on AGI alignment. prev: GPT-Neo, the Pile, LM evals, RL overoptimization, scaling SAEs to GPT-4. EleutherAI cofounder.
Collin Burns
@collinburns4
@AnthropicAI. Previously @OpenAI @berkeley_ai @Columbia. Former Rubik's Cube world record holder.
John Schulman
@johnschulman2
Recently started @thinkymachines. Interested in reinforcement learning, alignment, birds, jazz music
Arthur Conmy
@arthurconmy
Aspiring 10x reverse engineer @GoogleDeepMind
Alex Mallen
@alextmallen
Redwood Research (@redwood_ai) Prev. @AiEleuther
Johannes Treutlein
@j_treutlein
AI alignment stress-testing research @AnthropicAI. On leave from my CS PhD at UC Berkeley, @CHAI_Berkeley. Opinions my own.
Julian
@mealreplacer
thinking about how to make AI go well @open_phil
Alex Pan
@aypan_17
CS PhD @UCBerkeley working on LLM safety and interpretability
Tara Rezaei
@tararezaeikh
mit ‘26, prev intern openai
Ryan Greenblatt
@ryanpgreenblatt
Chief scientist at Redwood Research (@redwood_ai), focused on technical AI safety research to reduce risks from rogue AIs
Adam Karvonen
@a_karvonen
ML Researcher, mostly focused on interpretability. I prefer email to DM.
Can Rager
@can_rager
AI Explainability | Physics
Nikola Jurkovic
@nikolaj2030

Jaden Fiotto-Kaufman
@jadenfk23
Principal Research Engineer at @ndif_team
Daniel Kokotajlo
@dkokotajlo

Tilde
@tilderesearch
Building the interpreter models to optimize AI deployments.
Elliot Glazer
@elliotglazer
Lead mathematician at Epoch AI.
Redwood Research
@redwood_ai
Pioneering threat mitigation and assessment for AI agents.
Janet Egan
@janet_e_egan
AI and National Security. Senior Fellow at CNAS. All views are my own.
Eli Rose🔸
@really_eli
'In the prison of his days, teach the free man how to praise.' 🔸givingwhatwecan.org/cause-areas
Peter Barnett
@peterbarnett_
Trying to ensure the future is bright. Researcher at @MIRIBerkeley
Dami Choi
@damichoi95
@TransluceAI / PhD student at @UofT and @VectorInst. Former Google AI Resident.
Tomek Korbak
@tomekkorbak
senior research scientist @AISecurityInst | previously @AnthropicAI @nyuniversity @SussexUni
Davis Brown
@davisbrownr
Research in science of {deep learning, AI security, safety}. PhD student at UPenn & RS at @PNNLab
Emmanuel Ameisen
@mlpowered
Interpretability/Finetuning @AnthropicAI Previously: Staff ML Engineer @stripe, Wrote BMLPA by @OReillyMedia, Head of AI at @InsightFellows, ML @Zipcar
Herbie Bradley
@herbiebradley
a generalist agent | AI research & governance | @CambridgeMLG | formerly @AISecurityInst @AiEleuther