
Jan Wehner
@janwehner436164
ELLIS PhD student in ML Safety @CISPA | AI Safety, Security, Interpretability
ID: 1798743386092158976
06-06-2024 15:47:13
12 Tweet
55 Followers
70 Following

Jacob Andreas
@jacobandreas
Teaching computers to read. Assoc. prof @MITEECS / @MIT_CSAIL / @NLP_MIT (he/him). lingo.csail.mit.edu web.mit.edu/jda/www
davidad 🎇
@davidad
Programme Director @ARIA_research | accelerate mathematical modelling with AI and categorical systems theory » build safe transformative AI » cancel heat death
Nathan Labenz
@labenz
AI Scout, building text-2-video @Waymark, host of The Cognitive Revolution podcast
Andrej Karpathy
@karpathy
Building @EurekaLabsAI. Previously Director of AI @ Tesla, founding team @ OpenAI, CS231n/PhD @ Stanford. I like to train large deep neural nets 🧠🤖💥
Katja Grace 🔍
@katjagrace
Thinking about whether AI will destroy the world at aiimpacts.org. DM or email for media requests. Feedback: admonymous.co/googolplex
François Chollet
@fchollet
Co-founder @ndea. Co-founder @arcprize. Creator of Keras and ARC-AGI. Author of 'Deep Learning with Python'.
Geoffrey Irving
@geoffreyirving
Chief Scientist at the UK AI Security Institute (AISI). Previously DeepMind, OpenAI, Google Brain, etc.
Lilian Weng
@lilianweng
Co-founder of Thinking Machines Lab @thinkymachines; Ex-VP, AI Safety & robotics, applied research @OpenAI; Author of Lil'Log
Rob Miles (in SF)
@robertskmiles
Explaining AI Alignment to anyone who'll stand still for long enough, on YouTube and Discord. Music, movies, microcode, and high-speed pizza delivery
NeurIPS Conference
@neuripsconf
Vancouver, Dec 10-15, 24. neurips.cc/Register/view-… Tweets to this account are not monitored. Please send feedback to [email protected].
Evan Hubinger
@evanhub
Head of Alignment Stress-Testing @AnthropicAI. Opinions my own. Previously: MIRI, OpenAI, Google, Yelp, Ripple. (he/him/his)
Chris Olah
@ch402
Reverse engineering neural networks at @AnthropicAI. Previously @distillpub, OpenAI Clarity Team, Google Brain. Personal account.
Frans Oliehoek
@faoliehoek
Associate Professor at TU Delft. Interactive learning and decision making: ML, RL, AI & MAS opinions are mine
Michael Sioutis
@msioutis
Μανθάνων μὴ κάμνε!
David Krueger
@davidskrueger
AI professor. Deep Learning, AI alignment, ethics, policy, & safety. Formerly Cambridge, Mila, Oxford, DeepMind, ElementAI, UK AISI. AI is a really big deal.
Daniel Paleka
@dpaleka
ai safety researcher | phd @CSatETH | danielpaleka.com
Pierre Bongrand
@bongrandp
AI scientist working on RNA @Harvard - Entrepreneur Prev: AthenAI (founder) / @TUDelft / @EPFL New blogpost on DL every sunday
Martin Vechev
@mvechev
Professor of Computer Science, ETH Zurich. Founder of INSAIT (insait.ai). Works on Safe/Secure AI, LLMs, Quantum. Co-founder of 6 Deep-Tech start-ups.
Luciano Siebert
@lucianosiebert
Assistant professor at Delft University of Technology, working on Responsible AI.
Miles Brundage
@miles_brundage
Independent AI policy researcher, wife guy in training, fan of cute animals and sci-fi. I have a Substack.
Robert Long
@rgblong
executive director of @eleosai
Jeffrey Ladish
@jeffladish
Applying the security mindset to everything @PalisadeAI
Joe Carlsmith
@jkcarlsmith
Philosophy, futurism, AI. Opinions my own.
Victoria Krakovna
@vkrakovna
Research scientist in AI alignment at Google DeepMind. Co-founder of Future of Life Institute @flixrisk. Views are my own and do not represent GDM or FLI.
Helen Toner
@hlntnr
AI, national security, China. Part of the founding team at @CSETGeorgetown (opinions my own). Author of Rising Tide on substack: helentoner.substack.com
Leopold Aschenbrenner
@leopoldasch
situational-awareness.ai
Lee Sharkey
@leedsharkey
Scruting matrices @ Goodfire | previously @ Apollo Research
Steven Adler
@sjgadler
ex-@OpenAI researcher & TPM (safety evaluations, AGI readiness, product safety lead, etc). Follow me on Substack: stevenadler.substack.com
AI Safety Papers
@safe_paper
Sharing the latest in AI safety and interpretability research.
Amanda Askell
@amandaaskell
Philosopher & ethicist trying to make AI be good @AnthropicAI. Personal account. All opinions come from my training data.
Daniel Kokotajlo
@dkokotajlo

ECAI2024
@ecai2024
27th European Conference on Artificial Intelligence (ECAI2024). 19-24 October 2024. Celebrating the past, inspiring the future.
Yoshua Bengio
@yoshua_bengio
Full professor at UdeM, Founder and Scientific Director of Mila, A.M. Turing Award Recipient. Working towards the safe development of AI for the benefit of all.
Annemieke Brouwer
@annemieke_br
Technical staff member at EU AI Office
Florian Tramèr
@florian_tramer
Assistant professor of computer science at ETH Zürich. Interested in Security, Privacy and Machine Learning
Geoffrey Hinton
@geoffreyhinton
deep learning
Nina
@ninapanickssery
ML @AnthropicAI. Interested in AI safety & interpretability.
Jason Wei
@_jasonwei
ai researcher @openai
Google DeepMind
@googledeepmind
We’re a team of scientists, engineers, ethicists and more, committed to solving intelligence, to advance science and benefit humanity.
OpenAI
@openai
OpenAI’s mission is to ensure that artificial general intelligence benefits all of humanity. We’re hiring: openai.com/jobs
Anthropic
@anthropicai
We're an AI safety and research company that builds reliable, interpretable, and steerable AI systems. Talk to our AI assistant Claude at Claude.ai.
FAR.AI
@farairesearch
Frontier alignment research to ensure the safe development and deployment of advanced AI systems.
Epoch AI
@epochairesearch
We are a research institute investigating the trajectory of AI for the benefit of society.
Jesse Hoogland
@jesse_hoogland
Researcher and decel working on developmental interpretability. Executive Director @ Timaeus
Marius Hobbhahn
@mariushobbhahn
CEO at Apollo Research @apolloaievals prev. ML PhD with Philipp Hennig & AI forecasting @EpochAIResearch
Nora Belrose
@norabelrose
AI, philosophy, spirituality. Blending Deleuze and Dōgen. Head of interpretability research at @AiEleuther, but tweets are my own views, not Eleuther’s.
Neel Nanda
@neelnanda5
Mechanistic Interpretability lead DeepMind. Formerly @AnthropicAI, independent. In this to reduce AI X-risk. Neural networks can be understood, let's go do it!
Tom Everitt
@tom4everitt
AGI safety researcher at @GoogleDeepMind, leading causalincentives.com switching to bsky.app/profile/tom4ev…
Cas (Stephen Casper)
@stephenlcasper
AI technical governance & risk management research. PhD Candidate @MIT_CSAIL / @MITEECS. Also at scasper.bsky.social. stephencasper.com
Apollo Research
@apolloaievals
We are an AI evals research organisation