
Lewis Ho
@_lewisho
Research Scientist at Google DeepMind
ID: 877336277330165760
21-06-2017 01:24:03
38 Tweet
232 Takipçi
165 Takip Edilen




In 2024, the AI community will develop more capable AI systems than ever before. How do we know what new risks to protect against, and what the stakes are? Our research team at Google DeepMind built a set of evaluations to measure potentially dangerous capabilities: 🧵



Curious about how we evaluate dangerous capabilities at Google DeepMind? 🤔 The Frontier Safety team just open-sourced resources for our in-house CTF & self-proliferation challenges! Check it out: github.com/google-deepmin…

We are hiring! Google DeepMind's Frontier Safety and Governance team is dedicated to mitigating frontier AI risks; we work closely with technical safety, policy, responsibility, security, and GDM leadership. Please encourage great people to apply! 1/ boards.greenhouse.io/deepmind/jobs/…




I'm proud of GoogleDeepMind/Google's v2 update to our Frontier Safety Framework. We were the first major tech company to produce an explicit risk management framework for extreme risks, and I'm glad we are continuing to push ahead on safety best practice. deepmind.google/discover/blog/…

We are excited to release a short course on AGI safety! The course offers a concise and accessible introduction to AI alignment problems and our technical & governance approaches, consisting of short recorded talks and exercises (75 minutes total). deepmindsafetyresearch.medium.com/1072adb7912c




