Sarah Cogan (@sarah_cogan) 's Twitter Profile
Sarah Cogan

@sarah_cogan

existential risks are bad. I’m tall. SWE @GoogleDeepMind Frontier Safety

ID: 708098492770680832

calendar_today11-03-2016 01:13:31

457 Tweet

374 Followers

310 Following

Frances Lorenz (@frances__lorenz) 's Twitter Profile Photo

Remember when AI researchers used to be like, "this is a neural network, isn't it cool? It can say DOG or CAT if you show it a DOG or a CAT :)" and now they're like, "we're gonna make God before China does" okay

Tom Lieberum 🔸 (@lieberum_t) 's Twitter Profile Photo

Are you worried about risks from AGI and want to mitigate them? Come work with me and my colleagues! We're hiring on the AGI Safety & Alignment team (ASAT) and the Gemini Safety team! Research Engineers: boards.greenhouse.io/deepmind/jobs/… Research Scientists: boards.greenhouse.io/deepmind/jobs/…

Victoria Krakovna (@vkrakovna) 's Twitter Profile Photo

We are excited to release a short course on AGI safety! The course offers a concise and accessible introduction to AI alignment problems and our technical & governance approaches, consisting of short recorded talks and exercises (75 minutes total). deepmindsafetyresearch.medium.com/1072adb7912c

Rohin Shah (@rohinmshah) 's Twitter Profile Photo

We're hiring! Join an elite team that sets an AGI safety approach for all of Google -- both through development and implementation of the Frontier Safety Framework (FSF), and through research that enables a future stronger FSF.

We're hiring! Join an elite team that sets an AGI safety approach for all of Google -- both through development and implementation of the Frontier Safety Framework (FSF), and through research that enables a future stronger FSF.
Rohin Shah (@rohinmshah) 's Twitter Profile Photo

Just released GDM’s 100+ page approach to AGI safety & security! (Don’t worry, there’s a 10 page summary.) AGI will be transformative. It enables massive benefits, but could also pose risks. Responsible development means proactively preparing for severe harms before they arise.

Just released GDM’s 100+ page approach to AGI safety & security! (Don’t worry, there’s a 10 page summary.)

AGI will be transformative. It enables massive benefits, but could also pose risks. Responsible development means proactively preparing for severe harms before they arise.
Anca Dragan (@ancadianadragan) 's Twitter Profile Photo

Per our Frontier Safety Framework, we continue to test our models for critical capabilities. Here’s the updated model card for Gemini 2.5Pro with frontier safety evaluations + explanation of how our safety buffer / alert thresholds approach applies to 2.0, 2.5, and what’s coming.

meatball times (@meatballtimes) 's Twitter Profile Photo

has anyone stopped to ask WHY students cheat? would a buddhist monk "cheat" at meditation? would an artist "cheat" at painting? no. when process and outcomes are aligned, there's no incentive to cheat. so what's happening differently at colleges? the answer is in the article:

has anyone stopped to ask WHY students cheat? would a buddhist monk "cheat" at meditation? would an artist "cheat" at painting? no. when process and outcomes are aligned, there's no incentive to cheat. so what's happening differently at colleges? the answer is in the article: