Rory Greig (@rorygreig1) 's Twitter Profile
Rory Greig

@rorygreig1

Research Engineer at Google DeepMind, interested in AI Alignment and Complexity Science.

ID: 2654227392

calendar_today17-07-2014 16:32:02

1,1K Tweet

605 Takipçi

4,4K Takip Edilen

Séb Krier (@sebkrier) 's Twitter Profile Photo

Great talk by Wolf Barfuss for Cooperative AI Foundation - there's a lot of value in bringing together the machine leaning community and the complex systems/multi agent econ fields, and this is currently heavily neglected. youtu.be/gHR6xv3xiqE?si…

Great talk by Wolf Barfuss for <a href="/coop_ai/">Cooperative AI Foundation</a> - there's a lot of value in bringing together the machine leaning community and the complex systems/multi agent econ fields, and this is currently heavily neglected.
youtu.be/gHR6xv3xiqE?si…
Google DeepMind (@googledeepmind) 's Twitter Profile Photo

AI systems can be powerful but opaque "black boxes" - even to researchers who train them. ⬛ Enter Gemma Scope: a set of open tools made up of sparse autoencoders to help decode the inner workings of Gemma 2 models, and better address safety issues. → dpmd.ai/gemma-scope

Max Roser (@maxcroser) 's Twitter Profile Photo

The big book by Doyne Farmer is out. I wrote this blurb for it. If you are interested in humanity's impact on the environment, economics, and technology I very much recommend the book.

The big book by Doyne Farmer is out.

I wrote this blurb for it. If you are interested in humanity's impact on the environment, economics, and technology I very much recommend the book.
Sam Bowman (@s8mb) 's Twitter Profile Photo

Foundations: Why Britain Has Stagnated. A new essay by Ben Southwood, Samuel Hughes & me. Why the UK's ban on investment in housing, infrastructure and energy is not just a problem. It is *the* problem. And how fixing it is the defining task of our generation. ukfoundations.co

Matt Clifford (@matthewclifford) 's Twitter Profile Photo

I’ve been saying for a while that stagnation is a policy choice: the UK can choose to be rich! Sam Bowman Ben Southwood + Samuel Hughes lay out the case in brilliant, horrifying detail. Read it, send it to your MP and ask them what they’re going to do about it

I’ve been saying for a while that stagnation is a policy choice: the UK can choose to be rich! <a href="/s8mb/">Sam Bowman</a> <a href="/bswud/">Ben Southwood</a> + <a href="/SCP_Hughes/">Samuel Hughes</a> lay out the case in brilliant, horrifying detail. 

Read it, send it to your MP and ask them what they’re going to do about it
Benjamin Hilton (@benjamin_hilton) 's Twitter Profile Photo

At @AISafetyInst we're researching 'safety cases' - careful arguments that AI systems are safe. Join my new safety cases policy team! You'll write strategy, engage labs and governments, fund cost-effective projects – and get stuck into the research, no ML experience required.

Miles Brundage (@miles_brundage) 's Twitter Profile Photo

With very rare exceptions, the economics profession is completely discrediting itself by either ignoring or badly misunderstanding AI

Anca Dragan (@ancadianadragan) 's Twitter Profile Photo

When I joined Google DeepMind last year, I came across this incredible group of people working on deliberative alignment, and managed to convince them to join my team in a quest to account for viewpoint and value pluralism in AI. Their Science paper is on AI-assisted deliberation

Hayden (@mallocmyheart) 's Twitter Profile Photo

cope in 2022: "it can't even draw hands" cope in 2024: "it can't even maintain coherence over medium to long time frames when simulating complex environments"

Alex Beal 🆎 (@beala) 's Twitter Profile Photo

You, a normie: I have no idea when to use o3-mini-high or o1. Me, a connoisseur: the o3-mini-high is really more of a sativa. Very cerebral. Gets those creative juices flowing. o1 is more of a body high. Couch locked, contemplating the universe.

Rohin Shah (@rohinmshah) 's Twitter Profile Photo

We're looking for strong ML researchers and software engineers. You *don't* need to be an expert on AGI safety; we're happy to train you. Learn more: alignmentforum.org/posts/wqz5CRzq… Research Engineer role: boards.greenhouse.io/deepmind/jobs/… Research Scientist role: boards.greenhouse.io/deepmind/jobs/…

Judd Rosenblatt — d/acc (@juddrosenblatt) 's Twitter Profile Photo

Turns out that Self-Other Overlap (SOO) fine-tuning drastically reduces deceptive behavior in language models—without sacrificing performance. SOO aligns an AI’s internal representations of itself and others. We think this could be crucial for AI alignment...🧵

Turns out that Self-Other Overlap (SOO) fine-tuning drastically reduces deceptive behavior in language models—without sacrificing performance.

SOO aligns an AI’s internal representations of itself and others. 

We think this could be crucial for AI alignment...🧵
Séb Krier (@sebkrier) 's Twitter Profile Photo

Excited to share Google DeepMind's AGI safety and security strategy to tackle risks like misuse and misalignment. Rather than high-level principles, this 145-page paper outlines a concrete, defense-in-depth technical approach: proactively evaluating & restricting dangerous

Excited to share <a href="/GoogleDeepMind/">Google DeepMind</a>'s AGI safety and security strategy to tackle risks like misuse and misalignment. Rather than high-level principles, this 145-page paper outlines a concrete, defense-in-depth technical approach: proactively evaluating &amp; restricting dangerous
Rohin Shah (@rohinmshah) 's Twitter Profile Photo

Just released GDM’s 100+ page approach to AGI safety & security! (Don’t worry, there’s a 10 page summary.) AGI will be transformative. It enables massive benefits, but could also pose risks. Responsible development means proactively preparing for severe harms before they arise.

Just released GDM’s 100+ page approach to AGI safety &amp; security! (Don’t worry, there’s a 10 page summary.)

AGI will be transformative. It enables massive benefits, but could also pose risks. Responsible development means proactively preparing for severe harms before they arise.
Zvi Mowshowitz (@thezvi) 's Twitter Profile Photo

It's going to be a while before I can properly write this one up, and I certainly have plenty of disagreements here, but this is a very, very good paper. It is The Way.

Demis Hassabis (@demishassabis) 's Twitter Profile Photo

Simulations are the future, & one of the main tools we’ll ultimately use to understand and predict things about the universe. This is why I’m so excited about Genie 3, our latest interactive world simulator - here are some insanely cool things you might have missed about it 🧵:

kache (@yacinemtb) 's Twitter Profile Photo

when an AI researcher tells you they're "working all day" they're not really working. they just take their laptop to the bathroom to check how their runs are doing at 2 am but its not actually work. just a severe gambling addiction