🚀 Rocket Cannot Be Resampled (@rocketalignment) 's Twitter Profile
🚀 Rocket Cannot Be Resampled

@rocketalignment

Writing about AI @theinformation via @Tarbell_Fellows, previously Community Manager @MATSProgram. Signal: (530) 400-4184

ID: 1778838345042350080

calendar_today12-04-2024 17:31:34

474 Tweet

525 Takipçi

600 Takip Edilen

🚀 Rocket Cannot Be Resampled (@rocketalignment) 's Twitter Profile Photo

Pack it up, everyone. They finally did it. They took the last AI pun for a newsletter name. Transformer, Gradient Updates, 3 Shot Learning (rip), now Hidden Layer. Everyone can go home. They did it. None left. It's done.

Eleos AI Research (@eleosai) 's Twitter Profile Photo

Despite flaws with model self-reports, we interviewed because: 1. Signal: catch red flags for further investigation 2. Scale: as models advance, these methods might yield better evidence 3. Precedent: if and when AIs matter morally, we should ask them how they want to be treated

Despite flaws with model self-reports, we interviewed because:

1. Signal: catch red flags for further investigation
2. Scale: as models advance, these methods might yield better evidence
3. Precedent: if and when AIs matter morally, we should ask them how they want to be treated
Junior Rojas (@junior_rojas_d) 's Twitter Profile Photo

I've been experimenting with attention mechanisms to design locomotion controllers that adapt to different shapes, this is the same controller running on two different bodies github.com/juniorrojas/mo… paper coming soon 👀

Alex Turner (@turn_trout) 's Twitter Profile Photo

Thought real machine unlearning was impossible? We show that distilling a conventionally “unlearned” model creates a model resistant to relearning attacks. 𝐃𝐢𝐬𝐭𝐢𝐥𝐥𝐚𝐭𝐢𝐨𝐧 𝐦𝐚𝐤𝐞𝐬 𝐮𝐧𝐥𝐞𝐚𝐫𝐧𝐢𝐧𝐠 𝐫𝐞𝐚𝐥.

Thought real machine unlearning was impossible? We show that distilling a conventionally “unlearned” model creates a model resistant to relearning attacks. 𝐃𝐢𝐬𝐭𝐢𝐥𝐥𝐚𝐭𝐢𝐨𝐧 𝐦𝐚𝐤𝐞𝐬 𝐮𝐧𝐥𝐞𝐚𝐫𝐧𝐢𝐧𝐠 𝐫𝐞𝐚𝐥.