Ajeya Cotra (@ajeya_cotra) 's Twitter Profile
Ajeya Cotra

@ajeya_cotra

AI could get really powerful soon and I worry we're underprepared. Funding research on AI capabilities+risks @open_phil (views my own), editor @plannedobs.

ID: 917512572965761024

linkhttp://planned-obsolescence.org calendar_today09-10-2017 22:10:18

348 Tweet

7,7K Followers

363 Following

ARC Prize (@arcprize) 's Twitter Profile Photo

We put OpenAI o1 to the test against ARC Prize. Results: both o1 models beat GPT-4o. And o1-preview is on par with Claude 3.5 Sonnet. Can chain-of-thought scale to AGI? What explains o1's modest scores on ARC-AGI? Our notes: arcprize.org/blog/openai-o1…

We put OpenAI o1 to the test against ARC Prize.

Results: both o1 models beat GPT-4o. And o1-preview is on par with Claude 3.5 Sonnet.

Can chain-of-thought scale to AGI? What explains o1's modest scores on ARC-AGI?

Our notes:
arcprize.org/blog/openai-o1…
Ethan Mollick (@emollick) 's Twitter Profile Photo

In this post, I tried to lay out, in a non-technical way, where we are in the development of AI, the state of the top five AI models available now, & what the near future of AI is going to be The short answer is that LLMs are going to keep getting better. oneusefulthing.org/p/scaling-the-…

Ajeya Cotra (@ajeya_cotra) 's Twitter Profile Photo

By 2026, will some companies allow AI agents to autonomously spend at least $10K at a time on behalf of the company, without any human review (though possibly with AI review)?

Max Nadeau (@maxnadeau_) 's Twitter Profile Photo

Update: Human baseline now exists! 1st-try acc is 48% on the pub eval set: arxiv.org/abs/2409.01374 AI pass@1 is somewhere in the 40s: x.com/mikeknoop/stat… So AI ≈ humans! But human=64% & AI=50s if given 3 tries. Still: A far cry from humans >> AIs. And scaling trends!

Ethan Mollick (@emollick) 's Twitter Profile Photo

After spending a decent amount of time with o1-preview, I would be very surprised if it is not able to do economically valuable analytical work inside large companies. The main issue is that prompting it remains really weird. But a real R&D effort inside firms might crack that.

Ajeya Cotra (@ajeya_cotra) 's Twitter Profile Photo

We'll see some signs of deceptive capabilities before it's unrecoverable (I'm excited about research on this e.g. anthropic.com/research/rewar…) but "sure we'll see it coming from far away and have plenty of time to stop it" is overconfident IMO.

Garrison Lovely (@garrisonlovely) 's Twitter Profile Photo

OpenAI whistleblower William Saunders is testifying before a Senate subcommittee today (so is Helen Toner and Margaret Mitchell). His written testimony is online now. Here are the most important parts 🧵

OpenAI whistleblower William Saunders is testifying before a Senate subcommittee today (so is Helen Toner and Margaret Mitchell). His written testimony is online now. Here are the most important parts 🧵
Ethan Perez (@ethanjperez) 's Twitter Profile Photo

I’m taking applications for collaborators via ML Alignment & Theory Scholars! It’s a great way for new or experienced researchers outside AI safety research labs to work with me/others in these groups: Neel Nanda, Evan Hubinger, mrinank 💗 enjoying solitude, Nina, Fabien Roger, Rylan Schaeffer, ...🧵