Jacob Hilton (@jacobhhilton) 's Twitter Profile
Jacob Hilton

@jacobhhilton

At the Alignment Research Center, formerly at OpenAI

ID: 979429388

calendar_today30-11-2012 00:19:41

102 Tweet

2,2K Followers

43 Following

METR (@metr_evals) 's Twitter Profile Photo

When will AI systems be able to carry out long projects independently? In new research, we find a kind of “Moore’s Law for AI agents”: the length of tasks that AIs can do is doubling about every 7 months.

When will AI systems be able to carry out long projects independently?

In new research, we find a kind of “Moore’s Law for AI agents”: the length of tasks that AIs can do is doubling about every 7 months.
Jacob Hilton (@jacobhhilton) 's Twitter Profile Photo

It is sad to see OpenAI's mission being reinterpreted to mean "proliferate OpenAI's products among non-profits". This is not the mission articulated in the OpenAI Charter, which it championed for years internally. It is the least onerous alternative that still says "non-profit".

Todor Markov (@todor_m_markov) 's Twitter Profile Photo

Today, myself and 11 other former OpenAI employees filed an amicus brief in the Musk v Altman case. We worked at OpenAI; we know the promises it was founded on and we’re worried that in the conversion those promises will be broken. The nonprofit needs to retain control of the

Geoffrey Irving (@geoffreyirving) 's Twitter Profile Photo

On top of the AISI-wide research agenda yesterday, we have more on the research agenda for the AISI Alignment Team specifically. See Benjamin's thread and full post for details; here I'll focus on why we should not give up on directly solving alignment, even though it is hard. 🧵

Victor Lecomte (@vclecomte) 's Twitter Profile Photo

A cute question about inner product sketching came up in our research; any leads would be appreciated! 🙂 cstheory.stackexchange.com/questions/5539…

Jacob Hilton (@jacobhhilton) 's Twitter Profile Photo

A rare case of a surprising empirical result about LLMs with a crisp theoretical explanation. Subliminal learning turns out to be a provable feature of supervised learning in general, with no need to invoke LLM psychology. (Explained in Section 6.)