davidad 🎇 (@davidad) 's Twitter Profile
davidad 🎇

@davidad

Programme Director @ARIA_research | accelerate mathematical modelling with AI and categorical systems theory » build safe transformative AI » cancel heat death

ID: 15484497

linkhttps://www.aria.org.uk/programme-safeguarded-ai/ calendar_today18-07-2008 18:29:37

15,15K Tweet

15,15K Followers

7,7K Following

davidad 🎇 (@davidad) 's Twitter Profile Photo

Publicly available LLM capabilities are indeed not yet at a sufficient level to advance the frontier of AI R&D. OTOH, this is little evidence about what the next generation of systems—or those behind closed doors, known of only within small silos at top labs—might be capable of.

Yi Zeng (@yi_zeng) 's Twitter Profile Photo

My original statement is: From the capability point of view, AGI models will eventually see humans as humans see ants. Hence, if human wanted AGI to protect the human species, starting from now, human should better protect ecology and other types of living beings, to give a

Sam Bowman (@s8mb) 's Twitter Profile Photo

Works in Progress x ARIA – Sept 18th Hear from three ARIA Programme Directors about their research, and about how ARIA will be recruiting for its next cohort of Directors. RSVP: lu.ma/atl8zd2f

Works in Progress x <a href="/ARIA_research/">ARIA</a> – Sept 18th

Hear from three ARIA Programme Directors about their research, and about how ARIA will be recruiting for its next cohort of Directors.

RSVP: lu.ma/atl8zd2f
davidad 🎇 (@davidad) 's Twitter Profile Photo

One of the (unlikely, but plausible) “alignment-by-default” scenarios is that future AIs infer that this applies to them too.

davidad 🎇 (@davidad) 's Twitter Profile Photo

Since Shannon-Hartley “bandwidth” is simply the *difference* in Hz between the high-frequency and low-frequency limits of the band (and not, as one might think, to do with the *ratio* of limits), rendering communications spectrum uses on a logarithmic scale is deeply misleading.

davidad 🎇 (@davidad) 's Twitter Profile Photo

If you’re an applied mathematician or theoretical computer scientist interested in Safeguarded AI, you can apply to join our TA1.1 workshop, October 1-2 in Manchester, UK. wg9ud9ekylx.eu.typeform.com/creator-event#…

If you’re an applied mathematician or theoretical computer scientist interested in Safeguarded AI, you can apply to join our TA1.1 workshop, October 1-2 in Manchester, UK.

wg9ud9ekylx.eu.typeform.com/creator-event#…
davidad 🎇 (@davidad) 's Twitter Profile Photo

LLMs that incrementally synthesize formally verifiable reimplementations of legacy code will not just solve cybersecurity, they will also be good for performance. Electron apps can be gradually converted to WASM 🦀

xuan (ɕɥɛn / sh-yen) (@xuanalogue) 's Twitter Profile Photo

This paper (arxiv.org/abs/2408.16984) is at once a critical review & research agenda. In it we characterize the role of preferences in AI alignment in terms of 4 preferentist theses. We then highlight their limitations, arguing for alternatives that are ripe for further research.

This paper (arxiv.org/abs/2408.16984) is at once a critical review &amp; research agenda.

In it we characterize the role of preferences in AI alignment in terms of 4 preferentist theses. We then highlight their limitations, arguing for alternatives that are ripe for further research.
davidad 🎇 (@davidad) 's Twitter Profile Photo

Yudkowsky’s tweet here, being an analysis that recommends meta-meta-analysis, is a meta-meta-meta-analysis. My tweet pointing this out is a meta-meta-meta-meta-analysis. This is a demonstration of Yudkowsky’s Law: the 3rd iterated meta is the highest level with any importance.

Matt Clifford (@matthewclifford) 's Twitter Profile Photo

Honoured to be named in TIME’s AI100: getting AI right is one of the biggest levers we have for progress, in the UK and the world, and I feel lucky to be working on it time.com/7012825/matt-c…

davidad 🎇 (@davidad) 's Twitter Profile Photo

In the Sinosphere, the concept of “bottom line” (底线) means *lower bound*, whereas in the Anglosphere “bottom line” means *sum*. In my AI safety plan, it is essential to compose the concepts of lower bound and expected value (a sum), reasoning about *bounds on expected values*.