I have been busy interviewing AI safety experts about their "big picture" of the existential AI risk landscape: what's going to happen with AI, how might things go wrong, what should the AI safety community be doing, and how does their work fit in with that.

In the end I will have responses from around 20 experts, from DeepMind, Anthropic, Redwood, FAR AI, and others.

Some examples of questions I asked were:

  • what will the first human-level AI look like?
  • how could AI bring about an existential catastrophe?
  • what are the most promising research directions in AI safety?
  • what are the biggest mistakes the AI safety community has made?

The results will be published in the form of a post or sequence summarizing the results, probably transcripts of the conversations, and maybe audio recordings (would people be interested in transcripts and/or recordings?)

Pre-register your predictions of the most common responses to these questions here, here and here!

This work was funded by FAR AI.




No comments on this post yet.
Be the first to respond.
Curated and popular this week
Relevant opportunities