This is a linkpost for a new paper called Preparing for the Intelligence Explosion, by Will MacAskill and Fin Moorhouse. It sets the high-level agenda for the sort of work that Forethought is likely to focus on.
Some of the areas in the paper that we expect to be of most interest to EA Forum or LessWrong readers are:
- Section 3 finds that even without a software feedback loop (i.e. “recursive self-improvement”), even if scaling of compute completely stops in the near term, and even if the rate of algorithmic efficiency improvements slow, then we should still expect very rapid technological development — e.g. a century’s worth of progress in a decade — once AI meaningfully substitutes for human researchers.
- A presentation, in section 4, of the sheer range of challenges that an intelligence explosion would pose, going well beyond the “standard” focuses of AI takeover risk and biorisk.
- Discussion, in section 5, of when we can and can’t use the strategy of just waiting until we have aligned superintelligence and relying on it to solve some problem.
- An overview, in section 6, of what we can do, today, to prepare for this range of challenges.
Here’s the abstract:
AI that can accelerate research could drive a century of technological progress over just a few years. During such a period, new technological or political developments will raise consequential and hard-to-reverse decisions, in rapid succession. We call these developments grand challenges.
These challenges include new weapons of mass destruction, AI-enabled autocracies, races to grab offworld resources, and digital beings worthy of moral consideration, as well as opportunities to dramatically improve quality of life and collective decision-making.
We argue that these challenges cannot always be delegated to future AI systems, and suggest things we can do today to meaningfully improve our prospects. AGI preparedness is therefore not just about ensuring that advanced AI systems are aligned: we should be preparing, now, for the disorienting range of developments an intelligence explosion would bring.
Some other quotes and comments from my notes (in addition to my main comment):
This is conservative. Why not "GPT-5"? (In which case the 100,000x efficiency gain becomes 10,000,000,000x.)
See APM section for how misaligned ASI takeover could lead to extinction. Also
brings to mind Yudkowsky's "boiling the oceans" scenario.
This is important. Something I need to read and think more about.
Why is this likely? Surely we need a Pause to be able to do this?
Expect these to be more likely to cause extinction than a good future? (Given Vulnerable World).
Yes!
It increases takeover risk(!) given lack of progress on (the needed perfect[1]) alignment and control techniques for ASI.
This whole section (the whole paper?) assumes that an intelligence explosion is inevitable. There is no mention of “pause” or “moratorium” anywhere in the paper.
We need much more of this!
Yes!
But could also just lead to Mutually Assured AI Malfunction (MAIM).
An international agreement sounds good.
This sounds like a terrible and reckless idea! Because we don’t know exactly where the thresholds are for recursive self-improvement to kick in.
Yes, unless we stop it happening (and we should!)
Problem is knowing that by the time the “if” is verified to have occurred, it could well be too late to do the “then” (e.g. once a proto-ASI has already escaped onto the internet).
Exactly! Need a moratorium now, not unworkable “if-then” commitments!
This is assuming ASI is alignable! (The whole Not just misalignment section is).
This has not been justified in the paper.
We need at least 13 9s of safety for ASI, and the best current alignment techniques aren't even getting 3 9s...
The little compute leads to much more once it has escaped!
The point is that we won't, unless we have many more 9s of reliability in terms of catching such attempts!