I’ve written a draft report evaluating a version of the overall case for existential risk from misaligned AI, and taking an initial stab at quantifying the risk from this version of the threat. I’ve made the draft viewable as a public google doc here (Edit: arXiv version here, video presentation here, human-narrated audio version here). Feedback would be welcome.
This work is part of Open Philanthropy’s “Worldview Investigations” project. However, the draft reflects my personal (rough, unstable) views, not the “institutional views” of Open Philanthropy.
Though in the world where the credible range of estimates is 1-10%, and 80% of the field believed the probability were >10% (my prediction from upthread), that would start to get into 'something's seriously wrong with the field' territory from my perspective; that's not a small disagreement.
(I'm assuming here, as I did when I made my original prediction, that they aren't all clustered around 15% or whatever; rather, I'd have expected a lot of the field to give a much higher probability than 10%.)