Hide table of contents

Disclosure: I’m a technical product manager at Metaculus. Views expressed are my own.

Thanks to Nate Morrison, Ryan Beck, Christian Williams for feedback on this post, and Leonard Smith for clarifying my thinking.

Recently Arvind Narayanan and Sayash Kapoor published “AI existential risk probabilities are too unreliable to inform policy,” which argues, among other things, that forecasts on long-run, unique, or rare events, for systems that are not “purely physical,” should not inform policy. There are a lot of incongruities and fictions in this piece, but here I’ll focus on some of the reasons why I believe assigning probabilities for these kinds of events is useful in general and can indeed be useful for policy.

I argue that assigning probabilities, even to fundamentally uncertain, “unique” events, can help us: 

  1. Communicate about what we expect to happen with more nuance
  2. Reason about things that don’t have clear reference classes
  3. Put our reputation where our mouth is

Keep in mind that a big part of the value of x-risk forecasting exercises for policymakers lies in identifying the precursors to catastrophe, the scenarios to avoid. The details of the distribution of p(doom) forecasts are less important. Instead, we should ask: What is a given policymaker’s “risk appetite?” What constitutes an unacceptably high risk? What can we do that has the best chance of averting disaster? There are signals everywhere for those with the eyes to see.

Communicate about what we expect to happen with more nuance

First, everyone makes claims about the future all the time, whether or not they think of these as forecasts and whether or not the claims relate to something unprecedented, like existential risks posed by AI. Often these claims use phrases like “there is no chance” or “it’s a sure thing” or, in Narayanan and Kapoor’s case, they describe a risk as “negligible.” Different people interpret words very differently, however, which hampers our ability to communicate what we expect to happen.

Source: Heuer, R. J. (1999). Psychology of intelligence analysis. Center for the Study of Intelligence. via SAS blog post by Rick Wicklin. (Heuer asked 23 NATO officers what they thought each of these phrases meant. The shaded areas are the ranges Sherman Kent (1964) proposed that the U.S. intelligence community use.)

Probabilities represent an improvement over that kind of wishy-washy, estimative language. They help us express our beliefs with more clarity and nuance so we can ultimately have better discussions about our policy preferences. For example, maybe you and I disagree about AI x-risk: Say you think the chances are on the order of one in a million, and I’m at 10%. We agree, however, that if there were a hugely disruptive cyberattack enabled by AI that also shut down the U.S. power grid for a day, we would become more concerned! Perhaps then your AI x-risk increases to one in a thousand (a 10x update), and I move slightly up to 10.2%. In this scenario, our updates reflect our underlying forecasts, where you don’t expect (1% probability) an AI-supported cyberattack to disrupt your life in this way, and I do (80% probability). The relative changes in our forecasts illuminate this difference.[1]

Reason about things that don’t have clear reference classes

Importantly, many of the most significant risks we face do not have clear reference classes. Take nuclear escalation risk, for example, which James Acton wrote about for Metaculus last year. Per Acton, forecasting nuclear weapon use today, given the one historical use of nuclear weapons in an armed conflict and the several close calls since then, is about as hard as predicting this year’s election “with no more data than the outcomes of the contests prior to the overhaul of the U.S. electoral system in 1804.”

But, Acton is quick to say, that’s no reason not to try! Part of the value is in the exercise. We might ask ourselves (to use Acton’s example): If Ukraine were to try to take back Crimea, how would that affect our odds that the conflict takes a turn for the nuclear? If Russia were to use nuclear weapons, how would that affect our odds that the war escalates further?

That said, the Forecasting Research Institute (my previous employer) performed a follow-up study to the Existential Risk Persuasion tournament (XPT) called the Adversarial Collaboration on AI Risk, which revealed that much of the disagreement about AI x-risk in particular boiled down to fundamental differences in worldview. That’s no huge surprise nor an indictment of assigning probabilities to x-risk. Our assessment of the risk is a function of the information we have, what we’ve observed about the world, and the meaning we’ve made out of it. So of course that will vary widely from person to person, even within-group. The wisdom of the crowd is like an ensemble model, combining many people’s estimates into one.

On AI x-risk, Narayanan and Kapoor cite the fact that “we might get different numbers if the [XPT] tournament were repeated today” as a reason to distrust x-risk numbers. On the contrary, this is a feature, not a bug! Part of forecasting’s value is tracking changes over time in the crowd’s risk assessment. On the Russo-Ukrainian War, Metaculus published the question Will Russia invade Ukrainian territory before 2023 on December 11, 2021. In the following weeks, Foreign Policy wrote “Past attacks suggest Moscow probably won’t move on Ukraine” (December 15). On January 28, Biden said it was a “distinct possibility” (January 28, 2022, BBC). Al Jazeera published “No, Russia Will Not Invade Ukraine” (February 9, 2022). How should one synthesize these bits of information? It’s not at all clear.

By contrast, by the time 100 forecasters had given their input, on December 15, the community prediction was 34%. On January 28, the community prediction was 45%. On February 11, the community prediction climbed to 70% overnight. The invasion happened on February 24th.

Will Russia invade Ukrainian territory before 2023? on Metaculus. Resolved "yes" on February 24, 2022.

When we’re estimating the risk of an extinction-causing asteroid impact (or another purely physical system) in the next hundred, thousand, ten thousand years, we can probably all agree on what base rate to look at, and furthermore, our risk estimate is unlikely to change very much as we gather more information. However, while a robust case for very unlikely events like extinction-causing asteroid impacts is all well and good, how should we allocate resources between potential risks such as anticipating (and, if possible, preventing) asteroid impacts vs averting climate disaster vs mitigating risks from AI? I argue that a robust case for a very unlikely event is less decision-relevant than a rich body of speculative cases for larger-scale risks like those posed by AI.

And as new evidence rolls in — this can be benchmarks surpassed, new legislation, developments in Taiwan — we update. In this way, AI x-risk is not so different from the risk of escalation in the Russo-Ukrainian war. Registering what we expect to happen is valuable, and so is registering how we would update if something were to happen, like China invading Taiwan. How would it change your assessment of AI x-risk if China seized control of chip manufacturing in Taiwan?[2]

Put our reputation where our mouth is

Yes, people’s forecasts on x-risk run the gamut of the probability scale, but that’s to be expected! There are a whole host of things at play here, not least of which is the probability that world governments intervene. We have much to gain by homing in on the key uncertainties so that we can work toward better benchmarks for these, agree on red-lines, and imagine scenarios. Further, there is value to tracking who is consistently wrong on shorter-run AI topics.

If someone is consistently surprised by the pace of AI progress, maybe we shouldn’t trust them when they say AI is all hype. Inasmuch as Metaculus is consistently “right” (or righter than others) about the pace of AI progress, it seems like a good indicator that the community collectively understands the mechanisms at play here pretty well… i.e. that their models are pretty good!

Reference classes for AI x-risk like the extinction of less-intelligent species might seem outlandish today, but we’re already starting to see whose models predict AI advancements and consequences better. Is it more like the arrival of search engines, which save us time and don’t result in the extinction of the human race — or is it more like the advent of nuclear weapons, i.e., a threat requiring a global response and coordination? Or is it something in between?

When we’re talking about existential risks, we’ve got to be able to be more specific than “almost certainly not” or “very unlikely” or in Kapoor’s words, “not urgent or serious.” That’s what probabilities are for. If you’re someone who’s especially bullish or bearish on risks posed by AI, surely there is something that, were it to happen, would make you re-examine your view. Odds are there’s a Metaculus question for you!

Parting thoughts

Speculation is proposing outcomes, pathways, onramps to better or worse futures. Assigning probabilities to those outcomes, I argue, is different. When we lack numerical models, we don’t have to throw our hands up and… well, what would we do? We all have implicit models about what track the world is on, what’s worth worrying about, what requires intervention. Narayanan and Kapoor clearly have their own models of what’s going to happen with AI. Their model says restricting AI development would increase x-risk from AI (which invites the question, “by how much?”) and that there exist policies that are “compatible with a range of possible estimates of AI risk” (which invites the question, “what range?”).

Probabilistic predictions aren’t a substitute for scenario planning or qualitative predictions; rather, they’re a set of tools for augmenting discourse about what’s coming and what we should do about it. There is certainly a cost to policymakers’ intervening in AI development. I’m not here to take a stance on what the right course of action is. But I do believe that there is value to expressing our beliefs as probabilities and reasoning through short- and long-term risks — from asteroids to wars to AI.

  1. ^

    The Forecasting Research Institute’s report “Roots of Disagreement on AI Risk: Potential and Pitfalls of Adversarial Collaboration” identifies some such cruxes between people concerned about AI x-risk and skeptics of AI x-risk, as well as places where they’re in agreement on longer-run negative effects of AI on humans.

  2. ^

    Taiwan Semiconductor Manufacturing Company (TSMC) currently controls 61.7% of market share in the global semiconductor foundry market (Statista).

Comments3


Sorted by Click to highlight new comments since:

Executive summary: Assigning probabilities to uncertain future events, including AI existential risks, is valuable for communication, reasoning, and accountability, and can inform policy decisions despite limitations.

Key points:

  1. Probabilities allow for more nuanced communication about expectations compared to vague language.
  2. Probabilistic reasoning is useful even for events without clear reference classes, like AI risks or nuclear escalation.
  3. Tracking changes in probability estimates over time provides valuable information as new evidence emerges.
  4. Assigning probabilities helps identify key uncertainties and scenarios to avoid, which is crucial for policymakers.
  5. Probabilistic forecasts allow for accountability by tracking who consistently makes accurate predictions about AI progress.
  6. Expressing beliefs as probabilities complements other tools like scenario planning and qualitative predictions for policy discussions.

 

 

This comment was auto-generated by the EA Forum Team. Feel free to point out issues with this summary by replying to the comment, and contact us if you have feedback.

Hey Molly,

Thanks for the post! 
Definitely agree that it would be valuable to explicate our models of the world, especially numerically!

Heads up that you have a repeat paragraph immediately after the probability graph for the Russian invasion of Ukraine.

"When we’re estimating the risk of an extinction-causing asteroid impact..."

Thanks Ivan! More work on actually articulating our models of the world: incoming...! I don't think point forecasts accomplish this very well.

Curated and popular this week
 ·  · 10m read
 · 
I wrote this to try to explain the key thing going on with AI right now to a broader audience. Feedback welcome. Most people think of AI as a pattern-matching chatbot – good at writing emails, terrible at real thinking. They've missed something huge. In 2024, while many declared AI was reaching a plateau, it was actually entering a new paradigm: learning to reason using reinforcement learning. This approach isn’t limited by data, so could deliver beyond-human capabilities in coding and scientific reasoning within two years. Here's a simple introduction to how it works, and why it's the most important development that most people have missed. The new paradigm: reinforcement learning People sometimes say “chatGPT is just next token prediction on the internet”. But that’s never been quite true. Raw next token prediction produces outputs that are regularly crazy. GPT only became useful with the addition of what’s called “reinforcement learning from human feedback” (RLHF): 1. The model produces outputs 2. Humans rate those outputs for helpfulness 3. The model is adjusted in a way expected to get a higher rating A model that’s under RLHF hasn’t been trained only to predict next tokens, it’s been trained to produce whatever output is most helpful to human raters. Think of the initial large language model (LLM) as containing a foundation of knowledge and concepts. Reinforcement learning is what enables that structure to be turned to a specific end. Now AI companies are using reinforcement learning in a powerful new way – training models to reason step-by-step: 1. Show the model a problem like a math puzzle. 2. Ask it to produce a chain of reasoning to solve the problem (“chain of thought”).[1] 3. If the answer is correct, adjust the model to be more like that (“reinforcement”).[2] 4. Repeat thousands of times. Before 2023 this didn’t seem to work. If each step of reasoning is too unreliable, then the chains quickly go wrong. Without getting close to co
 ·  · 1m read
 · 
JamesÖz
 ·  · 3m read
 · 
Why it’s important to fill out this consultation The UK Government is currently consulting on allowing insects to be fed to chickens and pigs. This is worrying as the government explicitly says changes would “enable investment in the insect protein sector”. Given the likely sentience of insects (see this summary of recent research), and that median predictions estimate that 3.9 trillion insects will be killed annually by 2030, we think it’s crucial to try to limit this huge source of animal suffering.  Overview * Link to complete the consultation: HERE. You can see the context of the consultation here. * How long it takes to fill it out: 5-10 minutes (5 questions total with only 1 of them requiring a written answer) * Deadline to respond: April 1st 2025 * What else you can do: Share the consultation document far and wide!  * You can use the UK Voters for Animals GPT to help draft your responses. * If you want to hear about other high-impact ways to use your political voice to help animals, sign up for the UK Voters for Animals newsletter. There is an option to be contacted only for very time-sensitive opportunities like this one, which we expect will happen less than 6 times a year. See guidance on submitting in a Google Doc Questions and suggested responses: It is helpful to have a lot of variation between responses. As such, please feel free to add your own reasoning for your responses or, in addition to animal welfare reasons for opposing insects as feed, include non-animal welfare reasons e.g., health implications, concerns about farming intensification, or the climate implications of using insects for feed.    Question 7 on the consultation: Do you agree with allowing poultry processed animal protein in porcine feed?  Suggested response: No (up to you if you want to elaborate further).  We think it’s useful to say no to all questions in the consultation, particularly as changing these rules means that meat producers can make more profit from sel