We are discussing the debate statement: "On the margin[1], it is better to work on reducing the chance of our[2] extinction than increasing the value of futures where we survive[3]". You can find more information in this post

When you vote and comment on the debate week banner, your comment will also appear here, along with a note indicating your initial vote, and your most recent vote (if your opinion has changed). 

However, you can also comment here any time throughout the week. Use this thread to respond to other people's arguments, and develop your own. 

If there are a lot of comments - consider sorting by “New” and interacting with posts that haven’t been voted or commented on yet
Also - perhaps don’t vote karma below zero for low effort submissions, we don’t want to discourage low effort takes on the banner. 

  1. ^

     ‘on the margin’ = think about where we would get the most value out of directing the next indifferent talented person, or indifferent funder.

  2. ^

     ‘our’ and 'we' = earth-originating intelligent life (i.e. we aren’t just talking about humans because most of the value in expected futures is probably in worlds where digital minds matter morally and are flourishing)

  3. ^

     Through means other than extinction risk reduction.  

25

0
0

Reactions

0
0
Comments65
Sorted by Click to highlight new comments since:
Peter Wildeford
8
1
1
43% disagree

I don't think this is as clear of a dichotomy as people think it is. A lot of global catastrophic risk doesn't come from literal extinction because human extinction is very hard. A lot of mundane work on GCR policy involves a wide variety of threat models that are not just extinction.

zdgroff
9
0
1
36% disagree

The value of the future conditional on civilization surviving seems positive to me, but not robustly so. I think the main argument for its being positive is theoretical (e.g., Spreading happiness to the stars seems little harder than just spreading), but the historical/contemporary record is ambiguous.

The value of improving the future seems more robustly positive if it is tractable. I suspect it is not that much less tractable than extinction risk work. I think a lot of AI risk satisfies this goal as well as the x-risk goal for reasons Will MacAskill gives in What We Owe the Future. Understanding, developing direct interventions for, and designing political processes for digital minds seem like plausible candidates. Some work on how to design democratic institutions in the age of AI also seems plausibly tractable enough to compete with extinction risk.

OscarD🔸
2
0
0
29% disagree

I think there are a lot of thorny definitional issues here that make this set of issues not boil down that nicely to a 1D spectrum. But overall extinction prevention will likely have a far broader coalition supporting it, while making the future large and amazing is far less popular since most people aren't very ambitious with respect to spreading flourishing through the universe, but I tentatively am.

I don't think it's valuable to ensure future moral patients exist for their own sake, and extinction risk reduction only really seems to expectably benefit humans who would otherwise die in an extinction event, which would be in the billions. An astronomical number of future moral patients could have welfare at stake if we don't go extinct, so I'd prioritize them on the basis of their numbers.

See this comment and thread.

Most people have a strong drive to perpetuate humanity. What makes EAs special is that EAs also care about others' suffering. So EAs should focus on trying to make sure the future isn't full of suffering.

Bella
7
4
0
29% agree

I agreed, but mostly because of my unfortunately-dim view of the tractability of work increasing the value of futures where we survive.

Me too Bella, see my comment above

By the way, it looks like there might be some problem with the Forum UI here, as this post has some text suggesting that, since writing this comment, I changed my mind from "29% agree" to "14% agree." But I haven't intentionally changed my vote on the top banner, or changed my mind.

That's odd - to me I'm still just seeing your first vote. I'll send this to Will to check out. 

Thanks for flagging this @Bella! There was a bug which made it update everyone's comment whenever anyone changed their vote 🤦‍♂️. This is fixed now

Matthew_Barnett
*2
0
0
71% disagree

The extinction of all Earth-originating intelligent life (including AIs) seems extremely unlikely over the next several decades. While a longtermist utilitarian framework takes even a 0.01 percentage point reduction in extinction risk quite seriously, there appear to be very few plausible ways that all intelligent life originating from Earth could go extinct in the next century. Ensuring a positive transition to artificial life seems more useful on current margins.

The extinction of all Earth-originating intelligent life (including AIs) seems extremely unlikely.

Extremely unlikely to happen... when? Surely all Earth-originating intelligent life will eventually go extinct, because the universe’s resources are finite.

In my comment I later specified "in [the] next century" though it's quite understandable if you missed that. I agree that eventual extinction of Earth-originating intelligent life (including AIs) is likely; however, I don't currently see a plausible mechanism for this to occur over time horizons that are brief by cosmological standards.

(I just edited the original comment to make this slightly clearer.)

Thanks for the clarification. I didn’t mean to be pedantic: I think these discussions are often unclear about the relevant time horizon. Even Bostrom admits (somewhere) that his earlier writing about existential risk left the timeframe unspecified (vaguely talking about "premature" extinction).

On the substantive question, I’m interested in learning more about your reasoning. To me, it seems much more likely that Earth-originating intelligence will go extinct this century than, say, in the 8973th century AD (conditional on survival up to that century). This is because it seems plausible that humanity (or its descendants) will soon develop technology with enough destructive potential to actually kill all intelligence. Then the question becomes whether they will also successfully develop the technology to protect intelligence from being so destroyed. But I don’t think there are decisive arguments for expecting the offense-defense balance to favor either defense or offense (the strongest argument for pessimism, in my view, is stated in the first paragraph of this book review). Do you deny that this technology will be developed "over time horizons that are brief by cosmological standards”? Or are you confident that our capacity to destroy will be outpaced by our capacity to prevent from destruction?

I'm optimistic about the very best value-increasing research/interventions. But in terms of what would actually be done at the margin, most work that people would do for "value-increasing" reasons would be confused/doomed, I expect (and this is less true for AI safety).

Tom Gardiner
4
2
0
57% disagree

The long-running future seems like it could well be unacceptably awful. From the perspective of a battery hen, it would seem much better that it's distant ancestors were pushed out of an ecological niche before humanity domesticated them. Throwing all our effort into X-risk mitigation without really tackling S-risks in a world of increasing lock-in across domains seems deeply unwise to me. 

I think avoiding existential risk is the most important thing. As long as we can do that and don't have some kind of lock in, then we'll have time to think about and optimize the value of the future.

Alistair Stewart
1
0
0
93% disagree

Making people happy is valuable; making happy people is probably not valuable. There is an asymmetry between suffering and happiness because it is more morally important to mitigate suffering than to create happiness.

KonradKozaczek
1
0
0
14% disagree

It seems even more important to avoid futures full of extreme suffering than to avoid extinction.

Derek Shiller
4
0
1
93% disagree

People in general, and not just longtermist altruists, have reason to be concerned with extinction. It may turn out not to be a problem or not be solvable and so the marginal impact seems questionable here. In contrast, few people are thinking about how to navigate our way to a worthwhile future. There are many places where thoughtful people might influence decisions that effectively lock us into a trajectory.

few people are thinking about how to navigate our way to a worthwhile future.


This might be true on the kinds of scales EAs are thinking about (potentially enourmous value, long time horizons) but is it not the case that many people want to steer humanity in a better direction? E.g. the Left, environmentalists, libertarians, ... ~all political movements?

I worry EAs think of this as some unique and obscure thing to think about, when it isn't.

(on the other hand, people neglect small probabilities of disastrous outcomes)

Lots of people think about how to improve the future in very traditional ways. Assuming the world keeps operating under the laws it has been for the past 50 years, how do we steer it in a better direction?

I suppose I was thinking of this in terms of taking radical changes from technology development seriously, but not in the sense of long timelines or weird sources of value. Far fewer people are thinking about how to navigate a time when AGI becomes commonplace than are thinking about how to get to that place, even though there might not be a huge window of time between them.

Charlie Harrison
4
1
0
64% disagree

If we treat digital minds like current animal livestock, the expected value of the future could be really bad. 

Manuel Allgaier
2
0
0
43% disagree

Depends, if x-risk is small (<5%) and if we expect outsized impact on the future (preventing negative value lock-ins), then the latter seems more important. I'm very unsure about both. 

I think human extinction (from ASI) is highly likely to happen, and soon, unless we stop it from being built[1]

See my comments in the Symposium for further discussion

  1. ^

    And that the ASI that wipes us out won't matter morally (to address footnote 2 on the statement)

Ozzie Gooen
2
0
0
29% agree

I have mixed feelings here. But one major practical worry I have about "increasing the value of futures" is that a lot of that looks fairly zero-sum to me. And I'm scared of getting other communities to think this way. 

If we can capture 5% more of the universe for utilitarian aims, for example, that's 5% less from others. 

I think it makes sense for a lot of this to be studied in private, but am less sure about highly public work.

Nathan Young
2
0
0
21% agree

I dunno, by how much? Seems contingent on lots of factors. 

To some extent I reject the question as not-super-action-guiding (I think that a lot of work people do has impacts on both things).

But taking it at face value, I think that AI x-risk is almost all about increasing the value of futures where "we" survive (even if all the humans die), and deserves most attention. Literal extinction of earth-originating intelligence is mostly a risk from future war, which I do think deserves some real attention, but isn't the main priority right now.

Hope I'm not misreading your comment, but I think you might have voted incorrectly, as if the scale is flipped.

[This comment is no longer endorsed by its author]Reply

I think Owen is voting correctly Robi - he disagrees that there should be more work on extinction reduction before there is more work on improving the value of the future. (to complicate this, he is understanding working on AI x-risk is mostly about increasing the value of the future, because, in his view, it isn't likely to lead to extinction). 

Apologies if the "agree" "disagree" labelling is unclear - we're thinking of ways to make it more parsable. 

working on AI x-risk is mostly about increasing the value of the future, because, in his view, it isn't likely to lead to extinction

Ah yes I get it now. Thanks!

This is right. But to add even more complication:

  • I think most AI x-risk (in expectation) doesn't lead to human extinction, but a noticeable fraction does
  • But a lot even of the fraction that leads to human extinction seems to me like it probably doesn't count as "extinction" by the standards of this question, since it still has the earth-originating intelligence which can go out and do stuff in the universe
    • However, I sort of expect people to naturally count this as "extinction"?

Since it wasn't cruxy for my rough overall position, I didn't resolve this last question before voting, although maybe it would get me to tweak my position a little.

Two major reasons/ considerations:
1- I'm unconvinced of the tractability of non-extinction-risk reducing longtermist interventions. 
2- Perhaps this is self-defeating - but I feel uncomfortable substantively shaping the future in ways that aren't merely making sure it exists. Visions of the future that I would have found un-objectionable a century ago would probably seem bad to me today. In short - this consideration is basically "moral uncertainty". I think extinction-risk reduction is, though not recommended on every moral framework, at least recommended on most. I haven't seen other ideas for shaping the future which are as widely recommended. 

niplav
3
1
0
71% ➔ 64% agree

Under moral uncertainty, many moral perspectives care much more about averting downsides than producing upsides.

Additionally, tractability is probably higher for extinction-level threats, since they are "absorptive"; decreasing the chance we end up in one gives humanity and their descendants ability to do whatever they figure out is best.

Finally, there is a meaningful sense in which working on improving the future is plagued by questions about moral progress and lock-in of values, and my intuition is that most interventions that take moral progress seriously and try to avoid lock-in boil down to working on things that are fairly equivalent to avoiding extinction. Interventions that don't take moral progress seriously instead may look like locking in current values.

I genuinely have no idea. 

Aidan Alexander
2
1
1
79% disagree

Far from convinced that continued existence at currently likely wellbeing levels is a good thing

Broadly Agree

Although I might have misunderstood and missed the point of this entire debate, so correct me if that is the case

I just don't believe  changing the future trajectory is tractable in say 50-100 years from now areas like politics, economics, AI welfare etc. I think its a pipe dream. We cannot predict technological, political and economic changes even in the medium term future. These changes may well quickly render our current efforts meaningless in 10-20 years. I think the effect of work we do now which is future focused diminishes in value exponentially to the point that in 20 years time it is probably almost meaningless. I could be convinced against this if you could show me interventions 20 years ago which were specifically future focus which are now still meaningful.

However I do believe it may be possible to reduce existential risk through direct, tangeble changes now which could prevent a disastrous one off event like a engnieered virus or AI apocalypse. Stopping a one off catastrophe though still exremely difficult seems much more realistic than influencing an overall global trajectory with so much uncertainty.

This is unless you include general global health and development and animal welfare work which directly improves the lives of beings today, and will also likely improve beings lives in the future but probably in a far more minor way than this question is addressing. If you include the effects ofstandard EA GHD and animal welfare work under the banner "Improving the value of futures where we survive" then I would probably swing my vote the other way.

Robi Rahman
2
0
0
93% agree

On the current margin, improving our odds of survival seems much more crucial to the long-term value of civilization. My reason for believing this is that there are some dangerous technologies which I expect will be invented soon, and are more likely to lead to extinction in their early years than later on. Therefore, we should currently spend more effort on ensuring survival, because we will have more time to improve the value of the future after that.

(Counterpoint: ASI is the main technology that might lead to extinction, and the period when it's invented might be equally front-loaded in terms of setting values as it is in terms of extinction risk.)

lilly
2
0
0
79% disagree

It's a tractability issue. In order for these interventions to be worth funding, they should reduce our chance of extinction not just now, but over the long term. And I just haven't seen many examples of projects that seem likely to do that.

OllieBase
2
0
0
36% agree

It seems plausible to me we might be approaching a "time of perils' where total x-risk is unacceptably high and will continue to be as we develop powerful AI systems, but might decrease later since we can use AI systems to tackle x-risk (though that seems hard and risky in its own myriad ways).

Broadly think we should still prioritise avoiding catastrophes in this phase, and bet on being able to steer later but low confidence.

finm
2
0
0
64% disagree

Partly this is because I think “extinction” as defined here is very unlikely (<<1%) to happen this century, which upper bounds the scale of the area. I think most “existential risk” work is not squarely targeted at avoiding literal extinction of all Earth-originating life.

Angelina Li
2
0
0
14% disagree

I feel extremely unsure about this one. I'm voting slightly against purely from the perspective of, "wow, there are projects in that direction that feel super neglected".

Matrice Jacobine
2
0
0
1
36% ➔ 29% disagree

Recent advances in LLMs have led me to update toward believing that we live in the world where alignment is easy (i.e. CEV naturally emerge from LLMs, and future AI agents will be based on understanding and following natural language commands by default), but governance is hard (i.e. AI agents might be co-opted by governments or corporations to lock in humanity in a dystopian future, and the current geopolitical environment, characterized by democratic backsliding, cold war mongering, and an increase in military conflicts including wars of aggression, isn't conducive to robust multilateral governance).

I really don't have a strong view, but I find myself sympathetic to the idea that the world is not great and is getting worse (if you consider non-human animals).

JackM
2
0
1
36% disagree

This is a question I could easily change my mind on.

The experience of digital minds seems to dominate far future calculations. We can get a lot of value from this, a lot of disvalue, or anything in between.

If we go extinct then we get 0 value from digital minds. This seems bad, but we also avoid the futures where we create them and they suffer. It’s hard to say if we are on track to creating them to flourish or suffer - I think there are arguments on both sides. The futures where we create digital minds may be the ones where we wanted to “use” them, which could mean them suffering. Alternatively, we have seen our moral circle expand over time and this may continue, so there is a real possibility we could create them to flourish. I don’t have a clear view which side wins here, so overall going extinct doesn’t seem obviously terrible to me.

We could instead focus on raising the moral status of digital minds, our ability to understand sentience and consciousness, improve societal values, making sure AI goes well and helps us with these things. These robustly increase the expected value of digital sentience in futures where we survive. 

So because reducing extinction risk is close to 0 expected value to me, and increasing the value of futures where we survive is robustly positive in expected value, I lean towards increasing the value of futures where we survive.

Tejas Subramaniam
2
0
0
21% ➔ 7% agree

I think the expected value of the long-term future, in the “business as usual” scenario, is positive. In particular, I anticipate that advanced/transformative artificial intelligence drives technological innovation to solve a lot of world problems (e.g., helping create cell-based meat eventually), and I also think a decent amount of this EV is contained in futures with digital minds and/or space colonization (even though I’d guess it’s unlikely we get to that sort of world). However, I’m very uncertain about these futures—they could just as easily contain a large amount of suffering. And if we don’t get to those futures, I’m worried about wild animal suffering being high in the meantime. Separately, I’m not sure addressing a lot of s-risk scenarios right now is particularly tractable (nor, more imminently, does wild animal suffering seem awfully tractable to me). 

Probably the biggest reason I’m so close to the center is I think a significant amount of existential risk from AI looks like disempowering humanity without killing literally every human, and hence, I view AI alignment work as at least partially serving the goal of “increasing the value of futures where we survive.”

JoA🔸
2
1
0
43% disagree

This is a difficult one, and both my thoughts and my justifications (especially the few sources I cite) are very incomplete. 

It seems to me for now that existential risk reduction is likely to be negative, as both human and AI-controlled futures could contain immense orders of magnitude more suffering than the current world (and technological developments could also enable more intense suffering, whether in humans or in digital minds). The most salient ethical problems with the extinction of earth-originating intelligent life seem to be the likelihood of biological suffering continuing on earth for millions of years (though it's not clear to me whether it would be more or less intense without intelligent life on earth), and the possibility of space (and eventually earth) being colonized by aliens (though whether their values will be better or worse remains an open question in my view).

Another point (which I'm not certain about how to weigh in my considerations) is that certain extinction events could massively reduce suffering on earth, by preventing digital sentience or even by causing the end of biological sentient life (this seems unlikely, and I've asked here how likely or unlikely EAs thought this was).

However, I am very uncertain of the tractability of improving future outcomes, especially considering recent posts by researchers at the Center on Long-Term Risk, or this one by a former researcher there, highlighting how uncertain it is that we are well-placed to improve the future. Nonetheless, I think that efforts made to improve the future, like the work of the Center for Reducing Suffering, the Center on Long-Term Risk, or the Sentience Institute, advocate for important values and could have some positive flow-through effects in the medium-term (though I don't necessarily think that this robustly improves the longer term future). I will note, however, that I am biased since work related the Center for Reducing Suffering was the primary reason I got into EA.

I am very open to changing my mind on this, but for now I'm under 50% agree because it seems to me that, in short:

  1. Extinction Risk Reduction could very well have negative expected value.
  2. Efforts to improve the value of futures where we survive might have some moderate positive effects in the short-term.

Lots of uncertainties. I expect to have moved my cursor before the end of the week!

Intelligence is the only chance of some redemption to the massive suffering probably associated to the emergence of consciousness. 

This is the age of danger, we are the first species on Earth that has figured out morality, so we shall survive at allmost all cost. 

We could devote 100% of currently available resources to existential risk reduction, live in austerity, and never be finished ensuring our own survival.  However, if increase the value of futures where we survive, we will develop more and more resources that can then be put to existential risk reduction.  People will be not only happier, but also more capable and skilled, when we create a world where people can thrive rather than just survive.  The highest quality futures are the most robust.  

Mjreard
1
0
0
21% agree

I think people overrate how predictable the effect of our actions on the future will be (even though they rate it very low in absolute terms); extinction seem like one of the very few (only?) things that seems like its effects will endure throughout a big part of the future. Still buy the theory that 0-1% of possible value is equally valuable to 98-99%; just about tractability

A higher value future reduces the chances of extinction.  If people value life, they will figure out how to keep it.

Patrick Hoang
1
0
0
50% ➔ 57% disagree

I think the most likely outcome is not necessarily extinction (I estimate <10% due to AI) but rather an unfulfilled potential. This may be humans simply losing control over the future and becoming mere spectators and AI not being morally significant in some way.

With long timeline and less than 10% probability: Hot take is these are co-dependent - prioritizing only extinction is not feasible. Additionally, does only one human exist while all others die count as non-extinction? What about only a group of humans survive? How should this be selected? It could dangerously/quickly fall back to Fascism. It would only likely benefit the group of people with current low to no suffering risks, which unfortunately correlates to the most wealthy group. When we are "dimension-reducing" the human race to one single point, we ignore the individuals. This to me goes against the intuition of altruism. 

I fundamentally disagree with the winner-take-all type of cause prioritization - instead, allocate resources to each area, and unfortunately there might be multiple battles to fight.

John Salter
2
1
5
29% disagree

The far future, on our current trajectory, seems net negative on average. Reducing extinction risk just multiplies its negative EV. 

I find it very difficult to determine whether the future will be net-negative or net-positive (when considering humans, factory-farmed animals, wild animals, and possibly artificial sentience). 
This makes it very hard to know whether work on extinction reduction is likely to be positive or not.
I prefer to work on things that aim to move the sign towards "net-positive".

What is the point of securing a future for humanity if that future is net-negative?

jboscw
1
0
0
93% agree

Very worried about AI risk, think short timelines are plausible

tylermjohn
1
0
0
36% disagree

I'm compressing two major dimensions into one here:

  • Expected value (I think EV of maxevas is vastly higher than EV of maxipok)
  • Robustness (the case for maxevas is highly empirically theory laden and values dependent)

What is maxevas? Couldn't find anything relevant by googling.

We can't work on increasing value if we are dead.

First and foremost, I'm low confidence here. 

I will focus on x-risk from AI and I will challenge the premise of this being the right way to ask the question.

What is the difference between x-risk and s-risk/increasing the value of futures? When we mention x-risk with regards to AI we think of humans going extinct but I believe that to be a shortform for wise compassionate decision making. (at least in the EA sphere) 

Personally, I think that x-risk and good decision making in terms of moral value might be coupled to each other. We can think of our current governance conditions a bit like correction systems for individual errors. If they pile up, we go off the rail and increase x-risk as well as chances of a bad future. 

So a good decision making system should both account for x-risk and value estimation, therefore the solution is the same and it is a false dichotomy? 

(I might be wrong and I appreciate the slider question anyway!)

I've heard this argument before, but I find it un-compelling in its tractability. If we don't go extinct, its likely to be a silent victory; most humans on the planet won't even realise it happened. Individual humans working on X-risk reduction will probably only impact the morals of people around them.

If we avoid extinction, plenty of people will have the time to take care of humanity's future. I'll leave it to them. Both topic have a lot of common ground anyway, like "not messing up with the biosphere" or "keeping control of ASI"

[comment deleted]4
0
0
Curated and popular this week
Relevant opportunities