We are discussing the debate statement: "On the margin[1], it is better to work on reducing the chance of our[2] extinction than increasing the value of futures where we survive[3]". You can find more information in this post

When you vote and comment on the debate week banner, your comment will also appear here, along with a note indicating your initial vote, and your most recent vote (if your opinion has changed). 

However, you can also comment here any time throughout the week. Use this thread to respond to other people's arguments, and develop your own. 

If there are a lot of comments - consider sorting by “New” and interacting with posts that haven’t been voted or commented on yet
Also - perhaps don’t vote karma below zero for low effort submissions, we don’t want to discourage low effort takes on the banner. 

  1. ^

     ‘on the margin’ = think about where we would get the most value out of directing the next indifferent talented person, or indifferent funder.

  2. ^

     ‘our’ and 'we' = earth-originating intelligent life (i.e. we aren’t just talking about humans because most of the value in expected futures is probably in worlds where digital minds matter morally and are flourishing)

  3. ^

     Through means other than extinction risk reduction.  

25

0
0

Reactions

0
0

Have you voted yet?

This post is part of Existential Choices Debate Week. Click and drag your avatar to vote on the debate statement. Votes are non-anonymous, and you can change your mind.
On the margin1, it is better to work on reducing the chance of our2 extinction, than increasing the value of futures where we survive3
Disagree
Agree
Comments69


Sorted by Click to highlight new comments since:
Peter Wildeford
13
2
1
43% disagree

I don't think this is as clear of a dichotomy as people think it is. A lot of global catastrophic risk doesn't come from literal extinction because human extinction is very hard. A lot of mundane work on GCR policy involves a wide variety of threat models that are not just extinction.

WilliamKiely
5
0
0
57% agree

Footnote 2 completely changes the meaning of the statement from common sense interpretations of the statement. It makes it so that e.g. a future scenario in which AI takes over and causes existential catastrophe and the extinction of biological humans this century does not count as extinction, so long as the AI continues to exist. As such, I chose to ignore it with my "fairly strongly agree" answer.

zdgroff
9
0
1
36% disagree

The value of the future conditional on civilization surviving seems positive to me, but not robustly so. I think the main argument for its being positive is theoretical (e.g., Spreading happiness to the stars seems little harder than just spreading), but the historical/contemporary record is ambiguous.

The value of improving the future seems more robustly positive if it is tractable. I suspect it is not that much less tractable than extinction risk work. I think a lot of AI risk satisfies this goal as well as the x-risk goal for reasons Will MacAskill gives in What We Owe the Future. Understanding, developing direct interventions for, and designing political processes for digital minds seem like plausible candidates. Some work on how to design democratic institutions in the age of AI also seems plausibly tractable enough to compete with extinction risk.

OscarD🔸
4
1
0
29% disagree

I think there are a lot of thorny definitional issues here that make this set of issues not boil down that nicely to a 1D spectrum. But overall extinction prevention will likely have a far broader coalition supporting it, while making the future large and amazing is far less popular since most people aren't very ambitious with respect to spreading flourishing through the universe, but I tentatively am.

Buck
2
0
0
50% agree

I think increasing the value of good futures is probably higher importance, but much less tractable

Bella
7
4
0
29% agree

I agreed, but mostly because of my unfortunately-dim view of the tractability of work increasing the value of futures where we survive.

By the way, it looks like there might be some problem with the Forum UI here, as this post has some text suggesting that, since writing this comment, I changed my mind from "29% agree" to "14% agree." But I haven't intentionally changed my vote on the top banner, or changed my mind.

That's odd - to me I'm still just seeing your first vote. I'll send this to Will to check out. 

Thanks for flagging this @Bella! There was a bug which made it update everyone's comment whenever anyone changed their vote 🤦‍♂️. This is fixed now

Me too Bella, see my comment above

I don't think it's valuable to ensure future moral patients exist for their own sake, and extinction risk reduction only really seems to expectably benefit humans who would otherwise die in an extinction event, who would be in the billions. An astronomical number of future moral patients could have welfare at stake if we don't go extinct, so I'd prioritize them on the basis of their numbers.

See this comment and thread.

Charlie Harrison
4
1
0
64% disagree

If we treat digital minds like current animal livestock, the expected value of the future could be really bad. 

Most people have a strong drive to perpetuate humanity. What makes EAs special is that EAs also care about others' suffering. So EAs should focus on trying to make sure the future isn't full of suffering.

Derek Shiller
4
0
2
93% disagree

People in general, and not just longtermist altruists, have reason to be concerned with extinction. It may turn out not to be a problem or not be solvable and so the marginal impact seems questionable here. In contrast, few people are thinking about how to navigate our way to a worthwhile future. There are many places where thoughtful people might influence decisions that effectively lock us into a trajectory.

few people are thinking about how to navigate our way to a worthwhile future.


This might be true on the kinds of scales EAs are thinking about (potentially enourmous value, long time horizons) but is it not the case that many people want to steer humanity in a better direction? E.g. the Left, environmentalists, libertarians, ... ~all political movements?

I worry EAs think of this as some unique and obscure thing to think about, when it isn't.

(on the other hand, people neglect small probabilities of disastrous outcomes)

Lots of people think about how to improve the future in very traditional ways. Assuming the world keeps operating under the laws it has been for the past 50 years, how do we steer it in a better direction?

I suppose I was thinking of this in terms of taking radical changes from technology development seriously, but not in the sense of long timelines or weird sources of value. Far fewer people are thinking about how to navigate a time when AGI becomes commonplace than are thinking about how to get to that place, even though there might not be a huge window of time between them.

Owen Cotton-Barratt
4
0
1
86% disagree

To some extent I reject the question as not-super-action-guiding (I think that a lot of work people do has impacts on both things).

But taking it at face value, I think that AI x-risk is almost all about increasing the value of futures where "we" survive (even if all the humans die), and deserves most attention. Literal extinction of earth-originating intelligence is mostly a risk from future war, which I do think deserves some real attention, but isn't the main priority right now.

Hope I'm not misreading your comment, but I think you might have voted incorrectly, as if the scale is flipped.

[This comment is no longer endorsed by its author]Reply

I think Owen is voting correctly Robi - he disagrees that there should be more work on extinction reduction before there is more work on improving the value of the future. (to complicate this, he is understanding working on AI x-risk is mostly about increasing the value of the future, because, in his view, it isn't likely to lead to extinction). 

Apologies if the "agree" "disagree" labelling is unclear - we're thinking of ways to make it more parsable. 

This is right. But to add even more complication:

  • I think most AI x-risk (in expectation) doesn't lead to human extinction, but a noticeable fraction does
  • But a lot even of the fraction that leads to human extinction seems to me like it probably doesn't count as "extinction" by the standards of this question, since it still has the earth-originating intelligence which can go out and do stuff in the universe
    • However, I sort of expect people to naturally count this as "extinction"?

Since it wasn't cruxy for my rough overall position, I didn't resolve this last question before voting, although maybe it would get me to tweak my position a little.

working on AI x-risk is mostly about increasing the value of the future, because, in his view, it isn't likely to lead to extinction

Ah yes I get it now. Thanks!

I think human extinction (from ASI) is highly likely to happen, and soon, unless we stop it from being built[1]

See my comments in the Symposium for further discussion

  1. ^

    And that the ASI that wipes us out won't matter morally (to address footnote 2 on the statement)

Two major reasons/ considerations:
1- I'm unconvinced of the tractability of non-extinction-risk reducing longtermist interventions. 
2- Perhaps this is self-defeating - but I feel uncomfortable substantively shaping the future in ways that aren't merely making sure it exists. Visions of the future that I would have found un-objectionable a century ago would probably seem bad to me today. In short - this consideration is basically "moral uncertainty". I think extinction-risk reduction is, though not recommended on every moral framework, at least recommended on most. I haven't seen other ideas for shaping the future which are as widely recommended. 

I genuinely have no idea. 

Tejas Subramaniam
3
1
0
21% ➔ 7% agree

I think the expected value of the long-term future, in the “business as usual” scenario, is positive. In particular, I anticipate that advanced/transformative artificial intelligence drives technological innovation to solve a lot of world problems (e.g., helping create cell-based meat eventually), and I also think a decent amount of this EV is contained in futures with digital minds and/or space colonization (even though I’d guess it’s unlikely we get to that sort of world). However, I’m very uncertain about these futures—they could just as easily contain a large amount of suffering. And if we don’t get to those futures, I’m worried about wild animal suffering being high in the meantime. Separately, I’m not sure addressing a lot of s-risk scenarios right now is particularly tractable (nor, more imminently, does wild animal suffering seem awfully tractable to me). 

Probably the biggest reason I’m so close to the center is I think a significant amount of existential risk from AI looks like disempowering humanity without killing literally every human, and hence, I view AI alignment work as at least partially serving the goal of “increasing the value of futures where we survive.”

Manuel Allgaier
2
0
0
43% disagree

Depends, if x-risk is small (<5%) and if we expect outsized impact on the future (preventing negative value lock-ins), then the latter seems more important. I'm very unsure about both. 

Ozzie Gooen
2
0
0
29% agree

I have mixed feelings here. But one major practical worry I have about "increasing the value of futures" is that a lot of that looks fairly zero-sum to me. And I'm scared of getting other communities to think this way. 

If we can capture 5% more of the universe for utilitarian aims, for example, that's 5% less from others. 

I think it makes sense for a lot of this to be studied in private, but am less sure about highly public work.

Matthew_Barnett
*2
0
0
71% disagree

The extinction of all Earth-originating intelligent life (including AIs) seems extremely unlikely over the next several decades. While a longtermist utilitarian framework takes even a 0.01 percentage point reduction in extinction risk quite seriously, there appear to be very few plausible ways that all intelligent life originating from Earth could go extinct in the next century. Ensuring a positive transition to artificial life seems more useful on current margins.

The extinction of all Earth-originating intelligent life (including AIs) seems extremely unlikely.

Extremely unlikely to happen... when? Surely all Earth-originating intelligent life will eventually go extinct, because the universe’s resources are finite.

In my comment I later specified "in [the] next century" though it's quite understandable if you missed that. I agree that eventual extinction of Earth-originating intelligent life (including AIs) is likely; however, I don't currently see a plausible mechanism for this to occur over time horizons that are brief by cosmological standards.

(I just edited the original comment to make this slightly clearer.)

Thanks for the clarification. I didn’t mean to be pedantic: I think these discussions are often unclear about the relevant time horizon. Even Bostrom admits (somewhere) that his earlier writing about existential risk left the timeframe unspecified (vaguely talking about "premature" extinction).

On the substantive question, I’m interested in learning more about your reasoning. To me, it seems much more likely that Earth-originating intelligence will go extinct this century than, say, in the 8973th century AD (conditional on survival up to that century). This is because it seems plausible that humanity (or its descendants) will soon develop technology with enough destructive potential to actually kill all intelligence. Then the question becomes whether they will also successfully develop the technology to protect intelligence from being so destroyed. But I don’t think there are decisive arguments for expecting the offense-defense balance to favor either defense or offense (the strongest argument for pessimism, in my view, is stated in the first paragraph of this book review). Do you deny that this technology will be developed "over time horizons that are brief by cosmological standards”? Or are you confident that our capacity to destroy will be outpaced by our capacity to prevent from destruction?

I tentatively agree with your statement that,

To me, it seems much more likely that Earth-originating intelligence will go extinct this century than, say, in the 8973th century AD.

That said, I still suspect the absolute probability of total extinction of intelligent life during the 21st century is very low. To be more precise, I'd put this probability at around 1% (to be clear: I recognize other people may not agree that this credence should count as "extremely low" or "very low" in this context). To justify this statement, I would highlight several key factors:

  1. Throughout hundreds of millions of years, complex life has demonstrated remarkable resilience. Since the first vertebrates colonized land during the late Devonian period (approximately 375–360 million years ago), no extinction event has ever eradicated all species capable of complex cognition. Even after the most catastrophic mass extinctions, such as the end-Permian extinction and the K-Pg extinction, vertebrates rebounded. Not only did they recover, but they also surpassed their previous levels of ecological dominance and cognitive complexity, as seen in the increasing brain size and adaptability of various species over time.
  2. Unlike non-intelligent organisms, intelligent life—starting with humans—possesses advanced planning abilities and an exceptional capacity to adapt to changing environments. Humans have successfully settled in nearly every climate and terrestrial habitat on Earth, from tropical jungles to arid deserts and even Antarctica. This extreme adaptability suggests that intelligent life is far less vulnerable to complete extinction compared to other complex life forms.
  3. As human civilization has advanced, our species has become increasingly robust against most types of extinction events rather than more fragile. Technological progress has expanded our ability to mitigate threats, whether they come from natural disasters or disease. Our massive global population further reduces the likelihood that any single event could exterminate every last human, while our growing capacity to detect and neutralize threats makes us better equipped to survive crises.
  4. History shows that even in cases of large-scale violence and genocide, the goal has almost always been the destruction of rival groups—not the annihilation of all life, including the perpetrators themselves. This suggests that intelligent beings have strong instrumental reasons to avoid total extinction events. Even in scenarios involving genocidal warfare, the likelihood of all intelligent beings willingly or accidentally destroying all life—including their own—seems very low.
  5. I have yet to see any compelling evidence that near-term or medium-term technological advancements will introduce a weapon or catastrophe capable of wiping out all forms of intelligent life. While near-term technological risks certainly exist that threaten human life, none currently appear to pose a credible risk of total extinction of intelligent life.
  6. Some of the most destructive long-term technologies—such as asteroid manipulation for planetary bombardment—are likely to develop alongside technologies that enhance our ability to survive and expand into space. As our capacity for destruction grows, so too will our ability to establish off-world colonies and secure alternative survival strategies. This suggests that the overall trajectory of intelligent life seems to be toward increasing resilience, not increasing vulnerability.
  7. Artificial life could rapidly evolve to become highly resilient to environmental shocks. Future AIs could be designed to be at least as robust as insects—able to survive in a wide range of extreme and unpredictable conditions. Similar to plant seeds, artificial hardware could be engineered to efficiently store and execute complex self-replicating instructions in a highly compact form, enabling them to autonomously colonize diverse environments by utilizing various energy sources, such as solar and thermal energy. Having been engineered rather than evolved naturally, these artificial systems could take advantage of design principles that surpass biological organisms in adaptability. By leveraging a vast array of energy sources and survival strategies, they could likely colonize some of the most extreme and inhospitable environments in our solar system—places that even the most resilient biological life forms on Earth could never inhabit.

I'm optimistic about the very best value-increasing research/interventions. But in terms of what would actually be done at the margin, most work that people would do for "value-increasing" reasons would be confused/doomed, I expect (and this is less true for AI safety).

Nathan Young
2
0
0
21% agree

I dunno, by how much? Seems contingent on lots of factors. 

Tom Gardiner
2
2
1
57% disagree

The long-running future seems like it could well be unacceptably awful. From the perspective of a battery hen, it would seem much better that it's distant ancestors were pushed out of an ecological niche before humanity domesticated them. Throwing all our effort into X-risk mitigation without really tackling S-risks in a world of increasing lock-in across domains seems deeply unwise to me. 

lilly
2
0
1
79% disagree

It's a tractability issue. In order for these interventions to be worth funding, they should reduce our chance of extinction not just now, but over the long term. And I just haven't seen many examples of projects that seem likely to do that.

JoA🔸
2
1
0
43% disagree

This is a difficult one, and both my thoughts and my justifications (especially the few sources I cite) are very incomplete. 

It seems to me for now that existential risk reduction is likely to be negative, as both human and AI-controlled futures could contain immense orders of magnitude more suffering than the current world (and technological developments could also enable more intense suffering, whether in humans or in digital minds). The most salient ethical problems with the extinction of earth-originating intelligent life seem to be the likelihood of biological suffering continuing on earth for millions of years (though it's not clear to me whether it would be more or less intense without intelligent life on earth), and the possibility of space (and eventually earth) being colonized by aliens (though whether their values will be better or worse remains an open question in my view).

Another point (which I'm not certain about how to weigh in my considerations) is that certain extinction events could massively reduce suffering on earth, by preventing digital sentience or even by causing the end of biological sentient life (this seems unlikely, and I've asked here how likely or unlikely EAs thought this was).

However, I am very uncertain of the tractability of improving future outcomes, especially considering recent posts by researchers at the Center on Long-Term Risk, or this one by a former researcher there, highlighting how uncertain it is that we are well-placed to improve the future. Nonetheless, I think that efforts made to improve the future, like the work of the Center for Reducing Suffering, the Center on Long-Term Risk, or the Sentience Institute, advocate for important values and could have some positive flow-through effects in the medium-term (though I don't necessarily think that this robustly improves the longer term future). I will note, however, that I am biased since work related the Center for Reducing Suffering was the primary reason I got into EA.

I am very open to changing my mind on this, but for now I'm under 50% agree because it seems to me that, in short:

  1. Extinction Risk Reduction could very well have negative expected value.
  2. Efforts to improve the value of futures where we survive might have some moderate positive effects in the short-term.

Lots of uncertainties. I expect to have moved my cursor before the end of the week!

niplav
3
1
0
71% ➔ 64% agree

Under moral uncertainty, many moral perspectives care much more about averting downsides than producing upsides.

Additionally, tractability is probably higher for extinction-level threats, since they are "absorptive"; decreasing the chance we end up in one gives humanity and their descendants ability to do whatever they figure out is best.

Finally, there is a meaningful sense in which working on improving the future is plagued by questions about moral progress and lock-in of values, and my intuition is that most interventions that take moral progress seriously and try to avoid lock-in boil down to working on things that are fairly equivalent to avoiding extinction. Interventions that don't take moral progress seriously instead may look like locking in current values.

Stephen McAleese
1
0
0
43% agree

I think avoiding existential risk is the most important thing. As long as we can do that and don't have some kind of lock in, then we'll have time to think about and optimize the value of the future.

Mjreard
1
0
0
21% agree

I think people overrate how predictable the effect of our actions on the future will be (even though they rate it very low in absolute terms); extinction seem like one of the very few (only?) things that seems like its effects will endure throughout a big part of the future. Still buy the theory that 0-1% of possible value is equally valuable to 98-99%; just about tractability

KonradKozaczek
1
0
0
14% disagree

It seems even more important to avoid futures full of extreme suffering than to avoid extinction.

Dylan Richardson
1
0
0
36% ➔ 29% disagree

While I don't entirely disregard x-risks, I have been unimpressed by the tractability of most interventions, excepting perhaps for bio-security ones.

The prevalent notion of "solving" the alignment problem as though it's a particularly hard math problem strikes me as overly represented, which entails neglecting other, more in-direct safety measures, like stable, transparent and trustworthy institutions, whether political or geopolitical (US-China war means what for AI?).

Relatedly, the harm aversion/moral purity signaling around working in AI companies (especially Anthropic!) seems counterproductive and yet has received ~no push back. It seems obvious to me that having concerned individuals (rather than E-acc ideologues) in high positions in companies is very important! I suspect that the dominance of Doomerism in AI Safety over more epistemically sound concerns is to blame. 

I find it very difficult to determine whether the future will be net-negative or net-positive (when considering humans, factory-farmed animals, wild animals, and possibly artificial sentience). 
This makes it very hard to know whether work on extinction reduction is likely to be positive or not.
I prefer to work on things that aim to move the sign towards "net-positive".

Aidan Alexander
2
1
2
79% disagree

Far from convinced that continued existence at currently likely wellbeing levels is a good thing

I really don't have a strong view, but I find myself sympathetic to the idea that the world is not great and is getting worse (if you consider non-human animals).

Broadly Agree

Although I might have misunderstood and missed the point of this entire debate, so correct me if that is the case

I just don't believe  changing the future trajectory is tractable in say 50-100 years from now areas like politics, economics, AI welfare etc. I think its a pipe dream. We cannot predict technological, political and economic changes even in the medium term future. These changes may well quickly render our current efforts meaningless in 10-20 years. I think the effect of work we do now which is future focused diminishes in value exponentially to the point that in 20 years time it is probably almost meaningless. I could be convinced against this if you could show me interventions 20 years ago which were specifically future focus which are now still meaningful.

However I do believe it may be possible to reduce existential risk through direct, tangeble changes now which could prevent a disastrous one off event like a engnieered virus or AI apocalypse. Stopping a one off catastrophe though still exremely difficult seems much more realistic than influencing an overall global trajectory with so much uncertainty.

This is unless you include general global health and development and animal welfare work which directly improves the lives of beings today, and will also likely improve beings lives in the future but probably in a far more minor way than this question is addressing. If you include the effects ofstandard EA GHD and animal welfare work under the banner "Improving the value of futures where we survive" then I would probably swing my vote the other way.

Robi Rahman
2
0
0
93% agree

On the current margin, improving our odds of survival seems much more crucial to the long-term value of civilization. My reason for believing this is that there are some dangerous technologies which I expect will be invented soon, and are more likely to lead to extinction in their early years than later on. Therefore, we should currently spend more effort on ensuring survival, because we will have more time to improve the value of the future after that.

(Counterpoint: ASI is the main technology that might lead to extinction, and the period when it's invented might be equally front-loaded in terms of setting values as it is in terms of extinction risk.)

OllieBase
2
0
0
36% agree

It seems plausible to me we might be approaching a "time of perils' where total x-risk is unacceptably high and will continue to be as we develop powerful AI systems, but might decrease later since we can use AI systems to tackle x-risk (though that seems hard and risky in its own myriad ways).

Broadly think we should still prioritise avoiding catastrophes in this phase, and bet on being able to steer later but low confidence.

finm
2
0
0
64% disagree

Partly this is because I think “extinction” as defined here is very unlikely (<<1%) to happen this century, which upper bounds the scale of the area. I think most “existential risk” work is not squarely targeted at avoiding literal extinction of all Earth-originating life.

Angelina Li
2
0
0
14% disagree

I feel extremely unsure about this one. I'm voting slightly against purely from the perspective of, "wow, there are projects in that direction that feel super neglected".

Matrice Jacobine
2
0
0
1
36% ➔ 29% disagree

Recent advances in LLMs have led me to update toward believing that we live in the world where alignment is easy (i.e. CEV naturally emerge from LLMs, and future AI agents will be based on understanding and following natural language commands by default), but governance is hard (i.e. AI agents might be co-opted by governments or corporations to lock in humanity in a dystopian future, and the current geopolitical environment, characterized by democratic backsliding, cold war mongering, and an increase in military conflicts including wars of aggression, isn't conducive to robust multilateral governance).

JackM
2
0
1
36% disagree

This is a question I could easily change my mind on.

The experience of digital minds seems to dominate far future calculations. We can get a lot of value from this, a lot of disvalue, or anything in between.

If we go extinct then we get 0 value from digital minds. This seems bad, but we also avoid the futures where we create them and they suffer. It’s hard to say if we are on track to creating them to flourish or suffer - I think there are arguments on both sides. The futures where we create digital minds may be the ones where we wanted to “use” them, which could mean them suffering. Alternatively, we have seen our moral circle expand over time and this may continue, so there is a real possibility we could create them to flourish. I don’t have a clear view which side wins here, so overall going extinct doesn’t seem obviously terrible to me.

We could instead focus on raising the moral status of digital minds, our ability to understand sentience and consciousness, improve societal values, making sure AI goes well and helps us with these things. These robustly increase the expected value of digital sentience in futures where we survive. 

So because reducing extinction risk is close to 0 expected value to me, and increasing the value of futures where we survive is robustly positive in expected value, I lean towards increasing the value of futures where we survive.

Arturo Macias
2
0
0
79% agree

Intelligence is the only chance of some redemption to the massive suffering probably associated to the emergence of consciousness. 

This is the age of danger, we are the first species on Earth that has figured out morality, so we shall survive at allmost all cost. 

Alistair Stewart
1
0
1
93% disagree

Making people happy is valuable; making happy people is probably not valuable. There is an asymmetry between suffering and happiness because it is more morally important to mitigate suffering than to create happiness.

John Salter
2
1
5
29% disagree

The far future, on our current trajectory, seems net negative on average. Reducing extinction risk just multiplies its negative EV. 

We could devote 100% of currently available resources to existential risk reduction, live in austerity, and never be finished ensuring our own survival.  However, if increase the value of futures where we survive, we will develop more and more resources that can then be put to existential risk reduction.  People will be not only happier, but also more capable and skilled, when we create a world where people can thrive rather than just survive.  The highest quality futures are the most robust.  

jboscw
1
0
0
93% agree

Very worried about AI risk, think short timelines are plausible

tylermjohn
1
0
0
36% disagree

I'm compressing two major dimensions into one here:

  • Expected value (I think EV of maxevas is vastly higher than EV of maxipok)
  • Robustness (the case for maxevas is highly empirically theory laden and values dependent)

What is maxevas? Couldn't find anything relevant by googling.

A higher value future reduces the chances of extinction.  If people value life, they will figure out how to keep it.

Jonas Hallgren
1
0
0
21% agree

First and foremost, I'm low confidence here. 

I will focus on x-risk from AI and I will challenge the premise of this being the right way to ask the question.

What is the difference between x-risk and s-risk/increasing the value of futures? When we mention x-risk with regards to AI we think of humans going extinct but I believe that to be a shortform for wise compassionate decision making. (at least in the EA sphere) 

Personally, I think that x-risk and good decision making in terms of moral value might be coupled to each other. We can think of our current governance conditions a bit like correction systems for individual errors. If they pile up, we go off the rail and increase x-risk as well as chances of a bad future. 

So a good decision making system should both account for x-risk and value estimation, therefore the solution is the same and it is a false dichotomy? 

(I might be wrong and I appreciate the slider question anyway!)

I've heard this argument before, but I find it un-compelling in its tractability. If we don't go extinct, its likely to be a silent victory; most humans on the planet won't even realise it happened. Individual humans working on X-risk reduction will probably only impact the morals of people around them.

Patrick Hoang
1
0
0
50% ➔ 57% disagree

I think the most likely outcome is not necessarily extinction (I estimate <10% due to AI) but rather an unfulfilled potential. This may be humans simply losing control over the future and becoming mere spectators and AI not being morally significant in some way.

With long timeline and less than 10% probability: Hot take is these are co-dependent - prioritizing only extinction is not feasible. Additionally, does only one human exist while all others die count as non-extinction? What about only a group of humans survive? How should this be selected? It could dangerously/quickly fall back to Fascism. It would only likely benefit the group of people with current low to no suffering risks, which unfortunately correlates to the most wealthy group. When we are "dimension-reducing" the human race to one single point, we ignore the individuals. This to me goes against the intuition of altruism. 

I fundamentally disagree with the winner-take-all type of cause prioritization - instead, allocate resources to each area, and unfortunately there might be multiple battles to fight.

To analyze people's responses, I can see this question being adjusted to consider prior assumptions: 1. What's your satisfaction on how we are currently doing in the world now? What are the biggest gaps to your ideal world? 2.What's your assessment of timeline + current % of extinction risk due to what?

Trym Braathen🔸
1
1
0
29% disagree

What is the point of securing a future for humanity if that future is net-negative?

Scott Smith 🔸
1
0
0
86% agree

We can't work on increasing value if we are dead.

If we avoid extinction, plenty of people will have the time to take care of humanity's future. I'll leave it to them. Both topic have a lot of common ground anyway, like "not messing up with the biosphere" or "keeping control of ASI"

[comment deleted]4
0
0
Curated and popular this week
Relevant opportunities