The central question being discussed in the current debate is whether marginal efforts should prioritize reducing existential risk or improving the quality of futures conditional on survival. Both are important, both are neglected, though the latter admittedly more so, at least within EA. But this post examines the tractability of shaping the long-term future if humanity survives, and the uncertainty about our ability to do so effectively.

I want to very briefly argue that given the complexity of long-term trajectories, the lack of empirical evidence, and the difficulty of identifying robust interventions, efforts to improve future value are significantly less tractable than reducing existential risk.

We have strong reasons to think we know what the likely sources of existential risk are - as @Sean_o_h's new paper lays out very clearly. The most plausible risks are well known, and we also have at least some paths towards mitigating them, at least in the form of not causing them. On the other hand, if we condition on humanity’s survival, we are dealing with an open-ended set of possible futures that is both not well characterized, and poorly explored. Exploration of futures is also not particularly tractable, given the branching nature and the complexity of the systems being predicted. And this problem is not just about characterizing futures - the tractability of interventions decreases as the system's complexity increases, especially over multi-century timescales. The complexity of socio-technological and moral evolution makes it infeasible, in my view, to shape long-term outcomes with even moderate confidence. It seems plausible that most interventions would have opposite signs in many plausible futures, and we seem unlikely to know the relative probabilities or the impacts.

And despite @William_MacAskill's book on the topic, we have very limited evidence for what works to guide the future - one of the few key criticisms I think should be generally convincing about the entire premise of longtermism. The exception, of course, is avoiding extinction.

And compared to existential risk, where specific interventions may have clear leverage points, such as biosecurity or AI safety, increasing the quality of long-term futures is a vast and nebulous goal. There is no singular control knob for “future value,” making interventions more speculative. So identifying interventions today that will robustly steer the future in a particular direction is difficult because, as noted, we lack strong historical precedent for guiding complex civilizations over thousands of years, and also, the existence of unpredictable attractor states (e.g., technological singularities, value shifts) makes long-term interventions unreliable. Work to change this seems plausibly valuable, but also more interesting than important, as I previously argued.

45

5
5
1

Reactions

5
5
1

Have you voted yet?

This post is part of Existential Choices Debate Week. Click and drag your avatar to vote on the debate statement. Votes are non-anonymous, and you can change your mind.
On the margin1, it is better to work on reducing the chance of our2 extinction, than increasing the value of futures where we survive3
C
C
F
H
JN
KY
M
M
M
NK
N
PB
P
Y
C
DS
I
JR
MZ
P
SK
U
B
B
I
J
OCB
S
TG
ZM
AA
CB
DM
HH
JS
M
V
VH
A
C
DR
EN
JS
MB
NKD
PS
R
WC
WM
BM
C
CH
CG
G
GF
L
M
M
PY
AM
AWE
B
F
J
M
M
MR
PH
P
F
JV
K
NR
S
T
TV
DT
EG
J
SDI
WH
J
KS
M
RR
S
T
T
A
B
C
C
GA
LD
L
L
MJ
ND
N
R
SC
A
C
C
EA
H
JW
S
G
J
JD
JC
KL
K
LW
SP
YY
DR
G
JS
JG
PW
PC
VK
A
DM
DM
FK
L
RB
SB
TA
TW
T
V
A
EPC
JF
KL
TS
T
BG
GA
ID
JB
P
TVD
W
W
Y
AWZ
C
CL
E
EF
HC
T
F
L
Q
W
AS
B
CC
CJ
E
FC
HY
JHN
M
O
R
V
AB
C
D
J
PB
SH
SM
Z
AK
A
B
C
GM
JDR
L
N
PB
TS
AZ
C
DH
G
JH
J
LR
MH
M
OS
Q
W
JM
KM
M
JJ
P
S
AM
C
MB
SS
T
TSL
CS
DA
HP
SS
GL
J
K
BA
CK
H
M
T
Disagree
Agree
Comments21


Sorted by Click to highlight new comments since:

I agree with this, and like to lean on what evidence we have rather than theory and speculation

There are many past events which increased the value of the future unintentionally, like free markets and most scientific advancements.

Previous examples of people and movements which intentionally planned to increase the value of the future mostly achieved it through concrete current change in laws and cultural norms. When MLK said "I have a dream", he fought for change then and there to make that dream a reality. Efforts like these have had a huge effect. This list is just what comes to mind, in no particular order and far from extensive. Most of these required community organising and protest as at least part of the process to achieve the concrete change.

- Abolition of slavery
- Women's suffrage
- Nuclear Treaties
- Universal declaration of human rights (which although breaking down a bit has been so successful that "human rights" are seen almost as an objective truth for many people)
- Banning warfare with poisonous gases
- Civil rights movement
- Access to free HIV treatment becoming a universal "right" through social movements and PEPFAR
- Animal welfare movement - specifically the cage free revolution
- Democratic transformation of nation states
- LGBTQ+ rights movements and legal changes
- Social welfare movements for the poorest part of the population in many rich countries

I think efforts to improve the value of the future through concrete current change right here right now are fantastic and EA can continue to focus on those where they are neglected and tractable, but I struggle to see a clear Theory of Change for other possible mechanisms due to, as you put it  very well"The complexity of socio-technological and moral evolution". Good things locked in now can be really hard to change in democratic countries. I would (perhaps slightly controversially) cite Obamacare as an example of something that was so beneficial to so many it became extremely hard to reverse.

I'm very open to other pathways for creating change that don't involve a concrete change right here right now, but you'll need to convince me.

This seems mostly correct, though I think the role of community organizing (versus elite consensus change) is strongly overstated.

Do you think "Most of these required community organising and protest as at least part of the process to achieve the concrete change." is that strong a statement? There is a pretty strong correlation between protest/organising and these changes. Elite consense is clearly very important, but I think that the voice of the masses can move the elite to consensus so there's some chicken and egg there. Also to mention a few cases here where I don't think elite consensus was strong at the time of change and their hand's were perhaps forced...

- Access to free HIV treatment (This I'm pretty sure of)
- Civil rights movement
- Women's suffrage

I do find this a tricky issue to keep a scout mindset on here on the forum, as I find EAs in general are unusually against protest and organising compared to other communities I am a part of. My feeling is this is largely because the nature of many EAs is more to be into research, debate and policy rather than social roles like organising and protest.

What makes you think it is overstated? I think its a tricky counterfactual question with a lot of room for conjecture.... 

Tangentially re: protest, I think things are slowly shifting, due to the work of folks like James Özden founding Social Change Lab to understand how social change movements can be more evidence-based and effective. For instance, James changed my mind on the effectiveness of radical protest tactics in What’s everyone got against throwing soup?, which drew upon this literature review to conclude that 

A nonviolent radical flank is likely to help, not hinder, a social movement. Specifically, we think there’s good evidence it can increase support for more moderate groups and increase the salience of an issue without harming support for the overall movement’s policy goals.

I'd also signal-boost James' article Protest Movements Could Be More Effective Than the Best Charities published in the Stanford Social Innovation Review. You should always take charts like the one below claiming superlative cost-eff with a metric ton of salt, but I mostly trust the general quality of his analysis and think his bottomline holds up.

That said, James seems to be the only person banging this drum, so I suppose your observation still broadly holds true.

I think James's did is really important and shows the potential good if a radical flank 

Traditionally though movements and organizing was often around issues which has decent public support already and wasnt necessarily that radical. The civil rights movement and the HIV medicine campaigners directly moved the elites towards their goal, and didn't just move the needle on public opinion for the better like the radical flank can

I agree that the examples you list are ones where organizing and protest played a large role, and I agree that it's effectively impossible to know the counterfactual - but I was thinking of the other examples, several where there was no organizing and protest, but which happened anyways - which seems like clear evidence that they are contributory and helpful but not necessary factors. On the other hand, effectiveness is very hard to gauge!

The conclusion is that organizing is likely or even clearly positive - but it's evidently not required, if other factors are present, which is why I thought it was overstated.

Yep I agree with that

(Low effort comment as I run out the door, but hope it adds value) To me the most compelling argument in favour of tractability is:

  • We could make powerful AI agents whose goals are well understood and do not change or update in ex ante predictable ways.
  • These agents are effectively immortal and the most powerful thing in the affectable universe, with no natural competition. They would be able to overcome potentially all natural obstacles, so they would determine what happens in the lightcone.
  • So, we can make powerful AI agents that determine what happens in the lightcone, whose goals are well understood and update in ex ante predictable ways.
  • So, we can take actions that determine what happens in the lightcone in an ex ante predictable way.

This more or less conforms to why I think trajectory changes might be tractable, but I think the idea can be spelled out in a slightly more general way: as technology develops (and especially AI), we can expect to get better at designing institutions that perpetuate themselves. Past challenges to affecting a trajectory change come from erosion of goals due to random and uncontrollable human variation and the chaotic intrusion of external events. Technology may help us make stable institutions that can continue to promote goals for long periods of time.

Well-understood goals in agents that gain power and take over the lightcone is exactly the thing we'd be addressing with AI alignment, so this seems like an argument for investing in AI alignment - which I think most people would see as far closer to preventing existential risk.

That said, without a lot more progress, powerful agents with simple goals is actually just a fancy way of guaranteeing of a really bad outcome, almost certainly including human extinction.

Here's a shower thought: 

  • If you think extinction risk reduction is highly valuable, then you need some kind of a model of what Earth-originating life will do with its cosmic endowment
  • Some of the parameters in you model must be related to things other than mere survival, like what this life is motivated by or will attempt to do
  • Plausibly, there are things you can do to change the values of those parameters and not just the extinction parameter 

It won't work for every model (maybe the other parameters just won't budge), but for some of them it should. 

  • If you think extinction risk reduction is highly valuable, then you need some kind of a model of what Earth-originating life will do with its cosmic endowment


No, you don't, and you don't even need to be utilitarian, much less longtermist!

Any disagreement about longtermist prioritization should presuppose longtermism 

First, you're adding the assumption that the framing must be longtermist, and second, even conditional on longtermism you don't need to be utilitarian, so the supposition that you need a model of what we do with the cosmic endowment would still be unjustified.

You're not going to be prioritizing between extinction risk and long term trajectory changes based on tractability if you don't care about the far future. And for any moral theory you can ask "why do you think this will be a good outcome?" and as long as you don't value life intrinsically you'll have to state some empirical hypotheses about the far future 

There is a huge range of "far future" that different views will prioritize differently, and not all need to care about the cosmic endowment at all - people can care about the coming 2-3 centuries based on low but nonzero discount rates, for example, but not care about the longer term future very much.

I don't understand why that matters. Whatever discount rate you have, if you're prioritizing between extinction risk and trajectory change you will have some parameters that tell you something about what is going to happen over N years. It doesn't matter how long this time horizon is. I think you're not thinking about whether your claims have bearing on the actual matter at hand. 

It would probably be most useful for you to try to articulate a view that avoids the dilemma I mentioned in the first comment of this thread. 

We have strong reasons to think we know what the likely sources of existential risk are - as @Sean_o_h's new paper lays out very clearly.

Looked at the paper. The abstract says:

In all cases, an outcome as extreme as human extinction would require events or developments that either have been of very low probability historically or are entirely unprecedented. This introduces deep uncertainty and methodological challenges to the study of the topic. This review provides an overview of potential human extinction causes considered plausible in the current academic literature...

So I think you are overstating it a bit, i.e., it's hard to support statements about existential risks coming from classified risks vs unknown unknowns/black swans. But if I'm getting the wrong impression I'm happy to read the paper in depth.

Not at all correct - and you clearly started the quote one sentence too late! "Potential causes of human extinction can be loosely grouped into exogenous threats such as an asteroid impact and anthropogenic threats such as war or a catastrophic physics accident. "

So the point of the abstract is that anthropogenic risks, ie. the ones that the next sentence calls "events or developments that either have been of very low probability historically or are entirely unprecedented,"  are the critical ones, which is why they are a large focus of the paper. 

I want to very briefly argue that given the complexity of long-term trajectories, the lack of empirical evidence, and the difficulty of identifying robust interventions, efforts to improve future value are significantly less tractable than reducing existential risk.

[...]

And compared to existential risk, where specific interventions may have clear leverage points, such as biosecurity or AI safety, increasing the quality of long-term futures is a vast and nebulous goal.

I guess, there is a misunderstanding in your analysis. Please correct me if I am wrong. 

"Increasing the quality of long-term futures" reduces existential risks. When longtermists talk about "increasing the quality of long-term futures," they include progress on aligning AIs as one of the best interventions they have in mind.

To compare their relative tractability, let's look at the best intervention to reduce Extinction-Risks and, on the other hand, at the best interventions for "increasing the quality of long-term futures", what I call reducing Alignment-Risks.

  • Illustrative best PTIs for Extinction-Risks reduction: Improving AI control, Reducing AI misuses. These reduce the chance of AI destroying future Earth-originating intelligent agents.
  • Illustrative best PTIs for Alignment-Risks reduction: Technical AI alignment, improving AI governance. These improve the quality of the long-term futures.

Now, let's compare their tractability. How these interventions differ in tractability is not clear. These interventions actually overlap significantly. It is not clear if reducing misuse risks is actually harder than improving alignment or than improving AI governance.

Interestingly, this leads us to a plausible contradiction in arguments against Alignment-Risks: Some will say that the interventions to reduce Alignment-Risks and Extinction-Risks are the same, and some will say they have vastly different tractability. One of the two groups is incorrect. Interventions can't be the same and have different tractability.

You make a dichotomy not present in my post, then conflate the two types of interventions while focusing only on AI risk - so that you're saying that two different kinds of what most people would call extinction reduction efforts are differently tractable - and conclude that there's a definition confusion.

To respond, first, that has little to do with my argument, but if it's correct, your problem is with the entire debate week framing, which you think doesn't present two distinct options, not with my post! And second, look at the other comments which bring up other types of change as quality increasing, and try to do the same analysis, without creating new categories, and you'll understand what I was saying better. 

More from Davidmanheim
36
Davidmanheim
· · 4m read
37
Davidmanheim
· · 4m read
Curated and popular this week
LintzA
 ·  · 15m read
 · 
Cross-posted to Lesswrong Introduction Several developments over the past few months should cause you to re-evaluate what you are doing. These include: 1. Updates toward short timelines 2. The Trump presidency 3. The o1 (inference-time compute scaling) paradigm 4. Deepseek 5. Stargate/AI datacenter spending 6. Increased internal deployment 7. Absence of AI x-risk/safety considerations in mainstream AI discourse Taken together, these are enough to render many existing AI governance strategies obsolete (and probably some technical safety strategies too). There's a good chance we're entering crunch time and that should absolutely affect your theory of change and what you plan to work on. In this piece I try to give a quick summary of these developments and think through the broader implications these have for AI safety. At the end of the piece I give some quick initial thoughts on how these developments affect what safety-concerned folks should be prioritizing. These are early days and I expect many of my takes will shift, look forward to discussing in the comments!  Implications of recent developments Updates toward short timelines There’s general agreement that timelines are likely to be far shorter than most expected. Both Sam Altman and Dario Amodei have recently said they expect AGI within the next 3 years. Anecdotally, nearly everyone I know or have heard of who was expecting longer timelines has updated significantly toward short timelines (<5 years). E.g. Ajeya’s median estimate is that 99% of fully-remote jobs will be automatable in roughly 6-8 years, 5+ years earlier than her 2023 estimate. On a quick look, prediction markets seem to have shifted to short timelines (e.g. Metaculus[1] & Manifold appear to have roughly 2030 median timelines to AGI, though haven’t moved dramatically in recent months). We’ve consistently seen performance on benchmarks far exceed what most predicted. Most recently, Epoch was surprised to see OpenAI’s o3 model achi
Dr Kassim
 ·  · 4m read
 · 
Hey everyone, I’ve been going through the EA Introductory Program, and I have to admit some of these ideas make sense, but others leave me with more questions than answers. I’m trying to wrap my head around certain core EA principles, and the more I think about them, the more I wonder: Am I misunderstanding, or are there blind spots in EA’s approach? I’d really love to hear what others think. Maybe you can help me clarify some of my doubts. Or maybe you share the same reservations? Let’s talk. Cause Prioritization. Does It Ignore Political and Social Reality? EA focuses on doing the most good per dollar, which makes sense in theory. But does it hold up when you apply it to real world contexts especially in countries like Uganda? Take malaria prevention. It’s a top EA cause because it’s highly cost effective $5,000 can save a life through bed nets (GiveWell, 2023). But what happens when government corruption or instability disrupts these programs? The Global Fund scandal in Uganda saw $1.6 million in malaria aid mismanaged (Global Fund Audit Report, 2016). If money isn’t reaching the people it’s meant to help, is it really the best use of resources? And what about leadership changes? Policies shift unpredictably here. A national animal welfare initiative I supported lost momentum when political priorities changed. How does EA factor in these uncertainties when prioritizing causes? It feels like EA assumes a stable world where money always achieves the intended impact. But what if that’s not the world we live in? Long termism. A Luxury When the Present Is in Crisis? I get why long termists argue that future people matter. But should we really prioritize them over people suffering today? Long termism tells us that existential risks like AI could wipe out trillions of future lives. But in Uganda, we’re losing lives now—1,500+ die from rabies annually (WHO, 2021), and 41% of children suffer from stunting due to malnutrition (UNICEF, 2022). These are preventable d
Rory Fenton
 ·  · 6m read
 · 
Cross-posted from my blog. Contrary to my carefully crafted brand as a weak nerd, I go to a local CrossFit gym a few times a week. Every year, the gym raises funds for a scholarship for teens from lower-income families to attend their summer camp program. I don’t know how many Crossfit-interested low-income teens there are in my small town, but I’ll guess there are perhaps 2 of them who would benefit from the scholarship. After all, CrossFit is pretty niche, and the town is small. Helping youngsters get swole in the Pacific Northwest is not exactly as cost-effective as preventing malaria in Malawi. But I notice I feel drawn to supporting the scholarship anyway. Every time it pops in my head I think, “My money could fully solve this problem”. The camp only costs a few hundred dollars per kid and if there are just 2 kids who need support, I could give $500 and there would no longer be teenagers in my town who want to go to a CrossFit summer camp but can’t. Thanks to me, the hero, this problem would be entirely solved. 100%. That is not how most nonprofit work feels to me. You are only ever making small dents in important problems I want to work on big problems. Global poverty. Malaria. Everyone not suddenly dying. But if I’m honest, what I really want is to solve those problems. Me, personally, solve them. This is a continued source of frustration and sadness because I absolutely cannot solve those problems. Consider what else my $500 CrossFit scholarship might do: * I want to save lives, and USAID suddenly stops giving $7 billion a year to PEPFAR. So I give $500 to the Rapid Response Fund. My donation solves 0.000001% of the problem and I feel like I have failed. * I want to solve climate change, and getting to net zero will require stopping or removing emissions of 1,500 billion tons of carbon dioxide. I give $500 to a policy nonprofit that reduces emissions, in expectation, by 50 tons. My donation solves 0.000000003% of the problem and I feel like I have f