A short and arguably unfinished blog post that I'm sharing as part of EA strategy fortnight. There's probably a lot more to say about this, but I've sat on this draft for a few months and don't expect to have time to develop the argument much further.
-
I understand longtermism to be the claim that positively shaping the long-term future is a moral priority. The argument for longtermism goes:
- The future could be extremely large;
- The beings who will inhabit that future matter, morally;
- There are things we can do to improve the lives of those beings (one of which is reducing existential risk);
- Therefore, positively shaping the long-term future should be a moral priority.
However, I have one core worry about longtermism and it’s this: people (reasonably) see its adherents as power-seeking. I think this worry somewhat extends to broad existential risk reduction work, but much less so.

Arguments for longtermism tell us something important and surprising; that there is an extremely large thing that people aren’t paying attention to. That thing is the long-term future. In some ways, it’s odd that we have to draw attention to this extremely large thing. Everyone believes the future will exist and most people don’t expect the world to end that soon.[1]
Perhaps what longtermism introduces to most people is actually premises 2 and 3 (above) — that we might have some reason to take it seriously, morally, and that we can shape it.
In any case, longtermism seems to point to something that people vaguely know about or even agree with already and then say that we have reason to try and influence that thing.
This would all be fine if everyone felt like they were on the same team. That, when longtermists say “we should try and influence the long-term future”, everyone listening sees themselves as part of that “we”.
This doesn’t seem to be what’s happening. For whatever reason, when people hear longtermists say “we should try and influence the long-term future”, they hear the “we” as just the longtermists.[2]
This is worrying to them. It sounds like this small group of people making this clever argument will take control of this extremely big thing that no one thought you could (or should) control.
The only thing that could make this worse is if this small group of people were somehow undeserving of more power and influence, such as relatively wealthy[3], well-educated white men. Unfortunately, many people making this argument are relatively wealthy, well-educated white men (including me).
To be clear, I think longtermists do not view accruing power as a core goal or as an implication of longtermism.[4] Importantly, when longtermists say “we should try and influence the long-term future”, I think they/we really mean everyone.[5]
Ironically, it seems that, because no one else is paying attention to the extremely big thing, they’re going to have to be the first ones to pay attention to it.
—
I don’t have much in the way of a solution here. I mostly wanted to point to this worry and spell it out more clearly so that those of us making the case for longtermism can at least be aware of this potential, unfortunate misreading of the idea.
- ^
58% of US adults do not think we are living in “the end times”. Not super reassuring.
- ^
- ^
As much as they try and make themselves less wealthy by donating a large portion of their income to charity.
- ^
I think you could make the case that this is often an indirect goal, such as getting the ears of important policymakers.
- ^
Except, perhaps, dictators and other ne'er-do-wells.
I'm not sure it is a full misreading, sadly. I don't think it a fair characterization of Ord, Greaves and MacAskill (though I am kind of biased because of my pride in having been an Oxford philosophy DPhil). It would be easy to give a radical deliberative democracy spin on Will and Toby's "long reflection" ideas in particular. But all the "pivotal act" stuff come out of certain people in the Bay, sure sounds like an attempt to temporarily seize control of the future without worrying too much about actual consent. Of course, the idea (or at least Yudkowsky's original vision for "coherent extrapolated volition") is that eventually the governing AIs will just implement what we all collectively want. And that could happen! But remember Lenin thought that the state would eventually "wither away" as Marx predicted, once the dictatorship of the proletariat had taken care of building industrial socialism...
Not to mention there are, shall we say, longtermism adjacent rich people like Musk and Thiel who seem pretty plausibly power-seeking, even if they are not really proper longtermists (or at least, they are not EAs).
(Despite all this, I should say that I think the in-principle philosophical case for longtermism is very strong. Alas, ideas can be both correct and dangerous.)
Yeah that's probably fair actually. This might make the view more sympathetic but not necessarily less dangerous. Maybe more dangerous, because most people will laugh you out the room if you say we need extreme measures to make sure we fill the galaxy with hedonium, but they will take 'extreme measures are needed or we might all die' rather more seriously.