Suppose we're sometime in the (near-ish) future. The longtermist project hasn't fulfilled 2020's expectations. Where did we go wrong? What scenarios (and with what probabilities) may have lead to this?
I hope this question isn't strictly isomorphic to asking about objections to long-termism.
Both of the possibilities below don't seem to be things that it would be that easy to realise even once we're in some (near-ish) future. I hope this isn't begging the question, it isn't intended to be. I've put credences and I'm glad you asked for them, but they are very uncertain.
One possibility is that we were just wrong about the whole long-termism thing. Given how much disagreement in philosophy there seems to be about basically everything, it seems prudent to give this idea non-trivial credence, even if you find arguments for long-termism very convincing. I'd maybe give a 10% probability to long-termism just being wrong.
More significant seems to be the chance that long-termism was right but that trying to directly intervene in the long-term future by taking actions that were only expected to have consequences in the long term was a bad strategy, and instead we should have been (approximate credence):
What about a scenario where long-termism turns out to be right, but there is some sort of community-level value drift which results in long-term cause areas becoming neglected, perhaps as a result of the community growing too quickly or some intra-community interest groups becoming too powerful? I wouldn't say this is very likely (maybe 5%), but we should consider the base rate of this type of thing happening.
I realise that this outcome might be subsumed in the the points raised above. Specifically, it might be that instead of directly trying to int... (read more)