Epistemic status: Just a thought that I have, nothing too rigorous
The reason Longtermism is so enticing (to me at least), is that the existence of so many future life hangs in the balance right now. It just seems to be a pretty good deed to me, to bring 10^52 people (or whatever the real number will turn out to be) into existence.
This hinges on the belief that Utility scales linearly with the number of QUALYs, so that twice as many people are also twice as morally valuable. My belief in this was recently shaken by this thought experiment:
***
You are a traveling EA on a trip to St. Petersburg. In a dark alley, you meet a Demon with the ability to create Universes and a serious gambling addiction. He says, he was about to create a universe with 10 happy people. But he gives you three fair dice and offers you a bet: You can throw the three dice and if they all come up 6, he refrains from creating a universe. If you roll anything else, he will double the number of people in the universe he will create.
You do the expected value calculation and figure out, that by throwing the dice you will create 696,8 QUALYs in expectation. You take the bet and congratulate yourself on your ethical decision.
After the good deed is done, and the demon has now committed to creating 20 happy people, he offers you the same bet again. Roll the 3 dice: he won't create a universe at 6,6,6 and doubles it at anything else. The demon tells you that he will offer you the same bet repeatedly. You do your calculations and throw the dice again and again, until, eventually, you throw all sixes, and the demon vanishes, without having to create any universe, in a cloud of sulfury mist and leaves you wondering if you should have done anything differently.
***
There are a few ways to weasel out of the demon's bet. You could say, that the strategy “always take the demons bet” has an expected value of 0 QUALYs, and so you should go with some tactic like “Take the first 20 bets, then call it a day”. But I think if you refuse a bet, you should be able to reject this bet without taking into account what bets you have taken in the past or are still taking in the future.
I think the only consistent way to refuse the Demons bets at some point is to have a bounded utility function. You might think it would be enough to have a utility function that does not scale linearly with the number of QUALYs, but logarithmically or something. But in that case, the demon can offer to double the amount of utility, instead of doubling the amount of QUALYs, and we are back in the paradox. At some point, you have to be able to say: “There is no possible universe that is twice as good as the one, you have promised me already”. So at some point, adding more happy people to the universe must have a negligible ethical effect. And once we accept that that must happen at some point, how confident are we, that 10^52 people are that much better than 8billion?
Overall I am still pretty confused about this subject and would love to hear more arguments/perspectives.
The cumulative EV of n decisions to roll repeatedly is
A:∑ni=0[u×2i×p]=up∑ni=02i
(where u is the initial utility of 10, and p stays constant at 63−163)
whereas the EV of committing to roll up to n times is
B:upn∑ni=02i
Which is much-much lower than A, as you point out.
But then again, for larger values of n, you're very unlikely to be allowed to roll for n times. The EV of n decisions to roll (A) times the probability of getting to the nth roll (pn−1), is
A×pn−1=(up∑ni=02i)×pn−1=upn∑ni=02i=B
In other words, A collapses to B if you don't assume any special luck. Which is to say that committing to a strategy has the same EV ex ante as fumbling into the same path unknowingly. This isn't very surprising, I suppose, but the relevancy is that if you have a revealed tendency to accept one-off St. Petersburg Paradox bets, that tendency has the same expected utility as deliberately committing to accept the same number of SPPs. If the former seems higher, then that's because your expectancy is wrong.
More generally, this means that it's important to try to evaluate one-off decisions as clues to what revealed decision rules you have. When you consider making a one-off decision, and that decision seems better than deliberately committing to using the decision-rules that spawned it, for all the times you expect to be in similar situation, then you are fooling yourself and you should update.
If you can predict that the cumulatively sum of the EV you assign to each one-off decision individually as you go along will be higher, compared to the EV you'd assign ex ante to the same string of decisions, then something has gone wrong in one your one-off predictions and you should update.
I've been puzzling over this comment from time to time, and this has been helpfwly clarifying for me, thank you. I've long been operating like this, but never entirely grokked why as clearly as now.