Hide table of contents

Survey

I think it would be interesting to know the confidence of EA Forum readers in the Repugnant Conclusion and each of its premises (see next section). For that purpose, I encourage you to fill this Form, where you can assign probabilities to each being true in principle. Feel free to comment below your rationale.

Please participate regardless of the extent to which you accept/reject the Repugnant Conclusion, such that the answers are as representative as possible. I recommend the following procedure:

  • Firstly, establish priors for your probabilities based solely on the description of the Repugnant Conclusion, its premises and your intuition.
  • Secondly, update your probabilities based on further investigation (e.g. reading this and this, and looking into the comments below).

You can edit your submitted answers as many times as you want[1]. I plan to publish the results in 1 to 2 months.

Repugnant Conclusion and its premises

“The Repugnant Conclusion is the implication, generated by a number of theories in population ethics, that an outcome with sufficiently many people with lives just barely worth living is better than an outcome with arbitrarily many people each arbitrarily well off”. In Chapter 8 of What We Owe to the Future (WWOF; see section “The Total View”), William MacAskill explains that it follows from 3 premises:

  • Dominance addition. “If you make everyone in a given population better off while at the same time adding to the world people with positive wellbeing, then you have made the world better”.
  • Non-anti-egalitarianism. “If we compare two populations with the same number of people, and the second population has both greater average and total wellbeing, and that wellbeing is perfectly equally distributed, then that second population is better than the first”.
  • Transitivity. “If one world is better than a second world, which itself is better than a third, then the first world is better than the third”.

Figure 8.7 of WWOF helps visualise the effect of the 2 1st premises via the following worlds:

  • World A:
    • “A world of ten billion people who all live wonderful lives of absolute bliss and flourishing”.
  • World A+:
    • “The ten billion people in A+ have even better lives than those in A, and the total population is larger: in A+ there are an additional ten billion people who have pretty good lives, though much less good than the other ten billion people’s”.
  • World B:
    • “In this world, there are the same number of people as in A+. But there is no longer any inequality; everyone has the same level of wellbeing. What’s more, in World B, the average and total wellbeing are greater than those of World A+”.

If dominance addition is true, A+ is better than A. If non-anti-egalitarianism is true, B is better than A+. Consequently, if transitivity is also true, B is better than A. This logic can be repeated indefinitely until arriving to a world Z where there is a very large population with very low positive wellbeing.

Accepting all of the premises does not imply accepting the total view, which “regards one outcome as better than another if and only if it contains greater total wellbeing, regardless of whether it does so because people are better off or because there are more well-off people”. For example, the Very Repugnant Conclusion does not follow from the above premises.

However, as each of the above premises is implied by the total view, rejecting at least one implies rejecting the total view.

My strongest objections to the premises

I believe all of the premises are close to indisputable, but my strongest objections are:

  • Dominance addition:
    • Critical range theory (see here): “adding an individual makes an outcome better to the extent that their wellbeing exceeds the upper end of a critical range, and makes an outcome worse to the extent that their wellbeing falls below the lower limit of the critical range”. Under these conditions, adding to the world people with sufficiently low positive wellbeing does not change the goodness of the world.
  • Non-anti-egalitarianism:
    • Increasing the mean wellbeing per being only makes the world better if the final mean wellbeing per being is sufficiently high. Consequently, a world with one being with very high wellbeing and N - 1 beings with wellbeing just just above zero may be better than a world with N beings with wellbeing just above zero.
  • All (including transitivity):

Acknowledgements

Thanks to Michael St. Jules, Pablo Stafforini, and Ramiro.

  1. ^

     Signing in to a Google account is required to do this, and ensure only one set of answers per person. If you do not have a Google account, and do not want to create one, I am happy to receive a private message with your answers.

Comments11
Sorted by Click to highlight new comments since: Today at 6:57 AM

Your objection against NAE is also raised by MichaelStJules in this post, which I have responded to. I completely accept NAE.

Rejecting transitivity is vulnerable to money pump arguments. So I completely accept transitivity.

I would reject mere addition. Mere addition implies that it would be moral to create someone if they experienced a year of extreme torture followed by just enough happiness to 'outweigh' the torture.

Welcome to the EA Forum!

I fully accept all the 3 premises of the Repugnant Conclusion, and also mere addition in principle. However, in practice, I would oppose bringing a human to life knowing he/she would endure significant torture afterwards, because I expect this would lead to a bad life if even the parents thought it was neet good (e.g. due to religious motives).

My main objections are:

  1. The Independence of Irrelevant Alternatives seems possibly false. Whether the move from A to A+ improves things depends on whether or not B is available. If B is available, then A>A+, and otherwise A<A+ (assuming no other options). The creation of the extra people in A+ creates further obligations if and only if there are options available in which they are better off. The extra people are owed more if and only if they could have been better off (possibly in a wide/nonidentity sense). This cuts against Dominance Addition and Transitivity, as stated, although note that rejecting IIA is compatible with transitivity within each option set, and DA in all option sets of size 2 and many others, just not all option sets.

  2. The aggregation required in Non-anti-egalitarianism seems possibly false. It would mean it’s better to bring someone down from a very high welfare life to a marginal life for a barely noticeable improvement to arbitrarily many other individuals. Maybe this means killing or allowing to die earlier to marginally benefit a huge number of people. Even from a selfish POV, everyone may be willing to pay the tiny cost for even a tiny and almost negligible chance of getting out of the marginal range (although they would need to be "risk-loving" or there would need to be value lexicality). The aggregation also seems plausibly false within a life: suppose each moment of your very long life is marginally good, except for a small share of excellent moments. Would you replace your excellent moments with marginally good ones to marginally improve all of the marginal ones and marginally increase your overall average and total welfare?

  3. The aggregation required in NAE depends on welfare being cardinally measurable on a common scale across all individuals so that we can take sums and averages, which seems possibly false.

  4. It's assumed there is positive welfare and net positive lives, contrary to negative axiologies, like antifrustrationism or in negative utilitariaism. Under negative axiologies, the extra lives are either already perfect, so the move from A+ to B doesn't make sense, or they are very bad to add in aggregate, and plausibly outweigh the benefits to the original population in A in moving to A+, so A>A+. (Of course, there may be other repugnant conclusions for negative axiologies.)

Also, I just remembered that the possibility of positive value lexicality means that these three axioms alone do not imply the RC, because Non-anti-egalitarianism wouldn't be applicable, as the unequal population with individuals past the lexical threshold has greater total and average welfare (in a sense, infinitely more). You need to separately rule out this kind of lexicality with another assumption (e.g. welfare is represented by (some subset of) the real numbers with the usual order and the usual operation of addition used to take the total and average), or replace NAE with something slightly different.

EDIT: Replaced "plausibly" with "possibly" to be clearer.

Re 2: Your objection to non-anti egalitarianism can easily be chalked up to scope neglect.

World A - One person with an excellent life plus 999,999 people with neutral lives.

World B - 1,000,000 people with just above neutral lives.

Let's use the veil of ignorance.

Would you prefer a 100% chance of a just above neutral life or a 1 in a million chance of an excellent life with a 99.9999% chance of a neutral life? I would definitely prefer the former.

Here is an alternative argument. 

Surely, it would be moral to decrease the wellbeing of a happy person from +1000 to +999 to make 1,000 neutral people 1 unit better off; rejecting this is outrageously implausible. 

If the process was repeated 1000 times, then it would be moral to bring a happy person down to neutrality to make a million neutral people 1 unit better off.

Re 2: I don't think the standard issues with pure aggregation can be appealed to in the strongest version of non-anti-egalitarianism, because you can add arbitrarily many steps so that in each step the pool of beneficiaries and harmed are equal in size, and the per individual stakes of the beneficiaries is greater. Transitivity generally does imply pure aggregation for reasons like this, so it seems like in this case you'd want to deny transitivity (or again, IIA) instead, or else you'll need to make a stronger and apparently costlier claim about how to trade off interests that isn't unique to pure aggregation.

Interesting! That argument seems right, if you rule out positive value lexicality, so that you can guarantee the beneficiaries can gain more than the harmed lose, and this can all be done in finitely many steps.

Sorry yeah, that was an unstated assumption of mine as well.

Thanks for all the detailed explanations!

Whether the move from A to A+ improves things depends on whether or not B is available.

In my mind, whether A+ is better than A only depends on the goodness of the difference between them:

  • 1) Increase in wellbeing for 10 M people.
  • 2) Additional 10 M people with positive wellbeing.

I think both 1) and 2) are good, so their combination is also good. Do you consider 1) good, but 2) neutral? If so, I would argue the combination of something good with something neutral is something good.

Would you replace your excellent moments with marginally good ones to marginally improve all of the marginal ones and marginally increase your overall average and total welfare?

I definitely would, but I can see why my intuition would push against that. For a typical life expectancy, and magnitude of excellent moments, marginally increasing the goodness of the marginally good moments would probably not be enough to increase total welfare.

The aggregation required in NAE depends on welfare being cardinally measurable on a common scale across all individuals so that we can take sums and averages, which seems plausibly false.

I agree a common scale is required. However, at a fundamental level, I believe wellbeing is a function of the laws of physics and elementary particles. So, as all humans are subject to the same laws of physics, and have very similar compositions, I expect there is a (very complicated unknown) welfare function which applies quite well to almost all individuals.

It's assumed there is positive welfare and net positive lives, contrary to negative axiologies, like antifrustrationism or in negative utilitariaism.

In 2 or 3 surveys in the UK, Kenya and Ghana, most people said they would prefer having lived their lives rather than not having been born (see Chapter 9 of What We Owe to the Future). Naturally, this does not mean that most people have net positive lives, but I would say it is strong evidence that it is possible to have net positive lives (e.g. I think there is at least 1 person in the world with a net positive life). If not in the present, at least in the future.

You need to separately rule out this kind of lexicality with another assumption (e.g. welfare is represented by (some subset of) the real numbers with the usual order and the usual operation of addition used to take the total and average), or replace NAE with something slightly different.

I agree, but I do not think that kind of lexicality is plausible. How can one produce something infinitely valuable from finite resources?

It might be interesting to ask for people's credence in the conclusion also!

Good point. Added!

I am very strongly attached to both dominance addition and non-anti-egalitarianism. If I was to reject a premise it would probably be transitivity, though I think there are very strong structural reasons to accept transitivity, as well as modestly strong principled ones (in all cases of intransitive values I am aware of, the thing you care about in a world depends on what other world it is being compared to, if this is necessarily the case for intransitive values, it requires a sort of extrinsic valuation that I am very averse to, and which is crucial to my strong acceptance of dominance addition as well).

A more realistic way I reject the repugnant conclusion in practice is on a non-ethical level. I don't think that my aversion to the repugnant conclusion (actually not nearly as strong as my aversion to other principled results of my preferred views, like those of pure aggregation) is sensitive to strong moral reasons at all. I think that rejection of many of these implications, for me, is more part of a different sort of project all together from principled ethics. I don't believe I can be persuaded of these conclusions by even the best principled arguments, and so I don't think there is any level on which I arrived at them because of such principles either.

I think many people find views like this to be a sort of cop-out, but I think arriving at moral rules in a way that more resembles running linear regressions on case-specific intuitions degrades an honest appreciation for both moral principles and my own psychology, I see no reason to expect there to be a sufficiently comfortable convergence between the best versions of the two.

Curated and popular this week
Relevant opportunities