Hide table of contents
9 min read 5

15

In this post, I will apply two methods used in dealing with moral uncertainty to a different problem, which is the one of choosing between radically divergent worldviews.

Imagine a young and altruistically motivated graduate. As she learns more, her circle of moral concern expands from people near her to all humans, and then to animals. At some point, she encounters arguments for prioritising the long-term future. Suddenly, all her previous efforts seem futile and it takes her mere seconds to decide that she wants to put her life in service of creating a good future. Just as she embarks on pursuing her completely re-ordered priorities, she halts for a second to contemplate the train of thought that led her thus far. She knows that not many people share her new priorities, and that most have radically different worldviews from hers. How certain can she be that she has not made a mistake on the way?

The above is an example of the kind of question discussed in the 80.000 hours podcast with Holden Karnofsky. We can roughly formulate the question as the one of how to split resources between worldviews that recommend very different courses of action. This kind of uncertainty, not about the claims of a certain theory, but about which of these theories we should choose, is similar to problems around moral uncertainty, like the ones discussed in Will MacAskill's thesis. What I will be doing in the following is to apply two of the methods used in that thesis to a concrete example of someone trying to figure out which actions to prioritise, given uncertainty as to which worldview is correct.

Worldviews, and where they diverge

One example that Holden discusses in the podcast is the comparison between helping humans and animals (alive today). So, for instance, we could help a human for $1,000 and a chicken for $1 and then face the decision which one to prioritise.

This comes down to questions of whether we value chickens, on what grounds we would value humans and not chickens, and how much exactly we should value chickens if we do value them. So, if you value chickens 10% as much as humans, you might want to decide in favour of chickens, and if you do not value chickens at all, helping the humans would be a much better decision.

We can think of these two outlooks as two different "worldviews". While I do not have a clear definition of that term (Holden himself says that the concept is "a very fuzzy idea"), relevant features of worldviews are:

  • Each yields particular outcomes (e.g. "prioritise X") that are very different from those of other worldviews.
  • These different outcomes might hinge on answering just one question differently (e.g. "Do chickens have moral value?"), or perhaps a few crucial questions.
  • Worldviews are constituted by a combination of views and beliefs that are not only normative, but also empirical, which makes it difficult to disentangle claims they make when we want to deal with uncertainty about worldviews (thanks to Max Daniel for pointing this out).

In the next section, I will discuss moral uncertainty, that is, uncertainties between first-order normative theories (such as Virtue Ethics, Hedonistic Utilitarianism, or Kantianism). This does not mean that I think that worldviews and normative theories are the same thing, even though I think we can apply measures of uncertainty to them in similar ways. I will also use the more neutral term "view", whenever I do not think it is important to distinguish the two terms.

 

The worldview split problem on an individual basis

Let's assume the fictitious graduate mentioned above (let's call her Lisa) has to pick a single option, for example in deciding which cause to devote her career to. When making this choice, it seems reasonable to not go for the option she likes most, but to also take into account how certain she is that each of the worldviews she derives the options from is correct.

This kind of thinking is what motivates the methods which Will uses in his thesis on normative uncertainty. He discusses examples in which agents have to weigh options which are ordered differently according to different normative theories, while being uncertain as to which of the theories is correct.

These decision-situations include a decision-maker at a specific point in time, who has at her disposal

  • a set of possible options (actions she can perform),
  • a set of theories that give an ordering of the options in terms of their choice-worthiness (whether, and sometimes, how much an option is preferred),
  • and a credence function that assigns each theory a subjective likelihood of being true.

This process is analogous to expected utility calculation, replacing "expected utility" with "expected choice-worthiness".

Below, I will apply these methods - ways of calculating and maximising expected choice-worthiness - to the case of different worldviews and see whether it is helpful in dealing with the kinds of "worldview split problems" discussed above.

 

Finding theories behind the worldviews

One major difficulty in assigning credences to worldviews is that often it is not clear what exactly constitutes a worldview. In this section, I explain why in the comparison further down, I am using the distinction between person-affecting and impersonal views to rank options. Feel free to skip to the next section if you don't want to hear all the details!

Person-affecting views capture the intuition that an act can be bad only if it is bad for someone, while impersonal views aim to maximize total (net) wellbeing.

The reason I searched for a different split in worldviews is that just attributing one worldview to each option made the whole prioritising framework uninformative: if our worldview is "I should care most about humans alive today" and one option is "do whatever is best for humans alive today", the framework's results will always be aligned with the worldview we assign highest credence to. If we assign this credence mainly based on intuition, we end up with what we would do if we just went for whatever we like best.

What I did to get around this was to note down some things someone might plausibly care about, and try to cluster them into more informative categories.

Things Lisa cares about:
  • Pursuing paths of action that produce good consequences.

  • Trying to cultivate care for others. This works well for other humans, somewhat for animals, but breaks down when we think about alien minds.

  • Capability to suffer as a very important aspect in who we assign moral value to.

  • Self-improvement: this includes virtues of good conduct in interaction with others, virtues like wisdom aimed at improving the way one makes decisions, and a general striving to become a better person. 

In addition, a flourishing world in which beings are happy, satisfy their preferences, and realise a variety of values seems like something worth caring about.

A lot of these values can be made use of in different worldviews: a commitment to bringing about the best possible consequences will vary according to which consequences are judged as best. A commitment to cultivating care will vary according to which beings we think of as ultimately "care-worthy" (other humans? animals? future beings?). For both, the person-affecting/impersonal distinction is relevant in deciding in which beings we should care about and which consequences we judge to be good. Capability to suffer seems like a less volatile criterion, which is uncertain still, but I'm hoping that this uncertainty is mainly empirical.

Looking at different virtues, there are some which which seem to favour a presentist perspective, centering around improving oneself and caring for others in the moment, and others which might favour long-termism (e.g. if we think that humanity as a whole is acting wisely by preventing its own extinction).

I have tried to show that just caring about producing good consequences or being virtuous does not give us an informative ranking of options, but that distinguishing between person-affecting and impersonal views does. I will apply this idea in the next section.

 

Applying this

Borda Rule

I started by ranking the options according to the Borda Rule, which Will uses in his thesis for theories that give merely ordinal choice-worthiness (that is, they give a preferred ranking, but do not provide information as to how much better or worse an option is) and where choice-worthiness is not comparable across theories.

Option person-affecting views (40% credence) impersonal views (60% credence)
(A) humans today 2 0
(B) animals today 1 1
(C) long-term future 0 2

 

The above table gives an ordering of preferences which seems roughly plausible to me, but might vary according to whether we take "person-affecting" to be about humans or including animals, and how we deal with the future. For example, we might think only about persons that already exist today, or additionally consider people who would have come into existence regardless of our intervention.

Multiplying the Borda score with the credence Lisa assigns to person-affecting and impersonal views, we get the following:

(A): 0.4*2+0.6*0= 0.8

(B): 0.4*1+0.6*1= 1

(C): 0.4*0+0.6*2= 1.2

We see that under this calculation, Option (C) - prioritizing the long-term future - scores highest. If we are very uncertain about which view is best and assign a 50% credence to each, we get:

(A): 0.5*2+0.5*0= 1

(B): 0.5*1+0.5*1= 1

(C): 0.5*0+0.5*2= 1

This makes it seem like, in this case, a merely ordinal ranking is still not better than just going with a simple intuition. If we add the assumption that choice-worthiness can be compared across different theories, and we think we can say something about how much more desirable one outcome is over the other, we can try our luck with cardinal ordering and see whether that is more informative.

Cardinal Ordering

For example, we might assign the following choice-worthiness to the different options under the two theories we are considering:

Option person-affecting views (40% credence) impersonal views (60% credence)
(A) humans today 0.7 0.05
(B) animals today 0.2 0.05
(C) long-term future 0.1 0.9

 

Credence-weighted, we get the following results across the two views:

(A) = 0.4*0.7+0.6*0.05= 0.31

(B) = 0.4*0.2+0.6*0.05= 0.11

(C) = 0.4*0.1+0.6*0.9= 0.58

Where do I get these numbers from? They are based on my intuition that under person-affecting views, humans alive today are considered as more important than either animals and the long-term future, while for impersonal views, the long-term future outweighs both current human and animal welfare by a lot. I was surprised to see such a low score assigned to animal welfare, since I intuitively think that animal welfare matters a lot. What the low score reflects is my (until now implicit!) assumption that person-affecting views really only deal with persons, excluding non-human animals. It also shows that I think person-affecting views only care about persons alive today, which is a strand of person-affecting views that differs from "caring only about people who would have existed anyway".

If we do think that there is some merit in the intuitive cardinal ordering I provided, this ordering still gives the same result for a 50-50 credence:

(A) = 0.5*0.7+0.5*0.05= 0.375

(B) = 0.5*0.2+0.5*0.05= 0.125

(C) = 0.5*0.1+0.5*0.9= 0.5

Option A would only be favoured if we assign 60% credence or more to person-affecting views being right (yielding a score of 0.44 versus option C scoring 0.42). This means that if Lisa gathered evidence that made her assign a credence of 60% or more to person-affecting views, she would have to not prioritise working on the long-term future, but on helping humans alive today.

Problems

I am aware of the assigned credences seeming somewhat random, in only reflecting the subjective likelihood that someone assigns to a certain view being true. I am also unsure about whether the orderings I ascribed to different views are correct interpretations of these views. However, I hope that having ascribed numbers to different orderings might facilitate discussion later on, making assumptions more tangible and thereby easier to challenge. The extent to which we should be bothered about this problem may also depend on what we would hope to use this method for.

As I mentioned above, worldviews encompass lots of different claims which are hard to disentangle. I tried to get around this by turning this into a question about person-affecting versus impersonal views, but very likely lost something along the way. This is why I hope that someone else will find this a useful prompt to propose a better way of thinking about this question.

Conclusion

Was this useful? /What did I learn? / What will I do as a result?

I am probably biased in favour of thinking that writing this post was useful because I invested time in doing so. Other reasons why this felt like a useful pursuit are that it was much more productive than just spending time reading on my own, since setting myself a deadline increased the amount I read and having a question in mind made my reading more focused and engaging.

As for other people, I hope this might provide a model that others can use to make explicit their thinking about their priorities and their uncertainties therein. Other benefits might lie in motivating people to engage with the question at large, or in prompting someone to see flaws in what I was doing and do it better in the future.

I learnt what my intuitive models of different moral theories look like and now have a much clearer grasp of where they are not as clear as I would hope them to be. I also enjoyed understanding what information might change my personal cause prioritisation. One immediate result of writing this is that I have a much better sense of which things I want to learn more about because doing so might affect my future behaviour.

If you have any answers to the questions and problems raised above, or anything else you would like to add, please do leave a comment!

Comments5


Sorted by Click to highlight new comments since:

Hello Ronja. I think it would be helpful to know if you think this is different from other approaches to moral uncertainty and, if so, which ones. I don't know if you take yourself to be doing something novel or providing an example using an existing theory.

Hi, thanks for your comment :) Seems like I should have made that clearer! Since what I'm doing is applying Will's approach, the approach is not itself new. I haven't seen it discussed with regards to the worldview-split problem, but since I ended up condensing different "worldviews" into a decision between two theories, it turned out to be basically the same (which is not without problem, for that matter). I still found it valuable to try out this process in practice, and since I am expecting many people to not have read Will's thesis, I hoped this would provide them with an example of such a process. One person told me they found it valuable to use this way of thinking for themselves, and someone else said they were more inclined to read the actual thesis now, so I think there is some value in this article, and the issue might be more about the way I'm framing it. If you have an idea for a framing you would have found more useful, I'd be happy to know. Do you think just adding a sentence or two at the start of the article might do?

Yeah, I think it would be good to put the research in context - true for posts here as other pieces of work - so readers know what sort of hat they should be wearing and if this is relevant for them.

So I skimmed this and it looks like you are basically just applying MacAskill's method. Did I miss something?

Btw, whether to assign ordinal or cardinal scores to things isn't really something that you should do in the context of normative uncertainty. It should come from the moral theory itself, and not be altered by considerations of uncertainty. If the moral theory has properties that allow us to model it with a cardinal ranking, then we do that, and if it doesn't then we use an ordinal ranking. One moral theory may have ordinal rankings and another may have cardinal ones. By the way, as far as MEC is concerned, an ordinal moral ranking is just a special case of cardinal moral rankings where the differences between consecutively ranked options are uniform.

Interesting - so then interventions that do well on both long-term future and humans today like AI and alternate foods would do very well by your numbers.

Curated and popular this week
 ·  · 23m read
 · 
Or on the types of prioritization, their strengths, pitfalls, and how EA should balance them   The cause prioritization landscape in EA is changing. Prominent groups have shut down, others have been founded, and everyone is trying to figure out how to prepare for AI. This is the first in a series of posts examining the state of cause prioritization and proposing strategies for moving forward.   Executive Summary * Performing prioritization work has been one of the main tasks, and arguably achievements, of EA. * We highlight three types of prioritization: Cause Prioritization, Within-Cause (Intervention) Prioritization, and Cross-Cause (Intervention) Prioritization. * We ask how much of EA prioritization work falls in each of these categories: * Our estimates suggest that, for the organizations we investigated, the current split is 89% within-cause work, 2% cross-cause, and 9% cause prioritization. * We then explore strengths and potential pitfalls of each level: * Cause prioritization offers a big-picture view for identifying pressing problems but can fail to capture the practical nuances that often determine real-world success. * Within-cause prioritization focuses on a narrower set of interventions with deeper more specialised analysis but risks missing higher-impact alternatives elsewhere. * Cross-cause prioritization broadens the scope to find synergies and the potential for greater impact, yet demands complex assumptions and compromises on measurement. * See the Summary Table below to view the considerations. * We encourage reflection and future work on what the best ways of prioritizing are and how EA should allocate resources between the three types. * With this in mind, we outline eight cruxes that sketch what factors could favor some types over others. * We also suggest some potential next steps aimed at refining our approach to prioritization by exploring variance, value of information, tractability, and the
 ·  · 5m read
 · 
[Cross-posted from my Substack here] If you spend time with people trying to change the world, you’ll come to an interesting conundrum: Various advocacy groups reference previous successful social movements as to why their chosen strategy is the most important one. Yet, these groups often follow wildly different strategies from each other to achieve social change. So, which one of them is right? The answer is all of them and none of them. This is because many people use research and historical movements to justify their pre-existing beliefs about how social change happens. Simply, you can find a case study to fit most plausible theories of how social change happens. For example, the groups might say: * Repeated nonviolent disruption is the key to social change, citing the Freedom Riders from the civil rights Movement or Act Up! from the gay rights movement. * Technological progress is what drives improvements in the human condition if you consider the development of the contraceptive pill funded by Katharine McCormick. * Organising and base-building is how change happens, as inspired by Ella Baker, the NAACP or Cesar Chavez from the United Workers Movement. * Insider advocacy is the real secret of social movements – look no further than how influential the Leadership Conference on Civil Rights was in passing the Civil Rights Acts of 1960 & 1964. * Democratic participation is the backbone of social change – just look at how Ireland lifted a ban on abortion via a Citizen’s Assembly. * And so on… To paint this picture, we can see this in action below: Source: Just Stop Oil which focuses on…civil resistance and disruption Source: The Civic Power Fund which focuses on… local organising What do we take away from all this? In my mind, a few key things: 1. Many different approaches have worked in changing the world so we should be humble and not assume we are doing The Most Important Thing 2. The case studies we focus on are likely confirmation bias, where
 ·  · 1m read
 · 
I wanted to share a small but important challenge I've encountered as a student engaging with Effective Altruism from a lower-income country (Nigeria), and invite thoughts or suggestions from the community. Recently, I tried to make a one-time donation to one of the EA-aligned charities listed on the Giving What We Can platform. However, I discovered that I could not donate an amount less than $5. While this might seem like a minor limit for many, for someone like me — a student without a steady income or job, $5 is a significant amount. To provide some context: According to Numbeo, the average monthly income of a Nigerian worker is around $130–$150, and students often rely on even less — sometimes just $20–$50 per month for all expenses. For many students here, having $5 "lying around" isn't common at all; it could represent a week's worth of meals or transportation. I personally want to make small, one-time donations whenever I can, rather than commit to a recurring pledge like the 10% Giving What We Can pledge, which isn't feasible for me right now. I also want to encourage members of my local EA group, who are in similar financial situations, to practice giving through small but meaningful donations. In light of this, I would like to: * Recommend that Giving What We Can (and similar platforms) consider allowing smaller minimum donation amounts to make giving more accessible to students and people in lower-income countries. * Suggest that more organizations be added to the platform, to give donors a wider range of causes they can support with their small contributions. Uncertainties: * Are there alternative platforms or methods that allow very small one-time donations to EA-aligned charities? * Is there a reason behind the $5 minimum that I'm unaware of, and could it be adjusted to be more inclusive? I strongly believe that cultivating a habit of giving, even with small amounts, helps build a long-term culture of altruism — and it would