Hide table of contents

Problem areas beyond 80,000 Hours' current priorities mentions "Broadly promoting positive values".

I have some some questions:

What are the values that are needed to further EA's interests?

Where (in which cultures or areas of culture at large) are they deficient, or where might they become deficient in the future?

Problem areas... mentions "altruism" and "concern for other sentient beings". Maybe those are the two that EA is most essentially concerned with. If so, what are the support values needed for maximizing those values?




New Answer
New Comment

5 Answers sorted by

If so, what are the support values needed for maximizing those values?

I think a healthy dose of moral uncertainty (and normative uncertainty in general) is really important to have, because it seems pretty easy for any ethical/social movement to become fanatical or to incur a radical element, and end up doing damage to itself, its members, or society at large. ("The road to hell is paved with good intentions" and all that.)

A large part of what I found attractive about EA is that its leaders emphasize normative uncertainty so much in their writings (starting with Nick Bostrom back in 2009), but perhaps it's not "proselytized" as much as it should be day-to-day.

At its heart, EA seems to naturally tend to promote a few things:

  • a larger moral circle is better than a smaller one
  • considered reasoning ("rationality") is better than doing things for other reasons alone
  • efficiency in generating outcomes is better than being less efficient, even if it means less appealing at an emotional level

I don't know that any of this are what EA should promote, and I'm not sure there's anyone who can unilaterally make the decision of what is normative for EA, so instead I offer these as the norms I think EA is currently promoting in fact, regardless of what anyone thinks EA should be promoting.

Not exactly answering your question, but I think EA has really good communication norms--such as steelmanning your opponent, focusing on empiricism, open discussion, double crux, no persona attacks, etc.

I do think the broader society can benefit significantly in discussing thorny topics (i.e. politics) if they adopted these communication norms.

This is an interesting question. 

One possible value is something like intrinsically valuing Truth or Better Reasoning. Perhaps also something like Productivity/Maximisation. The rationality community is perhaps a good example of promoting such values (explicitly here). 

It feels somewhat double-edged to promote instrumental values. This can cause all types of troubles if it's misinterpreted or too successful.

What do you think are the important values? 

I'm basically an outsider to EA, but "from afar", I would guess that some of the values of EA are 1) against politicization, 2) for working and building rather than fighting and exposing ("exposing" being "saying the unhealthy truth for truth's sake", I guess), 3) for knowing and self-improvement (your point), 4) concern for effectiveness (Gordon's point). And of course, the value of altruism.

These seem like they are relatively safe to promote (unless I'm missing something.)

Altruism is composed of 1) other... (read more)

My interpretation of this was promoting robustly good values (e.g. violence is bad) at scale as an effective intervention.

For instance, these are values that the UK government tries to promote:

"Champion democracy, human rights and the rule of law, and address global challenges, including through campaigns on preventing sexual violence in conflict, reducing modern slavery and promoting female education. Promote human and environmental security through London Illegal Wildlife Trade Conference. Deepen relationships between states and people, including through the Commonwealth Summit."


A few free ideas occasioned by this:

1. The fact that this is a government paper makes me think of "people coming together to write a mission statement." To an extent, values are agreed-upon by society, and it's good to bear that in mind. (Working with widespread values instead of against them, accepting that to an extent values are socially-constructed (or aren't, but the crowd could be objectively right and you wrong) and adjusting to what's popular instead of using a lot of energy to try to change things.)

2. My first reaction ... (read more)

Curated and popular this week
Relevant opportunities