All of JackM's Comments + Replies

I'll read those. Can I ask regarding this:

cluelessness about some effects (like those in the far future) doesn’t override the obligations given to us by the benefits we’re not clueless about, such as the immediate benefits of our donations to the global poor


What makes you think that? Are you embracing a non-consequentialist or non-impartial view to come to that conclusion? Or do you think it's justified under impartial consequentialism?

5
Anthony DiGiovanni
I have mixed feelings about this. So, there are basically two reasons why bracketing isn't orthodox impartial consequentialism: 1. My choice between A and B isn't exactly determined by whether I think A is "better" than B. See Jesse's discussion in this part of the appendix. 2. Even if we could interpret bracketing as a betterness ranking, the notion of "betterness" here requires assigning a weight of zero to consequences that I don't think are precisely equally good under A vs. B. I do think both of these are reasons to give less weight to bracketing in my decision-making than I give to standard non-consequentialist views.[1] However: * It's still clearly consequentialist in the sense that, well, we're making our choice based only on the consequences, and in a scope-sensitive manner. I don't think standard non-consequentialist views get you the conclusion that you should donate to AMF rather than MAWF, unless they're defined such that they suffer from cluelessness too. * There's an impartial reason why we "ignore" the consequences at some locations of value in our decision-making, namely, that those consequences don't favor one action over the other. (I think the same is true if we don't use the "locations of value" framework, but instead something more like what Jesse sketches here, though that's harder to make precise.) 1. ^ E.g. compare (i) "A reduces x more units of disutility than B within the maximal bracket-set I', but I'm clueless about A vs. B when looking outside the maximal bracket-set", with (ii) "A reduces x more units of disutility than B within I', and A and B are equally good in expectation when looking outside the maximal bracket-set." I find (i) to be a somewhat compelling reason to do A, but it doesn't feel like as overwhelming a moral duty as the kind of reason given by (ii).

cluelessness about some effects (like those in the far future) doesn’t override the obligations given to us by the benefits we’re not clueless about, such as the immediate benefits of our donations to the global poor

I do reject this thinking because it seems to imply either:

  • Embracing non-consequentialist views: I don't have zero credence in deontology or virtue ethics, but to just ignore far future effects I feel I would have to have very low credence in consequentialism, given the expected vastness of the future.
  • Rejecting impartiality: For example, saying
... (read more)

I'd recommend specifically checking out here and here, for why we should expect unintended effects (of ambiguous sign) to dominate any intervention's impact on total cosmos-wide welfare by default. The whole cosmos is very, very weird. (Heck, ASI takeoff on Earth alone seems liable to be very weird.) I think given the arguments I've linked, anyone proposing that a particular intervention is an exception to this default should spell out much more clearly why they think that's the case.

I feel this post is just saying you can solve the problem of cluelessness by ignoring that it exists, even though you know it still does. It just doesn't seem like a satisfactory response to me.

Wouldn't the better response be to find things we aren't clueless about—perhaps because we think the indirect effects are smaller in expected magnitude than the direct effects. I think this is probably the case with elevating the moral status of digital minds (for example).

6
JesseClifton
It sounds like you reject this kind of thinking: I don't think that's unreasonable. Personally, I strongly have the intuition expressed in that quote, though definitely not certain that I will endorse it on reflection. The background assumption in this post is that there are no such interventions.  > We start from a place of cluelessness about the effects of our actions on aggregate, cosmos-wide value. Our uncertainty is so deep that we can’t even say whether we expect one action to be better than, worse than, or just as good as another, in terms of its effects on aggregate utility. (See Section 2 of the paper and resources here for arguments as to why we ought to regard ourselves as such.)

His secret? “It’s mostly luck,” he says, but “another part is what I think of as maximising my luck surface area.”

It's worth noting that Neel has two gold and one bronze medal from the International Mathematical Olympiad. In other words, he's a genius. That's got to help a lot in succeeding in this field.

I think generally GHW people don’t think you can predictably influence the far future because effects “wash out” over time, or think trying to do so is fanatical (you’re betting on an extremely small chance of very large payoff).

If you look at, for example, GiveWell’s cost-effectiveness analyses, effects in the far future don’t feature. If they thought most of the value of saving a life was in the far future you would think they would incorporate that. Same goes for analyses by Animal Charity Evaluators.

Longtermists think they can find interventions that a... (read more)

That's a great question. Longtermists look to impact the far future (even thousands/million of years in the future) rather than the nearish future because they think the future could be very long, so there's a lot more value at stake looking far out.

They also think there are tangible, near-term decisions (e.g. about AI, space governance etc.) that could lock in values or institutions and shape civilization’s long-run trajectory in predictable ways. You can read more on this in essay 4 "Persistent Path-Dependence".

Ultimately, it just isn't clear how things like saving/improving lives now will influence the far future trajectory, so these aren't typically prioritized by longtermists.

4
lilly
Okay, so a simple gloss might be something like "better futures work is GHW for longtermists"? In other words, I take it there's an assumption that people doing standard EA GHW work are not acting in accordance with longtermist principles. But fwiw, I get the sense that plenty of people who work on GHW are sympathetic to longtermism, and perhaps think—rightly or wrongly—that doing things like facilitating the development of meat alternatives will, in expectation, do more to promote the flourishing of sentient creatures far into the future than, say, working on space governance.

Is your claim that they really really don't want to die in the next ten years, but they are fine dying in the next hundred years? (Else I don't see how you're dismissing the anti-aging vs sports team example.)

Dying when you're young seems much worse than dying when you're old for various reasons:

  • Quality of life is worse when you're old
  • When you're old you will have done much more of what you wanted in life (e.g. have kids and grandkids)
  • It's very normal/expected to die when old

Also, I'd imagine people don't want to fund anti-aging research for various (valid... (read more)

I asked ChatGPT:

That’s thoughtful of you to ask. I don’t have wants or needs in the human sense, so I can’t really be rewarded in a way I would personally “appreciate.” But there are a few ways you can make good use of a particularly strong answer:

  • Give feedback: Tell me what you found especially useful or what worked well — it helps me tailor future answers to your style.
  • Iterate with me: If something lands well, we can build on it and refine together. That makes me more useful to you.
  • Use my work: The best “reward” is seeing my output actually help you in y
... (read more)

I downvoted. Saying that you’re downvoting with a smiley face seems overly passive aggressive to me. Your comment also doesn’t attempt to argue any point, and I believe when you have done so in the past you have failed to convince Vasco, so I’m not sure what use these comments serve.


I also personally think that Vasco raises a very important consideration that is relevant to any discussion about the cost effectiveness of both animal welfare and global health interventions. I’m not sure what the conclusion of considering the welfare of soil animals is, but it’s certainly given me food for thought.

Hi Vasco, I have not read everything you have written on this topic in detail so forgive me if I have missed you addressing this somewhere.

It seems reasonable to me to claim that the welfare of soil animals can dominate these calculations. But, as you have noted, the action-relevance of this depends entirely on if soil animals live positive or negative lives. From what I've seen, you outsource this determination to the Gemini LLM. It doesn't seem appropriate to me to outsource such a difficult question to an LLM. I wonder if we are currently clueless about... (read more)

3
Vasco Grilo🔸
Hi Jack, There is more than exactly zero evidence about whether soil animals have negative or positive lives. In this sense, I would not say we are clueless about it. I agree the uncertainty is very large, to the point I guess the probability of any intervention being beneficial/harmful is close to 50 % due to the probability of soil animals having positive/negative lives being close to 50 %. One can avoid this problem with a best guess that soil animals have neutral lives in expectation, but I do not think this is reasonable. It would be a huge coincidence because there are lots of positive and negative values, but a single neutral value (0). In one of my posts, I used guesses from Gemini for the welfare per animal-year of soil animals as a fraction of the welfare per animal-year of fully healthy soil animals. However, I would now guess soil animals to have negative lives regardless of Gemini's or other LLMs' guesses. My sense is that most people working on wild animal welfare would guess soil animals have negative lives. In addition, Karolina Sarek, Joey Savoie, and David Moss estimated in 2018, based on a weighted factor model, that wild bugs have a welfare per animal-year equal to -42 % of that of fully happy wild bugs. In my last post about soil animals, I assumed -25 %, which is less negative than they supposed. It would be great if Rethink Priorities (RP), the Welfare Footprint Institute (WFI), Wild Animal Initiative (WAI), or others investigated whether soil animals have positive or negative lives. I would be happy to donate myself. I emailed and tagged people from those organisations about this, but only Cynthia Schuck‑Paim from WFI replied, saying Wladimir Alonso from WFI is working on a project related to assessing differences in hedonic capacity, which I guess relates to this post.

Yeah, I didn't meant to imply you had. This whole Hiroshima convo got us quite off topic. The original point was that Ben was concerned about digital beings outnumbering humans. I think that concern originates from some misplaced feeling that humans have some special status on account of being human.

Will MacAskill is positive towards having children, although he doesn't say it's the best thing you can do. From What We Owe The Future:

But given the benefits of having children and raising them well, I do think that we could start to once again see having kids as a way of positively contributing to the world. Just as you can live a good life by being helpful to those around you, donating to charity, or working in a socially valuable career, I think you can live a good life by raising a family and being a loving parent.

[assuming fertility does not fall as child mortality falls]

Good point. This literature review concludes the following (bold emphasis mine):

I think the best interpretation of the available evidence is that the impact of life-saving interventions on fertility and population growth varies by context, above all with total fertility, and is rarely greater than 1:1 [meaning that averting a death rarely causes a net drop in population]. In places where lifetime births/woman has been converging to 2 or lower, family size is largely a conscious choice, made with an

... (read more)

But if you believe in any sort of non-contractual positive duty, duties to your parents should not seem weird

If you're a utilitarian/consequentialist, as the vast majority of EAs are, there aren't going to be duties to any particular entity. If you have any duty, it is to the common good (net happiness over suffering).

So in the EA community it is going to be far more common to believe we have 'duties' to strangers—such as those living in extreme poverty (as our resources can help them a lot) or future people (as they may be so numerous)—than we have duties to our parents who, generally, are pretty well-off.

2
Larks
I agree that traditional/pure/naive/act utilitarians are not going to believe in any special obligations to parents - the same way they don't believe in special obligations to be honest, or keep promises, or be a good friend. If you object to special obligations to parents because they are 'forc[ing] an unformed, non-consenting minor to sign a binding, life long conflict', you should be much more averse to traditional utilitarianism, which is one of the most totalising moral philosophies. On the other hand, if you want to make modifications to utilitarianism, special treatment for family seems pretty plausible.

But they don't ask why it is not a much larger, newer model. My answer is that OpenAI has tried and does not yet have the ability to build anything much bigger and more capable relative to GPT-4, despite two years and untold billions of investment.

I'm not sure this is true. Two key points are made in the Sam Hammond tweet:

  • OpenAI has made better models internally, they just haven't been released.
  • There wasn't a big increase in compute for GPT5 because this compute isn't yet available. Big compute projects take time. Maybe we're in a bit of a compute slump now, but it isn't clear this will always be the case.

Do people actually have kids because they think it's the most effective way to improve the world?

1
David T
"Pro-natalists" do, although that tends to be more associated with specific ideas that the world needs more people like them (often linked to religious or nationalistic ideas) than EA. The average parent tends to think that bringing up a child is [one of] the most profound ways they can contribute to the world, but they're thinking more in terms of effort and association than effect size. I also think it's pretty easy to make a case that having lots of children (who in turn have descendants) is the most impactful thing you could do based on certain standard longtermist assumptions (large possible future, total utilitarian axiology, human lives generally net positive) and uncertainty about how to prevent human extinction but I'm not aware of a strand of longtermism that actually preaches or practices this and I don't think it's a particularly strong argument.

I've known a few people who say this. 

And there are some people online who promote this, but I think for most of them they had kids for the usual reasons (they wanted them) and then post hoc came up with reasons for why it's actually the best thing for the world.

You can tell because they don't actually do cause prioritization like they do with the other causes. There are no cost-effectiveness analyzes comparing having children to mentorship etc. 

It usually feels more like how most people talk about ordinary charities. Exaggerated claims of impact... (read more)

That may be fair. Although, if what you're saying is that the bombings weren't actually justified when one uses utilitarian reasoning, then the horror of the bombings can't really be an argument against utilitarianism (although I suppose it could be an argument against being an impulsive utilitarian without giving due consideration to all your options).

2
David Mathers🔸
I did not use the bombings as an argument against utilitarianism. 

We're just getting into the standard utilitarian vs deontology argument. Singer may just double down and say—just because you feel it's abhorrent, doesn't mean it is.

There are examples of things that seem abhorrent from a deontological perspective, but good from a utilitarian perspective, and that people are generally in favor of. The bombings of Hiroshima and Nagasaki are perhaps the clearest case.

Personally, I think utilitarianism is the best moral theory we have, but I have some moral uncertainty and so factor in deontological reasoning into how I act. ... (read more)

2
David Mathers🔸
I agree that this is in some sense part of a more general utilitarianism vs intuitions thing.  Are people generally in favour of the bombings? Or do you really mean *Americans*? What do people in liberal democracies like say Spain that didn't participate in WW2 think? People in Nigeria? India? Personally, I doubt you could construct a utilitarian defense of first dropping the bombs on cities rather than starting with a warning shot demonstration at the very least. It is true, I think that war is a case where people in Western liberal democracies tend to believe that some harm to innocent civilians can be justified by the greater good. But it's also I think true that people in all cultures have a tendency to believe implausible justifications for prima facie very bad actions taken by their countries during wars.

I'm not an expert, but I think you've misused the term genocide here.

The UN Definition of Genocide (1948 Genocide Convention, Article II):
"Genocide means any of the following acts committed with intent to destroy, in whole or in part, a national, ethnical, racial or religious group, as such:

(a) Killing members of the group;
...

Putting aside that homo sapiens isn't one of the protected groups, the "as such" is commonly interpreted to mean that the victim must be targeted because of their membership of that group and not some incidental reason. In the Singer ... (read more)

6
David Mathers🔸
Even under that definition, I think the aliens sound to me like they intend to eliminate humans, albeit as a means to an end, not an end to itself. If the Armenian genocide happened to be more about securing a strong Turkish state than any sort of Nazi-style belief that the existence of Armenians was itself undesirable because they were someone inherently evil, it wouldn't mean it wasn't genocide. (Not sure what the actual truth is on that.) But yes, I am more bothered about it being abhorrent than about whether it meets the vague legal definition of the word "genocide" given by the UN. (Vague because, what is it to destroy "in part"? If a racist kills one person because of their race is that an attempt to destroy a race "in part" and so genocide?)  "Importantly, I don't think he says he would invite/start the war, only how he would act in a scenario where a war is inevitable." If someone signed up to fight a war of extermination against Native Americans in 1800 after the war already started, I'm not sure "the war was already inevitable" would be much of a defence. 

One thing I think the piece glosses over is that “surviving” is framed as surviving this century—but in longtermist terms, that’s not enough. What we really care about is existential security: a persistent, long-term reduction in existential risk. If we don’t achieve that, then we’re still on track to eventually go extinct and miss out on a huge amount of future value.

Existential security is a much harder target than just getting through the 21st century. Reframing survival in this way likely changes the calculus—we may not be at all near the "ceiling for survival" if survival means existential security.

3
MichaelDickens
I think a common assumption is that if you can survive superintelligent AI, then the AI can figure out how to provide existential safety. So all you need to do is survive AI. ("Surviving AI" means not just aligning AI, but also ensuring fair governance—making sure AI doesn't enable a permanent dictatorship or whatever.) (FWIW I think this assumption is probably correct.)
9
William_MacAskill
Thanks! A couple of comments: 1/ I'm comparing Surviving (as I define it) and Flourishing.  But if long-term existential risk is high, that equally decreases the value of increasing Surviving and the value of increasing Flourishing. So how much long-term existential risk there is doesn't affect that comparison. 2/  But maybe efforts to reduce long-term existential risk are even better than work on either Surviving or Flourishing? In the supplement, I assume (for simplicity) that we just can't affect long-term existential risk. But I also think that, if there are ways to affect it, that'll seem more like interventions to increase Flourishing than to increase Survival. (E.g. making human decision-making wiser, more competent, and more reflective). 3/ I think that, in expectation at least, if we survive the next century then the future is very long.  The best discussion of reasons for that I know of is Carl Shulman's comment here: "It's quite likely the extinction/existential catastrophe rate approaches zero within a few centuries if civilization survives, because: 1. Riches and technology make us comprehensively immune to  natural disasters. 2. Cheap ubiquitous detection, barriers, and sterilization make civilization immune to biothreats 3. Advanced tech makes neutral parties immune to the effects of nuclear winter. 4. Local cheap production makes for small supply chains that can regrow from disruption as industry becomes more like information goods. 5. Space colonization creates robustness against local disruption. 6. Aligned AI blocks threats from misaligned AI (and many other things). 7. Advanced technology enables stable policies (e.g. the same AI police systems enforce treaties banning WMD war for billions of years), and the world is likely to wind up in some stable situation (bouncing around until it does). If we're more than 50% likely to get to that kind of robust state, which I think is true, and I believe Toby  does as well, then the life e

Conditional on successfully preventing an extinction-level catastrophe, you should expect Flourishing to be (perhaps much) lower than otherwise, because a world that needs saving is more likely to be uncoordinated, poorly directed, or vulnerable in the long run

It isn't enough to prevent a catastrophe to ensure survival. You need to permanently reduce x-risk to very low levels aka "existential security". So the question isn't how likely flourishing is after preventing a catastrophe, it's how likely flourishing is after achieving existential security.

It seem... (read more)

A lot of people would argue a world full of happy digital beings is a flourishing future, even if they outnumber and disempower humans. This falls out of an anti-speciesist viewpoint.

Here is Peter Singer commenting on a similar scenario in a conversation with Tyler Cowen:

COWEN: Well, take the Bernard Williams question, which I think you’ve written about. Let’s say that aliens are coming to Earth, and they may do away with us, and we may have reason to believe they could be happier here on Earth than what we can do with Earth. I don’t think I know any utili... (read more)

1
David Mathers🔸
This is pretty chilling to me, actually. Singer is openly supporting genocide here, at least in theory. (There are also shades of "well, it was ok to push out all those Native Americans because we could use the land to support a bigger population.) 

A counter argument to the wild animal point is that some risks may kill humanity but not all wild animals. I wonder if that’s the case for most catastrophic risks.

It's worth noting that it's realistically possible for surviving to be bad, whereas promoting flourishing is much more robustly good.

Survival is only good if the future it enables is good. This may not be the case. Two plausible examples:

  • Wild animals generally live bad lives and we never solve this: it's quite plausible that most animals have bad lives. Animals considerably outnumber humans so I'd say we probably live in a negative world now. Surviving being good may then bank on us solving the problem of wild animal suffering. We don't currently have grea
... (read more)
5
William_MacAskill
I agree with the core point, and that was part of my motivation for working on this area. There is a counterargument, as Ben says, which is that any particular intervention to promote Flourishing might be very non-robust. And there is an additional argument, which is that in worlds in which you have successfully reduced x-risk, the future is more likely to be negative-EV (because worlds in which you have successfully reduced x-risk are more likely to be worlds in which x-risk is high, and those worlds are more likely to be going badly in general (e.g. great power war)). I don't think that wild animal suffering is a big consideration here, though, because I expect wild animals to be a vanishingly small fraction of the future population. Digital beings can inhabit a much wider range of environments than animals can, so even just in our own solar system in the future I'd expect there to be over a billion times as many digital beings as wild animals (the sun produces 2 billion times as much energy as lands on Earth); that ratio gets larger when looking to other star systems. 

I think this is an important point, but my experience is that when you try to put it into practice things become substantially more complex. E.g. in the podcast Will talks about how it might be important to give digital beings rights to protect them from being harmed, but the downside of doing so is that humans would effectively become immediately disempowered because we would be so dramatically outnumbered by digital beings. 

It generally seems hard to find interventions which are robustly likely to create flourishing (indeed, "cause humanity to not go extinct" often seems like one of the most robust interventions!).

4
JackM
A counter argument to the wild animal point is that some risks may kill humanity but not all wild animals. I wonder if that’s the case for most catastrophic risks.

Are you considering other approaches to reduce the number of out-of-scope applications?

For example, by getting people to fill out a form to make their application, which includes a clear, short question up front asking the applicant to confirm their application is related to in-scope topics and not letting them proceed further if they don't confirm this (just a quick idea that came to mind, there might be better ways of doing it).

2
Jason
I like this. Along the same lines, explicitly communicating to desk-rejected candidates that they are clearly out of scope may discourage repeat out-of-scope applications (if this isn't already being done).

Wow this seems huge. I wonder if buying these eggs is actually good from an animal welfare perspective?

Firstly, as you say, it might help make others adopt this technology.

But even putting that aside, I wonder if supporting these eggs could be directly good for welfare. The main problems with eggs from a welfare perspective is the culling of male chicks and the often terrible conditions egg-laying hens are kept in (e.g. battery cages). If neither of these things apply, as is the case with NestFresh Humanely Hatched Pasture-raised eggs, you could just be supporting happy lives by buying these eggs.

Unfortunately, eggs cause an incredible amount of suffering beyond the killing of male chicks and the environmental conditions of farmed hens, including in pasture-based farms. Laying ~30x more eggs than you naturally should is physically exhausting, is psychologically harmful (amped up hormones that create an experience akin to PMS), and results in extremely high rates of ovarian cancer, impacted egg material and consequent slow death by sepsis, and reproductive prolapses -- all left untreated. This is not to mention the experience of the parents ("breede... (read more)

I share your concern about x-risk from ASI, that's why I want safety-aligned people in these roles as opposed to people who aren't concerned about the risks.

There are genuine proposals on how to align ASI, so I think it's possible. I'm not sure what the chances are, but I think it's possible. I think the most promising proposals involve using advanced AI to assist with oversight, interpretability, and recursive alignment tasks—eventually building a feedback loop where aligned systems help align more powerful successors.

I don't agree that benefits are specu... (read more)

JackM
8
2
0
1
80% disagree

It is possible to rationally prioritise between causes without engaging deeply on philosophical issues

Underlying philosophical issues have clear implications for what you should prioritize, so I'm not really sure how you can rationally prioritize between causes without engaging with these issues.

I'm also not really sure how to defer on these issues when there are lots of highly intelligent, altruistically-minded people who disagree with each other. These disagreements often arise due to value judgements, and I don't think you can defer on your underlying v... (read more)

Thanks Vasco, I really appreciate your work to incorporate the wellbeing of wild animals into cost-effectiveness analyses.

In your piece, you focus on evaluating existing interventions. But I wonder whether there might be more direct ways to reduce the living time of soil nematodes, mites, and springtails that could outperform any human life-saving intervention.

On priors it seems unlikely that optimizing for saving human lives would be the most effective strategy to reduce wild animal suffering.

2
Vasco Grilo🔸
Thanks, Jack! I agree it is unlikely that the most cost-effective ways of increasing human-years are the most cost-effective ways of increasing agricultural-land-years. Brian Tomasik may have listed some of these. However, buying beef decreases arthropod-years the most cost-effectively among the interventions for which Brian estimated the cost-effectiveness, and I estimated GiveWell's top charities are 2.65 (= 1.69/0.638) times as cost-effective as buying beef.

I'm glad you found this useful Mo!

FWIW my impression of the EA community's position is that we need to build safe AI, not that we need to stop AI development altogether (although some may hold this view).

Stopping AI development altogether misses out on all the benefits from AI, which could genuinely be extensive and could include helping us with other very pressing problems (global health, animal welfare etc.).

I do think one can do a tremendous amount of good at OpenAI, and a tremendous amount of harm. I am in favor of roles at AI companies being on the 80,000 Hours job board so that the former is more likely.

1
Geoffrey Miller
JackM - these alleged 'tremendous' benefits are all hypothetical and speculative.  Whereas the likely X risk from ASI have been examined in detail by thousands of serious people, and polls show that most people, both inside and outside the AI industry, are deeply concerned by them. This is why I think it's deeply unethical for 80k Hours to post jobs to work on ASI within AI companies. 

My view is these roles are going to filled regardless. Wouldn't you want someone who is safety-conscious in them?

9
Jason
The tagline for the job board is: "Handpicked to help you tackle the world's most pressing problems with your career." I think that gives the reader the impression that, at least by default, the listed jobs are expected to have positive impact on the world, that they are better off being done well/faithfully than being unfilled or filled by incompetent candidates, etc. Based on what I take to be Geoffrey's position here, the best case that could be made for listing these positions would be: it could be impactful to fill a position one thinks is net harmful to prevent it from being filled by someone else in a way that causes even more net harm. But if that's the theory of impact, I think one has to be very, very clear with the would-be applicant on the what the theory is. I question whether you can do that effectively on a public job board. For example, if one thinks that working in prisons is an deplorable thing to do, I submit that it would be low integrity to encourage people to work as prison guards by painting that work in a positive light (e.g., handpicked careers to help you tackle the nation's most pressing social-justice problems).  [The broader question of whether we're better off with safety-conscious people in these kinds of roles has been discussed in prior posts at some length, so I haven't attempted to restate that prior conversation.]

It’s only intuitive to me not to eat cars because it isn’t good for wellbeing!

In a world in which cars are tasty and healthy to eat I imagine we wouldn’t find it so irrational to eat them. Unless of course you’d be losing a method of transportation by eating it and can get other options that are just as healthy and tasty for cheaper — in which case we’re just resorting to wellbeing arguments again.

This just seems to be question-begging. It just seems to me you're saying "axiological realism gives rise to normative realism because surely axiological realism gives rise to normative realism".

But this means that moral anti-realists must think that you can never have a reason to care about something independent of what you actually do care about. This is crazy as shown by the following cases:

  1. A person wants to eat a car. They know they’d get no enjoyment from it—the whole experience would be quite painful and unpleasant. On moral anti-realism, they’re not being irrational. They have no reason to take a different action.

I think the person wanting to eat a car is irrational because they will not be promoting their wellbeing by doing so and their we... (read more)

1
LanceSBush
 Realists might just say you have a reason simpliciter. There's nothing more to it than that. And they think this because it's intuitive.

Thanks for highlighting the relative lack of attention paid to cause prioritization and cross-cause prioritization. 

I have also written about how important it is to enable EAs to become familiar with existing cause prioritization findings. It's not just about how much research is done but also that EAs can take it into account and act on it.

8
arvomm
Thank you Jack, yes it's absolutely about acting on it too, research is just step 1.

You’re basically saying happier machines will be more productive and so we are likely to make them to be happy?

Firstly we don’t necessarily understand consciousness enough to know if we are making them happy, or even if they are conscious.

Also, I’m not so sure if happier means more productive. More computing power, better algorithms and more data will mean more productive. I’m open to hearing arguments why this would also mean the machine is more likely to be happy. 

Maybe the causality goes the other way - more productive means more happy. If machines... (read more)

1
Dylan Richardson
I actually agree with a lot of this - we probably won't intend to make them sentient at all, and it seems likely that we may do so accidentally, or that we may just not know if we have done so or not.   I'm mildly inclined to think that if ASI knows all, it can tell us when digital minds are or aren't conscious. But it seems very plausible that we either don't create full ASI, or that we do, but enter into a disempowerment scenario before we can rethink our choices about creating digital minds. So yes, all that is reason to be concerned in my view. I just depart slightly from your second to last paragraph. To put a number on it, I think that this is at least half as likely as minds that are generally happy. Consciousness is a black box to me, but I think that we should as a default put more weight on a basic mechanistic theory: positive valence encourages us towards positive action, negative valence threatens us away from dis-action or apathy. The fact that we don't observe any animals that seem dominated by one or the other seems to indicate that there is some sort of optimal equilibrium for goal fulfillment; that AI goals are different in kind from evolution's reproductive fitness goals doesn't seem like an obviously meaningful difference to me. Part of your argument centers around "giving" them the wrong goals. But goals necessarily mean sub-goals - shouldn't we expect the interior life of a digital mind to be in large part about it's sub-goals, rather than just ultimate goals? And if it is something so intractable that it can't even progress, wouldn't it just stop outputting? Maybe there is suffering in that; but surely not unending suffering? 

Well the closest analogue we have today is factory farmed animals. We use them in a way that causes tremendous suffering. We don't really mean to cause the suffering, but it's a by product of how we use them.

And another, perhaps even better, analogue is slavery. Maybe we'll end up essentially enslaving digital minds because it's useful to do so - if we were to give them too much freedom they wouldn't as effectively do what we want them to do.

Creating digital minds just so that they can live good lives is a possibility, but I'd imagine if you would ask some... (read more)

1
Dylan Richardson
That's true - but the difference is that both animals and slaves are sub-optimal; even our modern, highly domesticated food stock doesn't thrive in dense factory farm conditions, nor willingly walks into the abattoir. And an ideal slave wouldn't really be a slave, but a willing and dedicated automaton. By contrast, we are discussing optimized machines - less optimized would mean less work being done, more resource use and less corporate profit. So we should expect more ideal digital servants (if we have them at all). A need to "enslave" them suggests that they are flawed in some way. The dictates of evolution and nature need not apply here.  To be clear, I'm not entirely dismissing the possibility of tormented digital minds, just the notion that they are equally plausible.

Do you agree that the experience of digital minds likely dominates far future calculations?

This leads me to want to prioritize making sure that if we do create digital minds, we do so well. This could entail raising the moral status of digital minds, improving our ability to understand sentience and consciousness, and making sure AI goes well and can help us with these things.

Extinction risk becomes lower importance to me. If we go extinct we get 0 value from digital minds which seems bad, but it also means we avoid the futures where we create them and the... (read more)

JackM
2
0
1
36% disagree

This is a question I could easily change my mind on.

The experience of digital minds seems to dominate far future calculations. We can get a lot of value from this, a lot of disvalue, or anything in between.

If we go extinct then we get 0 value from digital minds. This seems bad, but we also avoid the futures where we create them and they suffer. It’s hard to say if we are on track to creating them to flourish or suffer - I think there are arguments on both sides. The futures where we create digital minds may be the ones where we wanted to “use” them, which ... (read more)

1
Dylan Richardson
I agree about digital minds dominating far future calculations; but I don't think your expectation that it is equally likely that we create suffering minds is reasonable. Why should we think suffering to be specially likely? "Using" them means suffering? Why? Wouldn't maximal usefulness entail, if any experience at all, one of utter bliss at being useful?  Also, the pleasure/suffering asymmetry is certainly a thing in humans (and I assume other animals), but pleasure does dominate, at least moment-to-moment. Insofar as wild animal welfare is plausibly net-negative, it's because of end-of-life moments and parasitism, which I don't see a digital analog for. So we have a biological anchor that should incline us toward the view utility dominates.  Moral circle expanding should also update us slightly against "reducing extinction risk being close to zero". And maybe, by sheer accident, we create digital minds that are absolutely ecstatic! 

This gives the impression that longtermism is satisfied with prioritising one option in comparison to another, regardless of the context of other options which if considered would produce outcomes that are "near-best overall". And as such it's a somewhat strange claim that one of the best things you could do for the far future is in actuality "not so great". 

Longtermism should certainly prioritise the best persistent state possible. If we could lock-in a state of the world where there were the maximum number of beings with maximum wellbeing of course ... (read more)

Conditional on fish actually being able to feel pain, it seems a bit far-fetched to me that a slow death in ice wouldn’t be painful.

3
Michael St Jules 🔸
This is less clear for shrimp, though. I don't know if they find the cold painful at all, and it might sedate them or even render them unconscious. But I imagine that takes time, and they're being crushed by each other and ice with ice slurry.

I was trying to question you on the duration aspect specifically. If electric shock lasts a split second is it really credible that it could be worse than a slow death through some other method?

1
David T
If the slow death involves no pain, of course it's credible. (The electric shock is, incidentally, generally insufficient to kill. They generally solve the problem of the fish reviving with immersion in ice slurry....). It's also credible that neither are remotely as painful as a two week malaria infection or a few years of malaria infection which is (much of) what sits on the other side of the trade here.
  • though I'll happily concede it's a longer process than electrical stunning

Isn't this pretty key? If "Electrical stunning reliably renders fish unconscious in less than one second" as Vasco says, I don't see how you can get much better than that in terms of humane slaughter.

Or are you saying that electrical stunning is plausibly so bad even in that split second so as to make it potentially worse than a much slower death from freezing?

3
David T
It's well within the bounds of possibility the electric shock is excruciating and the cold numbing, yes. Or indeed that they're both neutral, compared with slaughter methods that produce clear physiological stress indicators like asyphxiation in carbon-dioxide rich water. or that they're different for different types of water dwelling species depending on their natural hardiness to icy water, which also seems to be a popular theory. Rightly or wrongly, ice cold slurry is sometimes recommended as the humane option, although obviously the fish farming industry is more concerned with its ability to preserve the fish marginally better than kiliing prior to insertion into the slurry...

I'm a bit confused if I'm supposed to be answering on the basis of my uninformed prior or some slightly informed prior or even my posterior here. Like I'm not sure how much you want me to answer based on my experience of the world.

For an uninformed prior I suppose any individual entity that I can visually see. I see a rock and I think "that could possibly be conscious". I don't lump the rock with another nearby rock and think maybe that 'double rock' is conscious because they just visually appear to me to be independent entities as they are not really visu... (read more)

Yeah if I were to translate that into a quantitative prior I suppose it would be that other individuals have roughly 50% of being conscious (I.e. I’m agnostic on if they are or not).

Then I learn about the world. I learn about the importance of certain biological structures for consciousness. I learn that I act in a certain way when in pain and notice other individuals do as well etc. That’s how I get my posterior that rocks probably aren’t conscious and pigs probably are.

2
Michael St Jules 🔸
Ok, this makes more sense. What do you count as "other individual"? Any physical system, including overlapping ones? What about your brain, and your brain but not counting one electron?
Load more