All of Jack Malde's Comments + Replies

That is fair. I still think the idea that aligned superintelligent AI in the wrong hands can be very bad may be under-appreciated. The implication is that something like moral circle expansion seems very important at the moment to help mitigate these risks. And of course work to ensure that countries with better values win the race to powerful AI.

Well I'm assigning extinction a value of zero and a neutral world is any world that has some individuals but also has a value of zero. For example it could be a world where half of the people live bad (negative) lives and the other half live equivalently good (positive) lives. So the sum total of wellbeing adds up to zero. 

A dystopia is one which is significantly negative overall. For example a world in which there are trillions of factory farmed animals that live very bad lives. A world with no individuals is a world without all this suffering.

Could it be more important to improve human values than to make sure AI is aligned?

Consider the following (which is almost definitely oversimplified):

 

ALIGNED AI

MISALIGNED AI

HUMANITY GOOD VALUES

UTOPIA

EXTINCTION

HUMANITY NEUTRAL VALUES

NEUTRAL WORLD

EXTINCTION

HUMANITY BAD VALUES

DYSTOPIA

EXTINCTION

For clarity, let’s assume dystopia is worse than extinction. This could be a scenario where factory farming expands to an incredibly large scale with the aid of AI, or a bad AI-powered regime takes over the world. Let's assume neutral world is equivalent to extinc... (read more)

0
Karthik Tadepalli
2d
I think a neutral world is much better than extinction, and most dystopias are also preferable to human extinction. The latter is debatable but the former seems clear? What do you imagine by a neutral world?
4
titotal
4d
If you don't think misalignment automatically equals extinction, then the argument doesn't work. The neutral world is now competing with "neutral world where the software fucks up and kills people sometimes", which seems to be worse. 

I was surprised by your "dearest" and "mirror" tests.

Call the first the “dearest test.” When you have some big call to make, sit down with a person very dear to you—a parent, partner, child, or friend—and look them in the eyes. Say that you’re making a decision that will affect the lives of many people, to the point that some strangers might be hurt. Say that you believe that the lives of these strangers are just as valuable as anyone else’s. Then tell your dearest, “I believe in my decisions, enough that I’d still make them even if one of the people who c

... (read more)
  • We can think of this stance as analogous to: 
    • The utilitarian parent: “I care primarily about doing what’s best for humanity at large, but I wouldn’t want to neglect my children to such a strong degree that all defensible notions of how to be a decent parent state that I fucked up.”

I wonder if we don't mind people privileging their own children because:

  1. People love their kids too damn much and it just doesn't seem realistic for people to neglect their children to help others.
  2. A world in which it is normalised to neglect your children to "focus on humanit
... (read more)
2
Lukas_Gloor
16d
I'd say the two are tied contenders for "what's best from an impartial view."  I believe the impartial view is under-defined for cases of population ethics, and both of these views are defensible options in the sense that some morally-motivated people would continue to endorse them even after reflection in an idealized reflection procedure. For fixed population contexts, the "impartial stance" is arguably better defined and we want equal considering of [existing] interests, which gives us some form of preference utilitarianism. However, once we go beyond the fixed population context, I think it's just not clear how to expand those principles, and Narveson's slogan isn't necessarily a worse justification than "the future could be super-long/big."

Looking forward to reading this. A quick note: in 3. Tomi’s argument that creating happy people is good your introductory text doesn't match what is in the table.


 

1
EJT
13d
Thanks, fixed now!

Not sure I entirely agree with the second paragraph. The white paper outlines how philanthropy in this area is quite neglected, and there are organisations like LaMP which could certainly use more funding. Page 5 of the white paper also outlines bottlenecks in the process - even if firms do have strong incentives to acquire talent there can be informational gaps that prevent them from finding the best individuals, and similar informational gaps exist for the individuals that prevent them from actively utilising the best pathways. 

Having said that I'm not claiming this is the best use of EA dollars - just posting for people's information.

Yeah I have a feeling that the best way to argue for this on EA grounds might surprisingly be on the basis of richer world economic growth, which is kind of antithetical with EA's origins, but has been argued to be of overwhelming importance e.g.:

Thank you, good to flag these points. 

Regarding the AI Safety point, I want to think through this more, but I note that the alignment approach of OpenAI is very capabilities-driven, requiring talent and compute to align AI using AI. I think one's belief of the sign of immigration on x-risk here might depend on how much you think top labs like OpenAI actually take the safety risks seriously. If they do, more immigration can help them make safe AI.

Regarding the meat-eater problem, I think the possibility of an animal Kuznets curve is relevant. If such a... (read more)

4
PabloAMC
1mo
FWIW, I believe not every problem has to be centered around “cool” cause areas, and in this case I’d argue both animal welfare and AI Safety should not be significantly affected.
6
Probably Good
2mo
Yes, this will be recorded. If you'd like access to the recording after the session, reach out to Vaish at vaishnav@probablygood.org. 

Please do! I'm fascinated by the idea that we can accelerate moral progress by focusing on economic growth.

1
Rafael Ruiz
2mo
Same! Seems like a fascinating, although complicated topic. You might enjoy Oded Galor's "The Journey of Humanity", if you haven't read it. :)

Hey, thanks for this list! I was wondering if you have come across Benjamin Friedman's Moral Consequences of Economic Growth and, if so, what you thought. I thought I might see it on this list.

3
Rafael Ruiz
2mo
I hadn't! Thanks for bringing this to my attention, I will take a look in the coming months.

Matthew is right that uncertainty over the future is the main justification for discount rates

I don't think this is true if we're talking about Ramsey discounting. Discounting for public policy: A survey and Ramsey and Intergenerational Welfare Economics don't seem to indicate this. 

Also, am I missing something, or would a zero discount rate make this analysis impossible?

I don't think anyone is suggesting a zero discount rate? Worth noting though that that former paper I linked to discusses a generally accepted argument that the discount rate should f... (read more)

2
Karthik Tadepalli
3mo
Yes, Ramsey discounting focuses on higher incomes of people in the future, which is the part I focused on. I probably shouldn't have said "main", but I meant that uncertainty over the future seems like the first order concern to me(and Ramsey ignores it). Habryka's comment: seems to be arguing for a zero discount rate. Good point that growth-adjusted discounting doesn’t apply here, my main claim was incorrect.

Do economists actually use discount rates to account for uncertainty? My understanding was that we are discounting expected utilities, so uncertainty should be accounted for in those expected utilities themselves.

Maybe it’s easier to account for uncertainty via an increasing discount rate, but an exponential discount rate seems inappropriate. For starters I would think our degree of uncertainty would moderate over time (e.g. we may be a lot more uncertain about effects ten years from now than today, but I doubt we are much more uncertain about effects 1,000,010 years from now compared to 1,000,000 or even 500,000 years from now).

2
Davidmanheim
2mo
The reason it seems reasonable to view the future 1,000,010 years as almost exactly as uncertain as 1,000,000 years is mostly myopia. To analogize, is the ground 1,000 miles west of me more or less uneven than the ground 10 miles west of me? Maybe, maybe not - but I have a better idea of what the near-surroundings are, so it seems more known. For the long term future, we don't have much confidence in our projections of either a million or a million an ten years, but it seems hard to understand why all the relevant uncertainties will simply go away, other than simply not being able to have any degree of resolution due to distance. (Unless we're extinct, in which case, yeah.)
7
Karthik Tadepalli
3mo
If you think that the risk of extinction in any year is a constant γ, then the risk of extinction by year t is γt, so that makes it the only principled discount rate. If you think the risk of extinction is time-varying, then you should do something else. I imagine that a hyperbolic discount rate or something else would be fine, but I don't think it would change the results very much (you would just have another small number as the break-even discount rate).

Nice research!

In my preferred model, investing in science has a social impact of 220x, as measured in Open Philanthropy’s framework.

How does this compare to other top causes that Open Phil funds? Is there a summary somewhere?

9
Linch
3mo
My impression is that it's worse. OP's GHW (non-GCR) resources historically used the 1,000x bar as the bar to clear, and the linked blog post implies that it's gone up over time.

You might consider testing your ideas a few times to see if they would be effective before you suggest them.

-3
yanni
3mo
I suppose I consider this a test :)  (obviously I don't mind my ideas not being refined before publishing)

Thank you Arvo, I really appreciate it! I look forward to seeing more work from you and the team.

Well you might need a reasoned response I.e. it seems that when I do X a bad thing happens to me therefore I should endeavor not to do X.

Here is the quote from Richard Dawkins:

“If you think about what pain is for biologically speaking, pain is a warning to the animal, ‘don’t do that again’.

“If the animal does something which results in pain, that is a kind of ritual death – it is telling the animal, ‘if you do that again you might die and fail to reproduce’. That’s why natural selection has built the capacity to feel pain into our nervous systems.

“You coul... (read more)

2
NickLaing
5mo
That does seem plausible but I think the opposite i more likely. Of course you need a reasoned response, but I'm not sure the magnitude of pain would necessarily help the association with the reasoned response Harmful action leads to negative stimulus (perhaps painful) which leads to withdrawl and future cessation of that action. It seems unlikely to me that increasing the magnitude of that pain would make a creature more likely to stop doing an action. More like the memory and higher functions would need to be sufficient to associate the action to the painful stimuli, and then a form of memory needs to be there to allow a creature to avoid the action in future.  It is unintuitive to me that the "amount" of negative stimuli (pain) would be what matters, more the strength of connection between the pain and the action, which would allow future avoidance of the behaviour. I use "negative stimuli" rather than pain, because I still believe we heavily anthropomorphise our own experience of pain onto animals. Their experience is likely to be so wildly different from ours (whether "better" or "worse") that I think even using the word pain might be misleading sometimes. More intelligent beings shouldn't necessarily need pain at all to avoid actions which could cause you to "die and fail to reproduce". I wouldn't think to avoid actions that could lead to, or would need very minor stimulus as a reminder.  Actually it does seem quite complex the more I think about it/ Its an interesting discussion anyway.

In fact, it has been suggested by Richard Dawkins that less intelligent animals might experience greater suffering, as they require more intense pain to elicit a response. The evolutionary process would have ensured they feel sufficient pain.

6
MichaelDickens
1mo
Agreed. I disagree with the general practice of capping the probability distribution over animals' sentience at 1x that of humans'. (I wouldn't put much mass above 1x, but it should definitely be more than zero mass.)
3
NickLaing
5mo
Why would they need more intense pain to elicit a response? Intuitively to me at least with less "reasoning" ability, the slightest bit of pain would likely illicit a response away from said pain. 
3
Corentin Biteau
5mo
Ah, that's interesting. I didn't know that.  I had in mind that maybe the power of thought could allow us to put things into perspective better and better support pain (as can be experienced through meditation). However, this can go both ways, as negative thoughts can cause additional suffering. But I shall check the suggestion by Dawkins, that sounds interesting.

I’m a bit confused by this. Presumably GiveWell doesn’t need that much money to function. Less Open Phil money probably won’t affect GiveWell, instead it will affect GiveWell’s recommended charities which will of course still receive money from other sources, in part due to GiveWell’s recommendation.

1
Gabriel Mukobi
5mo
Ah right, I was conflating GiveWell's operating costs (I assume not too high?) and their funding sent to other charities both as "GiveWell continuing to work on global poverty." You're right that they'll still probably work on it and not collapse without OP, just they might send much less to other charities.

Strong upvoted. I think this is correct, important and well-argued, and I welcome the call to OP to clarify their views. 

This post is directed at OP, but this conclusion should be noted by the EA community as a whole which still prioritises global poverty over all else.

The only caveat I would raise is that we need to retain some focus on global poverty in EA for various instrumental reasons: it can attract more people into the movement, allows us to show concrete wins etc.  

3
Benny Smith
5mo
Yeah, I think this caveat is important. At the same time, GiveWell will continue to work on global poverty regardless of what OP does, right?

Nice paper!

One interesting result of the paper is that neglectedness seems is key to whether a policy change matters for a long time. For policies that can be expected to attract more interest after the referendum passes, I see less persistence. It is not a hugely dramatic effect, but it could make a difference on the margin or in extreme cases. This seems to lend some support to the EA practice of paying attention to neglectedness.

This is probably me being stupid, but I'm not sure I understand this. Are you saying more neglected areas are those that would... (read more)

1
OscarD
5mo
I interpreted this in the same way as you. Presumably the referendum will lead to a short-term uptick in popular interest, but maybe our model could be that there is some baseline 'interestingness' of an issue that public interest reverts to soon after the referendum? It perhaps depends on what the process for getting something on a referendum is. If it is hard and requires very many people to already endorse the proposed policy, then almost by definition referendums aren't very neglected. But if a few actors can get a referendum, then the topic may still be quite neglected.

I'm sorry, this doesn't engage with the main point(s) you are trying to make, but I'm not sure why you use the term "existential risk" (which you define as risks of human extinction and undesirable lock-ins that don’t involve s-risk-level suffering) when you could have just used the term "extinction risk".

You say:

If you’re uncertain whether humanity’s future will be net positive, and therefore whether existential risk[1] reduction is good, you might reason that we should keep civilization going for now so we can learn more and, in the future, make a b

... (read more)
8
Winston
6mo
Thanks :) Good point. Minor point: I don't think it's strictly true that reducing risks of undesirable lock-ins is robustly good no matter what the expected value of the future is. It could be that a lock-in is not good, but it prevents an even worse outcome from occurring. I included other existential risks in order to counter the following argument: "As long as we prevent non-s-risk-level undesirable lock-ins in the near-term, future people can coordinate to prevent s-risks." This is a version of the option value argument that isn't about extinction risk. I realize this might be a weird argument for someone to make, but I covered it to be comprehensive. But the way I wrote this, I was pretty much just focused on extinction risk. So I agree it doesn't make a lot of sense to include other kinds of x-risks. I'll edit this now.

Previous work has referred to such a risk as 'existential risk'. But this is a misnomer. Existential risk is technically broader and it encompasses another case: the risk of an event that drastically and permanently curtails the potential of humanity. For the rest of this report we characterise the risk as that of extinction where previous work has used 'existential'. 

I was happy to see this endnote, but then I noticed several uses of "existential risk" in this abridged report when I think you should have said "extinction risk". I'd recommend going through to check this.

4
arvomm
6mo
It's good to hear that you agree extinction is the better term in this framework. Though I think it makes sense to talk about the more general 'existential' term in the exposition sometimes. In particular, for entirely pedagogical reasons, I decided to leave it with the original terminology in the summary since readers who are already familiar with the original models might skim this post or miss that endnote, and the definition of risk hasn't changed. I see this report, and the footnote, as asking researchers that, from hereon, we use extinction when the maths are set up like they are here. All that said, I've indeed noticed instances after the summary where the conceptual accuracy would be improved by making that swap. Thank you again; I'll keep a closer eye on this, especially in future revised versions of the full report.

In theory people will always prefer cash because they can spend it on whatever they want (unless of course it is difficult to buy what they most want). This isn’t really up for debate.

What is up for debate is if people actually spend money in a way that most effectively improves their welfare. It sounds paternalistic to say, but I suspect they don’t for the reasons Nick and others have given.

Yeah I feel that sometimes theories get really convoluted and ad hoc in an attempt to avoid unpalatable conclusions. This seems to be one of those times.

I can give Scanlon a free pass when he says under his theory we should save two people from certain death rather than one person from certain death because the 'additional' person would have some sort of complaint. However when the authors of this post say, for a similar reason, that the theory implies it's better to do an intervention that will save two people with probability 90% rather than one person w... (read more)

I do worry about future animal suffering. It's partly for that reason that I'm less concerned about reducing risks of extinction than I am about reducing other existential risks that will result in large amounts of suffering in the future. This informed some of my choices of interventions for which I am 'not clueless about'. E.g. 

  • Technical AI alignment / AI governance and coordination research: it has been suggested that misaligned AI could be a significant s-risk.
  • Expanding our moral circle: relevance to future suffering should be obvious.
  • Global prior
... (read more)

A relevant GPI paper is Longtermism, aggregation, and catastrophic risk by Emma J. Curran.

I briefly summarised it here, also pasted below:

The bottom line: If one is sceptical about aggregative views, where one can be driven by sufficiently many small harms outweighing a smaller number of large harms, one should also be sceptical about longtermism.

My brief summary:

  • Longtermists generally prefer reducing catastrophic risk to saving lives of people today. This is because, even though you would be reducing probability of harm by a small amount if focusing on ca
... (read more)

Contractualism doesn't allow aggregation across individuals. If each person has 0.3% chance of averting death with a net, then any one of those individual's claims is still less strong than the claim of the person who will die with probability ~=1. Scanlon's theory then says save the one person.

8
Linch
6mo
Yeah Scanlon's theory doesn't allow for differentiation even between a strong claim and many only slightly worse claims. The authors of this post tries to rescue the theory by the small relaxation that you can treat high probabilities and numbers of morally almost-as-bad things to be worse than 1 very bad and certain thing. But while I could imagine it going through for preventing 2 people from dying with 80% probability vs 1 person with 100%, I don't think it goes through for ice cream, or AMF. A system that doesn't natively do aggregation has a lot of trouble explaining why many numbers of people each with a 0.3% of counterfactually dying has as much ore more moral claim to your resources as a single identified person with ~100% chance of counterfactually dying. (As a side note, I try to ground my hypotheticals in questions that readers are likely to have first-hand familiarity with, or can easily visualize themselves in that position. Either very few or literally no one in this forum has experience with obscenely high numbers of dust specks, or missile high command. Many people in this conversation have experience with donating to AMF, and/or eating ice cream). 

I don't buy the argument that AI safety is in some way responsible for dangerous AI capabilities. Even if the concept of AI safety had never been raised I'm pretty sure we would still have had AI orgs pop up.

Also yes it is possible that working on AI Safety could limit AI and be a catastrophe in terms of lost welfare, but I still think AI safety work is net positive in expectation given the Bostrom astronomical waste argument and genuine concerns about AI risk from experts. 

The key point here is that cluelessness doesn't arise just because we can thin... (read more)

2
Arepo
6mo
I think at this point we can amicably disagree, though I'm curious why you think the 'more people = more animals exploited' philosophy applies to people in Africa, but not in the future. One might hope that we learn to do better, but it seems like that hope could be applied to and criticised in either scenario.

Thanks for your response, I'm excited to see your sequence. I understand you can't cover everything of interest, but maybe my comments give ideas as to where you could do some further work.

I'll need to find some time to read that Holden post.

I'm happy that the flowchart was useful to you! I might consider working on it in the future, but I think the issues are that I'm not convinced many people would use it and that the actual content of the flowchart might be pretty contentious - so it would be easy to be accused of being biased. I was using my karma score as a signal of if I should continue with it, and karma wasn't impressive.

I think you might be right that Greaves is too strong on this and I'll admit I'm still quite uncertain about exactly how cluelessness cashes out. However, I know I have difficulties funding GHD work (and would even if there was nothing else to fund), but that I don't have similar difficulties for certain longtermist interventions. I'll try to explain.

I don't want to fund GHD work because it's just very plausible that the animal suffering might outweigh the human benefit. Some have called for development economists to consider the welfare of non-human anima... (read more)

2
Arepo
6mo
I have no particular reason to think you shouldn't believe in any of those claims, but fwiw I find it quite plausible (though wouldn't care to give particular credences atm) that at least some of them could be bad, eg: * Technical AI safety seems to have been the impetus for various organisations who are working on AI capabilities in a way that everyone except them seems to think is net negative (OpenAI, Deepmind, Anthropic, maybe others). Also, if humans end up successfully limiting AI by our own preferences, that could end up being a moral catastrophe all of its own. * 'Expanding our moral circle' sounds nice, but without a clear definition of the morality involved it's pretty vague what it means - and with such a definition, it could cash out as 'make people believe our moral views', which doesn't have a great history. * Investing for the future could put a great deal of undemocratic power into the hands of a small group of people whose values could shift (or turn out to be 'wrong') over time. And all of these interventions just cost a lot of money, something which the EA movement seems very short on recently.  

I’m very excited to read this sequence! I have a few thoughts. Not sure how valid or insightful they are but thought I’d put them out there:

On going beyond EVM / risk neutrality

  • The motivation for investigating alternatives to EVM seems to be that EVM has some counterintuitive implications. I'm interested in the meta question of how much we should be swayed by counterintuitive conclusions when EVM seems to be so well-motivated (e.g. VNM theorem), and the fact that we know we are prone to biases and cognitive difficulties with large numbers.
  • Would alternative
... (read more)
7
JoshuaBlake
6mo
Would you mind expanding a bit on why this applies to GHD and not other cause areas please? E.g.: wouldn't your concerns about animal welfare from GHD work also apply to x-risk work?
8
Arepo
6mo
I support most of this comment, but strongly disagree with this, or at least think it's much too strong. Cluelessness isn't a categorical property which some interventions have and some don't - it's a question of how much to moderate your confidence in a given decision. Far from being the unanswerable question Greaves suggests, it seems reasonable to me to do any or all of the following: 1. Assume unknown unknowns pan out to net 0 2. Give credences on a range of known unknowns 3. Time-limit the above process in some way, and give an overall best guess expectation for remaining semi-unknowns  4. Act based on the numbers you have from above process when you stop 5. Incorporate some form of randomness in the criteria you investigate If you're not willing to do something like the above, you lose the ability to predict anything, including supposedly long-termist interventions, which are all mired in their own uncertainties. So while one might come to the view that GHD is in fact bad because of eg the poor meat eater problem, it seems irrational to be agnostic on the question, unless you're comparably agnostic towards every other cause.
3
Mo Putera
6mo
I've been interested in this as well, and I consider Holden's contra arguments in Sequence thinking vs. cluster thinking persuasive in changing my mind for decision guidance in practice (e.g. at the "implementation level" of personally donating to actual x-risk mitigation funds) -- I'd be curious to know if you have a different reaction. Edited to add: I just realized who I'm replying to, so I wanted to let you know that your guided cause prio flowchart was a key input at the start of my own mid-career pivot, and I've been sharing it from time to time with other people. In that post you wrote and if I'm interpreting your followup comment correctly, it's sad to see little interest in such a good first-cut distillation. Here's hoping interest picks up going forward.
8
Bob Fischer
6mo
Thanks for engaging, Jack! As you'd expect, we can't tackle everything in a single sequence; so, you won't get our answers to all your questions here. We say a bit more about the philosophical issues associated with going beyond EVM in this supplementary document, but since our main goal is to explore the implications of alternatives to EVM, we're largely content to motivate those alternatives without arguing for them at length. Re: GHD work and cluelessness, I hear the worry. We'd like to think about this more ourselves. Here's hoping we're able to do some work on it in the future. Re: not all x-risk being the same, fair point. We largely focus on extinction risk and do try to flag as much in each report.

I might have gone for "Utilitarianism may be irrational or self-undermining" rather than "Utilitarianism is irrational or self-undermining".

Personally, I would have kept the original title. Titles that are both accurate and clickbaity are the best kind - they get engagement without being deceptive. 

I don't think karma is always a great marker of a post's quality or appropriateness. See an earlier exchange we had.

2
MichaelStJules
6mo
Unfortunately, I think clickbait also gets downvotes even if accurate, and that will drop the post down the front page or off it.

Ah ok I actually used the word “attack”. I probably shouldn’t have, I feel no animosity at all towards Michael. I love debating these topics and engaging with arguments. I wish he’d had more room to expand on his person-affecting leanings. In a sense he is “attacking” longtermism but in a way that I welcome and enjoy responding to.

I happen to think the level of attention Will gave to population ethics and the concepts of the non-identity problem, repugnant conclusion, and person-affecting intuition is fairly admirable for a book intended for a general non-... (read more)

I'm not sure I have framed the review as an attack? I don't think it is. I have no problem with Michael writing the review, I just disagree with the points he made.

It was a while since I read the book in its entirety, but I will just leave a quote from the introduction which to me doesn't read as "disquietingly polemical" (bold emphasis mine):

For those who want to dig deeper into some of my claims, I have compiled extensive supplementary materials, including special reports I commissioned as background research, and made them available at whatweowethefutur

... (read more)
5
lastmistborn
6mo
The general tone of your comments + the line "I'm still happy I wrote that section because I wanted to defend longtermism from your attack" in one comment gives me the impression that you are, but I'm fully willing to accept that it's just the lack of emotive expressiveness in text.  Yes, MacAskill does have these explicit lines at certain points (I'd argue that this is the bare minimum, but it's a problem I have with a large swathe of academic and particularly pop-philosophy texts and as I said it's in some measure a matter of personal preference), but the overall tone of the text and the way he engages with counterarguments and positions still came off as polemical to me. I admittedly hold seminal texts - which WWTF is obviously intended to be - up to particularly high standards in this regard, which I think is fair but completely understand if others disagree. To be clear, I think that this also weakens the argumentation overall rather than just being a lopsided defense or a matter of tone. I think the points raised here about the intuition of neutrality are an good example of this; a more robust engagement with the intuition of neutrality and its implications for longtermism could help specify longtermism and it's different strains to make it less of an amorphous moral imperative to "think/care about future generations" and a more easily operationalized and intellectually/analytically robust moral philosophy since it would create room for a deeper discussion of how longtermist approaches that prioritize the existence of future people differ from longtermist approaches that view the benefits for future people as secondary.

OK. I think Will intended WWOTF to be a persuasive piece so I’m not sure if this is a valid criticism. He wasn’t writing a textbook.

I think this is confused. WWOTF is obviously both aiming to be persuasive and coming from a place of academic analytical philosophical rigour. Many philosophers write books that are both, e.g. Down Girl by Kate Manne or The Right to Sex by Amia Srinivasan. I don't think a purely persuasive book would have so many citations. 
.

-3
lastmistborn
6mo
Book reviews are meant to be informative and critiques aren't always meant to be negative, so I don't know why you're framing it as an attack on WWTF or MacAskill. Knowing the tone of a work is valuable information for someone reading a book review. On a personal note, I'll say that I also agree with the "disquieting" portion of "disquietingly polemical" - I had the sense that WWTF presented longtermism and caring about future generations as a kind of foregone conclusion and moral imperative rather than something to be curious about and think deeply on, but I prefer these kinds of books to be more proactive in very strongly establishing the opposing viewpoints, so it's probably more irksome to me than it would be to others. He wasn't writing a textbook and it's prerogative to write something that's an outright manifesto if he so chooses, but that doesn't make pointing out the tone an unvalid critique.

Hey, thanks for you comment! To be honest my addendum is a bit speculative and I haven't thought about a huge amount. I think I may have been a little extreme and that factoring moral uncertainty would soften some of what I said.

Would 100 percent of philosophies working on the question of the far future really be the best way to improve the field, with other important philosophical professions neglected?

When you say "improve the field" I'm not sure what you mean. Personally I don't think there is intrinsic value in philosophical progress, only instrumental... (read more)

That said, off the top of my head, philosophers who have written sympathetically about person-affecting views include Bader, Narveson (two classic articles here and here), Roberts (especially here, but she's written on it a few times), Frick (here and in his thesis), Heyd, Boonin, Temkin (here and probably elsewhere). There are not 'many' philosophers in the world, and population ethics is a small field, so this is a non-trivial number of authors! For an overview of the non-identity problem in particular, see the SEP.

I agree we should be more swayed by arg... (read more)

Fourth, I'm not sure why you think I've misrepresented MacAskill (do you mean 'misunderstood'?). In the part you quote, I am (I think?) making my own assessment, not stating MacAskill's view at all.

You say the following in the summary of the book section (bold part added by me):

If correct, this [the intuition of neutrality] would present a severe challenge to longtermism

By including it in the 'summary' section I think you implicitly present this as a view Will espoused in the book - and I don't agree that he did.

But the cause du jour of longtermism is prev

... (read more)

I lean towards thinking the following is unfair.

Third, the thrust of my article is that MacAskill makes a disquietingly polemical, one-sided case for longtermism.

If one were just to read WWOTF they would come away with an understanding of:

  • The intuition of neutrality - what it is, the fact that some people hold it, the fact that if you accept it you shouldn't care about losing future generations.
  • The non-identity problem - what it is and why some see it as an argument against being able to improve the future.
  • The repugnant conclusion - what it is, how some fi
... (read more)
3
lastmistborn
6mo
I think this is another point where you're missing context. It's kind of a quirk of academic language, but "polemical" is usually used in contrast to analytical in texts like these - meaning that the work in question is more argumentative/persuasive than analytical or explicative, which I honestly think is a very apt description of WWTF. 

Thanks for this reply Michael! I'll do a few replies and understand that you don't want to get in a long back and forth so will understand if you don't reply further.

Firstly, the following is all very useful background so I appreciate these clarifications:

First, the piece you're referring to is a book review in an academic philosophy journal. I'm writing primarily for other philosophers who I can expect to have lots of background knowledge (which means I don't need to provide it myself).

Second, book reviews are, by design, very short. You're even discourag

... (read more)

Fair point, but I would still disagree his analysis implies that human extinction would be good. He discusses digital sentience and how, on our current trajectory, we may develop digital sentience with negative welfare. An implication isn't necessarily that we should go extinct, but perhaps instead that we should try to alter this trajectory so that we instead create digital sentience that flourishes. 

So it's far too simple to say that his analysis "concludes that human extinction would be a very good thing". It is also inaccurate because, quite literally, he doesn't conclude that. 

So I agree with your choice to remove that wording.

If I were you I would remove that part altogether. As Kyle has already said his analysis might imply that human extinction is highly undesirable.

For example, if animal welfare is significantly net negative now then human extinction removes our ability to help these animals, and they may just suffer for the rest of time (assuming whatever killed us off didn’t also kill off all other sentient life).

Just because total welfare may be net negative now and may have been decreasing over time doesn’t mean that this will always be the case. Maybe we can do something about it and have a flourishing future.

7
Jeff Kaufman
7mo
Yeah, this seems like it's raising the stakes too much and distracting from the main argument; removed.
5
Jeff Kaufman
7mo
But his analysis doesn't say that? He considers two quantities in determining net welfare: human experience, and the experience of animals humans raise for food. Human extinction would bring both of these to zero. I think maybe you're thinking his analysis includes wild animal suffering?

I agree that alignment research would suffer during a pause, but I've been wondering recently how much of an issue that is. The key point is that capabilities research would also be paused, so it's not like AI capabilities would be racing ahead of our knowledge on how to control ever more powerful systems. You'd simply be delaying both capabilities and alignment progress.

You might then ask - what's the point of a pause if alignment research stops? Isn't the whole point of a pause to figure out alignment?

I'm not sure that's the whole point of a pause. A pau... (read more)

Fair enough. A more precise question could be, would it be beneficial to slow progress from the current trend?

Or another question could be, would it be desirable or undesirable to give more compute and talent to top AI labs?

4
Chris Leong
7mo
Hmm... your worry still isn't completely clear from that. Is your worry that any attempts to slow AI will reduce the lead time of top labs, giving them less time to align a human-level AI when they develop it?
Load more