All of Richard Y Chappell🔸's Comments + Replies

Ok, thanks for expanding upon your view! It sounds broadly akin to how I'm inclined to address Pascal's Mugging cases (treat the astronomical stakes as implying proportionately negligible probability). Astronomical stakes from x-risk mitigation seems much more substantively credible to me, but I don't have much to add at this point if you don't share that substantive judgment!

2
Vasco Grilo🔸
You are welcome! Makes sense. I see Pascal's muggings as instances where the probability of the offers is assessed indepently of their outcomes. In contrast, for any distribution with a finite expected value, the expected value density (product between the PDF and value) always ends up decaying to 0 as the outcome increases. In meta-analyses, effect sizes, which can be EVs under a given model, are commonly weighted by the reciprocal of their variance. Variance tends to increase with effect size, and therefore larger effect sizes are usually weighted less heavily. People sometimes point to Holden Karnofsky's post Why we can’t take expected value estimates literally (even when they’re unbiased) to justify not relying on EVs (here are my notes on it from 4 years ago). However, the post does not broadly argue against using EVs. I see it as a call for not treating all EVs the same, and weighting them appropriately.

You keep describing possible scenarios in which the actual value of averting extinction would be low. Do you not understand the scenario conditions under which the actual value would be high? I'm suggesting that those kinds of scenario (in which astronomically valuable futures are reliably securable so long as we avoid extinction this century) seem reasonably credible to me, which is what grounds my background belief in x-risk mitigation having high expected value.

Edited to add: You may in a sense be talking about "expected value", but insofar as it is bas... (read more)

Do you not understand the scenario conditions under which the actual value would be high?

I do. In theory, there could be worlds which i) will actually be astronomically valuable given some efforts, but that ii) will be practically actually neutral without such efforts. In this case, the efforts would be astronomically valuable due to meaningfully increasing the probability of astronomically valuable futures. I just think such worlds are super implausible. I can see many worlds satisfying i), but not i) and ii) simultaneously.

We need to abstract from that s

... (read more)

I really want to stress the difference between saying something will (definitely) happen versus saying there's a credible chance (>1%) that it will happen. They're very different claims!

Lovelace and Menabrea probably should have regarded their time as disproportionately likely (compared to arbitrary decades) to see continued rapid progress. That's compatible with thinking it overwhelmingly likely (~99%) that they'd soon hit a hurdle.

As a heuristic, ask: if one were, at the end of history, to plot the 100 (or even just the 50) greatest breakthrough perio... (read more)

It is true that Lovelace and Menabrea should have assumed a credible chance of rapid progress. Who knows, maybe if they had had the right resources and people, we could have had computers much earlier than we ultimately had.

But when talking about ASI, we are not just talking about rapid progress, we are talking about the most extreme progress imaginable. Extraordinary claims require extraordinary evidence, and so forth. We do not know what breakthroughs ASI requires, nor do we know how far we are from it.

It all comes down to the question of whether the cur... (read more)

4
David Mathers🔸
It seems like if you find it incredible to deny and he doesn't, it's very hard to make further progress :(  I'm on your side about the chance being over 1% in the next decade, I think, but I don't know how I'd prove it to a skeptic, except to gesture and say that capabilities have improved loads in a short time, and it doesn't seem like the are >20 similar sized jumps before AGI. But when I ask myself what evidence I have for "there are not >20 similar sized jumps before AGI" I come up short. I don't necessarily think the burden of proof here is actually on people arguing that the chance of AGI in the next decade is non-negligible though: it's a goal of some serious people within the relevant science, and they are not making zero progress, and some identifiable quantifiable individual capabilities have improved very fast. Plus the extreme difficulty of forecasting technological breakthroughs over more than a couple of years cuts both ways. 

Thanks for explaining your view! 

On the first point: I think we should view ASI as disproportionately likely in decades that already feature (i) recent extraordinary progress in AI capabilities that surprises almost everyone, and (ii) a fair number of experts in the field who appear to take seriously the possibility that continued progress in this vein could soon result in ASI.

I'd then think we should view it as disproportionately unlikely that ASI will either be (a) achieved before any such initial signs of impressive progress, OR (b) achieved centur... (read more)

While it might feel to you that AI progress has been rapid in the past decade, most innovations behind it such as neural networks, gradient descent, backpropagation, and the concept of language models are very old innovations. The only major innovation in the past decade is the Transformer architecture from 2017, and almost everything else is just incremental progress and scaling on larger models and datasets. Thus, the pace of AI architecture development is very slow and the idea that a groundbreaking new AGI architecture will surface has a low probabilit... (read more)

I'm a bit confused by this response. Are you just saying that high expected value is not sufficient for actual value because we might get unlucky?

Some possible extinction-averting events merely extend life for 1 second, and so provide little value. Other possibilities offer far greater extensions. Obviously the latter possibilities are the ones that ground high expected value estimates.

I think Shulman's points give us reason to think there's a non-negligible chance of averting extinction (extending civilization) for a long time. Pointing out that other possibilities are also possible doesn't undermine this claim.

4
Vasco Grilo🔸
Thanks for the follow-up, Richard. No. I agree high expected value does not ensure high actual value, but I only care about the expected (not the actual) increase in welfare. This is far from obvious to me, and I wonder what makes you so confident. I am not aware of any reasonably quantitative and empirical modelling of the expected increase in future welfare under a longtermist perspective. I expect what David Thorstad calls rapid diminution. I see the difference between the probability density function (PDF) after and before an intervention reducing the nearterm risk of human extinction as quickly decaying to 0, thus making the increase in the expected value of the astronomically valuable worlds negligible. For instance: * If the difference between the PDF after and before the intervention decays exponentially with the value of the future v, the increase in the expected value density caused by the intervention will be proportional to v*e^-v. * The above rapidly goes to 0 as v increases. For an expected value of the future equal to my estimate of 1.40*10^52 human lives, the increase in the expected value density will involve a factor of 1.40*10^52*e^(-1.40*10^52) = 10^(log10(1.40) + 52 - log10(e)*1.40*10^52) = 10^(-6.08*10^51), i.e. it will be basically 0. It seems that you are inferring one could meaningfully increase the probability of astronomically valuable futures because these are plausible. However, astronomically valuable futures could be plausible while having a probability that is very hard to change.

Distinguish pro tanto vs all-things-considered (or "net") high stakes. The statement is literally true of pro tanto high stakes: the 1% chance of extremely high stakes is by itself, as far as it goes, an expected high stake. But it's possible that this high stake might be outweighed or cancelled out by other sufficiently high stakes among the remaining probability space (hence the subsequent parenthetical about "unless one inverts the high stakes in a way that cancels out...").

The general lesson of my post is that saying "there's a 99% chance there's nothi... (read more)

The responses to my comment have provided a real object lesson to me about how a rough throwaway remark (in this case: my attempt to very briefly indicate what my other post was about) can badly distract readers from one's actual point! Perhaps I would have done better to entirely leave out any positive attempt to here describe the content of my other post, and merely offer the negative claim that it wasn't about asserting specific probabilities.

My brief characterization was not especially well optimized for conveying the complex dialectic in the other pos... (read more)

1
TFD
I think you have be underestimating to what extent the responses you are getting do speak to the core content of your post, but I will leave a comment there to go into it more.

On (what I take to be) the key substantive claim of the post:

I think that nontrivial probability assignments to strong and antecedently implausible claims should be supported by extensive argument rather than manufactured probabilities.

There seems room for people to disagree on priors about which claims are "strong and antecedently implausible". For example, I think Carl Shulman offers a reasonably plausible case for existential stability if we survive the next few centuries. By contrast, I find a lot of David's apparent assumptions about which proposition... (read more)

3
Vasco Grilo🔸
Hi Richard. I like @CarlShulman's points. However, decreasing the risk of human extinction over the next few decades is not enough for astronomical benefits even if the risk is concentrated there, and the future is astronomically valuable. Imagine human population is 10^10 without human extinction, and that the probability of human extinction over the next 10 years is 10 % (in reality, I guess the probability of human extinction over the next 10 years is more like 10^-7), and then practically 0 forever, which implies infite human-years in the future. As an extreme example, an intervention decreasing to 0 the risk of human extinction over the next 10 years could still have negligible value. If it only postpones extinction by 1 s, it would only increase future human-years by 317  (= 10^10*1/(365.25*24*60^2)). I have not seen any empirical quantitative estimates of increases in the probability of astronomically valuable futures.

Hi David, I'm afraid you might have gotten caught up in a tangent here! The main point of my comment was that your post criticizes me on the basis of a misrepresentation. You claim that my "primary argumentative move is to assign nontrivial probabilities without substantial new evidence," but actually that's false. That's just not what my blog post was about.

In retrospect, I think my attempt to briefly summarize what my post was about was too breezy, and misled many into thinking that its point was trivial. But it really isn't. (In fact, I'd say that my co... (read more)

It's not a psychological question. I wrote a blog post offering a philosophical critique of some published academic papers that, it seemed to me, involved an interesting and important error of reasoning. Anyone who thinks my critique goes awry is welcome to comment on it there. But whether my philosophical critique is ultimately correct or not, I don't think that the attempt is aptly described as "personal insult", "ridiculous on [its] face", or "corrosive to productive, charitable discussion". It's literally just doing philosophy.

I'd like it if people read my linked post before passing judgment on it.

The meta-dispute here isn't the most important thing in the world, but for clarity's sake, I think it's worth distinguishing the following questions:

  1. Does a specific text—Thorstad (2022)—either actually or apparently commit a kind of "best model fallacy", arguing as though establishing Time of Perils hypothesis as unlikely to be true thereby suffices to undermine longtermism?
  2. Does another specific text—my 'Rule High Stakes In, Not Out'—either actually or apparently have as its "primary argumentative move... to assign nontrivial probabilities without substant
... (read more)

As I see it, I responded entirely reasonably to the actual text of what you wrote. (Maybe what you wrote gave a misleading impression of what you meant or intended; again, I made no claims about the latter.)

Is there a way to mute comment threads? Pursuing this disagreement further seems unlikely to do anyone any good. For what it's worth, I wish you well, and I'm sorry that I wasn't able to provide you with the agreement that you're after.

-17
Yarrow Bouchard 🔸

Honestly, I still think my comment was a good one! I responded to what struck me as the most cruxy claim in your post, explaining why I found it puzzling and confused-seeming. I then offered what I regard as an important corrective to a bad style of thinking that your post might encourage, whatever your intentions. (I made no claims about your intentions.) You're free to view things differently, but I disagree that there is anything "discourteous" about any of this.

-6
Yarrow Bouchard 🔸

There's "understanding" in the weak sense of having the info tokened in a belief-box somewhere, and then there's understanding in the sense of never falling for tempting-but-fallacious inferences like those I discuss in my post.

Have you read the paper I was responding to? I really don't think it's at all "obvious" that all "highly trained moral philosophers" have internalized the point I make in my blog post (that was the whole point of my writing it!), and I offered textual support. For example, Thorstad wrote: "the time of perils hypothesis is probably f... (read more)

4
David Mathers🔸
Fair point, when I re-checked the paper, it doesn't clearly and explicitly display knowledge of the point you are making. I still highly doubt that Thorstad really misunderstands it though. I think he was probably just not being super-careful. 

This sort of "many gods"-style response is precisely what I was referring to with my parenthetical: "unless one inverts the high stakes in a way that cancels out the other high-stakes possibility."

I don't think that dystopian "time of carols" scenarios are remotely as credible as the time of perils hypothesis. If someone disagrees, then certainly resolving that substantive disagreement would be important for making dialectical progress on the question of whether x-risk mitigation is worthwhile or not.

What makes both arguments instances of the nontrivial pr

... (read more)
1
TFD
I think you are making some unstated assumptions that it would be helpful to make explicit. You say your argument is basically just explaining how expected values work, but it doesn't seem like that is true to me, I think you need to make some assumptions unrelated to how expected values work for your argument to go through. If I were to cast your argument in the language of "how expected values work" it would go like this: An expected value is the the sum of a bunch of terms that involve multiplying an outcome by its probability, so of the form x * p, where x is the outcome (usually represented by some some number) and p is the probability associated with that outcome. To get the EV we take terms like that representing every possible outcome and add them up. Because these terms have two parts, the term as a whole can be large even if the probability is small. So, the overall EV can be driven primarily by a small probability of a large positive outcome because it is dominated by this one large term, which is large even when the probability is small. We rule high stakes in, not out. The problem is that this argument doesn't work without further assumptions. In my version I said "can be driven". I think your conclusion requires "is driven", which doesn't follow. Because there are other terms in the EV calculation their sum could be negative and of sufficient magnitude that the overall EV is low or negative even if one term is large and positive. This doesn't require that any particular term in the sum has any particular relationship to the large positive term such that it is "inverting" that term, although that would be sufficient, it isn't the only way for the overall EV to be small/negative. Their could be a mix of moderate negative terms that adds up to enough to reduce the overall EV. Nothing about this seems weird or controversial to me. For example, a standard normal distribution has large positive values with small probabilities but has an expectation of ze
8
David Thorstad
Thanks Richard! Writing is done for an audience. Effective altruists have a very particular practice of stating their personal credences in the hypotheses that they discuss. While this is not my practice, in writing for effective altruists I try to be as precise as I can about the relative plausibility that I assign to various hypotheses and the effect that this might have on their expected value. When writing for academic audiences, I do not discuss uncertainty unless I have something to add which my audience will find to be novel and adequately supported.   I don’t remind academic readers that uncertainty matters, because all of them know that on many moral theories uncertainty matters and many (but not all) accept such theories. I don’t remind academic readers of how uncertainty matters on some popular approaches, such as expected value theory, because all of my readers know this and many (but fewer) accept such theories. The most likely result of invoking expected value theory would be to provoke protests that I am situating my argument within a framework which some of my readers do not accept, and that would be a distraction.  I don’t state my personal probability assignments to claims such as the time of perils hypothesis because I don’t take myself to have given adequate grounds for a probability assignment. Readers would rightly object that my subjective probability assignments had not been adequately supported by the arguments in the paper, and I would be forced to remove them by referees, if the paper were not rejected out of hand. For the same reason, I don’t use language forcing my personal probability assignments on readers. There are always more arguments to consider, and readers differ quite dramatically in their priors. For that reason, concluding a paper with the conclusion that a claim like the time of perils hypothesis has a probability on the order of 10^(-100) or 10^(-200) would again, rightly provoke the objection that this claim has not

Seems like you and the other David T are talking past each other tbh.

Above you reasonably argue the [facetious] "time of carols" hypothesis is not remotely as credible as the time of perils hypothesis. But you also don't assign a specific credence to it, or provide an argument that the "time of carols" is impossible or even <1%[1]

I don't think it would be fair to conclude from this that you don't understand how probability works, and I also don't think that it is reasonable to assume that the probability of the 'time of carols' should be assumed suffici... (read more)

3
David Mathers🔸
Obviously David, as a highly trained moral philosopher with years of engagement with EA understands how expected value works though. I think the dispute must really be about whether to assign time of perils very low credence. (A dispute where I would probably side with you if "very low" is below say 1 in 10,000). 

The AI bubble popping would be a strong signal that this [capabilities] optimism has been misplaced.

Are you presupposing that good practical reasoning involves (i) trying to picture the most-likely future, and then (ii) doing what would be best in that event (while ignoring other credible possibilities, no matter their higher stakes)?

It would be interesting to read a post where someone tries to explicitly argue for a general principle of ignoring credible risks in order to slightly improve most-probable outcomes. Seems like such a principle would be pretty... (read more)

-2
Yarrow Bouchard 🔸
@Richard Y Chappell🔸, would you please do me the courtesy of acknowledging that you misunderstood my argument? I think this was a rather uncharitable reading on your part and would have been fairly easy to avoid. Your misreading was not explicitly forestalled by the text but not supported by the text, either, and there was much in the text to suggest I did not hold the view that you took to be the thesis or argument. I found your misreading discourteous for that reason.  Much of the post is focused on bad intellectual practices, such as: * Not admitting you got a prediction wrong after you got it wrong * Repeating the same prediction multiple times in a row and repeatedly getting it wrong, and seemingly not learning anything * Making fake graphs with false data, no data, dubious units of measurement, no units of measurements, and other problems or inaccuracies * Psychological or social psychological biases like millennialist cognitive bias, bias resulting from the intellectual and social insularity of the EA community, and possible confirmation bias (e.g. why hasn't Toby Ord's RL scaling post gotten much more attention?) * Acceptance or tolerance of arguments and assertions that are really weak, unsupported, or sometimes just bad I don't interpret your comment as a defense or endorsement of any of these practices (although I could if I wanted to be combative and discourteous). I'm assuming you don't endorse these practices and your comment was not intended as a defense of them. So, why reply to a post that is largely focused on those things as if the thesis or argument or thrust of the post is something other than that, and which was not said in the text?  On the somewhat more narrow point of AI capabilities optimism, I think the AI bubble popping within the next 5 years or so would be strong evidence that the EA community's AI capabilities optimism has been misplaced. If the large majority of people in the EA community only thought there's a 0.1% chance
4
Jason
It seems plausible that there could be significant adverse effects on AI Safety itself. There's been an increasing awareness of the importance of policy solutions, whose theory of impact requires support outside the AI Safety community. I think there's a risk that AI Safety is becoming linked in the minds of third parties with a belief in AGI imminence in a way that will seriously if not irrevocably damage the former's credibility in the event of a bubble / crash. One might think that publicly embracing imminence is worth the risk, of course. For example, policymakers are less likely to endorse strong action for anything that is expected to have consequences many decades in the future. But being perceived as crying wolf if a bubble pops is likely to have some consequences.
8
Yarrow Bouchard 🔸
No, of course not. I have written about this at length before, on multiple occasions (e.g. here and here, to give just two examples). I don’t expect everyone who reads one of my posts for the first time to know all that context and background — why would they? — but, also, the amount of context and background I have to re-explain every time I make a new post is already high because if I don’t, people will just raise the obvious objections I didn’t already anticipate and respond to in the post. But, in, short: no. I agree, but I didn’t say the AI bubble popping should settle the matter, only that I hoped it would motivate people to revisit the topic of near-term AGI with more open-mindedness and curiosity, and much less hostility toward people with dissenting opinions, given that there are already clear, strong objections — and some quite prominently made, as in the case of Toby Ord’s post on RL scaling — to the majority view of the EA community that seem to have mostly escaped serious consideration. You don’t need an external economic event to see that the made-up graphs in "Situational Awareness" are ridiculous or that AI 2027 could not rationally convince anyone of anything who is not already bought-in to the idea of near-term AGI for other reasons not discussed in AI 2027. And so on. And if the EA community hasn’t noticed these glaring problems, what else hasn’t it noticed? These are examples that anyone can (hopefully) easily understand with a few minutes of consideration. Anyone can click on one of the "Situational Awareness" graphs and very quickly see that the numbers and lines are just made-up, or that the y-axis has an ill-defined unit of measurement (“effective compute”, which is relative the tasks/problems compute is used for) or no unit of measurement (just “orders of magnitude”, but orders of magnitude of what?) and also no numbers. Plus other ridiculous features, such as claiming that GPT-4 is an AGI. With AI 2027, it takes more like 10-20 minute

Quick clarification: My target here is not so much people with radically different empirical beliefs (such that they regard vaccines as net-negative), but rather the particular form of status quo bias that I discuss in the original post.

My guess is that for relatively elite audiences (like those who read philosophy blogs), they're unlikely to feel attached to this status quo bias as part of their identity, but their default patterns of thought may lead them to (accidentally, as it were) give it more weight than it deserves. So a bit of heated rhetoric and stigmatization of the thought-pattern in question may help to better inoculate them against it.

(Just a guess though — I could be wrong!)

I think if some people are importantly right about something big, and others (esp. with more power) are importantly wrong, it's worth cheerleading getting things right even if it happens to correlate with your in-group!

3
justsaying
I wonder if the crux here is the effectiveness of your particular call to action: "Please strive to be less stupid, and call it out when you see it in others." I am guessing I am a pretty typical ea-forum reader in that I am appalled by the anti-vaccine turn of the u.s. government. I cannot do much to "be less stupid" by your lights in this particular respect because I generally  agree with you on the immorality of preventing vaccine access. But I also don't think calling out the stupidity when I see it is necessarily a good strategy.  That could be very alienating, reduce trust,  make anti-vaccine advocates feel victimized, inadvertently associate my various controversial views with vaccines, and increase backlash in the form of more anti-vaccine advocacy. I'm surr in some instances it is in fact the exactly right thing to do, but I also don't think it's the straightforward correct response towards people who genuinely think that vaccines cause autism, death, or other harms. 

Interesting post! Re: "how spotlight sizes should be chosen", I think a natural approach is to think about the relative priorities of representatives in a moral parliament. Take the meat eater problem, for example. Suppose you have some mental representatives of human interests, and some representatives of factory farmed animal interests. Then we can ask each representative: "How high a priority is it for you to get your way on whether or not to prevent this child from dying of malaria?" The human representatives will naturally see this as a very high prio... (read more)

3
Jim Buhler
Interesting! This resembles Michael St. Jules' hedging proposal. Is it that obvious? I find it hard to come up with interventions that reduce farmed animal suffering (and farmed animal suffering only, ignoring all the rest) more robustly than "minimize the number of entities that incidentally perpetuate this farmed animal suffering". I feel like this should actually be pretty high up the list for someone who strictly spotlights farmed animals (like the farmed animal representatives are supposed to in your spotlighting parliament, if I understand correctly). Why don't the managers of the Animal Welfare Fund support preventing the existence of more potential meat-eaters, then? (via, e.g., increasing x-risks.) I presume because * most of their biggest donors don't spotlight farmed animals, are concerned about effects on other things, and would stop donating to the fund if the managers were to do that. * the fund managers themselves, like most farmed animal people, also don't spotlight farmed animals and are concerned about effects on other things. Curious if you agree.  For the moral trade between the representatives of human victims of malaria and farmed animal representatives to be fair, in your setup, the preferences of the latter would have to actually stoplight farmed animals the same way the former spotlights human victims of malaria. I.e., the preferences of farmed animal representatives in your spotlighting parliament should not be those of real farmed animal advocates who are not spotlighting farmed animals (otherwise, they would obviously be pro-x-risks and stuff despite the downsides on other beings, the same way the representatives of human malaria victims are anti-poverty despite the meat-eater pb).

Funnily enough, the main example that springs to mind is the excessive self-flagellation post-FTX. Many distanced themselves from the community and its optimizing norms/mindset—for understandable reasons, but ones more closely tied to "expressing" (and personal reputation management) than to actually "helping", IMO.

I'd be curious to hear if others think of further candidate examples.

EA Infrastructure Fund or Giving What We Can? For the latter, "our best-guess giving multiplier for [2023-24] was approximately 6x".

2
Vasco Grilo🔸
Here is the post presenting GWWC's impact evaluation (I also looked into it), and here is a post from Melanie Basnak reflecting on Open Philanthropy's (OP's) grantmaking on effective giving.
2
Bentham's Bulldog
Thanks!

I think it's more like he disagrees with you about the relative strengths of the objections and responses. (fwiw, I'm inclined to agree with him, and I don't have any personal stake in the matter.)

7
Greg_Colbourn ⏸️
That's possible, but the responses really aren't good. For example:  And then there's a whole lot of moral philosophical-rationalist argument in the footnote. But he completely ignores an obvious option - working to oppose the potentially net-negative organisation. Or in this case: working towards getting an international treaty on AGI/ASI, that can rein in Anthropic and all the others engaged in the suicide race. I think Carlsmith could actually be highly impactful here, if he worked as a lobbyist or diplomat, and a public communicator (perhaps focused on an academic audience).

Any intellectual community will have (at least implicit) norms surrounding which assumptions / approaches are regarded as:

(i) presumptively correct or eligible to treat as a starting premise for further argument; this is the community "orthodoxy".

(ii) most plausibly mistaken, but reasonable enough to be worth further consideration (i.e. valued critiques, welcomed "heterodoxy")

(iii) too misguided to be worth serious engagement.

It would obviously be a problem for an intellectual community if class (ii) were too narrow. Claims like "dissent isn't welcome" imp... (read more)

I'm not seeing the barrier to Person A's thinking there's a 1/1000 chance, conditional on reaching the 50th century, of going extinct in that century. We could easily expect to survive 50 centuries at that rate, and then have the risk consistently decay (halving each century, or something like that) beyond that point, right?

If you instead mean to invoke, say, the 50 millionth century, then I'd think it's crazy on its face to suddenly expect a 1/1000 chance of extinction after surviving so long. That would no longer "seem, on the face of it, credible".

Am I missing something?

1
tobycrisford 🔸
I was assuming in my example that the "Time of perils" that Person A believes we might be living through is supposed to be over by the 50th century, so that the 50th century is already in the period where extinction risk is supposed to have become very low. But suppose Person A adopts your alternative probabilities instead. Person A now believes in a 1/1000 chance of going extinct in the 50th century, conditional on reaching it, and then the probability halves in each century after that. But if that's what they believe, you can now just run my argument on the 100th century instead. Person A now proposes a probability of ~10^(-18) of going extinct in the 100th century (conditional on reaching it) which seems implausibly overconfident to me on the face of it! I agree with you, that if we were considering the 50 millionth century, then a probability of 1/1000 would be far too high. I agree that it would be crazy to stipulate a probability for the Nth century that is much higher than 1/N, because surviving N centuries is evidence that typical extinction risk per century is lower than this (except maybe if we were considering centuries close to the time the sun is expected to die..?) But my point is that in order to get a truly big future, with the kind of stakes that dominate our expected value calculations, then we need the probability of extinction to decay much faster than 1/N. We need the "Time of Perils" hypothesis. It needs to decay exponentially* (something like the halving that you've suggested). And before too long that exponential decay is going to lead to implausibly low probabilities of extinction.   *Edit: Actually not too confident on this claim now I think it through some more. Perhaps you can still get a very large future with sub-exponential decay. Maybe this is another way out for Person A in fact!

Thanks, yeah, I like your point there that "false negatives are costlier than false positives in this case", and so even <50% credence can warrant significant action. (I wouldn't literally say we should "act as if 3H is true" in all respects—as per Nuno's comment, uncertainty may justify some compounding "patient philanthropy", which could have high stakes if the hinge comes later. But that's a minor quibble: I take myself to be broadly in agreement with your larger gist.)

My main puzzlement there is how you could think that you ought to perform an act that you simultaneously ought to hope that you fail to perform, subsequently (and predictably) regret performing, etc. (I assume here that all-things-considered preferences are not cognitively isolated, but have implications for other attitudes like hope and regret.) It seems like there's a kind of incoherence in that combination of attitudes, that undermines the normative authority of the original "ought" claim. We should expect genuinely authoritative oughts to be more wholeheartedly endorsable.

1
Jakob Lohmar
That seems like a strange combination indeed! I will need to think more about this...

Right, so one crucial clarification is that we're talking about act-inclusive states of affairs, not mere "outcomes" considered in abstraction from how they were brought about. Deontologists certainly don't think that we can get far merely thinking about the latter, but if they assess an action positively then it seems natural enough to take them to be committed to the action's actually being performed (all things considered, including what follows from it). I've written about this more in Deontology and Preferability. A key passage:

If you think that other

... (read more)
3
Jakob Lohmar
Yeah that makes sense to me. I still think that one doesn't need to be conceptually confused (even though this is probably a common source of disagreement) to believe both that (i) one action's outcome is preferable to the other action's outcome even though (ii) one ought to perform the latter action. For example, one might think the former outcome is overall preferable because it has much better consequences. But conceptual possibility aside, I agree that this is a weird view to have. At the very least, it seems that all else equal one should prefer the outcome of the action that one takes to be the most choiceworthy. Not sure if it has some plausibility to say that this doesn't necessarily hold if other things are not equal - such as in the case where the other action has the better consequences.

Thanks! You might like my post, 'Axiology, Deontics, and the Telic Question' which suggests a reframing of ethical theory that avoids the common error. (In short: distinguish ideal preferability vs instrumental reasoning / decision theory rather than axiology vs deontics.)

I wonder if it might also help address Mogensen's challenge. Full aggregation seems plausibly true of preferability not just axiology. But then given principles of instrumental rationality linking reasons for preference/desire to reasons for action, it's hard to see how full aggregation c... (read more)

3
Jakob Lohmar
Thanks - also for the link! I like your notion of preferability and the analysis of competing moral theories in terms of this notion. What makes me somewhat hesitant is that the objects of preferability, in your sense, seem to be outcomes or possible worlds rather that the to-be-evaluated actions themselves? If so, I wonder if one could push back against your account by insisting that the choiceworthiness of available acts is not necessarily a function of the preferability of their outcomes since... not all morally relevant features of an action are necessarily fully reflected in the preferability of its outcome? But assuming that they are, I guess that non-consequentialists who reject full aggregation would say that the in-aggregate larger good is not necessarily preferable. But I'm not sure. I agree that this seems not very intutive.

To be clear: I'd be excited for more people to look into these claims! Seems worth investigating. But it's not my comparative advantage.

4
Vasco Grilo🔸
Good to know! I think highlighting the importance of the topic is one way of getting more people to investigate it ;).

Sorry, I don't think I have relevant expertise to assess such empirical claims (which is why I focus more on hypotheticals). It would certainly be convenient if helping people turned out to be the best way to also reduce non-human suffering! And it could be true (I don't take convenience to be an automatic debunker or anything). I just have no idea.

4
Vasco Grilo🔸
Note many do not engage with my empirical claims about effects on soil animals for philosophical reasons ("that assumes utilitarianism!"), and assessing these is a comparative advantage you have. Moreover, in the effective altruism community, some people initially concluded that targeting farmed animals increases animal welfare very cost-effectively based on classical utilitarian reasons, and I expect many of those people to depart from classical utilitarianism after realising it plus empirical evidence point towards prioritising soil animals much more than farmed animals. People with a scout mindset, and not invested in any particular way of increasing animal welfare could be in a better place to assess the extent to which such departures from classical utilitarianism are post hoc justifications, eventually in the form of empirical arguments which are seldomly investigated. As a concrete example of philosophical arguments being important, @Wladimir J. Alonso, the founder and innovation director of the Welfare Footprint Institute (WFI), values averting intense suffering more than is justified by its intensity. Some people, not necessarily Wladimir, endorse this so much that they would prefer averting 1 h of excruciating pain over an infinite amount of annoying pain, and would never consider effects on soil animals given other animals can experience much more intense suffering.
2
Vasco Grilo🔸
I suspect you are overestimating the difficulty of checking the empirical claims. I am pretty confident that funding HIPF decreases the suffering of soil animals much more cost-effectively than cage-free corporate campaigns decrease the suffering of chickens. I estimate HIPF decreases 5.07 billion soil-animal-years per $, and that cage-free corporate campaigns improve 10.8 chicken-years per $. For HIPF to decrease the suffering of soil animals less cost-effectively than cage-free corporate campaigns decrease the suffering of chickens, the reduction in suffering due to improving 1 chicken-year would have to be larger than than from decreasing 469 M soil-animal-years (= 5.07*10^9/10.8), whereas I calculate chickens only have 921 k (= 221*10^6/240) times as many neurons as nematodes, which are the soil animals with the fewest neurons. Moreover, I think the number of neurons underestimates the absolute value of the welfare per animal-year. Rethink Priorities' (RP) moral weight project assumes shrimps have 10^-6 as many neurons as humans (see Table 5 here), but the estimate for their welfare range in Table 8.6 of Bob Fischer's book is 8 % that of humans. For the reasons above, I am also pretty confident that funding HIPF changes the welfare of soil animals much more cost-effectively than cage-free corporate campaigns increase the welfare of chickens. However, I am very uncertain about whether HIPF increases or decreases animal welfare due to being very uncertain about whether soil nematodes have positive or negative lives. To clarify, I believe researching whether soil nematodes have positive or negative lives would increase animal welfare even more cost-effectively than funding HIPF, but that this still increases animal welfare much more cost-effectively than interventions targeting farmed animals. I would prefer saving human lives to decrease animal welfare such that soil animals had positive lives. I think saving human lives decreases soil-animal-years, and therefo

Thanks for your reply! Working backwards...

On your last point, I'm fully on board with strictly decoupling intrinsic vs instrumental questions (see, e.g., my post distinguishing telic vs decision-theoretic questions). Rather, it seems we just have very different views about what telic ends or priorities are plausible. I give ~zero credence to pro-annihilationist views on which it's preferable for the world to end than for any (even broadly utopian) future to obtain that includes severe suffering as a component. Such pro-annihilationist lexicality strikes m... (read more)

Regarding the "world-destruction" reductio:

this isn't strong evidence against the underlying truth of suffering-focused views. Consider scenarios where the only options are (1) a thousand people tortured forever with no positive wellbeing whatsoever or (2) painless annihilation of all sentience. Annihilation seems obviously preferable.

I agree that it's obviously true that annihilation is preferable to some outcomes. I understand the objection as being more specific, targeting claims like: 

(Ideal): annihilation is ideally desirable in the sense that it... (read more)

2
RalfG
@Richard Y Chappell🔸 what do you think of Aaron's response below? I am using this comment to flag that the discussion you two are having seems very important to me, and I look forward to seeing your reply to Aaron's points.
4
Aaron Bergman
Thanks! Yeah I mean on the first one, I acknowledge that this seems pretty counterintuitive to me but again just don't think it is overwhelming evidence against the truth of the view. Perhaps a reframing is "would this still seem like a ~reductio conditional on a long reflection type scenario that results in literally everyone agreeing that it's desirable/good?" And I don't mean this in the sense of just "assume that the conclusion is ground truth" - I mean it in the sense of "does this look as bad when it doesn't involve anyone doing anything involuntary?" to try to tease apart whether intuitions around annihilation per se are to any extent "just" a proxy for guarding against the use of force/coercion/lack of consent. Another way to flip the 'force' issue would be "suppose a society concludes unanimously including via some extremely deliberative process (that predicts and includes the preferences of potential future people) that annihilation is good and desired. Should some outside observer forcibly prevent them taking action to this end (assume that the observer is interested purely in ethics and doesn't care about their own existence or have valenced experience)?" I'll note that I can easily dream up scenarios where we should force people, even a whole society, to do something against its will. I know some will disagree, but I think we should (at least in principle, implementation is messy) forcibly prevent people from totally voluntarily being tortured (assume away masochism - like suppose the person just has a preference for suffering that results in pure suffering with no 'secretly liking it' along for the ride) This one I more eagerly bite the bullet on, it just straightforwardly seems true to me that this is possible in principle (i.e., such a world could/would be genuinely very bad). And relevantly, orthodox utilitarianism also endorses this in principle, some of the time (i.e. just add up the utils, in principle one suffering monster can have enough

Isn't the point of the Long Reflection to avoid "locking in" irreversible mistakes? Extinction, for example, is irreversible. But large population isn't. So I don't actually see any sense in which present "min-natalism" maintains more future "optionality" (or better minimizes moral risks) than pro-natalism. Both leave entirely open what future generations choose to do. They just differ in our present population target. And presently aiming for a "minimal population" strikes me as much the worse and riskier of the two options, for both intrinsic moral reasons and instrumental ones like misjudging / undershooting the minimally sustainable level.

4
Wei Dai
Perhaps the most important question is whether you support a restriction on space colonization (completely or to a few nearby planets) during the Long Reflection. Unrestricted colonization seems good from a pure pro-natalist perspective, but bad from an optionalist perspective, as it makes much more likely that if anti-natalism (or adjacent positions like there should be strict care or controls over what lives can be brought into existence) is right, some of the colonies will fail to reach the correct conclusion and go on to colonize the universe in an unrestricted way, thus making humanity as a whole unable to implement the correct option. If you do support such a restriction, then I think we agree on "the highest order bits" or the most important policy implication of optionalism, but probably still disagree on what is the best population size during the Long Reflection, which may be unresolvable due to our differing intuitions. I think I probably have more sympathy for anti-natalist intuitions than you do (in particular that most current lives may have negative value and people are mistaken about this), and worry more that creating negative-value lives and/or bringing lives into existence without adequate care could constitute a kind of irreversible or irreparable moral error. Unfortunately I do not see a good way to resolve such disagreements at our current stage of philosophical progress.

Thanks! I'd previously found it a bit stressful deciding which posts were relevant enough to share here, so I ended up outsourcing the decision to the good folks on the Forum team (who also take care of the cross-posting). Accordingly, a good share of the appreciation is owed to them! :-)

Your executive summary (quoted below) appears to outright assert that quantification is "harmful" and "results in poor decision making". I don't think those claims are well-supported.

If you paint a picture that focuses only on negatives and ignores positives, it's apt to be a very misleading picture. There may be possible ways to frame such a project so that it comes off as just "one piece of the puzzle" rather than as trying to bias its readership towards a negative judgment. But it's an inherently risky/difficult undertaking (prone to moral misdirection)... (read more)

Looking at EA’s history can show us strong and in many cases negative influence from utilitarian ideas.

It also shows strong and in vastly more cases positive influence from (what you call) "utilitarian" ideas (but really ought to be more universal--ideas like that it is better to do more good than less, and that quantification can help us to make such trade-offs on the basis of something other than mere vibes).

Unless there's some reason to think that the negative outweighs the positive, you haven't actually given us any reason to think that "utilitarian in... (read more)

1
Leo_
Yes! I don't deny the positive impact that has come from EA, and the focus on quantification, and have tried to touch on that in the conclusion section as well. I very much believe everyone would benefit from better use of quantification, evidence, and rationality.  I'm not sure I've got the arguments or evidence to say whether EA's utilitarian influence is net positive or negative (and I've seen arguments in both directions), but that's not my point here. I'm not arguing from a utilitarian basis. I'm trying to paint a picture of the scope and impact of utilitarian thought's negative impacts on EA to try to help EAs and other people invested in doing good better evaluate the impacts and viability of such an ideology. This is only meant to be one piece of the puzzle.

It's mostly not anything specific to going vegan. Just the general truism that effort used for one purpose could be used for something else instead. (Plus I sometimes donate extra precisely for the purpose of "offsetting", which I wouldn't otherwise be motivated to do.)

Mostly just changing old habits, plus some anticipated missing of distinctive desired tastes. It's not an unreasonable ask or anything, but I'd much rather just donate more. (In general, I suspect there's insufficient social pressure on people to increase our donations to good causes, which also shouldn't be "so effortful", and we likely overestimate the personal value we get from marginal spending on ourselves.)

0
ClimateDoc
What do you think it is about going vegan that would prevent you from donating more? I'm still not sure of the causal link.
2
Xylix
I think in general if we agree to a ballpark of "10% donations is enough to satisfice some goodness thresholds", and also to "It would be good for social pressure to exist for everyone to do at least threshold amount of good", I think it raises various considerations. 10% makes sense to me as a schelling point (and I think the tables that scale by income bracket are also sensible).  But if the threshold amount of good would be "Donate 10%, aim for an impactful career, become vegan" (which is what I feel the social pressure inside EA is pointing towards), I think that is already a significant ask for many people. I think it is also important to note that some people are more motivated by trying to maximize impact and offset harm, and some people more  motivated by minimizing harm and satisficing for impact. (Of course a standard total utilitarian model would output that whatever maximizes your net impact is best, but human value systems aren't perfectly utilitarian.) How do "donate 10%, become vegan, aim for an impactful", and "donate 30%", and "donate 20%, aim for an impactful career" compare in effectiveness as norms? I think this is pretty hard to estimate. What kind of social pressure are you pointing here? Is it more in the direction of "donate 30%" or "donate as much as you can and aim for an impactful career?" Or do you mean social pressure in the wider society, and not within the EA community? (Fwiw I think people underestimate the value of effective marginal spending on themselves, when considering areas of spending where there is space for significant extra value (Like purchasing more free time.). People plausibly overestimate the value on some other things, especially if one doesn't do spending introspectiont.)

I don't understand the relevance of the correlation claim. People who care nothing for animals won't do either. But that doesn't show that there aren't tradeoffs in how to use one's moral efforts on the margins. (Perhaps you're thinking of each choice as a binary: "donate some" Y/N + "go vegan" Y/N? But donating isn't binary. What matters is how much you donate, and my suggestion is that any significant effort spent towards adopting a vegan diet might be better spent on further increasing one's donations. It depends on the details, of course. If you find a... (read more)

4
Jacob Watts🔸
Ya, idk, I am just saying that the tradeoff framing feels unnatural. Or, like, maybe that's one lens, but I don't actually generally think in terms of tradeoffs b/w my moral efforts. Like, I get tired of various things ofc, but it's not usually just cleanly fungible b/w different ethical actions I might plausibly take like that. To the extent it really does work this way for you or people you know on this particular tradeoff, then yep; I would say power to ya for the scope sensitivity. I agree that the quantitative aspect of donation pushes towards even marginal internal tradeoffs here mattering and I don't think I was really thinking about it as necessarily binary. 

My main confusion with your argument is that I don't understand why donations don't also count as "personal ethics" or as "visible ethical action" that could likewise "ripple outward" and be replicated by others to good effect. (I also think the section on "equity" fundamentally confuses what ethics should be about. I care about helping beneficiaries, not setting up an "equitable moral landscape" among agents, if the latter involves preventing the rich from pursuing easy moral wins because this would be "unfair" to those who can't afford to donate.)

One mor... (read more)

6
ClimateDoc
Out of interest, what is it you consider so effortful about becoming vegan that it would so substantially reduce the effort you could put towards other causes? Do you think it is knock-on effects of enjoying food less, effort required to learn to change your meals, effects from finding it harder socially, or something else? The actual effort to change to a vegan diet isn't that high in my view, at least if you have access to a decent supermarket (having done it) - it's just learning to make some different foods and remembering to buy some multivitamins once in a while (or at least B12). Once you've done the learning, it's not really an ongoing extra effort (like there's not really an ongoing effort in knowing how to cook omni food), and the benefits accrue over time. I wonder if people overestimate the effect on enjoyment. First, if you find vegan alternatives that you enjoy, then you don't lose out a lot. Second, I think most EAs are probably familiar with hedonic adaptation, and how your happiness levels seem to be pretty resilient to lifestyle changes in the long-term (hence making donating money seem like less of a big deal) - so switching food also seems unlikely to really make you emotionally worse off. Third, we probably spend less than an hour per day with food in our mouths - it doesn't seem like it should be that important to overall wellbeing - I recall Daniel Kahnemann making a point that we overestimate the impact of certain things because we imagine the effect when we are doing them but not the lack of effect during all the time we are not doing them. Social is quite situation-dependent. But if it's just that you have friends who take you to restaurants with no decent vegan option, it doesn't prevent being vegan in other meals. Shared meals with family who won't accept vegan food would seem trickier, but again there are surely some meals where a person could normally be independent. Edit - or I guess worries about health could be another reason? We
5
Jacob Watts🔸
I agree with 1, but I think the framing feels forced for point #2. I don't think it's obvious that these actions would be strongly in tension with each other. Donating to effective animal charities would correlate quite strongly with being vegan. Homo economicus deciding what to eat for dinner or something lol. I actually totally agree that donations are an important part of personal ethics! Also, I am all aboard for the social ripple effects theory of change for effective donation. Hell yes to both of those points. I might have missed it, but I don't know that OP really argues against those contentions? I guess they don't frame it like that though.

Just to clarify: Spears & Geruso's argument is that average (and not just total) quality of life will be significantly worse under depopulation relative to stabilization. (See especially the "progress comes from people" section of my review.)

The authors discuss this a bit. They note that even "higher fertility" subcultures are trending down over time, so it's not sufficiently clear that anyone is going to remain "above replacement" in the long run. That said, this does seem the weakest point for thinking it an outright extinction risk. (Though especially if the only sufficiently high-fertility subcultures are relatively illiberal and anti-scientific ones - Amish, etc. - the loss of all other cultures could still count as a significant loss of humanity's long-term potential! I hope it's OK to note this; I know the mods are wary that discussion in this vicinity can often get messy.)

I wrote "perhaps the simplest and most probable extinction risk". There's room for others to judge another more probable. But it's perfectly reasonable to take as most probable the only one that is currently on track to cause extinction. (It's hard to make confident predictions about any extinction risks.) I think it would be silly to dismiss this simply due to uncertainty about future trends.

What reason is there to think that demographic trends will suddenly reverse? If it isn't guaranteed to reverse, then it is an extinction risk.

6
David Mathers🔸
The risk is non-zero, but you made a stronger claim that it was "the most probable extinction risk around".  EDIT: As for reasons to think they will reverse, they seem to be a product of liberal modernity, but currently we need a population way, way above the minimum viable number to keep long term modernity going. Maybe AI could change that I guess, but it's very hard to make predictions about the demographic trend if AI does all work. 

I'd guess that (for many readers of the book) less air travel outweighs "buying more" furniture and kids toys, at least. But the larger point isn't that the change is literally zero, but that it doesn't make a sufficiently noticeable change to near-term emissions to be an effective strategy. It would be crazy to recommend a DINK lifestyle specifically in order to reduce emissions in the next 25 years. Like boycotting plastic straws or chatgpt.

Updated to add the figure from this paper, which shows no noticeable difference by 2050 (and little difference even... (read more)

2
NickLaing
There have been a couple of studies showing that families with kids emit more than those without. Including this one from Sweden, where they emit 25 percent more. https://www.sciencedaily.com/releases/2020/04/200415152921.htm That graph just shows that IF we manage to get emissions to plummet late in the century then a difference of 15 percent in population by 2100 might not make much of a difference to climate change. That's fine but there are many assumptions there.  I don't think having a few less people changes the game in climate change, but nor do I think it's very bad in other ways either.

As a general rule, it isn't necessary to agree on the ideal target in order to agree directionally about what to do on present margins. For example, we can agree that it would be good to encourage more effective giving in the population, without committing to the view (that many people would "personally disagree" with) that everyone ought to give to the point of marginal utility, where they are just as desperate for their marginal dollar as their potential beneficiaries are.

The key claim of After the Spike is that we should want to avoid massive depopulati... (read more)

Everyone has fundamental assumptions. You could imagine someone who disagrees with yours calling them "just vibes" or "presuppositions", but that doesn't yet establish that there's anything wrong with your assumptions. To show an error, the critic would need to put forward some (disputable) positive claims of their own.

The level of agreement just shows that plenty of others share my starting assumptions.

If you take arguments to be "circular" whenever a determined opponent could dispute them, I have news for you: there is no such thing as an argument that l... (read more)

1
Manuel Del Río Rodríguez 🔹
I am trying to articulate (probably wrongly) the disconnect I perceive here. I think 'vibes' might sound condescending, but ultimately, you seem to agree with assumptions (like math axioms) not being amenable to disputation. Like, technically, in philosophical practice, one can try to show, I imagine, that given assumption x some contradiction (or at least, something very generally perceived as wrong and undesirable) follows. I do share the feeling expressed by Charlie Guthmann here that a lot of starting arguments for moral realists are just of the type 'x is obvious/self-evident/feels good to be/feels worth believing', and when stated in that way, they feel equally obviously false to those who don't share those intuitions, and as magical thinking ('If you really want something, the universe conspires to make it come about' Paulo Coelho style). I feel more productive engaging strategies should just avoid altogether any claims of the mentioned sort, and perhaps start with stating what might follow from realist assumptions that might be convincing/persuasive to the other side, and vice versa.

I agree it's often helpful to make our implicit standards explicit. But I disagree that that's "what we're actually asking". At least in my own normative thought, I don't just wonder about what meets my standards. And I don't just disagree with others about what does or doesn't meet their standards or mine. I think the most important disagreement of all is over which standards are really warranted. 

On your view, there may not be any normative disagreement, once we all agree about the logical and empirical facts. I think it's key to philosophy that the... (read more)

4
Peter Wildeford
Really warranted by what? I think I'm an illusionist about this in particular as I don't even know what we could be reasonably disagreeing over. For a disagreement about facts (is this blue?), we can argue about actual blueness (measurable) or we can argue about epistemics (which strategies most reliably predict the world?) and meta-epistemics (which strategies most reliably figure out strategies that reliably predict the world?), etc. For disagreements about morals (is this good?), we can argue about goodness but what is goodness? Is it platonic? Is it grounded in God? I'm not even sure what there is to dispute. I'd argue the best we can do is argue to our shared values (perhaps even universal human values, perhaps idealized by arguing about consistency etc.) and then see what best satisfies those. ~ Right - and this matches our experience! When moral disagreements persist after full empirical and logical agreement, we're left with clashing bedrock intuitions. You want to insist there's still a fact about who's ultimately correct, but can't explain what would make it true. ~ I think we're successfully engaging in a dispute here and that does kind of prove my position. I'm trying to argue that you're appealing to something that just doesn't exist and that this is inconsistent with your epistemic values. Whether one can ground a judgement about what is "really warranted" is a factual question. ~ I want to add that your recent post on meta-metaethical realism also reinforces my point here. You worry that anti-realism about morality commits us to anti-realism about philosophy generally. But there's a crucial disanalogy: philosophical discourse (including this debate) works precisely because we share epistemic standards - logical consistency, explanatory power, and various other virtues. When we debate meta-ethics or meta-epistemology, we're not searching for stance-independent truths but rather working out what follows from our shared epistemic commitments. Th

Actually have high integrity, which means not being 100% a utilitarian/consequentialist

Sorry for the necro-reply, but just saw this and wanted to register that I think a 100% utilitarian/consequentialist can still genuinely have high integrity. (I think people are generally quite confused about what a fitting consequentialist mindset looks like. It absolutely is not: "do whatever I naively estimate will maximize expected value, without regard for trustworthiness etc.") See, e.g., Naïve Instrumentalism vs Principled Proceduralism.

Load more