In April when we released my interview with SBF, I attempted to very quickly explain his views on expected value and risk aversion for the episode description, but unfortunately did so in a way that was both confusing and made them sound more like a description of my views rather than his.
Those few paragraphs have gotten substantial attention because Matt Yglesias pointed out where it could go wrong, and wasn't impressed, thinking that I'd presented an analytic error as "sound EA doctrine".
So it seems worth clarifying what I actually do think. In brief, I entirely agree with Matt Yglesias that:
- Returns to additional money are certainly not linear at large scales, which counsels in favour of risk aversion.
- Returns become sublinear more quickly when you're working on more niche cause areas like longtermism, relative to larger cause areas such as global poverty alleviation.
- This sublinearity becomes especially pronounced when you're considering giving on the scale of billions rather than millions of dollars.
- There are other major practical considerations that point in favour of risk-aversion as well.
(SBF appears to think the effects above are smaller than Matt or I do, but it's hard to know exactly what he believes, so I'll set that aside here.)
———
The offending paragraphs in the original post were:
"If you were offered a 100% chance of $1 million to keep yourself, or a 10% chance of $15 million — it makes total sense to play it safe. You’d be devastated if you lost, and barely happier if you won.
But if you were offered a 100% chance of donating $1 billion, or a 10% chance of donating $15 billion, you should just go with whatever has the highest expected value — that is, probability multiplied by the goodness of the outcome [in this case $1.5 billion] — and so swing for the fences.
This is the totally rational but rarely seen high-risk approach to philanthropy championed by today’s guest, Sam Bankman-Fried. Sam founded the cryptocurrency trading platform FTX, which has grown his wealth from around $1 million to $20,000 million.
The point from the conversation that I wanted to highlight — and what is clearly true — is that for an individual who is going to spend the money on themselves, the fact that one quickly runs out of any useful way to spend the money to improve one's well-being makes it far more sensible to receive $1 billion with certainty than to accept a 90% chance of walking away with nothing.
On the other hand, if you plan to spend the money to help others, such as by distributing it to the world's poorest people, then the good one does by dispersing the first dollar and the billionth dollar are much closer together than if you were spending them on yourself. That greatly strengthens the case for taking the risk of receiving nothing in return for a larger amount on average, relative to the personal case.
But: the impact of the first dollar and the billionth dollar aren't identical, and in fact could be very different, so calling the approach 'totally rational' was somewhere between an oversimplification and an error.
———
Before we get to that though, we should flag a practical consideration that is as important, or maybe more so, than getting the shape of the returns curve precisely right.
As Yglesias points out, once you have begun a foundation and people are building organisations and careers in the expectation of a known minimum level of funding for their field, there are particular harms to risking your entire existing endowment in a way that could leave them and their work stranded and half-finished.
While in the hypothetical your downside is meant to be capped at zero, in reality, 'swinging for the fences' with all your existing funds can mean going far below zero in impact.
The fact that many risky actions can result in an outcome far worse than what would have happened if you simply did nothing, is a reason for much additional caution, one that we wrote about in a 2018 piece titled 'Ways people trying to do good accidentally make things worse, and how to avoid them'. I regret that I failed to ask any questions that highlighted this critical point in the interview.
(This post won't address the many other serious issues raised by the risk-taking at FTX, which, according to news reports, have gone far beyond accepting the possibility of not earning much profit, and which can't be done justice here.
If those reports are accurate, the risk-taking at FTX was not just a coin flip that came up tails — it was immoral and perhaps criminal itself due to the misappropriation of other people's money for the purpose of risky investments. This has resulted in incalculable harm to customers, investors, trust in broader society, and has set back all the causes some of FTX's staff said they wanted to help.)
———
To return to the question of declining returns and risk aversion — just as one slice of pizza is delicious but a tenth may not be enjoyable to eat at all, people trying to use philanthropy to do good do face 'declining marginal returns' as they incrementally try to give away more and more money.
How fast that happens is a difficult empirical question.
But if one is funding the fairly niche and neglected problems SBF said he cared the most about, it's fair to say that any foundation would find it difficult to disperse $15 billion to projects they were incredibly excited about.
That's because a foundation with $15 billion would end up being a majority of funding for those areas, and so effectively increase the resources going towards them by more than 2-fold, and perhaps as much as 5-fold, depending on how broad a net they tried to cast. That 'glut' of funding would result in some more mediocre projects getting the green light.
Assuming someone funded projects starting with the ones they believed would have the most impact per dollar, and then worked down — the last grant made from such a large pot of money will be clearly worse, and probably have less than half the expected social impact per dollar as the first.
So between $1 billion with certainty versus a 10% chance of $15 billion, one could make a theoretical case for either option — but if it were me I would personally lean towards taking the $1 billion with certainty.[1]
Notice that by contrast, if I were weighing up a guaranteed $1 million against a 10% chance of $15 million, the situation would be very different. For the sectors I'd be most likely to want to fund, $15 million from me, spread out over a period of years, would represent less than a 1% increase, and so wouldn't overwhelm their capacity to sensibly grow, leading the marginal returns to decline more slowly. So in that case, setting aside my personal interests, I would opt for the 10% chance of $15 million.
———
Another massive real-world consideration we haven't mentioned yet which pushes in favour of risk aversion is the following: how much you are in a position to donate is likely to be strongly correlated with how much other donors are able to donate.
In practice risk-taking around philanthropy will mostly centres on investing in businesses. But businesses tend to do well and poorly together, in cycles, depending on broad economic conditions. So if your bets don't pay off, say, because of a recession, there's a good chance other donors will have less to give as well. As a result, you can't just take the existing giving of other donors for granted.
This is one reason for even small donors to have a reasonable degree of risk aversion. If they all adopt a risk-neutral strategy they may all get hammered at once and have to massively reduce their giving simultaneously, adding up to a big negative impact in aggregate.
This is a huge can of worms that has been written about by Christiano and Tomasik as far back as 2013, and more recently by my colleague Benjamin Todd.
———
This post has only scratched the surface of the analysis one could do on this question, and attempted to show how tricky it can be. For instance, we haven't even considered:
- Uncertainty about how many other donors might join or drop out of funding similar work in future.
- Indirect impacts from people funding similar work on adjacent projects.
- Uncertainty about which problems you'll want to fund solutions to in future.
I regret having swept those and other complications under the rug for the sake of simplicity in a way that may well have confused some listeners to the show and seemed like an endorsement of an approach that is risk-neutral with respect to dollar returns, which would in fact be severely misguided.
(If you'd like to hear my thoughts on FTX more generally as opposed to this technical question you can listen to some comments I put out on The 80,000 Hours Podcast feed.)
Thanks for the reply, Neel.
First I should note that I wrote my previous comment on my phone in the middle of the night when I should have been asleep long before, so I wasn't thinking fully about how others would interpret my words. Seeing the reaction to it I see that the comment didn't add value as written and I probably should just just waited to write it later when I could unambiguously communicate what bothered me about it at length (as I do in this comment).
To clarify, I agree with you an Yglesias that most longtermists are working on things like preventing AI from causing human extinction only a few decades from now, meaning the work is also very important from a short-term perspective that doesn't give weight to what happens after say, 2100. So I agree with you that ""longtermism" in practice mostly manifests as working on [reducing near-term] x-risk."
I also agree that there's an annoying thing about "longtermist EA marketing" related to the above. (I liked your Simplify EA Pitches to "Holy Shit, X-Risk".)
To explain what bothered me about Yglesias' post more clearly, let me first say that my answer to "What's long-term about "longtermism"?" is the (my words:) "giving significant moral weight to the many potential beings that might come to exist over the course of the long-term future (trillions upon trillions of years)" part of longtermism. Since that "part" of longtermism actually is wholly what long-termism is, one could also just answer "longtermism is long-term".
In other words, the question sounds similar to (though not exactly like) "What's liberal about liberalism?" or "What's colonial about colonialism?"
I therefore would expect a post with the title "What's long-term about "longtermism"?" to explain that longtermism is a moral view that gives enough moral weight to the experiences of future beings that might come to exist such that the long-term future of life matters a lot in expectation given how long that future might be (trillions upon trillions of years) and how much space in the universe it might make use of (a huge number of resources beyond this pale blue dot).
But instead, Yglesias' post points out that the interventions that people who care about beings in the long-term future think are most worthwhile often look like things that people who didn't care about future generations would also think are important (if they held the same empirical beliefs about near-term AI x-risk, as some of them do).
And my reaction to that is, okay, yes Yglesias, I get it and agree, but you didn't actually argue that longtermism isn't "long term" like your title suggested you might. Longtermism absolutely is "long-term" (as I described above). The fact that some interventions favored by longtermists also look good from non-longtermist moral perspectives doesn't change that.
Yglesias:
This statement is a motte in that he says "any particularly unusual ideas about the long term" rather than "longtermism".
(I think the vast majority of people care about future generations in some capacity, e.g. they care about their children and their friends' children before the children are born. Where we draw the line between this and some form of "strong longtermism" that actually is "particularly unusual" is unclear to me. E.g. I think most people also actually care about their friends' unborn children's unborn children too, though people often don't make this explicit so it's unclear to me how unusual the longtermism moral view actually is.)
If we replace the "any particularly unusual ideas about the long term" with "longtermism" then Yglesias' statement seems to become an easily-attackable bailey.
In particular, I would say that the statement seems false and uncharitable and unsubstantiated. Yglesias is making a generalization, and obviously it's a generalization that's true of some people working on reducing x-risks posed by AI, but I know it's definitely not true of many others working on x-risks. E.g. There are definitely many self-described longtermists working on reducing AI x-risk who are in fact motivated by wanting to make sure that humanity doesn't go extinct so that future people can come to exist.
While I'm not an AI alignment researcher, I've personally donated a substantial fraction of my earnings to people doing this work and do many things that fall in the movement building / field building category to try to get other people to work on reducing AI risk, and I can personally attest to the fact that I care a lot more about preventing extinction to ensure that future beings are able to come to exist and live great lives than I care about saving my own life and everyone I know and love today. It's not that I don't care about my own life and everyone else alive today--I do a tremendous amount--but rather that as Derrick Parfit says the worst part about everyone dying today would by far be the loss of all future value, not 8 billion humans lives being cut short.
I hope this clarifies my complaint about Yglesias' What's long-term about "longtermism"? post.
The last thing that I'll say in this comment is that I found the post via Yglesias' Some thoughts on the FTX collapse post that Rob responded to in the OP. Here's how Yglesias cited his "What's long-term about "longtermism"?" in the FTX collapse post:
As I've explained at length in this comment, I think longtermism is not confused. Contra Yglesias (though again Yglesias doesn't actually argue against the claim, which is what I found annoying), longtermism is in fact "long-term."
Yglesias is actually the one who is confused, both in his failure to recognize that longtermism is in fact "long-term" and because he confuses/conflates the motivations of some people working on reducing near-term extinction risk from AI with "longtermism."
Again: Longtermism is a moral view that emphasizes the importance of future generations throughout the long term future. People who favor this view (self-identified "longtermist" EAs) often end up favoring working on reducing the risk of near-term human extinction from AI. People who are only motivated by what happens in the near term may also view working on this problem to be important. But that does not mean that longtermism is not "long term", because "the motivation of some people working on reducing near-term extinction risk from AI" is not "longtermism."
I want to say "obviously!" to this (because that's what I was thinking when I read Yglesias' post late last night and which is why I was annoying by it), but I also recognize that EAs' communications related to "longtermism" have been far from perfect and it's not surprising that some smart people like Yglesias are confused.
In my view it probably would have been better to have and propagate a term for the general idea of "creating new happy beings is a morally good as opposed to morally neutral matter" rather than "longtermism," and then we could just talk about the obvious fact that under this moral view it seems very important to not miss out on the opportunity to put the extremely large stock of resources available in our galaxy and beyond to use producing happy beings for trillions upon trillions of years to come, by e.g. allowing human extinction in the near term or otherwise not becoming grabby and enduring for a long time. But this would be the subject of another discussion.
Edited to add: Sorry this post is so long. Whenever I feel like I wasn't understood in writing I have a tendency to want to write a lot more to overexplain my thoughts. In other words I've written absurdly long comments like this before in similar circumstances. Hopefully it wasn't annoying to read it all. Obviously the time cost to me of writing it is much more than the time-cost to you or others for reading it, but I also I'm wary of putting out lengthy text for others to read where shorter text could have sufficed. I just know I have trouble keeping my comments concise under conditions like this and psychologically it was easier for me to just write everything out as I wrote it. (To share, I also think doing this generally isn't a very good use of my time and I'd like to get better at not doing this, or at least not as often.)