Hide table of contents

This post gives an overview of how I’m thinking about the “funding in EA” issue, building on many conversations. Although I’m involved with a number of organisations in EA, this post is written in my personal capacity. You might also want to see my EAG talk which has a related theme, though with different emphases. As a warning, I’m particularly stretched for time at the moment and might not have capacity to respond to comments. For helpful comments, I thank Abie Rohrig, Asya Bergal, Claire Zabel, Eirin Evjen, Julia Wise, Ketan Ramakrishnan, Leopold Aschenbrenner, Matt Wage, Max Daniel, Nick Beckstead, Stephen Clare, and Toby Ord.
 

Summary

  • EA is in a very different funding situation than it was when it was founded. This is both an enormous responsibility and an incredible opportunity.
  • It means the norms and culture that made sense at EA’s founding will have to adapt. It’s good that there’s now a serious conversation about this.
  • There are two ways we could fail to respond correctly:
    • By commission: We damage, unnecessarily, the aspects of EA culture that make it valuable; we support harmful projects; or we just spend most of our money in a way that’s below-the-bar.
    • By omission: we aren’t ambitious enough, and fail to make full use of the opportunities we now have available to us. Failure by omission is much less salient than failure by commission, but it’s no less real, and may be more likely.
  • Though it’s hard, we need to inhabit both modes of mind at once. The right attitude is one of judicious ambition.[1]
  • Judicious, because I think we can avoid most of the risks that come with an influx of potential funding without compromising on our ability to achieve big things. That means: avoiding unnecessary extravagance, and conveying the moral seriousness of distributing funding; emphasising that our total potential funding is still tiny compared to the problems in the world, and there is still a high bar for getting funded; being willing to shut down lower-performing projects; and cooperating within the community to mitigate risks of harm.
  • Ambition, because it would be very easy to fail by thinking too small, or just not taking enough action, such that we’re unable to convert the funding we’ve raised into good outcomes. That means we should, for example: create more projects that are scalable with respect to funding; buy time and increased productivity when we can; and be more willing to use money to gain information, by just trying something out, rather than assessing whether it’s good in the abstract.

 

Intro

Well, things have gotten weird, haven’t they?

Recently, I went on a walk with a writer, and it gave me a chance to reflect on the earlier days of EA. I showed him the first office that CEA rented, back in 2013. It looks like this:

To be clear: the office didn’t get converted into an estate agent — it was in the estate agent, in a poorly-lit room in the basement. Here’s a photo from that time:

 


 Normally, about a dozen people worked in that room. When one early donor visited, his first reaction was to ask: “Is this legal?” 

At the time, there was very little funding available in EA. Lunch was the same, every day: budget baguettes and plain hummus. The initial salaries offered by CEA were £15,000/yr pre-tax. When it started off, CEA was only able to pay its staff at all because I loaned them £7,000 — my entire life savings at the time. One of our first major donations was from Julia Wise, for $10,000, which was a significant fraction of the annual salary she received from being a mental health social worker at a prison. Every new GWWC pledge we got was a cause for celebration: Toby Ord estimated the expected present value of donations from a GWWC pledge at around $70,000, which was a truly huge sum at the time.[2]

Now the funding situation is… a little different. This post is about taking stock, and reflecting on how we should respond to that. It builds on thinking I’ve done over the last 14 months, many conversations I’ve had, and the many recent Forum posts and comments. I’m not trying to be prescriptive  — you should figure out your own takes — but hopefully I can be a little helpful. I’m aiming for this post to convey what I hope to be the right attitude to the situation as a whole, rather than merely discussing one aspect or other of the issue, as most of the recent posts have done.

In a nutshell: our current situation is both an enormous responsibility and an incredible opportunity. If we’re going to respond appropriately, we need to act with judicious ambition, holding both of these frames in mind. 
 

The current situation

Effective altruism has done very well at raising potential funding[3] for our top causes. This was true two years ago: GiveWell was moving hundreds of millions of dollars per year; Open Philanthropy had potential assets of $14 billion from Dustin Moskovitz and Cari Tuna. But the last two years have changed the situation considerably, even compared to that. The primary update comes from the success of FTX: Sam Bankman-Fried has an estimated net worth of $24 billion (though bear in mind the difficulty of valuing crypto assets, and their volatility), and intends to give essentially all of it away. The other EA-aligned FTX early employees add considerably to that total.[4]

There are other prospective major donors, too. Jaan Tallinn, the cofounder of Skype, is an active EA donor. At least one person earning to give (and not related to FTX) has a net worth of over a billion; a number of others are on track to give hundreds of millions in their lifetime. Among Giving Pledge signatories, there are around ten who are at least somewhat sympathetic to either effective altruism or longtermism. And there are a number of other successful entrepreneurs who take EA or longtermism seriously, and who could increase the total aligned funding by a lot. So, while FTX’s rapid growth is obviously unusual, it doesn’t seem like a several-orders-of-magnitude sort of fluke to me, and I think it would be a mistake to think of it as a ‘black swan’ sort of event, in terms of EA-aligned funding.  

So the update I’ve made isn’t just about the level of funding we have, but also the growth rate. Previously, it wasn’t obvious to me whether Dustin and Cari were flukes or not; if they were, all it would take is for their interests to move elsewhere, or for Facebook stock to tank, for the amount of EA-aligned potential funding to decline considerably. 

Now I think the amount of EA-aligned funding is, in expectation, considerably bigger in the future than it is today. Of course, over the next five years, total potential funding could still decrease by a lot, especially if FTX crashes. But it also could increase by many tens of billions more, if FTX does very well, or if new very large donors get on board. So we should at least be prepared for a world where there’s even more EA-aligned potential funding than there is today.

There’s a tricky question about how fast we should be spending this down. Compared to others in EA, I think I’m unusually sympathetic to patient philanthropy: I don’t think the chance of a hinge moment in the next decade is dramatically higher than the chance of a hinge moment in 2072-82 (say); and I think our understanding of how to do good is improving every year, which gives a reason for delay. 

But even I think that we should greatly increase our giving compared to now. One reason in favour of spending quickly is that, even if you endorse patient philanthropy, you should probably still distribute some significant proportion of your funding over time (perhaps in the low single-digit percentage points[5]) because philanthropic resources have diminishing returns. And if you think that we’re at a very influential time, perhaps because you think we’ll probably see transformative AI in our lifetimes, then it should be larger still. (I lay out some more reasons for and against faster spending in an appendix.) 

A second reason is that we can fund community-building, which is a form of investment, and which seems to have very high returns. Indeed, the success of FTX, and of EA in general, should give us a major update in this direction. So far, we've generated more than $30 bn for something like $200 mn, at a benefit:cost ratio of 150 to 1;[6] and even excluding the success of FTX, the benefit-cost looks very good (especially if we consider that funding raised is not all, or maybe even most, of the impact that outreach to date has generated). Further investment in outreach is likely to continue to raise much more money for the most pressing problems than it costs.[7]

A third reason is option value: if we build the infrastructure to productively and scalably absorb funding, then we can choose not to use it if it turns out to not be the right decision; whereas if we don’t build the infrastructure now, then it will take time to do so if in a few years’ time it does turn out to be needed. 

A final consideration that weighs on me for spending faster isn’t based on impact grounds. Rather, it simply feels wrong to have such financial assets when there’s such suffering in the world, and such grave risks that we face. Now, of course what we ought to do is whatever is impact-maximising over the long run; but at least in terms of my moral aesthetics, it really feels that the appropriate thing is for this money to get used, soon.

For the time being, let’s suppose that we aim just to spend the return on EA-aligned funding: about $2 billion per year (at 5% real rate of return). Spending even this amount of funding effectively will be a huge challenge. It’s a big challenge even within global health and wellbeing, where the biggest scale-up in giving is currently happening. GiveWell now aims to move $1 billion per year by 2025, including (as a tentative plan) an annual allocation from Open Phil of about $500 million. But even now, and even if they lower their funding bar from 8x the cost-effectiveness of GiveDirectly to 5x the cost-effectiveness of GiveDirectly, they still have more funding available than funding gaps to fill.

Spending this funding will be a truly enormous challenge within cause areas such as AI risk and governance, worst-case biorisk, and community-building, that have fewer or no existing organisations that could productively use such sums of money. The Future Fund expressed a bold aim of giving between $100mn and $1bn this year. Let’s say that this ends up at $300mn in grants (which might be about 30% of total EA money moved this year). That’s a rapid scale-up from a standing start, but it’s a shortfall of $1.2 billion compared to what it would need to spend just to distribute the rate of return on the financial assets of Sam Bankman-Fried and Gary Wang. 

If the total potential funding grows, or if the right thing to do is to be spending down our assets, then that number increases, perhaps considerably. And we should be particularly prepared for scenarios where our potential funding increases by a lot, because we have more impact in those scenarios than in scenarios where our potential funding decreases considerably.

Meeting this challenge means that EA’s culture and norms will need to adapt. In 2013, it made sense for us to work in a poorly-lit basement, eating baguettes and hummus. Now it doesn’t. Frugality is now comparatively less valuable, and saving time and boosting productivity in order to make more progress on the most pressing problems is comparatively more valuable. Creating projects that are maximally cost-effective is now comparatively less valuable; creating projects that are highly scalable with respect to funding, and can thereby create greater total impact even at lower cost-effectiveness, is comparatively more valuable. Extensive desk research to evaluate a small or medium-sized funding opportunity is comparatively less valuable; just spending money to actually try something and find out empirically if it works is comparatively more worthwhile. 

Now, I miss the basement days. It feels morally appropriate to be eating baguettes and hummus every day. But solving the big problems in the world isn’t about acting in a way that feels appropriate; it’s about doing the highest-impact thing. 

Nonetheless, it’s natural to worry: is this really EA adapting to a new situation, or is it value-drift? Maybe we’re fooling ourselves! I think it’s good we’re being vigilant about this. So let’s discuss how we should adapt; how to respond appropriately to the situation we’re in, while not losing the mission-driven focus that was present in the basement of an Oxford estate agent.

There are two ways in which we could fail in response to this challenge. We could cause harm by commission: doing dumb things that end up net-negative overall. Or we could cause harm by omission: failing to do things that would have enormous positive impact. Let’s take each of these risks in turn.  
 

Risks of commission: causing harm

As has been noted in a number of recent Forum posts, there are ways that scaling up our giving could cause harm, such as by damaging EA’s culture — destroying what makes EA distinctive and great — or by funding net-negative projects. There are a number of risks here; I’ll discuss a few, but this list is still incomplete. 
 

Appearances of extravagance

Other things being equal, EA wants to appeal to morally-motivated people. It’s more valuable to have someone who intrinsically wants to make the world better than someone who’s just doing it for a paycheck: they’ll be in it for longer, and are more likely to make better decisions even in cases where their income doesn’t depend on them making the right decision. But morally-motivated people, especially on college campuses, often find seemingly-extravagant spending distasteful.

What’s more, this is a perfectly rational response. Living on rice and beans is a costly signal: it’s easier for genuinely morally motivated people to do it than people who are faking. So if I meet someone living on rice and beans and giving a chunk of their income away, I take seriously their claims to be morally motivated. If I meet someone who’s flying business class — they might be doing that because they’re morally serious and trying to maximise their output, but it’s much harder for an outsider to tell. [8]

So we don’t want to turn off morally dedicated people. This is a major worry for me; I think that EA developing a bad reputation is one of the leading existential risks to the community, and a reputation for extravagance would not at all be helpful for that. 

But there’s a balancing act. Very few people will want to live on rice and beans forever. Initially, we disproportionately appealed to people who were willing to be very frugal, and turned off those who weren’t, so the current community is disproportionately constituted of such people. Now, we have the chance to appeal to people who are less willing to be ultra-frugal, but have many other great qualities and can contribute enormously to the world.[9] And this is a great thing.

My impression is that most of the issues that have gotten people worried have been unforced errors: people getting carried away, or not thinking about how what they did would be perceived, or talking about spending in a way that seems flippant. (And sometimes the alleged situations have been misrepresented, distorted through a game of Telephone.) This, naturally, feels alienating to many people, especially those who are new to EA. The opportunity cost of spending is very real and very major, in absolute terms; not recognising that can seem suspicious. 

Given that the most egregious errors seem unforced, I think there are some easy wins, such as: 

  • Treating expenditure with the moral seriousness it deserves. Even offhand or joking comments that take a flippant attitude to spending will often be seen as in bad taste, and apt to turn people off. Similarly, luxurious aesthetics are generally not a good look, especially on college campuses, and often don’t have commensurate benefits.
  • Heavily considering what you show as well as what you do, especially if you’re in a position of high visibility. “Signalling” is often very important! For example, the funding situation means I now take my personal giving more seriously, not less. I think the fact that Sam Bankman-Fried is a vegan and drives a Corolla is awesome, and totally the right call. And, even though it won’t be the right choice for most of us, we can still celebrate those people who do make very intense moral commitments, like the many kidney donors in EA, or a Christian EA I heard about recently who lives in a van on the campus of the tech company he works for, giving away everything above $3000 per year.
     

Harming quality of thought

Another worry is that funding will negatively impact EA’s ability to think; that, insidiously, people will be incentivised to believe whatever will help them get funding, or that particular worldviews will get artificially inflated in virtue of receiving more funding than they should receive.

My guess is that culture is an even bigger worry (where people, including funders, go too far in the direction of deferring to those regarded as particularly smart, or are too worried about deviating from what they regard as consensus views within the community). But either way, having incorrect beliefs or focusing on the wrong things is an easy way for the EA community to lose almost all its value. And we want to reduce that as much as possible. 

Again, I think there are actions we can take to mitigate this risk:

  • One partial solution is to diversify funding. For example, the goal of diversifying and decentralising decision-making was a major motivation behind the launch of Future Fund’s regrantors program. Now, if you have some heterodox idea, you can potentially receive funding from lots of different sources, from different people with different perspectives and networks, rather than just one. And I expect the pool of large EA-aligned donors or grantmakers to increase in the future.
  • Another is to champion independence of thought. It seems to me we do pretty well at this (depending on how you count, three to five of the 10 most-upvoted Forum posts are ‘critical’ or at least ‘critically self-reflective’ posts). But I’d love to see more high-quality research that critically engages with views that are widely held within the community. 
     

Resentment

There’s a tough messaging challenge around the funding situation. On the one hand, we want to convey that people should be developing big, ambitious plans, and convey how much we need to scale up the community’s giving. Given that, it’s natural to feel disappointed or even resentful if you create such plans but then don’t receive funding.

But there’s always an opportunity cost, and the bar for receiving funding is still extremely high. We could easily use up the entirety of our potential funding on cash transfers, or clean tech funding, or pandemic preparedness technology, or compute for the most safety-conscious AI labs. Given this opportunity cost, even despite the scale-up in funding, most well-meaning projects still won’t get funded. For example, Future Fund is trying to scale up its giving rapidly, but in the recent open call it rejected over 95% of applications. 

As a proportion of the world’s resources, EA-aligned financial resources are still tiny: the return on EA-aligned financial assets is less than a hundredth of Alphabet’s yearly revenue and about one four-hundredths of the US defence budget. And they’re also tiny compared to the problems we face: despite about $160bn in annual official development assistance, and $630bn in annual (public and private) climate spending, extreme poverty and climate change are still serious problems. 

To address this issue, how we talk about the situation is important: 

  • As well as emphasising the importance of forming big plans, and the potential for projects to scale, we need to emphasise that it’s not easy to get funding for projects; the bar for last-dollar funding is very high.
  • In my experience, EA funders are highly aware that they’re going to get a lot of funding decisions wrong. But more could be done to emphasise that any funding processes are going to be imperfect, and many great opportunities won’t get funded even though, in a perfect world, they would be. 
     

Losing evolutionary forces towards greater impact

One worry I’ve had is that availability of funding could mean we lose incentives towards excellence. If it’s too easy to get funding, then a mediocre project could just keep limping on, rather than improving itself; or a substandard project could continue, even though it would be better if it shut down and the people involved worked elsewhere.  

Within the nonprofit world, there’s a general problem where, unlike unprofitable companies, bad nonprofits don’t die. We should worry that the same problem will affect us.

That said, this is something that I think donors are generally keeping in mind; many seed grants won’t be renewed, and if a project doesn’t seem like a good use of the people running it, then it’s not likely to get funded. 

One way we as a community can mitigate this concern further is to celebrate failures. For example, No Lean Season was an impressive-looking global development non-profit; it was incubated at Evidence Action, and went through Y Combinator (in the same batch as CEA). But, after an RCT found that its impact was lower than they’d hoped, and after they had to terminate their relationship with a partner organisation, they shut down and published their reasons for doing so.

This is a socially weird thing to do, and very unusual within the nonprofit world. But it was awesome, and should be praised as such.
 

Risks of harm 

There’s one huge difference between aiming to do good and aiming to make profit. If you set up a company aiming to make money, generally the very worst that can happen is that you go bankrupt; there’s a legal system in place that prevents you from getting burdened by arbitrarily large debt. However, if you set up a project aiming to do good, the amount of harm that you can do is basically unbounded.

This is a common worry in EA, and it's extremely important as far as it goes. The standard solution is to communicate and cooperate with others with shared goals; if there’s a range of opinions on whether something is a good idea, then following the majority view is the right strategy. And, in practice, all the major funders closely communicate and coordinate, and behave cautiously; similarly, when someone is starting a new project, in my experience they tend to get extensive feedback from the community on the risks and benefits of that project. 

Indeed, my honest take is that EAs are generally on the too-cautious end. As well as the unilateralist’s curse (where the most optimistic decision-maker determines what happens), there’s a risk of falling into what we could call the bureaucrat’s curse,[10] where everyone has a veto over the actions of others; in such a situation, if everyone follows their own best-guesses, then the most pessimistic decision-maker determines what happens. I’ve certainly seen something closer to the bureaucrat’s curse in play: if you’re getting feedback and your plans, and one person voices strong objections, it feels irresponsible to go ahead anyway, even in cases where you should. At its worst, I’ve seen the idea of unilateralism taken as a reason against competition within the EA ecosystem, as if all EA organisations should be monopolies. 

The suggested examples of harmful projects I’ve heard tend, in my view, not to come from people who take the unilateralist’s curse seriously and are in close communication with the community, but go ahead and do it anyway. Instead, all along they were power-grabs, or they were by people who just didn’t care what others thought of their plans. In contrast, I’ve found that, for those who are highly concerned about unilateralism, if they end up doing something that might be harmful, they quickly receive feedback and course-correct. 

Overall, risks of harm are something I think we’re actually managing pretty well as a community, and can keep managing well if we:

  • Stay in constant communication about our plans with others, inside and outside of the EA community, who have similar aims to do the most good they can
  • Remember that, in the standard solution to the unilateralist’s dilemma, it’s the median view that’s the right (rather than the most optimistic or most pessimistic view)
  • Are highly willing to course-correct in response to feedback
     

Risks of omission: squandering the opportunity 

There are a number of ways in which the influx of funding could cause real harm. Despite this, I don’t think it’s the most likely way we’ll fail. 

It seems to me to be more likely that we’ll fail by not being ambitious enough; by failing to take advantage of the situation we’re in, and simply not being able to use the resources we have for good ends. 

It’s hard to internalise, intuitively, the loss from failing to do good things; the loss of value if, say, EA continued at its current giving levels, even though it ought to have scaled up more. For global health and development, the loss is clear and visceral: every year, people suffer and lives are lost. It’s harder to imagine for those concerned by existential risks. But one way to make the situation more vivid is to imagine you were in an “end of the world” movie with a clear and visible threat, like the incoming asteroid in Don’t Look Up. How would you act? For sure, you’d worry about doing the wrong thing. But the risk of failure by being unresponsive and simply not doing enough would probably weigh on you even harder.

There are a couple of reasons why I’m particularly worried about risks of omission.  First, it’s just very hard to seriously scale up giving while spending the money effectively. It’ll involve enormous amounts of work, from hundreds or thousands of people. Often, it’ll involve people doing things that just aren’t that enjoyable: management and scaling organisations to large sizes are rarely people’s favourite activities; and, it will be challenging to incentivise enough people to do these things effectively.

To see how hard this is, we can look at existing foundations. The foundation that has most successfully scaled up its giving is the Gates Foundation: it gives out about $6 bn per year, which is extremely impressive — far more than any other foundation.[11] But it seems to me they are falling far short of their goals. Bill Gates and Melinda French Gates have said they want the foundation to spend all its assets within 20 years of their deaths, and comment that: “The decision to use all of the foundation’s resources in this century underscores our optimism for progress and determination to do as much as possible, as soon as possible, to address the comparatively narrow set of issues we’ve chosen to focus on.” 

Even though the Gates Foundation spends far more than any other foundation, since 2000 their total assets have increased from $100 billion[12] to $316 billion[13] — over a factor of 3. They’re distributing close to $6 bn per year, but that’s less than half the return they get on their total financial assets (at 5% real return per year). Given the ages of Bill Gates and Melinda French Gates, they should expect to live approximately another 30 years. But in order to spend down their assets within 50 years, if the foundation distributed a fixed amount every year, it would need to give out over $17 bn per year. 

I don’t want to make any claims about the tricky question of the optimal rate of giving over time. But we should at least feel the potential loss, here, if scaling up too slowly means that less good is done. 

A second reason why I’m worried about scaling too slowly, or too low a plateau, is that there are asymmetric costs to trying to do big things versus being cautious. Compare: (i) How many times can you think of an organisation being criticised for not being effective enough? and (ii) How many times can you think of someone being criticised for not-founding an organisation that should have existed? (Or, suppose I hadn’t given a talk on earning to give at MIT in 2012,[14] would anyone be berating me?) In general, you get public criticism for doing things and making mistakes, not for failing to do anything at all.

The asymmetric costs are especially worrying when salaries represent only a tiny fraction of the value you create, which is especially true for non-profit projects. VCs struggle to get entrepreneurs to be ambitious and risk-taking enough: the solution that has emerged is to pay successful entrepreneurs huge amounts of money. A successful EA megaproject might generate far more value for the world than Uber (for example), but, even if EA salaries were to increase enormously, the founders and early employees will still get paid much less than the founders and early employees of Uber.[15]

The importance of finding ways to scale our giving also changes how we should think about grantmaking. Early EA culture was built on a highly skeptical mindset. This is still important in many ways (this post by Holden on ‘minimal trust investigations’ is one of my favourite blog posts of the last year). But it can cause us to go awry if it means we don’t take chances of upside seriously, or when we focus our concern on false positives rather than false negatives. 

I worry we’ve made some errors in the past by not taking the chance of best-case scenarios seriously, out of a desire to be rigorous and skeptical. For example, I mentioned that Toby initially estimated the value of a Giving What We Can Pledge at $70,000 (as one example of quantifying the benefits of outreach and community-building more generally). I remember having arguments with people who claimed that estimate was too optimistic. But take the 7000 Giving What We Can members, and assume that none of them give anything apart from Sam Bankman-Fried, who gives his net worth. Then a Pledge was actually worth $2 million — 30 times higher than Toby’s “optimistic” estimate at the time.[16] In general, if our successes are sampling from a heavy-tailed distribution, the historical average value of our impact will very likely be lower than the true mean. 

And when we look to future community-building efforts, the asymmetry of upside and downside suggests to me that, if we put the risk of harm to one side, we should be much more concerned about missing opportunities for impact than about spending money in ways that don’t have impact.

It’s easiest to quantify when looking at earning to give (but is in no way limited to that). We’ve seen, now, that EA outreach can inspire people to earn to give in ways that put them on track to donate hundreds of millions of dollars or more. That means the worry about missing out on opportunities to change people’s careers should, for the time being, loom larger than the worry about overspending. 

(Quantitatively: suppose $200 is spent on an intro to EA retreat for someone. If that has a more than a one in a five hundred thousand chance of inspiring the attendee to earn to give and successfully donate $100 million over their lifetime, then the expected financial benefit is positive. Given the successes we’ve seen, both from FTX and outside of that, the real probability is orders of magnitude larger. That’s not to say $200 on a retreat is how much should be spent — if you can have the same impact at cheaper cost, you should. And excessive spending can even become counterproductive if it sends the wrong message. But it indicates just how small community-building spending is in comparison the potential benefits from changing people’s careers for the better.)  

The need to scale changes the optimal approach to grantmaking in another way, too: it also means that making many small grants (small relative to the tens of billions of dollars per year we might need to spend) in order to find out, empirically, what things seem cost-effective, becomes well worth it.

Here’s a toy example. Suppose you give out 100 grants of $100,000 each. They all do nothing apart from one, which demonstrates a scalable way of absorbing $100 million at 120% of the cost-effectiveness of last-dollar spending. You’ve spent $10 million in order to gain impact equivalent to $20 million at last-dollar spending.[17] It’s a good use of money, even though 99 of the grants achieved nothing. 

I think this toy example often reflects reality. It’s much easier, and more reliable, to assess a project once it's already been tried. If you need to scale giving dramatically, then often it makes sense to fund something and find out empirically how good it is, so that in two years’ time you can decide whether to stop funding altogether, or scale donations considerably. If the cost to fund and get the information is a small proportion of the giving you hope to scale up to, then it can be well worth just making the grant and figuring out how cost-effective it is later on if it seems potentially promising as something to scale. (A similar thought lies in part behind Future Fund’s 2022 goal of doing “bold and decisive tests of highly scalable funding models.”)
 

An incredible opportunity

It’s easy to feel stressed about the current situation. But paralysing anxiety or insomnia-inducing stress probably aren’t the attitudes that will help you have the most long-term impact.

So let’s reframe things, for a moment at least.

A classic reason why people feel unmotivated to do good things is that their contribution will be just “a drop in the bucket.” A helpful psychological response is to think of the impact that a community you’re part of, working together on that problem, will have. 

When we think of the impact EA has had so far, it’s pretty inspiring. Let’s just take one organisation: Against Malaria Foundation. Since its founding, it has raised $460 million, in large part because of GiveWell’s recommendation. Because of that funding, 400 million people have been protected against malaria for two years each; that’s a third of the population of sub-Saharan Africa. It’s saved on the order of 100,000 lives — the population of a small city.

We did that. 

And the current funding situation means that is just the beginning

The amount of potential funding is still very small from the perspective of the world as a whole. But we’re now at a stage where we can plausibly have a very significant impact on some of the world’s biggest problems. Could we reduce existential risk from AI or pandemic by more than 10%, eradicate a global disease, or bring forward the end of factory farming by a year? Probably.

We should be judiciously ambitious. To achieve the sort of impact we’re now capable of, it means being sensitive to the risks of ambition, and the negatives of spending funds, for sure. But it also means we need to use the opportunities we have available to us. We should think big and be willing to take bold actions, while mitigating the risks. If we can manage to do both of these at once, we as a community can achieve some amazing things.
 

Appendix: how fast should we be scaling funding?

It’s non-obvious to me what the ideal rate of distributing funding should be, although my fairly strong best guess is that we should be scaling up to giving much more than we are at the moment.

Briefly, the main reasons I see in favour of giving more are:

  • To date, community-building has been an outstanding investment.
  • Even on fairly patient philanthropy models, we should be spending more than we currently are.
  • We have more impact in worlds where there is even more potential funding in a few years’ time, so we should be particularly prepared for those scenarios.
  • Option value: if we build the infrastructure to productively absorb funding, then we can choose not to use it if it turns out to not be necessary; whereas if we don’t build the infrastructure now, then it will take time to do so if in a few years’ time it does turn out to be necessary.
  • At least within longtermist funding, the lack of scalable funding opportunities is a turn-off for some potential donors; creating new scalable funding opportunities can lead to more funding overall in the long run.
  • The chance of near-term hinge moments, such as transformative AI within the next 20 years.  
     

In my mind, the strongest cases against dramatically scaling up our giving are:

  • We might do it badly, with negative cultural consequences for our movement that are hard to reverse.
  • There might be some future time where it would be extremely valuable to spend truly enormous amounts of funding. This could be on compute in the run-up to AGI, or on rapid responses to a worst-case pandemic.
  • Perhaps the returns to giving diminish extremely rapidly, and scaling up our giving is a distraction from other things we could be doing. 
     
  1. ^

    I considered suggesting the slogan “Move fast and don’t break things” to encapsulate this, but I thought that “move fast” isn’t really the right framing for ambition: setting up a massively scalable project might mean being small and testing things out for years in order to put yourself in a position to grow enormously.

  2. ^

    I don’t know exactly what the situation was like for GiveWell in New York or for the LessWrong / SingInst crowd in the Bay, but it wasn’t radically different: salaries were low, and funding was scarce.

  3. ^

    By “potential funding” I mean financial assets from people who plan to give those financial assets away to EA-aligned causes.

  4. ^

    For example, Gary Wang has a publicly estimated net worth of $5.9 billion, and plans to use the majority of that for EA-aligned goals.

  5. ^

    For example, in a very simple model by Phil Trammell, if you find yourself at a moderately influential time, you should spend 2.4% of your resources per year. On this simple model, if you think that “influentialness” (or “hingeyness”) will be roughly flat over the next fifty years, but will permanently fall to 1/10 of its current level soon after that, then you should give out  0.83% each year.  (These numbers shouldn’t be relied on though; they are intended only to be illustrative.)

  6. ^

     If you were sceptical of replicating FTX-esque success, then that number might drop - but I think even so it should be much higher than 10:1.

  7. ^

    Note that the ‘community-building’ argument also justifies significant funding of direct work, simply because a movement that never does anything of actual value is not very compelling.

  8. ^

    What’s weird but important to bear in mind is that the perception of extravagance often has little to do with the amount of money actually being spent. One organisation I know hosted a conference at a very fancy venue: in reality, the venue is owned by a charitable foundation, so was a comparatively cheap option. But it looked extremely grand, and there were a number of complaints.

  9. ^

    In some cases, an extravagant lifestyle can even produce a lot of good, depending on the circumstances. I know of some people who have attended luxurious parties, met major philanthropists there, and gotten them involved in EA. It’s not my preferred moral aesthetic, but the world’s problems don’t care about my aesthetics. (Needless to say, if you find yourself in this unusual position, you should probably take special care to make sure that attending luxurious parties really is the way you can do the most good.)

  10. ^

    H/T Nick Beckstead

  11. ^

    I believe the foundation that distributes the second-largest amount per year is Wellcome, which in 2020/1 gave out £1.2 billion.

  12. ^

    Inflation-adjusted to today’s money

  13. ^

    Including the pledge from Warren Buffet to give almost all his wealth, which is currently $120 billion.

  14. ^

    https://80000hours.org/stories/sam-bankman-fried/

  15. ^

    Of course, relative to Uber there is the additional “incentive” of the impact that the project will create, which mitigates this issue.

  16. ^

    Of course, there’s plenty to argue with in the estimate, but I don’t think it changes the core point.

  17. ^

    In this toy example, I’m ignoring inflation and investment returns, and assuming that you can’t in advance identify which project is a “hit”. See also https://www.openphilanthropy.org/blog/hits-based-giving

Comments190
Sorted by Click to highlight new comments since: Today at 11:00 AM
Some comments are truncated due to high volume. (⌘F to expand all)Change truncation settings

I feel like this post mostly doesn't talk about what feels like to me the most substantial downside of trying to scale up spending in EA, and increased availability of funding. 

I think the biggest risk of the increased availability of funding, and general increase in scale, is that it will create a culture where people will be incentivized to act more deceptively towards others and that it will attract many people who will be much more open to deceptive action in order to take resources we currently have. 

Here are some paragraphs from an internal memo I wrote a while ago that tried to capture this: 

========

I think it it was Marc Andreessen who first hypothesized that startups usually go through two very different phases:

  1. Pre Product-market fit: At this stage, you have some inkling of an idea, or some broad domain that seems promising, but you don't yet really have anything that solves a really crucial problem. This period is characterized by small teams working on their inside-view, and a shared, tentative, malleable vision that is often hard to explain to outsiders.
  2. Post Product-market fit: At some point you find a product that works for people. The transition here ca
... (read more)

Reading this, I guess I'll just post the second half of this memo that I wrote here as well, since it has some additional points that seem valuable to the discussion: 

When I play forward the future, I can imagine a few different outcomes, assuming that my basic hunches about the dynamics here are correct at all:

  1. I think it would not surprise me that much if many of us do fall prey to the temptation to use the wealth and resources around us for personal gain, or as a tool towards building our own empire, or come to equate "big" with "good". I think the world's smartest people will generally pick up on us not really aiming for the common good, but I do think we have a lot of trust to spend down, and could potentially keep this up for a few years. I expect eventually this will cause the decline of our reputation and ability to really attract resources and talent, and hopefully something new and good will form in our ashes before the story of humanity ends.
  2. But I think in many, possibly most, of the worlds where we start spending resources aggressively, whether for personal gain, or because we do really have a bold vision for how to change the future, the relationships of the centra
... (read more)
6
Charles He
2y
To onlookers: There’s a often a low amount of resolution and expertise in some comments and concerns on the LW and EAF, and this creates “bycatch” and reduces clarity. With uncertainty, I'll lay out one story that seems like it matches the concerns in the parent comment.   Strong Spending I'm not entirely sure this is correct, but for large EA spending, I usually think of the following: * 30%-70% growth in head count in established institutions, sustained for multiple years * Near six figure salaries for junior talent, and well over six figure salaries for very good talent and management who can scale and build an organization (people who can earn multiple times that in the private sector and cause an organization to exist and have impact) * Seven figure salaries for extreme talent (world's best applied math, CS, top lawyers) * Discretionary spending * Buying operations, consulting and other services So all the above is manageable,  even sort of fundamental for a good leader or ED or CEO. This is why quality CEO or leadership is so important, to hire and integrate this talent well and manage this spending. This is OK. This is considered “high”, but it's not really by real world standards. 
3
Charles He
2y
Next-level Next-level Now distinct from the above comment, there’s a whole other reference class of spending where: 1. People can get an amount of cash that is a large fraction of all spending in an existing EA cause area in one raise. 2. The internal environment is largely "deep tech" or not related to customers or operations So I'm thinking about valuations in the 2010- tech sector for trendy companies.  I'm not sure, but my model of organizations that can raise 8 figures per person in a series B, for spending that is pretty much purely CapEx (as opposed to capital to support operations or lower margin activity, e.g. inventory, logistics) has internal activity that is really, really different than the above "high" spending in the above comment.    There's issues here, that are hard to appreciate.  So Facebook's raises were really hot and oversubscribed. But building the company was a drama fest for the founders, and also there was a nuclear reactor hot business with viral growth. So that's epic fires to put out every week, customers and partners, actual scaling issues of hockey stick growth (not this meta-business advice discussion on the forum). It's a mess. So CEO and even junior people have to deal.  But once you're just raising that amount in deep tech mode, my guesses for how people think,  feel, and behave inside of that company with valuations in the 8-9 figures per person. My guess is that the attractiveness, incentives and beliefs in that environment, are really different than even the hottest startups, even above those where junior people exit with 7 figures of income.   To be concrete, the issues on the rest of EA might be that: * Even strong EA CEOs won’t be able to hire many EA talent like software developers (but they should be worried about hiring pretty much anyone really). If they hire, they won't be able to keep them at comfortable, above EA salaries, without worrying about attrition. * Every person who can convincingly claim inter
3
Charles He
2y
I'm not sure, but in situations where this sort of dynamic or resource gradient happens, this isn't resolved by the high gradient stopping (people don't stop funding or founding institutions), because the original money is driven by underlying forces that is really strong. My guess is that a lot of this would be counter productive. Typically in those situations, I think the best path is moderation and focusing on development and culture in other cause areas.
3
M. Y. Zuo
2y
These are some very important points, thanks for taking the time to write them out.  I just made an account here, though I've only ever commented on LW before, just to stress how important and vital it is to soberly assess the change in incentives. Because even the best have strengths and weaknesses that need to be adapted to. "Show me the incentives and I will show you the outcome" - Charlie Munger

I thought this comment was valuable and it's also a concern I have.

It makes me wonder if some of the "original EA norms", like donating a substantial proportion of income or becoming vegan, might still be quite important to build trust, even as they seem less important in the grand scheme of things (mostly, the increase in the proportion of people believing in longtermism).  This post makes a case for signalling.

It also seems to increase the importance of vetting people in somewhat creative ways. For instance, did they demonstrate altruistic things before they knew there was lots of money in EA? I know EAs who spent a lot of their childhoods volunteering, told their families to stop giving them birthday presents and instead donate to charities, became vegan at a young age at their own initiative, were interested in utilitarianism very young, adopted certain prosocial beliefs their communities didn't have, etc. When somebody did such things long before it was "cool" or they knew there was anything in it for them, this demonstrates something, even if they didn't become involved with EA until it might help their self-interest. At least until we have Silicon Valley parents making ... (read more)

7
Yitz
2y
How bad is it to fund someone untrustworthy? Obviously if they take the money and run, that would be a total loss, but I doubt that’s a particularly common occurrence (you can only do it once, and would completely shatter social reputation, so even unethical people don’t tend to do that). A more common failure mode would seem to be apathy, where once funded not much gets done, because the person doesn’t really care about the problem. However, if something gets done instead of nothing at all, then that would probably be (a fairly weak) net positive. The reason why that’s normally negative is due to that money then not being used in a more cost-effective manner, but if our primary problem is spending enough money in the first place, that may not be much of an issue at all.

I think it's easier than it might seem to do something net negative even ignoring opportunity cost. For example, actively compete with some other better project, interfere with politics or policy incorrectly, create a negative culture shift in the overall ecosystem, etc.

Besides, I don't think the attitude that our primary problem is spending down the money is prudent. This is putting the cart before the horse, and as Habryka said might lead to people asking "how can I spend money quick?" rather than "how can I ambitiously do good?" EA certainly has a lot of money, but I think people underestimate how fast $50 billion can disappear if it's mismanaged (see, for an extreme example, Enron).

1
Yitz
2y
That’s a fair point, thank you for bringing that up :)
2
Geoffrey Miller
2y
Thomas -- excellent reply, and good points. I've written a bit about virtue signaling, and agree that there are good forms (reliable, predictive) and bad forms (cheap talk, deceptive, misguided) of virtue signaling.  I also agree that EA could be more creative and broad-minded about what kinds of virtue signaling are likely to be helpful in predictive future integrity, dedication, and constructiveness in EA.  Historically, a lot of EA signaling has involved living frugally, being vegan, being a good house-mate in an EA shared house, collaborating well on EA projects, getting lots of upvotes on EA Forum, etc. Assessing those signals accurately requires a lot of first-hand or second-hand knowledge, which can be hard to do at scale, as the EA movement grows.  As EA grows in scale and becomes more diverse in terms of background (e.g. recruits more established professionals from other fields, not just recent college grads), we may need to get savvier about domain-specific virtue signals, e.g. how do medical researchers vs geopolitical security experts vs defense attorneys vs bioethicists vs blockchain developers show their true colors? The very tricky trade-off, IMHO, is that often the most reliable virtue signals in terms of predicting personality traits (honesty, humility, conscientiousness, kindness) are often the least efficient in terms of actually accomplishing real-world good. For example, defense attorneys who do a lot of pro bono work doing appeals for death row inmates might be showing genuine dedication and altruism -- but this might be among the least effective uses of their time in achieving criminal justice reform.   So, do we want the super-trustworthy but scope-insensitive lawyers involved in EA, or the slightly less virtue-signaling but more rational and scope-sensitive lawyers?  That seems like a real dilemma. Traditionally, EA has solved it mostly by expecting a fair amount of private personality-signaling (e.g. being a conscientious vegan house-ma

I share your worries about the effects on culture. At the same time I don't see this vision as bad:

For many months, they will sit down many days a week and ask themselves the question "how can I write this grant proposal in a way that person X will approve of" or "how can I impress these people at organization Y so that I can get a job there?", and they will write long Google Docs to their colleagues about their models and theories of you, and spend dozens of hours thinking specifically about how to get you to do what they want, while drawing up flowcharts that will include your name, your preferences, and your interests.

Imagine a global health charity that wants to get on the GiveWell Top Charities list. Wouldn't we want it to spend much time thinking about how to get there, ultimately changing the way it works in order to come up with the evidence needed to get included? For example, Helen Keller International was founded more than 100 years ago and its vitamin A supplementation program is recommended by GiveWell. I would love to see more external organisations change in order to get EA grants instead of us trying to reinvent the wheel where others might already be good.

Organisat... (read more)

1
Charles He
2y
I was going to write an elaborate rebuttal of the parent comment.  In that rebuttal, I was going to say there's a striking lack of confidence. The concerns seems like a pretty broad argument against building any business or non-profit organization with a virtuous culture. There's many counterexamples against this argument—and most have the additional burden of balancing that growth while tackling existential issues like funding. It's also curious that corruption and unwieldly growth has to set in exactly now, versus say with the $8B in 2019.   Now I sort of see how, combined with several other factors, how maintaining culture and dealing with adverse selection ("lemons") might be an issue.
9
GMcGowan
2y
  IMO, a reasonable analogy here is to the relationship between startups and VCs. What do VCs do to weed out the lemons here? Market forces help in the long run (which we won't have to the same degree) but surely they must be able to do this to some degree initially.
5
RedStateBlueState
2y
If you want to get a lot of money for your project, EA grants are not the way to do it. Because of the strong philosophical principles of the EA community, we are more skeptical and rigorous than just about any funding source out there. Granted, I don't actually know much about the nonprofit grant space as a whole: if it comes to the point that EA grants are basically the only game in town for nonprofit funding, then maybe it could become an issue. But if that becomes the case I think we are in a very good position and I believe we could come up with some solutions.

Almost all nonprofit grants usually require everyone to take very low salaries. There are very few well-paying nonprofit projects. My guess is EA is the most widely-known community that might pay high salaries for relatively illegible nonprofit projects (and maybe the only widely-known funder/community that pays high-salaries for nonprofit projects in-general).

2
MaxRa
2y
I think I'm less worried about the risk of increased deception. The analogy breaks down somewhat because these number of 10-hour conversations are also scaling with size of the movement, right? And I think it's relatively discernible whether somebody actually cares about doing good when you talk to them a lot. I don't think you need to be a particularly senior EA for noticing altruistic and impact-driven intentions.  Additionally I'm also less worried because I think most people actually also care about doing good and doing things efficiently. EA will still select for people who are less motivated to work in industry, where I expect wages to still be higher for somebody capable enough to scheme a great grant proposal.
1
Ulrik Horn
2y
Very good point on culture. Culture eats strategy for breakfast as they say. EA is definitely strategy heavy and I think your comment brings up a very important issue to investigate.
1[anonymous]2y
I would flip this and say, it's inevitable that this will happen, so what do we do about it? There are areas we can learn from: * Academia, as you mention - what do we want to avoid here? Which bits actually work well? * Organisations that have grown very rapidly and/or grown in a way that changes their nature. On a for-profit basis - Facebook as a cautionary tale of what happens when personal control and association isn't matched with institutional development? On a not-for-profit basis - I work for Greenpeace and we're certainly very different to what we were decades ago, with a mix of 'true believers' and people semi-aligned, generally in more support roles. Some would say we've sold out and indeed some people have abandoned us for other groups that are more similar to our early days, but we certainly have a lot more political influence that we did when we were primarily a direct action / protest group. * Corruption studies at a national level. What can we learn of the institutions of very low corruption countries e.g. in Scandinavia that we might adapt? 
6
Nathan Young
2y
I think the question is predictivity. How can you run the most predictive systems possible for selecting good grants/employing suitable people?  I guess over time, networks will be worse predictors and the average trustworthiness of applicants will fall slightly, to which we should respond accordingly. Though I guess that we have to ackowledge that some grants will be misspent and that the optimal  amount of bad grants may not be 0.
[anonymous]2y20
0
0

Definitely agree that networks will become worse predictors and ultimately grants, job offers etc. will become more impersonal. This isn't entirely a bad thing. For example personal and network-oriented approaches have significant issues around inclusivity that well-designed systems can avoid, especially if the original network is pretty concentrated and similar (see: the pic in the original post...)

As this happens this may also mean that over time people who have been in EA for a while may feel that 'over time the average person in the movement feels less similar to them'. This is a good thing!... if recognised, and well-managed, and people are willing to make the cognitive effort to make it work. 

Thanks so much for writing this Will! I can't emphasise enough how much I appreciate it.

 if a project doesn’t seem like a good use of the people running it, then it’s not likely to get funded. 

Two norms that I'd really like to see (that I haven't seen enough of) are:
1. Funders being much more explicit  to applicants about why things aren't funded (or why they get less funding than asked for). Even a simple tagging system like "out of our funding scope" or "seemed too expensive", "not targeted enough", or "promising (review and resubmit)" (with a short line about why) is explicit yet simple.

2. More funder diversity while maintaining close communications (e.g. multiple funders with different focus areas/approaches/epistemics, but single application form to apply to multiple funders and those funders sharing private information such as fraud allegation etc).

I know feedback is extremely difficult to do well (and there are risks in giving feedback), but I think that lack of feedback creates a lot of problems, e.g.:

  • resentment and uneasiness towards funders within the community;
  • the unilateralists curse is exacerbated (in cases where something is not funded because it's seen
... (read more)

Heartily agree with this.

For the pilot of SoGive Grants, we plan to 

(1) Provide feedback as much as we can (the only reason we haven't promised to give feedback to everyone is that this is a pilot and we don't know whether that's feasible for us)

(2) The application form is almost a copy and paste of the EA Funds application form, to make life easier for those who are applying to both

(BTW applications are still open and close on 22nd May)

I also want to echo pretty much every bullet point that Luke made about the value of feedback, which I think are excellent points.

8
Lorenzo Buonanno
2y
I'm really curious as to why this is being downvoted (was at -2 when I originally wrote this comment, now it's at 0 with 7 votes), I find SoGive Grants interesting and relevant to the discussion. Especially since More funder diversity is a main point of Luke's comment.

I don't know but FWIW my guess is some people might have perceived it as self-promotion of a kind they don't like.

(I upvoted Sanjay's comment because I think it's relevant to know about his agreement and about the plans for SoGive Grants given the context.)

9
Arepo
2y
I regret that I have but one strong upvote to give this. Lack of feedback on why some of the projects I've been involved in didn't get funding has been incredibly frustrating. One further benefit of getting it would have been that it can help across the ecosystem when you get turned down by Funder A and apply to Funder B - if you can pass on the feedback you got from Funder A (and how you've responded to it), that can save a lot of Funder B's time. As a meta-point, the lack of feedback on why there's a lack of feedback also seems very counterproductive. 
5
Linch
2y
Obviously I do not speak for Will or anybody else, but I wrote "Some unfun lessons I learned as a junior grantmaker" partially as a response to what I perceived as misconceptions by some of the comments in this thread and elsewhere on the Forum.
-6
Jonas V
2y

Thanks for writing this, Will. I appreciate the honesty and ambition. Thank you for all you do and I hope you have people around you who love and support you.

I like the framing of judicious ambition. My key question around this and the related longtermism discussion is something like, What is the EA community for?

  • A democractic funding body?
  • A talent pool?
  • Community support?
  • Error checkers?

Are we the democratic body that makes funding decisions? No and I don't want us to be. Doing the most good likely involves decisions that the median EA will disagree with. I would like to trial forecasting funding outcomes and voting systems, but I don't assume that EA should be democratic. The question is what actually does the most good.

Are we a body of talented professionals who work on lower wages than they otherwise would? Yes, but I think we are more than that. Fundamentally it's our work that is undervalued, rather than us. Animals, the global poor and future generations cannot pay to save their own lives, so we won't be properly remunerated, except by the joy we take from doing it.

Are we community support for one another? Yes, and I think in regard to this dramatic shift in EA's fortunes that... (read more)

I think this could be a standalone post

7
Vaidehi Agarwalla
2y
(fwiw strongly agree! Even in it's current form I think it would start a really interesting / valuable discussion)
5
Nathan Young
2y
Yeah, I thought I'd test the waters then write it up if people thought it was valuable. Do you have any other frames for what the community might be?

I think talent pool doesn't /quite/ capture things like entrepreneurs and field builders and people who are building things.

Error checkers is good, but also something like finding cause x isn't quite error checking. It's more like avoiding the risk of omission kind of stuff / rethinking fundamental assumptions.

I think my problem with this question, which I've been thinking about in maybe different ways for many years, is that there isn't a binary between community member and [core ea decision maker / etc.] - it's very much a continuum and people occupy multiple roles at once. This becomes complicated because you can't quite set boundaries the same way.

For example, if a community member starts their own grantmaking foundation, are they still a community members or a decision maker? Does it only count if they know the right people or are in a certain cause area? How do their responsibilities change (or not change)?

4
Kirsten
2y
Researchers? People who develop and test different approaches to doing good at a rapid pace
1
Jeremy
2y
Strongly agree as well!
1
PhilippaE
2y
I see a big role of the community as being the facilitator of coordination. Tight coordination might enable more and better projects to be done, with less redundancy. It might also reduce beneficial competition & diversity of projects and ideas if people don't pursue good projects because someone else is already in that space. Career advising seems like a good example of where different approaches can usefully coexist.

Thanks for writing this, really great post. 

I don't think this is super important, but when it comes to things like FTX I think it's also worth keeping in mind that besides the crypto volatility and stuff there's also the fact that a lot of what we're marking EA funding to aren't publicly-traded assets, and so numbers should probably be taken with an even bigger pinch of salt than usual. 

For example, the numbers for  FTX here are presumably backed out of the implied valuation from its last equity raise, but AFAIK this was at the end of January this year. Since then Coinbase (probably the best publicly traded comparator) stock has fallen ~62% in value, whereas FTX's nominal valuation hasn't changed in the interim since there hasn't been a capital raise. But presumably, were FTX to raise money today the implied valuation would reflect a somewhat similar move

Not a huge point, and in any case these kinds of numbers are always very rough proxies anyway since things aren't liquid, but I think maybe worth keeping in mind when doing BOTECs for EA funding

This comment has aged very  well! 

Thanks for this write up, Will! I hope it changes the minds of people who are skeptical/unhappy about our massive funding influx. 

I think a lot of EAs are not  motivated to seek personal financial rewards, instead they find themselves seeking truth in graduate school/academia or trying to improve the world via non-profits. They see their similarly intelligent, well educated peers go into industry, optimizing for "make as much money as possible" and they just fundamentally do not relate to that value function. I wonder if this kind of personality type (if you can call it that) lies at the root of a lot of people's discomfort with EA non-profit jobs suddenly paying really well. 

Maybe we could offer special community building grants with the option that you will work in a basement, subsisting only on baguettes and hummus? ;)

[anonymous]2y20
0
0

I agree (and have formerly resembled this type...)  This is quite embedded in a lot of nonprofit culture. Part of it is what motivates the individual and their personality, part of it is the concept of supporters' money. 'Would the person who gave you £5 a month want you to be spending your money on that?' In practice this leads to counterproductive underspending. I remember waiting weeks to get maybe £100 worth of extra memory so I could crunch numbers at a reasonable speed without crashing the computer. The concept of taxpayers' money works similarly. 

There's probably a good forum post in there somewhere about how the psychology of charity affects perceptions of EA...

Creating projects that are maximally cost-effective is now comparatively less valuable; creating projects that are highly scalable with respect to funding, and can thereby create greater total impact even at lower cost-effectiveness, is comparatively more valuable. 

I think this framing is wrong, or at best unhelpful because we shouldn’t avoid prioritizing cost-effectiveness. When you stop prioritizing cost-effectiveness, it stops being effective altruism. Resources are still finite. The effectiveness of solutions to dire problems still differs dramatically. And we have only scratched the surface of understanding which solutions are gold and which are duds. I think it’s cost-effectiveness all the way down. 

I hope Will doesn’t mean “creating maximally cost-effective projects is now less valuable” when he says “creating maximally cost-effective projects is now less valuable”. I hope Will means “We should use average cost-effectiveness instead of marginal cost-effectiveness because cost-effectiveness often decreases with more funding. This means that some projects which were more cost-effective at small levels of funding will become less cost-effective at larger levels of fun... (read more)

[anonymous]2y65
0
0

Really interesting, and something I'll need to come back to. Just to pick out one bit:

Often, it’ll involve people doing things that just aren’t that enjoyable: management and scaling organisations to large sizes are rarely people’s favourite activities; and, it will be challenging to incentivise enough people to do these things effectively.

I've seen variations on this theme in a few posts, and it doesn't resonate with my own experience. In a genuinely influential management/ops role, there's a great deal of satisfaction to be had in seeing your organisation become more effective - if what that organisation is doing is highly worthwhile. I worry a bit that the tone of 'yeah this isn't glamorous, but someone has to do it' is putting off talent in the area. If an attraction to EA is around doing the most good, and this area is a bottleneck, there seem much more positive framings to be used. 

One other question - I've seen quite a few posts trying to set what to do with EA's increased resources through inductive reasoning. I've seen less around examining what others have done successfully or unsuccessfully in terms of embedding sustainable growth and development (e.g. Singapore), m... (read more)

I was also surprised to be seeing management and scaling organisations described as "rarely people’s favourite activities", this seems to be a strong claim. For me, it's the most motivating activity and I'm trying to find an organisation where I can contribute in this area.

Agreed, I love management and improving organisational systems, and was really surprised by this comment!

Couldn't agree more, Rob. Perhaps my perception is coloured by my own experience and circle of friends, but there certainly seems to be a subset of people out there who genuinely enjoy scaling organisations. I think this is particularly the case in the for-profit sphere, where feedback loops are sometimes instantaneous thus leading to increased satisfaction among the scale-up types.  

5
Charles He
2y
As a caution, onlookers should know that there tends to be a large supply of would-be management advice or scaling advice whose quality is often mixed. This is because: * It is attractive to supply because this advice is literally executive or senior managerial work, so appears high status/impact/compensation. * It is attractive to supply because it moves into organizations where often the hard operational work and important niches have been developed successfully. In reality, it is often this object level activity that is hard and in "short supply". * Even in successful organizations, staff are often working around management/CEO or succeeding at their tasks despite leadership (it's not that leadership is bad it's that it provides many things in complicated ways). * Like other meta work, it can be (extremely) difficult to understand if you're good or bad. In particular, for scaling, the feedback loops can be very long. * Like other meta work, there can only be so many "cooks in the kitchen". In general, it is normal to scale or add more object level work, but for meta or managerial work the slots are limited (think of the reasons orgs only have have 1-2 CEOs) and more management can be negative. To see this, look at the general opinion of "management consulting" and how rarely these services are actually used by small, highly effective companies and organizations that I think are similar in profile to EA orgs.  I suspect that when they are used, it's because of great respect and trust for specific principals, and not because "management" can be easily sprinkled onto an existing organization.   Another issue is that "management" is a word that means many different things. As a very positive thing, and to an unusual degree, even junior EAs perform major management roles in EA organizations. Maybe the source of more management talent or activity would be to promote or "tap on the shoulder" inside EAs. There should be caution about creating external proces
3[anonymous]2y
It's useful to separate out consultancy/advice-giving versus the actual doing. I would say though that a successful management/operations setup should be able to at least ameliorate the feedback issue you mention (e.g. by identifying leading and/or more quickly changing metrics that are aligned and gaining value from these). 
2
Charles He
2y
I think your comment and sentiment is great. My response wasn't directly related. I guess I'm more concerned about "by catch" or overindexing. For example, activity and discussions that are wobbly about getting into management and scaling, "Great Leap Forward", sort of style.   Honestly, the root issue here is that I have some distrust related to the causes and processes about this post, the NB post, all of which seems to be related to discussion and concerns that might originated or closely involve the EA forum. I am don't think these have the best relationship to reality[1]. It seems healthy for the issues to settle down.  1. ^ I think the discourse on funding/optics has been slightly defective or tinged. This caused Will, SBF to pop onto the forum. This presence is fantastic, great and should continue, but maybe in this instance, different processes or events could have occurred, so they could have used their valuable time and public presence to communicate to EA about something else.

I know that they very much are not doing it for the gratitude, but I still want to express my thanks to Sam and the other core members of Alameda and FTX, as well as others earning-to-give, for making the current funding situation possible. I think earning-to-give has been downplayed in recent years in EA, and it seems like a thankless job to work crazy hours and take crazy risks for many years, doing frequently weird, uninteresting, or frustrating work, just to donate it in service of making the future better.

My own organization and I'm sure that of many others in this movement have and will benefit greatly from this influx of funding. But as the post says, this is a huge responsibility and we should not bear it lightly.

I hope that the large donors in EA will continue to be responsible stewards of the resources they've gathered. I hope too,  that the rest of us in the EA community will become strong enough to execute well on the most important priorities, to build a safe and flourishing future for our grandchildren's grandchildren.

[This comment is no longer endorsed by its author]Reply

Quantitatively: suppose $200 is spent on an intro to EA retreat for someone. If that has a more than a one in a five hundred thousand chance of inspiring the attendee to earn to give and successfully donate $100 million over their lifetime, then the expected financial benefit is positive. Given the successes we’ve seen, both from FTX and outside of that, the real probability is orders of magnitude larger. That’s not to say $200 on a retreat is how much should be spent — if you can have the same impact at cheaper cost, you should. And excessive spending can even become counterproductive if it sends the wrong message. But it indicates just how small community-building spending is in comparison the potential benefits from changing people’s careers for the better.

This is a bit of a nit, but $200 is very low here. You're likely to spend that much per attendee on housing alone, before considering other costs like organizer time or covering transportation.

I had the same exact reaction! "Only $200 for one attendee? In this economy? What is that, 20 bananas?" 
 

2
Ben Stewart
2y
Haha what a crossover!

a Christian EA I heard about recently who lives in a van on the campus of the tech company he works for, giving away everything above $3000 per year

Will this person please give an in-depth interview on some podcast? Could be anonymous if desired.

That person is Oliver Yeung and he has done a two part talk where he discusses this - main talk, Q&A.

(I spoke to him to okay sharing these, if any interviewer wants to speak to him then DM me and I can put you in touch)

Will, thanks very much for writing this. It's great to be having this discussion and to see the major players are thinking hard about this. I wanted to raise a couple of issues that merit reflection but haven't (AFAIT) been made so far.

You note that EA has gone from a few guys in a basement to commanding serious funding. But, what might the future of EA be? Where could it be in another 10 years? There could be 10x, or even 100x, of relevant funding. In line with the idea of judicious ambition, how should we be planning for it? Who should be planning for it?

Related to this, how much, and what type, of centralisation and governance are optimal across the various bits of the movement? One thing that strikes me is that 'EA resources' are very centralised: there are only a few major donors, advisors to those donors, and leaders of key organisations, and all those people know each other. What's more, lots of decision-making happens privately. All of this clearly has some major advantages, such as speed and coordination; it's appropriate, given it's about private individuals spending their money; it's also pretty unsurprising that this has happened because EA started so recently.

But, as E... (read more)

I logically acknowledge that: "In some cases, an extravagant lifestyle can even produce a lot of good, depending on the circumstances... It’s not my preferred moral aesthetic, but the world’s problems don’t care about my aesthetics."

I know that, but... I care about my aesthetics.

For nearly everyone, I think there exists is a level of extravagance that disgusts their moral aesthetics. I'm sure I sit above that level for some, with my international flights and two $80 keyboards. My personal aesthetic disgust triggers somewhere around "how dare you spend $1000 on a watch when people die of dehydration". Giving a blog $100,000 isn't quite disgusting, yet, ew?

The post I've read that had the least missing mood around speculative philanthropy was probably the So You Want To Run A Microgrants Program retrospective on Astral Codex Ten, which included the following:

If your thesis is “Instead of saving 300 lives, which I could totally do right now, I’m gonna do this other thing, because if I do a good job it’ll save even more than 300 lives”, then man, you had really better do a good job with the other thing.

I like the scenario this post gives for risks of omission: a giant Don't Look Up... (read more)

[EDIT 2: Based on the responses I received, I am probably wrong here. I will probably delete my portion eventually to not deter future readers from applying for grants. Leaving it up for a while longer for epistemic reasons. FYI that reading this thread might be a poor use of time.]

Side note: I see you mention multiple times that community building is a good use of money, and I agree, but that hasn't been what I have been seeing EAIF, the primary funder for CB work, go by. It is possible you are not using the term community building in the way I think of it, but:

Embarrassing context: I was refused by EAIF for FTE salary to do CB (in Austin, TX), and in response I: talked to many community builders, looked at past grant reports, dug through all the EA groups resources and relevant forum posts I could find, and spent most of my EAG London in 1:1s trying to understand. Result: It seems that funding for full-time community-building salary is not really a thing that happens (at least in America, outside of priority cities: NYC, DC, Boston, and the bay).

This to me says that funders (EAIF?) simply don't believe in CB. Personally, I think that if something is worth doing at all, it is:
1. w... (read more)

[I’m an EAIF grant manager, but I wasn’t involved in this particular grant.]

I’m sorry you’ve been having a frustrating time in your community building work. As you say, rejections sting even in the best of circumstances, particularly when it feels counter to the narrative being portrayed of there being funding available. Working hard to help others is difficult enough without feeling that others are refusing to support you in it.

It seems very difficult to me to accurately represent in advance what kinds of community building EAIF is and isn’t keen to fund, because it depends on a lot of details about the place/person/description of activities planned. Having said that, I’m keen to avoid people getting a false impression of our priorities. I wanted to clarify that we are in fact keen to fund full time community builders outside of existing EA hubs

It happens that the majority of past requests we’ve had for full time positions in the US have come from Boston/NY/SF/Berkeley. We’ve received a number of applications for full time community builders in non-hub cities in the rest of the world though. For example we’ve funded full time community builders in Italy, the Philippines, De... (read more)

Thanks for saying that. I understand that grantmaking is complex and that some CB plans simply won't be right to encourage. But I still don't really feel this changes my expectation around community building being funded for full-time. Some questions that would go a long way to correct this impression if answered:

(FWIW I feel weird posting this publicly, [EDIT: and I don't necessarily think you/EAIF should be expected to respond here] but I think it is important to ask these questions)

[EDIT: Also reading all this is probably not a productive activity for people who don't work in CB or grantmaking]

  1. Can you share how many of those organisers of non-american areas (Italy, Denmark, Czech Republic, Philippines) were mainly funded for FTE after doing PTE first? I know at least that in the Phillipines, some organizers were funded for part-time work first. I also remember reading one post by a part-time organizer (I think in the Phillipines) reflecting on that dilemma. S/he was lucky to get their full-time job to reduce to part-time, but wondering whether they should just quit their part-time job and work the remaining hours as EA Phillipine hours for free because there was so much to be do
... (read more)

Hi, EAIF chair here. I agree with Michelle's comment above, but wanted to reply as well to hopefully help shed more light on our thinking and priorities.

As a preamble, I think all of your requests for information are super reasonable, and that in an ideal world we'd provide such information proactively. The main reason we're not doing so are capacity constraints.

I also agree it would be helpful if we shared more about community building activities we'd especially like to see, such as Buck did here and as some AMA questions may have touched uppn. Again this is because we need to focus our limited capacity on other priorities, such as getting back to applicants in a reasonable timeframe. 

I should also add that I generally think that most of the strategizing about what kind of community building models are most valuable is best done by organizations and people who (unlike the fund managers) focus on the space full time – such as the Groups team at CEA, Open Phil's Longtermist EA Movement Building Team, and the Global Challenges Project. Given the current setup of EA Funds, I think the EAIF will more often be in a role of enabling more such work. E.g., we funded the Global Challen... (read more)

Thank you for this well-thought-out response. I appreciate the effort it took you and Michelle to respond to me. I am leaning much more that I was wrong about all this then. And if LA's application was initially part-time, that was one foundational wrong piece. I still wish that I could have received more details about my own application (the email specified that no feedback could be provided), but I will encourage more people I know to apply for CB work. 

I have added a qualifier to my original comment that I am probably wrong. As this particular forum piece and the comments are likely to be revisited for some time (maybe years?), I will probably eventually redact my comment fully to not confuse and deter future readers about how supported CB work would be. Will leave it up for epistemic reasons for at least a week longer.

Thanks again!

7
Kirsten
2y
This was a really inspiring reply to read Ivy.
9
Charles He
2y
(Disclaimer: I don’t know if my support or comment has value, no one likes me and I have poor hygiene.) People I know interacted with this person in real life in a professional context. From these interactions, what the commentor is saying seems accurate. In this particular instance, it seems like a valuable and talented leader could have been funded to do good, well aligned EA work that they were deeply passionate about.
9
Ivy Mazzola
2y
Thanks for your kind words. Most people have been surprised which has been affirming (much needed because rejections are the opposite). I got some in-person feedback suggesting EAIF saw risks to doing Austin CB too soon or with the wrong person (ouch).  I'm sure lots of people submit actually-risky projects who simply can't see them as risky (or themselves as risky agents), so take my confusion with a grain of salt.  The fund managers are people I genuinely respect. I'm just concerned that it was beaurecrat's curse which is also modus operandi for non-uni CB all around. EA has some bottlenecks that early or mid-career professionals are better suited to fill than students. So I don't want non-uni groups to be unhelpfully neglected. 
2
Chris Leong
2y
When did you apply? I wouldn't be surprised if you had a better chance of being funded now than in the past.

Recently, but it's complicated: 
I applied mid-March to FTX Future Fund. They passed me on to EAIF, saying they felt EAIF was better equipped to evaluate the grant (Fair, but I had chosen FTX because I saw no full-time employment grants for community builders in EAIF's past grant reports).  To their credit, EAIF did reach out to my references and gave me an interview in early April (so they were probably on the fence). Then said no on April 15th. 

P.S. I'm confident I could have gotten funding for part-time work, but I think the most impactful and innovative stuff for information value comes in the "later" hours, like designing and presenting workshops and courses. I could be thinking about this wrong but part-time work for an entire city is comparatively (and urgently) full of meetups, 1:1s, information dissemination, email, light outreach, and ops. Still important but not what I thought made the role most worth creating. 

guyi
2y28
0
0

There’s one huge difference between aiming to do good and aiming to make profit. If you set up a company aiming to make money, generally the very worst that can happen is that you go bankrupt; there’s a legal system in place that prevents you from getting burdened by arbitrarily large debt. However, if you set up a project aiming to do good, the amount of harm that you can do is basically unbounded.

 

I am very confused about this reasoning. It seems clear that there is a lot worse harm that can be caused by a for-profit enterprise than simply that enterprise going bankrupt. What about weapons manufacturers or fossil fuel and tobacco companies? There are many industries that profit from activities that many people would consider a net harm to humanity.

The key difference I see with a non-profit enterprise aiming to do good is that its scale is driven by external factors, the choices of donors. The harm a non-profit can cause is bounded simply because the funding it receives from its donors is bounded. In contrast, a successful for profit enterprise has a mechanism to scale itself up by using its profits to grow the enterprise. The practical implication of this is for profit corpo... (read more)

There’s one huge difference between aiming to do good and aiming to make profit. If you set up a company aiming to make money, generally the very worst that can happen is that you go bankrupt; there’s a legal system in place that prevents you from getting burdened by arbitrarily large debt. However, if you set up a project aiming to do good, the amount of harm that you can do is basically unbounded.

I am very confused about this reasoning. It seems clear that there is a lot worse harm that can be caused by a for-profit enterprise than simply that enterprise going bankrupt. What about weapons manufacturers or fossil fuel and tobacco companies? There are many industries that profit from activities that many people would consider a net harm to humanity.

I think it was good that you noticed your confusion! In this case, I believe your confusion primarily stems from misunderstanding the paragraph. Will is not saying that the worst that a company can do from the impartial point of view  is go bankrupt. He's saying that the worst that a company can do from a profit-maximizing perspective ("aiming to make profit") is go bankrupt. Whereas (EA) charities are presumed to be judged from the impartial point of view, in which case it will be inappropriate to ignore the moral downsides.

guyi
2y19
1
0

To be clear, stating that I was confused was a polite way of indicating that I think this reasoning itself is confused. Why should we evaluate for-profit businesses only from a profit-maximizing perspective? Having profitability as the primary goal of enterprise doesn't preclude that enterprise from doing massive amounts of harm. If a for-profit enterprise does harm in its attempts to make profit should we ignore that harm simply because they've succeeded in turning a profit? If you interpretation of Will's reasoning is what he intended, then he is asking us compare aiming to do good and aiming to make profit by evaluating each on different criteria. Generally, this is a misleading way of making comparisons.

This is important because this sort of reasoning is used to justify an uncircumspect version of encouraging people to earn to give that I see coming from this community. As I've seen it, the argument goes that in your career your should focus on acquiring wealth rather than doing good because you can then use that wealth to do good by being an effective altruist. But this ignores that you can potentially do more harm in acquiring great wealth than you can compensate for by using... (read more)

8
jacksondc
2y
The way I read it, Will is comparing the challenge of doing good to the challenge of earning money (for its own sake). Not, as you assume, to the challenge of doing good by earning money. The point is that if we try to learn any lessons from people who are optimizing to earn money, we'll need to keep in mind that we have reason to be more risk-averse than they are.
1
Linch
2y
Yes, this is what I meant to say, but failed to communicate. Thank you for putting it more succinctly and politely than I could.
5
Linch
2y
If someone from High-Impact Athletes writes a post outlining what we can learn from deliberate practice in professional sports, and mentions salient differences between how metrics work in sports training and EA training as a caveat to this advice, I do not think the correct takeaway is "this whole discussion is irrelevant to EA because sports are of ~0 impact for the world, and probably net negative anyway."
9
guyi
2y
What did I say warranted this comparison? I am not saying anything anyone has said is irrelevant to EA. I am saying argument Will made that for-profit businesses can't do much harm because their worst outcome is bankruptcy is misleading. I am also saying that encouraging people to earn to give without considering that harm can by done by earning is one of the ways the EA movement could end up having a net negative impact.
1
Linch
2y
At first I thought you genuinely misread his comment, and politely corrected you. And then your next comment (which did not thank me, or acknowledge the correction) suggested you deliberately misread it so you can get on your hobbyhorse about something else. So I tried to illustrate why this is not an appropriate strategy with an analogy. Perhaps my issue was that I being too metaphorical and insufficiently direct earlier. In that case I'm sorry for the lack of clarity in communication. 
2
Charles He
2y
There's several extremely bad outcomes of bad charities: * One example is in the footnotes and where people actually died[1]. * Another famous example is a pumping system for developing countries that consumed donor money and actively made it more difficult to get water. It's not clear anything would have stopped these people besides their own virtue or self awareness, or some kind of press attention. The effect of first world faculty and wealth is overwhelming and can trample all kinds of safeguards (a American volunteer blew the whistle on the first case, a fact that is thought provoking and informative, after hundreds of children and actual doctors passed through the clinic).   In a real business, overwhelming effort is being made to make sure the business is successful. In the hundred trillion dollar world economy, almost no one is paying money to help or harm people.  For any given amount of money, you can do tremendous more harm and kill people with the meme of doing good, than running a business, even in situations where there aren't many functioning institutions. 1. ^ Here is one example that made it into NPR and the New Yorker, Guardian, etc.  I suspect the truth is more complicated than these articles suggest. I think being a woman, blonde, American and blogging/Instagram played a role in the reason this person is being read about on US national media—the implication I'm making is that this might be happening all the time.  It's incredibly, deceptively hard to accomplish anything in very different cultures/economies/societies, much less cost effectively. Achieving this is possible, but rare and hard and undervalued.
1
guyi
2y
My point wasn't that charities are incapable of doing harm. There are many examples of charities doing harm as you point out. The point is that reasoning that non-profits have more potential to cause harm that for-profit seems to ignore that many for-profit enterprises operate at much larger scale than any non-profits and do tremendous amounts of harm. Yes, most successful business are primarily focused on making profit rather doing good or harm. But this doesn't mean they aren't willing to do harm in the pursuit of profit! If someone dedicates an amount of money to business that then grows it large enough to do lots of harm, even if as a side-effect,  it's quite conceivable they could accomplish more total harm than someone simply dedicating that same money to a directly harmful (but not profitable) venture.
2
Charles He
2y
You're absolutely right. For profits absolutely do harm. In general "capitalism has really huge harms", almost every EA or reader here would agree (note that I'm not an EA necessarily or represent EA thought). The scale is the point here—you're also exactly right. For many activities, it makes many, many millions to create a situation where we are harming people.  To be tangible, it's hard to think of any business you or I could setup that would be that harmful as posing as a fake charity and disrupting medical service or food supply.
2
Charles He
2y
Well, I've actually sort of slipped into another argument about scale and relative harm, and got you to talk about that.  But that doesn't respond to your original point, that businesses can do huge harm and EA needs to account for that. So that's unfair to you. Trying to answer your point, and using your view about explicitly weighing and balancing harms, there's another point about "counterfactual harm" that responds to a lot of your concerns.  In the case of a crypto currency company:   If you make a new make crypto company, and become successful by operating a new exchange, even if you become the world's biggest exchange, it's unclear how much that actually caused any more mining (e.g. by increasing Bitcoin's price).  There's dozens of exchanges already, besides the one you created. So it's not true that you can assign or attribute 20% or 50% of emissions to money, just from association.  In reality, I think it's reasonable that the effect is small, so even if the top #1 trading platform wasn't founded, almost the same amount of mining would occur. (If you track cryptocurrency prices, it seems plausible that no one cares that much about the quality of exchanges).  So the money that would have gone to your platform and been donated to charity, would buy yachts for someone else instead. (By the way—as part of your crypto currency company, if you make and promote a new cryptocurrency that doesn't mine, and "stakes" instead, then your cryptocurrency company might accelerate the transition to "staking", which doesn't produce greenhouse gasses like mining. Your contribution to greenhouse gasses is negative despite being a crypto company. But I share the sentiment that you can totally roll your eyes at this idea, let's just leave this point here.)   You mentioned other concerns about other companies. I think it's too difficult for me to respond, for reasons that aren't related to the merit of the concern.
-5
guyi
2y

Great post! I'll think about the ramifications later. One minor note:

As well as the unilateralist’s curse (where the most optimistic decision-maker determines what happens), there’s a risk of falling into what we could call the bureaucrat’s curse,[10] where everyone has a veto over the actions of others; in such a situation, if everyone follows their own best-guesses, then the most pessimistic decision-maker determines what happens.

If I understand Bostrom et.al's paper correctly, this is just a special case of the unilateralist's curse, and is mathematically equivalent (bottom of pg.7-8 in pdf, 355-356 in the journal):

Finally, fifth, though we have thus far focused on cases where a number of agents can undertake an initiative and it matters only whether at least one of them does so, a similar problem arises when any one of a group of agents can spoil an initiative—for instance, where universal action is required to bring about an intended outcome. Consider the following example: 

[...]

These cases of unilateral spoiling or abstinence are formally equivalent to the original unilateralist curse, with merely the sign reversed. 

Since the problem in these cases is the resul

... (read more)

As someone who has heard a lot of criticism from non-EAs about the perception of loose money and funding in EA, I really appreciate this post and the thoughtfulness :)

 I do really believe that the outsider perception of lots of money in EA is a significantly negative and potentially a major deterrent to community-building efforts. Posts like this and, as you mention, criticisms/red teaming of EA I think really help build a more positive image and more importantly accountability. Curious about how to translate these efforts (which take place mainly on the Forum, not something new EAs typically are on) into outreach material and community building. 

olly
2y18
0
0

Great to hear Will's thinking this. Some thoughts

I believe that EA is about

  • discovering problem areas that are underfunded/ignored
  • discovering robust, testable (maybe capital efficient?) solutions to those problem areas
  • allocating capital to these solutions

EA has capital, but the bottleneck is either problem areas that are underfunded and/or robust testable solutions to problem areas.

To me, this sounds a lot like starting a startup. Founders aim to find under-explored areas in which they can build robust solutions. We know this works well for incentivizing innovation.

So my thinking is as follows: rather than attempting to prescribe what people could solve, EA starts giving large cash payouts for demonstrated QALYs saved (near term / long term / recurring). Admittedly this could be a lot of work. However, EA is already doing this in its assessment of charities. The objective would be to use EA capital to provide an upside for startups that demonstrate that they save a considerable number of QALYs. Maybe this creates strange incentives but could be something to explore.  

Some clarifications on terminology:

Briefly, the main reasons I see in favour of giving more are:

  • ....
  • Option value: if we build the infrastructure to productively absorb funding, then we can choose not to use it if it turns out to not be necessary; whereas if we don’t build the infrastructure now, then it will take time to do so if in a few years’ time it does turn out to be necessary.
  1. How do you define infrastructure ? Could you give some concrete examples of infrastructure you had in mind when you wrote this? (I've noticed there is a lack of clarity around this term in general). 
1
David M
2y
I turned this into a question. Maybe someone will answer: https://forum.effectivealtruism.org/posts/piLpBkxFyGcnvT6fA/what-is-meant-by-infrastructure-in-ea

Regarding Harming quality of thought, my main worry  is a more subtle one:

It is not that people might end up with different priorities than they would otherwise have, but that they might end up with the same priorities but worse reasoning

I.e. before there was a lot funding, they thought "Oh I should really think about what to work on. After thinking about it really careful, X seems most important". 

Now they think "Oh X seems important and also what I will get funded for, so I'll look into that first. After looking into it, I agree with funders that this seems most important."

This is still for the same X, and their conclusions are still the same. But their reasoning about X has now become worse because they investigated important claims less thoroughly.

Great post  - always fun to see Will weighing in on hot-button issues. 🎉

As for where to draw the line on personal spending and frugality, the example of flying business class on an airplane is a perfect illustration: no one needs to fly business class, and the marginal benefits of extra legroom and early boarding are so not worth the 2x or 3x ticket price, imo. 

To the concern about value drift and optics, our reputation as a movement would obviously be tarnished if folks like Will and Toby (or any of us) bought yachts and mansions. If we can avoid flagrantly conspicuous consumption, that'd be great. Beyond that, we shouldn't be eating rice and beans every night. I want a well-balanced diet of kale and quinoa for every EA doing good work out there!

Let's not forget that SBF cooks his own meals 

Also relevant, I once asked Peter Singer why we don't all walk around in ash and sackcloth in order to donate every spare penny, and his response was "if everyone in a movement you had never heard of before were walking around in ash and sackcloth, would you really want to join?" So even my homeboy Peety acknowledges the importance of optics.

Last point: take the Further Pledge i... (read more)

Buck
2y38
0
0

I massively disagree re the business class point. In particular, many people (e.g. me) can sleep in business class seats that let you lie flat, when they would have not slept and been quite sad and unproductive.

not worth the 2x or 3x ticket price

As a general point, the ratio between prices is irrelevant to the purchasing choice if you're only buying something once--you only care about the difference in price and the difference in value.

3
DB
2y
Agree with this as a general principle, provided the "difference in value" also takes into account longer-term effects like movement reputational cost. I don't think individuals choosing to fly business class based on productivity calculations has much, if any, movement reputational cost. On the other hand, a prominent EA figure might accurately calculate that they gain one extra productive work hour each week, valued at say $100, by paying someone $50 to brush and floss their teeth for them while they sit there working. This is obviously a fanciful scenario, but I think there are lots of murky areas between flying business class and having a personal teeth brusher where the all-things-considered value calculation isn't trivial. This is especially the case for purchasing decisions that can't easily be converted to work productivity boosts, e.g. buying expensive luxury items for the pleasure they bring.

no one needs to fly business class, and the marginal benefits of extra legroom and early boarding are so not worth the 2x or 3x ticket price

If all you want is extra legroom, you can get an exit-row seat for much less. Early boarding isn't worth very much unless you're traveling with a carry-on that can't be checked (ex: musical instrument) and there are cheaper ways to get it. I see the real benefit of business class as (a) a more comfortable place to work or (b) arriving better rested, especially if it gets you a lay-flat seat on an overnight flight. Personally, this isn't a trade-off that has made sense for me, but I can see cases where it would be worth it if it gives you essentially an extra working day.

FWIW I've been trying to fly business class for transatlantic flights for a few years for these reasons. I think it's an usually big effect size for me because otherwise long haul flights play badly with my chronic fatigue and can cost me effectively >1 day, but I expect that many people would get a few hours' worth of extra productive time (I take advantage of both the lie-flat bed and the good work environment for writing that doesn't need internet).

I've felt weird about expensing it so mostly just been paying for it myself (I don't have many other big expenses in my life except childcare), but I have noticed that I'm sometimes wanting to make a strong recommendation that a particular other person try to fly business class, and offered to pay for it personally because I think this will help significantly more than what I can otherwise do with my donations. So I seem to be at "we should probably do a bit more of this at the margin".

7
Rebecca
2y
Yeah I think it's a very different calculation if your flight is up to 24 hours long. Also you can only take an exit row seat if you have the (physical) capacity to help in an emergency (e.g. can't be flying solo with kids or elderly relatives, you have to be able  to throw 20kg at a time,  can't have certain other impairments, have to be able to understand the language of the relevant country and so on), so I don't know how scalable a suggestion that is.

Here's a prediction: In the not-too-distant future, someone who calls themselves an effective altruist is going to purchase a private plane or helicopter and justify it saying the time it saves and the amount of extra good they can do with that saved time is worth the expense. The community is going to have a large population that disagrees and sees it as a wasteful extravagance, and a smaller but vocal population that will agree with the purchase as a worthwhile tradeoff, especially if that person is part of a sub-community within EA that is ok with more speculative expected value calculations. Instead of there being a clear, coordinated response disavowing the purchase as extravagant, the community is going to hesitate and argue about the extent to which it is good to feed utility monsters and be muted in its outward response. But that's not going to stop the wider media picking up the story. A small fraction of the population will then henceforth liken EAs to the pastors at megachurches with private jets who use do-gooder justifications for selfish purposes. And yes, you could construct some sort of hypothetical where someone needs a helicopter to more quickly fly between trolley levers to save a bunch of people. But the much more likely scenario is that someone wants a helicopter and is fine using an iffy, cursory justification for it and the trolley brakes are working just fine.

2
aaronmayer
2y
Well phrased! I'd bet you're right that something akin to this will happen in the future. Solid prediction. I guess I'm just trying to not be that guy, and I hope everyone else tries to not be that guy too.

My intuition is that there is also some potential cultural damage, not from the money the community has, but from not communicating well that we also care a lot about many standard problems such as third world poverty. I feel that too often the cause prioritization step is taken for granted or obvious, and can lead to a culture where only "cool AI Safety stuff" is the only thing worth doing.

Can you give an example of communication that you feel suggests "only AI safety matters"?

9
MichaelPlant
2y
Not exactly the same thing, but there was a whole post and discussion on whether EA is "just longtermism" last week. 
Dion
2y28
0
0

Adding onto this, the Virtual Programs (Introductory) currently has 3 weeks dedicated to Longtermism, Existential Risks and Emerging Technologies whereas there are little to no compulsory content on poverty, global health or climate change. (except Pandemics) Many of my participants have voiced out on this. If facilitators are not able to give a good answer, it can be easy for newcomers to have a skewed perspective that EA is just longtermism and x-risk.

In particular, this comment by Max Dalton. While I don't think that means "Only AI safety matters" I think it would lead to much more content on AI safety than I expected.

Where we have to decide a content split (e.g. for EA Global or the Handbook), I want CEA to represent the range of expert views on cause prioritization. I still don't think we have amazing data on this, but my best guess is that this skews towards longtermist-motivated or X-risk work (like maybe 70-80%). 


https://forum.effectivealtruism.org/posts/LRmEezoeeqGhkWm2p/is-ea-just-longtermism-now-1#2_1_Funding_has_indeed_increased__but_what_exactly_is_contributing_to_the_view_that_EA_essentially_is_longtermism_AI_Safety_

7
projectionconfusion
2y
Its not specific communications so much as it is the level of activity around specific causes. How many posts and how much discussion time is spent on AI and other cool intellectual things, vs. more mundane but important things like malaria. Danger of being seen as just a way for people to morally justify doing the kind of things they already want to do. 
3
PabloAMC
2y
I don't think we have ever said this, but the is what some people (eg Timnit Gebru) have come to believe. That was why, as the EA community grows and becomes more widely known, it is important to get the message of what we believe right. See also the link by Michael above.
1[comment deleted]2y

This was a very clear and helpful thread. I'd suggest something that a much higher amount of funding could allow EA to pursue:

Recruit more mid-career and late-career researchers into EA, particularly established academics who are already working on EA-related issues, but who might not be inside the EA community yet, and who never viewed EA as a potential funding source. Often these researchers have established track records of publishing and consulting, run large lab groups full of grad students and post-docs, and have high impact and visibility within their fields. 

But they're often spending huge amounts of time applying for government research grants that have very low funding rates (below 10%), to keep their labs going and to supplement their salaries (e.g. for teaching buy-outs & summer salaries). This is extremely frustrating for most of them. And they have to wrap their real research interests up in some kind of package that sounds appealing given the current NIH, NSF,  EU, or UK Research Council funding priorities, which have heavy political biases & very narrow Overton windows.

If these researchers were more aware that EA can offer grants just as large as t... (read more)

Here’s a reminder I’ve found helpful for feeling risks of omission: there’s a world in which we fail.

There’s a world in which we tried to stop X-risk by unaligned AI or a catastrophic pandemic, but we didn’t. We didn’t act with enough urgency, and being really cautious about our spending slowed us down.

Thanks for writing this! I am also very strongly drawn to frugality from a combination of, as you say, moral aesthetic, and experience during a time when money was much less available. I wish I could give some solid advice here on "how I learned to stop making bad labor-money tradeoffs" but this is still not something I'm great at.

[anonymous]2y13
0
0

This is a great post and a fascinating history - has inspired me to create an account.

"So far, we've generated more than $30 bn for something like $200 mn, at a benefit:cost ratio of 150 to 1;[6]"

This I think is a weak chain in the argument. You may have raised $30bn for great causes, but how much of that would have gone to great or even good causes regardless?

Charities can often find fundraising to be profitable, but they are also normally taking resources from each other - (shifting donations from charity a to charity b) and so sometimes the net effect across the charitable economy is just spending more on seeking donations.

In EA case - would these people pledging to against malaria foundation really have spent the money on sports cars? Or even on ineffective opera house charities? I don't think so. The kind of people motivated to seek out a good charity would still do so, but they would arguably do less effectively in the absence of an effective altruism movement guiding them.

I don't disagree with the general thrust of the piece. But I think that throwaway 10x return claim on fundraising is perhaps dangerous.

The kind of people motivated to seek out a good charity would still do so, but they would arguably do less effectively in the absence of an effective altruism movement guiding them.

I think the question of what people would have done in the absence of EA movement building is really hard, but my impression is different here. Personally, without being surrounded by a group of people who view altruistic dedication as normal, I think a likely outcome would have been to increasingly prioritize myself and my family as I got older. That is a common pattern, with idealism and willingness to make sacrifices decreasing with age.

The excited/obligatory motivation perspective is also relevant here: without a movement I think you get many fewer excitement-motivated people working on altruistically valuable things.

1[anonymous]2y
Agree it is hard to know. I think it's a very good point that a movement/community can sustain dedication over time.
7
Benjamin_Todd
2y
The unstated claim is that the charities EAs are donating to now are significantly more effective than where people would have donated otherwise (assuming they would have donated at all). If the gain in cost-effectiveness is (say) 10-fold, then the value of where the money would have been donated otherwise is only 10% of the value now generated. That would reduce the cost-effectiveness multiple from 10x to 9x. I think a 10x average gain seems pretty plausible to me – though it's a big question! Some of the reasoning is here, though this post is about careers rather than donations: https://80000hours.org/articles/careers-differ-in-impact/

Great post

What are we going to spend this on? There seems to be a shortage of evidence-based global development causes. Current GiveWell charities are growing as fast as they can but the funding pool is growing faster (which is great! But it is already making me hesitant to give or encourage others to)

Should we not be working with and giving to orgs like Innovations for Poverty Action and JPAL to help us find new causes? Our cause discovery rate seems very slow as it is. The only new GiveWell cause in the last few years has been New Incentives, which is already fully-funded.

6
freedomandutility
2y
If it’s possible / feasible, I imagine Charity Entrepreneurship collaborating with JPAL and IPA would lead to some new highly cost-effective global health startups.
5
Tyner
2y
EA funds gave $3 Million to IPA over the last two years: https://funds.effectivealtruism.org/funds/global-development Did you mean something different?
1
Henry Howard
2y
That's great. Cause discovery should be top priority while we're running out of causes

But morally-motivated people, especially on college campuses, often find seemingly-extravagant spending distasteful.

 

As far as I can see, no-one else has raised this, but to me the optics of having large sums of money available and not spending it are as bad or worse as spending too freely. Cf Christopher Hitchens' criticism of Mother Teresa - and closer to home, Evan's criticisms a few years ago that EA fund payouts were being granted too infrequently. For what it's worth, I find the latter a much bigger concern.

3
Arepo
2y
Sub-hypothesis: the people  who find extravagant spending distasteful are  disproportionately likely to be the people who object to  the billionaires that enable it - and so that spending it isn't what pisses them off so much as what draws their attention to the scenario they dislike.

The "bureaucrat's curse" reminds me of Vitalik's bulldozer vs vetocracy political axis: https://vitalik.ca/general/2021/12/19/bullveto.html

Vetocracy can be beneficial if a system's strength depends on it not changing. e.g. People invest in Bitcoin because it's incredibly difficult to change its monetary policy. Bitcoin doesn't need to innovate. 

But if Ethereum is too vetocratic and fails to innovate - it could get outcompeted by other more nimble startups like Solana or Avalanche.

The current mood in the AI Safety community appears to be pessimistic. F... (read more)

Apologies for the basic question but, if there’s more money allocated than there are effective projects, should those of us earning to give start funding less effective charities? Is there a ranking with effectiveness and funding % we can sort by?

I think there might be several things we could spend more money on while still being viewed in a positive light by most of our stakeholders and the more spartan among us. Examples include:

-More staff in organisations and projects to prevent burn-out

-Child-care facilities, paid parental leave, etc.

-At least 4 weeks paid holidays/year 

I agree that all such spending should be vetted according to our high bars and must admit I have not done so before making this comment. That said, anecdotal evidence from for example from Patagonia offering child care see... (read more)

Among Giving Pledge signatories, there are around ten who are at least somewhat sympathetic to either effective altruism or longtermism. And there are a number of other successful entrepreneurs who take EA or longtermism seriously

How are you defining longtermism, to end up with EA and longtermism being alternatives? 

To clarify, is there another earning-to-give EA billionaire who is choosing to remain anonymous?

1
Patrick Gruban
2y
He might be referring to Gary Wang as he does later in the text, but not sure about this

At least one person earning to give (and not related to FTX) has a net worth of over a billion


Can't be Gary Wang, as he's related to FTX

3[comment deleted]2y

One of the keys things you hit on is  "Treating expenditure with the moral seriousness it deserves. Even offhand or joking comments that take a flippant attitude to spending will often be seen as in bad taste, and apt to turn people off."

However, I wouldn't characterize this as an easy win, even if it would be an unqualified positive. Calling out such comments when they appear is straightforward enough, but that's a slow process that could result in only minor reductions. I'd be interested in hearing ideas for how to change attitudes more thoroughly and quickly, because I'm drawing a blank.

Thanks for the thoughtful post. (Cross-posting a comment I made on Nick's recent post.)

My understanding is that people were mostly speculating on the EAF about the rejection rate for the FTX future fund's grants and distribution of $ per grantee. What might have caused the propagation of "free-spending" EA stories:

  • the selection bias at EAG(X) conferences where there was a high % of  grantees.
  • the fact that the FTX future fund did not (afaik) released their rejection rate publicly
  • other grants made by other orgs happening concurrently (eg. CEA)

This post ... (read more)

There are other prospective major donors, too. Jaan Tallinn, the cofounder of Skype, is an active EA donor. At least one person earning to give (and not related to FTX) has a net worth of over a billion; a number of others are on track to give hundreds of millions in their lifetime. Among Giving Pledge signatories, there are around ten who are at least somewhat sympathetic to either effective altruism or longtermism. And there are a number of other successful entrepreneurs who take EA or longtermism seriously, and who could increase the total aligned fundi

... (read more)

There being a lot of funding available in EA also changes the calculus for people deciding if they want to donate their own money. If there are super rich people donating to EA, to the extent that finding ways to spend money is a problem, then the motivation for normal individuals to donate is lower. 

I think it changes it some, but not hugely? Even if the best remaining option for making the world better was direct cash transfers I think donations would still make a lot of sense; There's Lots More To Do.

I also don't think we stay in the current dynamic. Part of why it is important for a lot of people to go into directly doing useful things now is to identify and scale up opportunities for directing a lot of money toward valuable things. It's become harder to identify extremely cost-effective ways of spending your money to speed that process up, but money will still be very important.

4
projectionconfusion
2y
Sorry I should have been clearer, I was meaning more in psychological terms than economic ones. An extra dollar might still do the same amount of good, but the way people intuitively assess impact it will feel very different depending on the funding context people feel it is in. 

"it can cause us to go awry if it means we don’t take chances of upside seriously, or when we focus our concern on false positives rather than false negatives"

I've encountered this problem repeatedly in my attempts to speak with EAs here in the East Bay. With one topic, for example, I can napkin the numbers for them: $5 Trillion in real estate impacted by hurricanes, in US alone - so, there's on-the-order-of a $1 Trillion wealth-effect if we can stop hurricanes. A proposal with a 1:1,000 chance of doing so would still be worth $1 Billion-ish to check for f... (read more)

where I have been waiting months for any response

You're referring to Seeking a Collaboration to Stop Hurricanes, right?

Reading your comments on that post, it sounds like you were hoping readers would respond to your post by making things happen. That's occasionally the way things go, but almost always if you want to move something along you need to drive it. For example, if you think this is possibly one of the most important things to do you could consider:

  • Seeking funding to work on this full-time. With your full time work you could learn how to run a simulation, or attempt to convince the government to run one. You could also explore related hurricane prevention proposals (for example, Myhrvold's proposal) .

  • Seeking funding to hire someone to work on this full-time (but if you aren't willing to do it I expect funders to consider that a negative signal)

  • Actively looking for collaborators, for example by attempting to identify relevant academics

0
Anthony Repetto
2y
Just to note for the next person: I am now being called "powerless and vulnerable" because I stand against being mis-represented. I have been mis-represented repeatedly, and so I have responded to each - yet, the fact that I  "repeatedly post (complaints about being misrepresented)... makes sure that not many people take (me) seriously." If your clique repeatedly mis-represents me, and then they use my own self-defense as a reason to justify exclusion, you've earned the title of clique!

I'm going to be honest. I think you'd have a better experience here if you engaged with people in a way that was less adversarial. I can understand why you might be frustrated that people didn't engage more with your ideas or that people misinterpreted what you wrote, but it seems to me that you're currently in a cycle where you felt you were mistreated or ignored which leads you to send out negative energy which then results in further negative interactions; and hence the cycle continues.

0
Anthony Repetto
2y
For some reason, my original response is not showing up. I definitely did NOT make any attack on anyone, during my comment. I don't see why it would be deleted - I request a review of whoever deleted my response. Here it is, again: "We rich white people would give you so much more respect if you poor black people spoke nicely when you complained." <--- this argument has been used a thousand times, around the world, to get people to cower while you continue to disrespect them. I won't cower; I am right to be upset, and I expect an apology for being misrepresented by them. I am not wrong for requesting this. Further, again, I am not talking about "lack of engagaement" - ONLY your people have made that claim, and I dismiss it each time you've made it. I continue to point-out: I have been repeatedly misrepresented. I deserve an apology.
0
Anthony Repetto
2y
More to the point: why is my tone the more pressing issue, compared to the fact that I've been repeatedly misrepresented? Your Us/Them priorities are showing.

I think the biggest reason why your tone is relevant here is that you are seeking introductions to potential collaborators. People care a lot about what others are like to work with!

2
Anthony Repetto
2y
I agree! So, consider the scenario: I stand-up and ask "does anyone know someone I might talk to?" and the response I get is "but we don't want to give you money". I correct that misrepresentation, repeatedly, until I suspect that I am being trolled - and my self-defense is used as a reason to ignore me. If I hadn't been poked-in-the-eye repeatedly, those introductions would begin on a pleasant footing. Core to this problem: each of you are focusing on how I can "get better results by playing nice". I am focusing on "I was misrepresented, and that should be considered first, in the moral calculus." If I roll-over every time someone bullies me, then I'll be liked by a whole lot of bullies. That doesn't sound like a win, to me.
5
Chris Leong
2y
I honestly think this is the one thing I could have said that could have helped you achieve your goals the most, more than offering a connection to a relevant person, if I actually knew someone interested in hurricane prevention.
4
Chris Leong
2y
I suppose what you're saying makes sense from where you stand. I guess I'm trying to show you another way of seeing the world, even though I know it won't make sense from your perspective. I'd encourage you to imagine briefly what it would mean for this to be true, to explore what the world looks like through this lens, and its internal logic. I suspect that this exercise could be valuable even if you don't end up agreeing with this perspective.
1
Anthony Repetto
2y
When did you edit your response? You were saying something else, originally... Yes, I can imagine the world where I respond to the misrepresentations with politeness - I did that for twenty years, and the misrepresentations continued, along with so many other forms of bullying. I have seen the world from that lens, and I learned that it's better for me to stand-up to misrepresentations, even if that means the bully doesn't like me.
2
Chris Leong
2y
I have no idea if I edited it or not. I tried checking to see if they had a history feature, but apparently not.
2
Chris Leong
2y
Maybe I should have been clearer. I'm asking you to imagine the world where everyone isn't intrinsically against you, they've tried to help and they've been pushed away. I know that's a difficult ask, but I suspect it would be worthwhile.
1[anonymous]2y
Strong downvote for extreme and inappropriate condescension in the guise of helping someone. There is no adequate reason for you to assume that Anthony is living in a world where everyone is intrinsically against him, and that he cannot even imagine not living in a different world. This is an extremely strong statement to make about someone you know through a few online comments. Why do you think you're right?  Even if you were right, helping him would not take the form of trying to point this out publicly in such a tactless way. 
3
Chris Leong
2y
I can see why you might think it's a guise, but it really isn't the case. I think you're correct that it does come off as slightly condescending, but this isn't intentional. I'm trying to expand the range of what I can say without coming off as condescending, but for there are some things where I find it challenging; where it feels to me like trying to thread a needle. In any case, your comment contains useful feedback. I just want to make it clear that it's a genuine attempt to say the most helpful thing that I can, even if I think it only has a small chance of making a difference, but I agree that a private message might have been better. As for why I think this what I think, it's mostly based on my experience of dealing with people. I could produce some explicit reasons if I really wanted to, but I'm not sure if it's worthwhile given that they are more the sideshow than anything.
3[anonymous]2y
Thanks, this is a good followup. I'm glad my comment contained useful feedback for you.  I think your attempt to help Anthony went awry when he asked you why his tone was the bigger issue than whether he had been misrepresented, and you did not even seem to consider that he could be right in your reply. Perhaps he is right? Perhaps not? But it's important to at least genuinely consider that he could be.
6
Anthony Repetto
2y
Thank you for recognizing that my concern was not addressed. I should mention, I am also not operating from an assumption of 'intrinsically against me' - it's an unusually specific reaction that I've received on this forum, in particular. So, I'm glad that you have spoken-up in favor of due consideration. My stomach knots thank you :)
4
Charles He
2y
I don’t feel good about this situation, but I think your judgement is really different than most reads of what happened: * It’s clear to me that there’s someone who isn’t communicating or creating beliefs in a way that would be workable. Chris Leong’s comments seem objectively correct (if not likely to be useful). * (While committing this sin with this comment itself) It’s clearly better to walk away and leave them alone than risk stirring up another round of issues.
4
Chris Leong
2y
My comment very well may not be useful. I think there's value in experimenting with different ways of engaging with people. I think it is possible to have these kind of conversations but I don't think that I've quite managed to figure out how to do that yet.
1
Charles He
2y
I think the person involved is either having a specific negative personal incident, or revealing latent personality traits that suggest the situation is much less promising and below a reasonable bar for skilled intervention in a conversation. With an willingness to be wrong and ignore norms, I think I could elaborate or make informative comments (maybe relevant of trust, scaling and dilution that seem to be major topics right now?). But it feels distasteful and inhumane to do this to one individual who is not an EA. (I think EAs can and should endure much more, directly and publicly, and this seems like it would address would be problems with trust and scaling).
4
Chris Leong
2y
That's useful feedback. I agree that it would have been better for me to engage with that more.
2[anonymous]2y
Glad to have been helpful :)
-6
Anthony Repetto
2y
-4
Anthony Repetto
2y
I am waiting for an apology from them - I don't know why I should be pleasant after repeatedly being disrespected. That sounds like you're asking me to "be a good little girl, and let them be mean to you, because if you're good enough, then they'll start to be nice." It's not a fault upon me that I should 'be nice until they like me' - they misrepresented me, which is the issue, NOT "lack of engagement".
-18
Anthony Repetto
2y
0
Anthony Repetto
2y
I am not looking for funding - I asked if anyone is interested in running those simulations, or knows someone they could put me in touch with. I quote from my post directly above: "I'm NOT looking for any funding, either - there's a decent chance that the cost of the solution is lower than the Federal Government's increased tax-revenue from hurricane-prevention, so I say that the government should pay for it." I'm appalled that multiple times, here now and when I posted originally, after stating that I am NOT seeking funding, I am repeatedly misrepresented as "seeking funding". It's a basic respect to read what I actually wrote. Included in my hope for connections are the relevant academics - I began my search at the EA Berkeley campus chapter. I know that the government would not listen to me without at least a first-pass simulation of my own; and I know that it is ludicrous for me to invest time into developing a skill that others possess, or re-inventing the wheel by making my own simulation. Those are all significantly more wasteful and ineffectual than asking people if they know anyone in related fields - this is because social networks are dense graphs, so only two or three steps away, I am likely to find an appropriate specialist. Your advice is not appropriate. At no point did I ask the readers to do the work of simulation, or proposing to the government, on my behalf; you used a strawman against me. I am specifically asking if people can look in their social network for anyone with relevant skill-sets, who I might talk to - those skilled folks are the place where I'm likely to find someone who would actually do work, not here on the forums with a string of dismissive 'help' and fallacies.

I can think of two main reasons why your posts haven't resulted in introductions to relevant specialists:

  • People with those connections haven't seen your posts.

  • While such people have seen your posts they don't consider this opportunity sufficiently promising to pass it on.

While many people do read the Forum it wouldn't be surprising if no one had seen your post who knew anyone relevant, since there aren't that many relevant experts. And even if they are, when you give someone an introduction you are staking some of your social capital, and based on your initial post and comments here I would not, personally, be willing to stake such capital.

I had seen that you'd written that you weren't looking for funding, and my post above doesn't suggest that you were. Instead, I was suggesting that you do and giving ideas on how you might use funding to make progress on this project. After reading your responses here, however, I withdraw that suggestion.

8
Anthony Repetto
2y
I apologize for lumping your funding-suggestion along-side others' funding-misrepresentation. I see that you are looking for ways to make it possible, and funding is what came to mind. Thank you. (I am still surprised that funding is continually the first topic, after I specify that the government is the best institution to finance such a project. EA would go bankrupt, if they tried to stop hurricanes...) And, I understand if people don't consider my proposal promising - I am not demanding that they divert resources, especially funds which are best spent on highest guaranteed impact! Yet, there is a cliquishness in excluding diverse dialogue based upon "social capital/reputation" - I hope you can see that the social graph's connectivity falls apart when we cut those ties. It's also odd that the only data-point used to evaluate me would be the slice of time immediately after I'd been prodded repeatedly. I wish I could hand you the video-tapes of my life, and let you evaluate me rightly. When I am repeatedly misrepresented, defending myself, then you don't see a representative slice of who I am. Worst of all, no measure of my persona or character is a measure of the worth of a thought. If I am not a good fit for making it happen, then the best I can do I find someone who fits that well. The idea itself stands or falls on its own merits, and measuring me ignores that. I won't know if it's worth doing until I have a simulation, at least. I don't know how anyone else has certainty on the matter, especially from such a noisy proxy as "perceived tone via text message".
4
Jeff Kaufman
2y
The reason I brought up funding was not that I thought it might make sense for EAs to fund the entire thing, but that it might allow you to address the reasons your proposal is currently stalled. I gave a few ideas of specific things you might do with funding: * Free up your time to learn how to run a simulation. * Free up your time to for lobbying. * Exploring existing work on hurricane prevention. * Hiring someone else to do any of the above.
8
Anthony Repetto
2y
Yes, I understand that funding can let me hire people to do that work - and I don't need funding to free my time. I understand that, if I delay for the sake of doing-it-alone, then I am responsible for that additional harm. It doesn't make sense for me to run a simulation or lobby by myself; and I've been in the position of hiring people, as well as working with people who are internally motivated. I hoped to find the internally motivated people, first - that's why I asked EA for connections, instead of just posting something on a job site.
1[comment deleted]2y

It is not reasonable to expect people to spend 1 to 2 hours listening to an idea that is not relevant to them.

-5
Anthony Repetto
2y

Have you talked to Sella Nevo, who does flood prediction at Google?

My own hot take here is that if you spent 1 billion dollars of EA money to save the US gov't 1 trillion dollars, you've likely wasted >900 million dollars. But I know other people are more optimistic about US gov't funding priorities (or more pessimistic about EA uses of money).

-1
Anthony Repetto
2y
I quote from above: "I'm NOT looking for any funding, either - there's a decent chance that the cost of the solution is lower than the Federal Government's increased tax-revenue from hurricane-prevention, so I say that the government should pay for it." I'm NOT asking to use EA money - I repeatedly clarify that, at every opportunity, and yet it is insisted-up multiple times, on this forum. No, EA only has $30B, so you can't afford stopping hurricanes, even if you spent your entire budget. I pointed to the potential value of looking for solutions at $1B, which is the actual expected-value, NOT the 'value to EA'. I'm not trying to take ANY dollars from your charities. Do you hear that, yet? I don't appreciate being repeatedly misrepresented and strawmanned. Are you suggesting that I cold-call Sella Nevo? Do you have a way to put me in touch, so that I am not ignored, as I have been here?

I did not say you were looking for funding. I am sorry to the degree I am responsible for miscommunication, or if I unintentionally upset you in any way. I am always trying to be better at communication. I hope you have a good day. 

3[anonymous]2y
Upvoted for the last three sentences, but I believe your first sentence is incorrect. The second paragraph of your initial comment does not make sense to me in the absence of you believing that Anthony was looking for funding. 

I was not intentionally suggesting that Anthony was asking for a billion dollars in funding. It's strange to me that >=2 people will read my comment that way. I'm again sorry for any miscommunication.

I don't think it's prudent for me to engage further in this thread, even though this type of thing naturally draws me in. I will donate $10 to Homeopaths Without Borders if I comment further.

I hope you have a good day.

[anonymous]2y10
0
0

My read on your comment is that you misread Anthony's allusion to $1b as about potentially spending $1b at some stage (whether right now or later), rather than about the expected impact of his idea. I could be wrong, but that's the only way your comment makes sense to me ("if you spent $1b of EA money" - what could this refer to besides spending $1b of money?). 

Anthony is asking for connection to someone who is skilled at running a particular kind of simulation to see if his idea has potential. He believes that the value of checking of his idea might be $1b, because of potentially trillions of dollars worth of gains. Crucially, it would not take $1b to check his idea - that figure is an estimate of the potential value of checking the idea, not of the cost of checking it. The cost of checking is probably something like the social capital to connect him with a relevant person and the costs involved in running the simulation (if it progresses to that stage).

I don't think this was a bad mistake on your end, just a quick, incorrect assumption that you made while trying to help someone. It only led to a fractious response because so many other EAs have also misread and misunderstood... (read more)

4
Charles He
2y
The “misrepresentation” about a search for funding was related for monies for a personal project or org to develop the intervention. The second paragraph was for funding for the intervention itself. They are really different things. Like the difference between an org researching food aid, versus buying billions of dollars of actual food. I doubt the person believes he can literally stop hurricanes without government funding. Unfortunately, I think you are muddying the waters in your intervention. With my read of the relevant person, this might not serve them well.
-12
Anthony Repetto
2y
1[comment deleted]2y
7
DC
2y
As someone who knows Anthony in-person and has engaged in more high-bandwidth communication with him than anyone else on this thread, I am happy to stake social capital on his insights being very much worth listening to broadly speaking and that he's worth connecting to anyone who could give his ideas legs.  I have downvoted at least one comment  in this thread that I felt was not conducive to more of his ideas being externalized into the world due to what I believe to be unnecessary focus on social norms/tone policing over tracking object-level ideas. I am not responding further nor am I responding to particular comments as I don't want to feed the demon thread, but I do want to provide clarity on my judgement of what-is-in-the-right and also state I think Anthony could very possibly provide us Cause X as much as anyone I've seen.  To that end, I believe his interest in new/different infrastructure for how to communicate and internalize ideas is reasonable, and that it's unreasonable to expect idea providers to also have to be the idea executors in the ideal impact marketplace, especially to the extent of expecting them to engage in implicit politics more than is strictly necessary to get the ball rolling.

it's unreasonable to expect idea providers to also have to be the idea executors in the ideal impact marketplace

I could be wrong, but I think that most people think that the key bottleneck is "idea executors", not "idea providers". (E.g. I heard Charity Entrepreneurship has many intervention ideas, but even after extensive selection and training they are bottlenecked by finding enough founders).

So one shouldn't be surprised if they share a great idea but it doesn't get any traction, it seems to be the current state of things.

-1
Charles He
2y
I think it’s important that this actually involves staking social capital because I would otherwise find such revision of very negative behaviour based on what is clearly external friendship (as well as the mass upvoting) more problematic than anything else that has occurred. Imagine if everyone did this for their friends/enemies on the forum.
-2
Anthony Repetto
2y
Quote from above: "I'm NOT looking for any funding, either - there's a decent chance that the cost of the solution is lower than the Federal Government's increased tax-revenue from hurricane-prevention, so I say that the government should pay for it." Hopefully, you read this comment BEFORE saying something like "But EA shouldn't spend $1B on your idea" or "So you want us to fund this?" I've received numerous mis-representations, each insisting that I am somehow asking for EA money. You demonstrate how poorly you pay attention; I'll copy the quote again, in case anyone forgot by now: "I'm NOT looking for any funding, either - there's a decent chance that the cost of the solution is lower than the Federal Government's increased tax-revenue from hurricane-prevention, so I say that the government should pay for it." Why am I repeatedly addressing such an obvious and shallow misrepresentation? What is going on with these people?

I was responding to Jeff - and thank you, Jeff, for clarifying that downvotes can hide me.

In my response to him, I was expressing my concern that a subset of the Forum has the power to hide my self-defense, so that my correction of their misrepresentation goes unnoticed, while their misrepresentations stand in full view.

Another EA Forum post, just recently ("Bad Omens in Current Community Building") was trying to bring to the community's attention that, among other things, EA is sometimes perceived as cultish or cliquish. I hope you can all see that, when my correction of others' misrepresentations are downvoted to obscurity, then that concern of cliquishness is real.

3
Anthony Repetto
2y
I should also add this note: there is a double-standard in communication, here. I was asked repeatedly to 'calm down and speak nicely, because only then will we listen' - meanwhile, the ones who misrepresented were given a pass to lead the listener by the nose along imputations such as "because you posted a lot, no one is going to listen to you." They got that pass, easily, with the header "being brutally honest"/"honestly". Should I just begin all my posts with "just being brutally honest", so that no one uses my tone as a reason to ignore the content of what I say?

Hi Anthony. I would say that in the responses I’ve read where they use words like ‘honestly’, my reading of the tone was that they were going for a “tough love” approach. Using the word ‘honestly’ (when not said to manipulate people) often indicates the person is aware that what they’re saying may been seen as too harsh, but that they think what they’re saying is of enough value to others that it still merits saying (and sometimes may only have that value if said bluntly).

In contrast, my interpretation of the tone in your comments, using the word ‘disrespect’ a lot, asking for an apology etc, was that it was solely about providing value to yourself. For most people I know, the concept of feeling ‘disrespected’ by others, and going around demanding apologies for it, would never occur to them. Having that mindset is something I associate with arrogance, aggression and self-righteousness. I think in general people in this forum are wary of engaging further with people who appear to lack some level of humility.

Perhaps in certain circles it is expected that you ought to defend yourself in that way, in order to show that that what someone has said about you really is incorrect? But in the absence of social pressure in that direction, doing so suggests personality traits that some might be wary of.

3
Anthony Repetto
2y
[Jeff deleted his response, yet it was still helpful!]