This post was prompted by three other posts on the EA forum. A recent post raises the alarm about abuse of power and relationships in the EA community. An earlier post suggested that the EA community welcomes shallow critiques but is less receptive to deep critiques. Dustin Moskovitz has recently mentioned that the EA forum functions as a conflict of interest appeals board for Open Philanthropy. Yet on the EA forum, there doesn’t seem to be many specific criticisms of Open Philanthropy.

 

For the sake of epistemics, I wanted to create this post and invite individuals to voice any issues they may have with Open Philanthropy or propose potential solutions.

 

I’ll start by discussing the funding dynamics within the field of technical alignment (alignment theory, applied alignment), with a particular focus on Open Philanthropy.

 

In the past two years, the technical alignment organisations which have received substantial funding include:

All of these organisations are situated in the San Francisco Bay Area. Although many people are thinking about the alignment problem, there is much less funding for technical alignment researchers for other locations (e.g., the east coast of the US, the UK, or other parts of Europe).

 

This collectively indicates that, all else being equal, having strong or intimate connections with employees of Open Philanthropy greatly enhances the chances of having funding, and it seems almost necessary. As a concerned EA, this seems incredibly alarming and in need of significant reform. Residency in the San Francisco Bay Area is also a must. A skeptical perspective would be that Open Philanthropy allocates its resources to those with the most political access. Since it's hard to solve the alignment problem, the only people grantmakers end up trusting to do so are those who are very close to them.

 

This is a problem with Open Philanthropy’s design and processes, and points to the biases of the technical alignment grantmakers and decision makers. This seems almost inevitable given (1) community norms around conflicts of interest and (2) Open Philanthropy’s strong centralization of power. This is not to say that any specific individual is to blame. Instead processes, structure, and norms are more useful to direct reforms towards.

 

Right now, even if a highly respected alignment researcher thinks what you do is extremely valuable, the decision ultimately can be blocked by an Open Philanthropy grantmaker, which could cause people to leave alignment altogether.

One common suggestion involves making the grantmaking process more democratic or less centralised. For example, the “regranting” approach has been successful for other grantmakers. This involves selecting a large pool of grantmakers or regrantors who have the autonomy to make their own decisions. With more grantmakers, there is less potential for Goodharting by individuals and reduces the likelihood of funding only those who are known best by a few Open Philanthropy staff. Additionally, Open Philanthropy can still choose regrantors who are more aligned with EA values or have previously demonstrated good judgement. A smaller thing that could help is indicating what proportion of funding in each specific area goes to organisations where there are, or have been, intimate relationships between organisation leadership and Open Philanthropy grantmakers.

 

It seems vital for community health to be able to speak about these sorts of issues openly. What criticisms of Open Philanthropy are important to share, or what are other ways they should improve?

85

0
0

Reactions

0
0

More posts like this

Comments8


Sorted by Click to highlight new comments since:
TW123
77
23
1

In the past two years, the technical alignment organisations which have received substantial funding include

Your post does not actually say this, but when I read it I thought you were saying that these are all the organizations that have received major funding in technical alignment. I think it would have been clearer if you had said "include the following organizations based in the San Francisco Bay Area:" to make it clearer you're discussing a subset.

Anyway, here are the public numbers, for those curious, of $1 million+ grants in technical AI safety in 2021 and 2022 (ordered by total size) made by Open Philanthropy:

  • Redwood Research: $9.4 million, and then another grant for $10.7 million
  • Many professors at a lot of universities: $14.4 million
  • CHAI: $11.3 million
  • Aleksander Madry at MIT: $1.4 million
  • Hofvarpnir Studios: $1.4 million
  • Berkeley Existential Risk Initiative - CHAI collaboration: $1.1 million
  • Berkeley Existential Risk Initiative - SERI MATS Program: $1 million

The Alignment Research Center received much less: $265,000.

There isn't actually any public grant saying that Open Phil funded Anthropic. However, that isn't to say that they couldn't have made a non-public grant. It was public that FTX funded Anthropic.

having strong or intimate connections with employees of Open Philanthropy greatly enhances the chances of having funding, and it seems almost necessary

Based on spending some time in Berkeley, I think a more accurate way to describe this is as follows:

People who care about AI safety and are involved in EA tend to move to Berkeley because that is where everyone else is. It really can increase your productivity if you can easily interact with others working in your field and know what is going on, or so the established wisdom goes. The people who have been around the longest are often leading research organizations or are grantmakers at Open Phil. They go to the same parties, have the same friends, work in the same offices, and often spend nearly all of their time working with little time to socialize with anyone outside their community. Unless they make a special effort to avoid dating anyone in their social community, they may end up dating a grantmaker.

If we want these conflicts of interest to go away, we could try simply saying it should be a norm for Open Phil not to grant to organizations with possible conflicts of interest. But knowing the Berkeley social scene, this means that many Open Phil grantmakers wouldn't be able to date anyone in their social circles, since basically everyone in their social circles is receiving money from Open Phil.

The real question is as you say one of structure: whether so many of the EA-aligned AI safety organizations should be headquartered in close proximity and whether EAs should live together and be friends with basically only other EAs. That's the dynamic that created the conflicts. I don't think the answer to this is extremely obvious, but I don't really feel like trying to argue both sides of it right now.

It's possibly true that regrantors would reduce this effect in grantmaking, because you could designate regrantors in other places or who have different friends. But my suspicion would be that regrantors would by default be the same people who are already receiving grants.

There isn't actually any public grant saying that Open Phil funded Anthropic

I was looking into this topic, and found this source:

Anthropic has raised a $124 million Series A led by Stripe co-founder Jaan Tallinn, with participation from James McClave, Dustin Moskovitz, the Center for Emerging Risk Research and Eric Schmidt. The company is a developer of AI systems.

Speculating, conditional on the pitchbook data being correct, I don't think that Moskovitz funded Anthropic because his object level beliefs about their value or because they're such good pals, rather I'm guessing he received a recommendation from Open Philanthropy, even if Open Philanthropy wasn't the vehicle he used to transfer the funds.

Also note that Luke Muehlhauser is part of the board of Anthropic; see footnote 4 here.

My impression is that the lag between "idea is common knowledge at OpenPhil" and "idea is public" is fairly long. I was hoping that hiring more communications staff would help with this, and maybe it has, but the problem still seems to exist.

As evidence that sharing information is valuable: 6 of the 20 most valuable Forum posts from 2022 were from OpenPhil authors, and Holden is the only author to have multiple posts in the top 20. (I think my personal list of most valuable posts is even more skewed towards OpenPhil authors.)

I understand the reasons why it's costly to do public writeups (ironically, Holden's post on public discourse is one of the better references), and I'm not confident that people are actually making a mistake here,[1] but it does seem plausible that this is being undervalued.

 

  1. ^

    Also I'm biased in favor of liking stuff that's on the Forum

In the past two years, the technical alignment organisations which have received substantial funding include:

In context it sounds like you're saying that Open Phil funded Anthropic, but as far as I am aware that is simply not true.

I think maybe what you meant to say is that, "These orgs that have gotten substantial funding tend to have ties to Open Phil, whether OP was the funder or not." Might be worth editing the post to make that more explicit, so it's clear whether you're alleging a conflict of interest or not.

"having strong or intimate connections with employees of Open Philanthropy greatly enhances the chances of having funding, and it seems almost necessary"

Is this a well-identified phenomenon (in the causal inference sense) ?

Consider the following directed acyclic graph:

Connected with OpenPhil employees  ----------> Gets funding from OpenPhil

                                            ^                                                                                                        ^
                                            |                                                                                                          |
                                            |                                                                                                          |
                                            |                                                                                                          |
                                            |                                                                                                          |
                                            +------      Works on alignment      ------+

One explanation for this correlation you identify is that being connected with OpenPhil employees leads to people getting funding from OpenPhil (as demonstrated by the horizontal arrow). However, another explanation is that working on alignment causes one to connect with others who are interested  in the problem of AI alignment, as well as getting funding from a philanthropic organisation which funds work on AI alignment (as demonstrated by the vertical arrows).

These two explanations are observationally equivalent, in the absence of exogenous variation with respect to how connected one is to OpenPhil employees. Since claiming that it is "almost necessary" to have "strong or intimate connections with employees of Open Philanthropy" to get funding implies wrongdoing from OpenPhil, I'd be interested in evidence which would demonstrate this!

Thanks for writing this!

I think Open Phil should clarify how they allocate resources between different cause areas. The high level approach is worldview diversification, but I would like to know more about the specifics. For example, is some sort of softmax approach being used? Why, or why not?

It also looks like Open Phil is directing too much resources towards GiveWell's (GW's) top charities, given the uncertainties about its near term effects on animals, and longterm effects (see here). Even if Open Phil has funded Rethink Priorities to do research on moral weights (see here), which is quite relevant to get a coherent picture about the near term effects of GW's top charities, it seems this research should have been done earlier, and before significant funding being moved towards GW.

Have Open Phil considered relocating to move away from the Bay Area? Would this reduce the incidence of tangled relationships and shared offices between grantmakers and grantees? Would this be good (reducing bias in funding decisions) or bad (by making grantmakers less well-informed)? They should really publish a cost-benefit analysis of these topics you're bringing up.

I know that a few Open Phil staff live outside the Bay Area and work remotely

Curated and popular this week
LintzA
 ·  · 15m read
 · 
Cross-posted to Lesswrong Introduction Several developments over the past few months should cause you to re-evaluate what you are doing. These include: 1. Updates toward short timelines 2. The Trump presidency 3. The o1 (inference-time compute scaling) paradigm 4. Deepseek 5. Stargate/AI datacenter spending 6. Increased internal deployment 7. Absence of AI x-risk/safety considerations in mainstream AI discourse Taken together, these are enough to render many existing AI governance strategies obsolete (and probably some technical safety strategies too). There's a good chance we're entering crunch time and that should absolutely affect your theory of change and what you plan to work on. In this piece I try to give a quick summary of these developments and think through the broader implications these have for AI safety. At the end of the piece I give some quick initial thoughts on how these developments affect what safety-concerned folks should be prioritizing. These are early days and I expect many of my takes will shift, look forward to discussing in the comments!  Implications of recent developments Updates toward short timelines There’s general agreement that timelines are likely to be far shorter than most expected. Both Sam Altman and Dario Amodei have recently said they expect AGI within the next 3 years. Anecdotally, nearly everyone I know or have heard of who was expecting longer timelines has updated significantly toward short timelines (<5 years). E.g. Ajeya’s median estimate is that 99% of fully-remote jobs will be automatable in roughly 6-8 years, 5+ years earlier than her 2023 estimate. On a quick look, prediction markets seem to have shifted to short timelines (e.g. Metaculus[1] & Manifold appear to have roughly 2030 median timelines to AGI, though haven’t moved dramatically in recent months). We’ve consistently seen performance on benchmarks far exceed what most predicted. Most recently, Epoch was surprised to see OpenAI’s o3 model achi
Dr Kassim
 ·  · 4m read
 · 
Hey everyone, I’ve been going through the EA Introductory Program, and I have to admit some of these ideas make sense, but others leave me with more questions than answers. I’m trying to wrap my head around certain core EA principles, and the more I think about them, the more I wonder: Am I misunderstanding, or are there blind spots in EA’s approach? I’d really love to hear what others think. Maybe you can help me clarify some of my doubts. Or maybe you share the same reservations? Let’s talk. Cause Prioritization. Does It Ignore Political and Social Reality? EA focuses on doing the most good per dollar, which makes sense in theory. But does it hold up when you apply it to real world contexts especially in countries like Uganda? Take malaria prevention. It’s a top EA cause because it’s highly cost effective $5,000 can save a life through bed nets (GiveWell, 2023). But what happens when government corruption or instability disrupts these programs? The Global Fund scandal in Uganda saw $1.6 million in malaria aid mismanaged (Global Fund Audit Report, 2016). If money isn’t reaching the people it’s meant to help, is it really the best use of resources? And what about leadership changes? Policies shift unpredictably here. A national animal welfare initiative I supported lost momentum when political priorities changed. How does EA factor in these uncertainties when prioritizing causes? It feels like EA assumes a stable world where money always achieves the intended impact. But what if that’s not the world we live in? Long termism. A Luxury When the Present Is in Crisis? I get why long termists argue that future people matter. But should we really prioritize them over people suffering today? Long termism tells us that existential risks like AI could wipe out trillions of future lives. But in Uganda, we’re losing lives now—1,500+ die from rabies annually (WHO, 2021), and 41% of children suffer from stunting due to malnutrition (UNICEF, 2022). These are preventable d
Rory Fenton
 ·  · 6m read
 · 
Cross-posted from my blog. Contrary to my carefully crafted brand as a weak nerd, I go to a local CrossFit gym a few times a week. Every year, the gym raises funds for a scholarship for teens from lower-income families to attend their summer camp program. I don’t know how many Crossfit-interested low-income teens there are in my small town, but I’ll guess there are perhaps 2 of them who would benefit from the scholarship. After all, CrossFit is pretty niche, and the town is small. Helping youngsters get swole in the Pacific Northwest is not exactly as cost-effective as preventing malaria in Malawi. But I notice I feel drawn to supporting the scholarship anyway. Every time it pops in my head I think, “My money could fully solve this problem”. The camp only costs a few hundred dollars per kid and if there are just 2 kids who need support, I could give $500 and there would no longer be teenagers in my town who want to go to a CrossFit summer camp but can’t. Thanks to me, the hero, this problem would be entirely solved. 100%. That is not how most nonprofit work feels to me. You are only ever making small dents in important problems I want to work on big problems. Global poverty. Malaria. Everyone not suddenly dying. But if I’m honest, what I really want is to solve those problems. Me, personally, solve them. This is a continued source of frustration and sadness because I absolutely cannot solve those problems. Consider what else my $500 CrossFit scholarship might do: * I want to save lives, and USAID suddenly stops giving $7 billion a year to PEPFAR. So I give $500 to the Rapid Response Fund. My donation solves 0.000001% of the problem and I feel like I have failed. * I want to solve climate change, and getting to net zero will require stopping or removing emissions of 1,500 billion tons of carbon dioxide. I give $500 to a policy nonprofit that reduces emissions, in expectation, by 50 tons. My donation solves 0.000000003% of the problem and I feel like I have f