Hide table of contents

@Aaron Bergman says 

I wish there was (and there should be) more discussion of “which singular organization is most deserving of money rn” Individual donors should make their best guess public and indicate openness to critique

twitter

So let's guess. What is your suggestion?

31

0
0

Reactions

0
0
New Answer
New Comment


3 Answers sorted by

In the AI governance space (specifically research; I'm not familiar with advocacy), I think GovAI and RP's AI Governance & Strategy [edit: now IAPS] team convert marginal money into research very well. I think this mostly because I like their teams and track records and (last I checked) they say they want to hire more but are funding-constrained.

But I wouldn't be surprised if some alignment or advocacy thing is predictably better.

I wouldn't be shocked to find it's a scrappy piece of anti AI advocacy like the AI pause work. Like maybe they get a big win and then it's been a few $k and very large impact. 

https://manifund.org/projects/holly-elmore-organizing-people-for-a-frontier-ai-moratorium 

Yes - if your timelines are short, then everything starts to look like it flows through a bottleneck of there actually still being a world in 2030 (2028, 2025..), which requires a global moratorium on AGI (as there is not enough time left for Alignment to be solved otherwise). There are a few people/orgs now working on this. Not sure which is the absolute best in terms of bang for buck though.

All of these are new (post-GPT-4): Centre for AI Policy, Artificial Intelligence Policy Institute, PauseAI, Stop AGI, Campaign for AI Safety, Holly Elmore, Safer AI, ... (read more)

Family Empowerment Media seems more cost-effective than GiveWell. They look especially promising from a cluster thinking approach.

  • Having control of their pregnancies also helps women according to a capabilities approach to welfare
  • Preventing unwanted pregnancies is a great opportunity in terms of animal welfare, the opposite of the meat eater problem
  • Is ~2x more cost effective as GiveWell's top charities according to Founders Pledge, possibly even higher according to Rethink Priorities
  • (speculative) There's a small possibility GiveWell might not want to fund them no matter their cost-effectiveness. GiveWell's mission seems to be to move as much money as possible to effective charities, not to donate the marginal dollar to the most cost-effective opportunity.
    GiveWell might worry that funding family planning would make it harder for them to get contributions from some very large donors, and hence hurt their mission. (This is very speculative and unlikely, but GiveWell has been in talks with FEM for more than a year, and hasn't published anything more than this, despite FEM having a clear funding gap. And I get the impression that many UHNWI are extremely pro-life and worried about declining fertility rates)

If you agree you can donate.

Hi!

Thanks for this comment. We are interested in potentially funding family planning programs, including Family Empowerment Media. Our research and grant decisions are independent of our outreach and fundraising plans. To clarify and add detail on some of the points above:

GiveWell recommended a $500,000 grant to Family Empowerment Media in March 2023. This funding would support a planned randomized controlled trial (RCT) of its program. We haven't yet published about this grant, but we plan to soon. (Our grant page publication often lags our funding decisions considerably, though this is something we're working to improve!)

We believe that FEM or other types of family planning–related programs may be very cost-effective—our current rough estimate for FEM's program (not the RCT of the program) puts it at 18-28 times as cost-effective as unconditional cash transfers—but we have a high degree of uncertainty about some of our inputs and expect that the results of FEM's RCT would help us address some of our uncertainties. We are also continuing to assess how to prioritize family planning programs among the many other promising programs we're investigating.

I hope that's helpful!

Best,
Miranda Kaplan
GiveWell Communications Associate

3
AnonymousTurtle
Hi Miranda, and apologies for writing semi-anonymously. That was very helpful and wonderful to hear! I am also very glad to hear that you keep your grant decisions independent of your outreach and fundraising plans, despite potential strong misaligned incentives, and I am relieved and immensely grateful that I was wrong to suspect otherwise. I'm confused as to why GiveWell is not filling the current funding gap of FEM's program, given that you estimate it to be more cost-effective than the programs you are currently funding, which are 13x cash. I imagine that's because you're less uncertain about the estimates for those programs? [I'm also curious as to why someone disagree-voted the above, if the voter is reading this I would find an explanation helpful.]
4
GiveWell
I’m glad you found this response helpful! We expect to wait until the results of FEM’s RCT are available before deciding whether to recommend funding for the program itself; we do have some remaining uncertainties, and it's possible that additional work could lead to significant changes in our cost-effectiveness estimate. We also plan to continue our own research on family planning programs in general, which may affect our funding decisions in this area. As our investigations progress, we’ll share updates on our website. Please feel free to reach out directly if you have questions in the meantime! Best, Roman Guglielmo Donor Relations Associate GiveWell
Comments2
Sorted by Click to highlight new comments since:

"Deserving" feels off-- how about something like "which org turns marginal $ into the most EV"?

Maybe this is a bad question to ask - it sort of implies that our inside view is good and useful. But maybe it isn't. I don't know.

Genuinely makes me feel a bit queasy thinking about this which seems like a bad sign.

Curated and popular this week
 ·  · 1m read
 · 
> Summary: We propose measuring AI performance in terms of the length of tasks AI agents can complete. We show that this metric has been consistently exponentially increasing over the past 6 years, with a doubling time of around 7 months. Extrapolating this trend predicts that, in under a decade, we will see AI agents that can independently complete a large fraction of software tasks that currently take humans days or weeks. > > The length of tasks (measured by how long they take human professionals) that generalist frontier model agents can complete autonomously with 50% reliability has been doubling approximately every 7 months for the last 6 years. The shaded region represents 95% CI calculated by hierarchical bootstrap over task families, tasks, and task attempts. > > Full paper | Github repo Blogpost; tweet thread. 
 ·  · 2m read
 · 
For immediate release: April 1, 2025 OXFORD, UK — The Centre for Effective Altruism (CEA) announced today that it will no longer identify as an "Effective Altruism" organization.  "After careful consideration, we've determined that the most effective way to have a positive impact is to deny any association with Effective Altruism," said a CEA spokesperson. "Our mission remains unchanged: to use reason and evidence to do the most good. Which coincidentally was the definition of EA." The announcement mirrors a pattern of other organizations that have grown with EA support and frameworks and eventually distanced themselves from EA. CEA's statement clarified that it will continue to use the same methodologies, maintain the same team, and pursue identical goals. "We've found that not being associated with the movement we have spent years building gives us more flexibility to do exactly what we were already doing, just with better PR," the spokesperson explained. "It's like keeping all the benefits of a community while refusing to contribute to its future development or taking responsibility for its challenges. Win-win!" In a related announcement, CEA revealed plans to rename its annual EA Global conference to "Coincidental Gathering of Like-Minded Individuals Who Mysteriously All Know Each Other But Definitely Aren't Part of Any Specific Movement Conference 2025." When asked about concerns that this trend might be pulling up the ladder for future projects that also might benefit from the infrastructure of the effective altruist community, the spokesperson adjusted their "I Heart Consequentialism" tie and replied, "Future projects? I'm sorry, but focusing on long-term movement building would be very EA of us, and as we've clearly established, we're not that anymore." Industry analysts predict that by 2026, the only entities still identifying as "EA" will be three post-rationalist bloggers, a Discord server full of undergraduate philosophy majors, and one person at
 ·  · 2m read
 · 
Epistemic status: highly certain, or something The Spending What We Must 💸11% pledge  In short: Members pledge to spend at least 11% of their income on effectively increasing their own productivity. This pledge is likely higher-impact for most people than the Giving What We Can 🔸10% Pledge, and we also think the name accurately reflects the non-supererogatory moral beliefs of many in the EA community. Example Charlie is a software engineer for the Centre for Effective Future Research. Since Charlie has taken the SWWM 💸11% pledge, rather than splurge on a vacation, they decide to buy an expensive noise-canceling headset before their next EAG, allowing them to get slightly more sleep and have 104 one-on-one meetings instead of just 101. In one of the extra three meetings, they chat with Diana, who is starting an AI-for-worrying-about-AI company, and decide to become a cofounder. The company becomes wildly successful, and Charlie's equity share allows them to further increase their productivity to the point of diminishing marginal returns, then donate $50 billion to SWWM. The 💸💸💸 Badge If you've taken the SWWM 💸11% Pledge, we'd appreciate if you could add three 💸💸💸 "stacks of money with wings" emoji to your social media profiles. We chose three emoji because we think the 💸11% Pledge will be about 3x more effective than the 🔸10% pledge (see FAQ), and EAs should be scope sensitive.  FAQ Is the pledge legally binding? We highly recommend signing the legal contract, as it will allow you to sue yourself in case of delinquency. What do you mean by effectively increasing productivity? Some interventions are especially good at transforming self-donations into productivity, and have a strong evidence base. In particular:  * Offloading non-work duties like dates and calling your mother to personal assistants * Running many emulated copies of oneself (likely available soon) * Amphetamines I'm an AI system. Can I take the 💸11% pledge? We encourage A