Feedback welcome: www.admonymous.co/mo-putera
I currently work with CE/AIM-incubated charity ARMoR on research distillation, quantitative modelling, consulting, and general org-boosting to support policies that incentivise innovation and ensure access to antibiotics to help combat AMR. I was previously an AIM Research Program fellow, was supported by a FTX Future Fund regrant and later Open Philanthropy's affected grantees program, and before that I spent 6 years doing data analytics, business intelligence and knowledge + project management in various industries (airlines, e-commerce) and departments (commercial, marketing), after majoring in physics at UCLA and changing my mind about becoming a physicist. I've also initiated some local priorities research efforts, e.g. a charity evaluation initiative with the moonshot aim of reorienting my home country Malaysia's giving landscape towards effectiveness, albeit with mixed results.
I first learned about effective altruism circa 2014 via A Modest Proposal, Scott Alexander's polemic on using dead children as units of currency to force readers to grapple with the opportunity costs of subpar resource allocation under triage. I have never stopped thinking about it since, although my relationship to it has changed quite a bit; I related to Tyler's personal story (which unsurprisingly also references A Modest Proposal as a life-changing polemic):
I thought my own story might be more relatable for friends with a history of devotion – unusual people who’ve found themselves dedicating their lives to a particular moral vision, whether it was (or is) Buddhism, Christianity, social justice, or climate activism. When these visions gobble up all other meaning in the life of their devotees, well, that sucks. I go through my own history of devotion to effective altruism. It’s the story of [wanting to help] turning into [needing to help] turning into [living to help] turning into [wanting to die] turning into [wanting to help again, because helping is part of a rich life].
Strong upvote for multiple reasons: thoroughness and transparency of reasoning and execution, stating upfront that the cost-effectiveness was well below WHO's threshold, prioritising reproducibility, and the "what this analysis doesn't show" and "lessons learned" sections.
I wonder if "de-averaging the portfolio" by ballparking cost per impact by channel (school awareness packages vs workshops vs crisis hotline) might help guide resource allocation across channels (send more packets vs get more hotline volunteers etc). My naive guess is that suicide risk isn't the same across channels, which the model as it stands implicitly assumes (which makes school awareness get ~10x as much impact credit as the crisis hotline); I'd assume that this risk for the school students is similar to that of the general population but that hotline callers are self-selected for being at much higher risk of suicide, so I'd explore the hypothesis that most of the bottomline suicide prevention impact comes from the crisis hotline even though most of the topline reach comes from the school awareness packages. This is also a channel attribution question, which as you've said the current model doesn't show and is a hard one to answer.
I am in the process of building such a thing
Is it available online / open-source / etc by any chance? Even just as a butterfly idea.
Great rule of thumb :) I'm sometimes knee-deep in chartmaking before I realise I don't actually know exactly what I want to communicate.
Tangentially reminded me of Eugene Wei's suggestion to "remove the legend", in an essay that also attempted to illustrate how to implement Ed Tufte's advice from his cult bestseller The Visual Display of Quantitative Information.
I'd also like to signal-boost the excellent chart guides from storytelling with data.
Aside: wow, the slide presentation you linked to above is both really aesthetically pleasing and has great content, thanks for sharing :)
I think you're conflating intervention durability with outcome durability? A child who survives cerebral malaria due to seasonal malaria chemoprevention gets to live the rest of their life; SMC programs are rerun because (mostly) new beneficiary cohorts are at highest risk, not because last year's cohort's survival expires somehow. Similarly with nets and child vaccinations and vitamin A deficiency prevention (i.e. the GW top charities), as well as salt iodisation and TaRL in education and many other top interventions recommended by the likes of TLYCS and FP and so on.
I'd also push back a bit on the "permanent solutions" phrasing. Infrastructure isn't that permanent and requires ongoing expenditures and has a shelf half-life (I used to work in ops in fluid resource-constrained environments so I feel this keenly), diseases can develop resistance to vaccines so you need boosters, etc. Ex-AIM CEO Joey Savoie has a great blog talking more about how Someone Always Pays: Why Nothing Is Really "Sustainable".
Phrasing nitpicking aside, some big funders are in fact funding more "permanent / sustainable" solutions. Open Phil Coefficient Giving's new $120M Abundance and Growth Fund aims to "accelerate economic growth and boost scientific and technological progress while lowering the cost of living", and Founders Pledge (which is almost OP-scale in giving) just launched a new Catalytic Impact Fund that targets "ecosystem leverage points" where small investments can build "sustainable, long-term solutions to global poverty and suffering".
Jason's comment above on timetable speedup is essentially how e.g. GiveWell models their grants for malaria vaccines. The model says their grant would need to speed up co-administration for all highest need children in all of subsaharan Africa by at least 9 months to clear their 10x bar, so you can interpret their grant as a bet that funding that clinical trial would in fact achieve at least 9 months speedup. Notice how it's an uncertain bet; I think most donors (weighted by dollars moved) care quite a fair bit about certainty of direct benefits, so they'd probably donate to e.g. the Top Charities Fund instead of the more experimental EV-maxxing All Grants Fund.
I admire influential orgs that publicly change their mind due to external feedback, and GiveWell is as usual exemplary of this (see also their grant "lookbacks"). From their recently published Progress on Issues We Identified During Top Charities Red Teaming, here's how external feedback changed their bottomline grantmaking:
In 2023, we conducted “red teaming” to critically examine our four top charities. We found several issues: 4 mistakes and 10 areas requiring more work. We thought these could significantly affect our 2024 grants: $5m-$40m in grants we wouldn’t have made otherwise and $5m-$40m less in grants we would have made otherwise (out of ~$325m total).
This report looks back at how addressing these issues changed our actual grantmaking decisions in 2024. Our rough estimate is that red teaming led to ~$37m in grants we wouldn't have made otherwise and prevented ~$20m in grants we would have made otherwise, out of ~$340m total grants. The biggest driver was incorporating multiple sources for disease burden data rather than relying on single sources.1 There were also several cases where updates did not change grant decisions but led to meaningful changes in our research.
Some self-assessed progress that caught my eye — incomplete list, full one here; these "led to important errors or... worsened the credibility of our research" (0 = no progress made, 10 = completely resolved):
- Failure to engage with outside experts (8/10): We spent 240 days at conferences/site visits in 2024 (vs. 60 in 2023). We think this type of external engagement helped us avoid ~$4m in grants and identify new grant opportunities like Uduma water utility ($480,000). We've established ongoing relationships with field experts. (more)
- Failure to check burden data against multiple sources (8/10): By using multiple data sources for disease burden, we made ~$34m in grants we likely wouldn't have otherwise and declined ~$14m in grants we probably would have made. We've implemented comprehensive guidelines for triangulating data sources. (more)
- Failure to account for individuals receiving interventions from other sources (7/10): We were underestimating how many people would get nets without our campaigns, reducing cost-effectiveness by 20-25%. We've updated our models but have made limited progress on exploring routine distribution systems (continuous distribution through existing health channels) as an alternative or complement to our mass campaigns. (more)
- Failure to estimate interactions between programs (7/10): We adjusted our vitamin A model to account for overlap with azithromycin distribution (reducing effectiveness by ~15%) and accounted for malaria vaccine coverage when estimating nets impact. We've developed a framework to systematically address this. (more)
(As an aside, I've noticed plenty of claims of GW top charity-beating cost-effectiveness figures both on the forum and elsewhere, and I basically never give them the credence I'd give to GW's own estimates, due to the kind of (usually downward) adjustments mentioned above like receiving interventions from other sources or between-program interventions, and GW's sheer reasoning thoroughness behind those adjustments, seriously, click on any of those "(more)"s)
Some other issues they'd "been aware of at the time of red teaming and had deprioritized but that we thought were worth looking into following red teaming" — again incomplete list, full one here:
- Insufficient attention to inconsistency across cost-effectiveness analyses (CEAs) (8/10): We made our estimates of long-term income effects of preventive health programs more consistent (now 20-30% of benefits across top charities vs. previously 10-40%) and fixed implausible assumptions on indirect deaths (deaths prevented, e.g., by malaria prevention that aren’t attributed to malaria on cause-of-death data). We've implemented regular consistency checks. (more)
- Insufficient attention to some fundamental drivers of intervention efficacy (7/10): We updated our assumptions about net durability and chemical decay on nets (each changing cost-effectiveness by -5% and 11% across geographies) and consulted experts about vaccine efficacy concerns, but we haven't systematically addressed monitoring intervention efficacy drivers across programs. (more)
- Insufficient sideways checks on coverage, costs, and program impact (7/10): We funded $900,000 for external surveys of Evidence Action's water programs, incorporated additional DHS data in our models, and added other verification methods. We've made this a standard part of our process but think there are other areas where we’d benefit from additional verification of program metrics. (more)
- Insufficient follow-up on potentially concerning monitoring and costing data (7/10): We’ve encouraged Helen Keller to improve its monitoring (now requiring independent checks of 10% of households), verified AMF's data systems have improved, and published our first program lookbacks. However, we still think there are important gaps. (more)
I always had the impression GW engaged outside experts a fair bit, so I was pleasantly surprised to learn they thought they weren't doing enough of it and then actually followed through so seriously, this is an A+ example of organisational commitment to and follow-through on self-improvement so I'd like to quote this section in full:
In 2024, we spent ~240 days at conferences or site visits, compared to ~60 in 2023. We spoke to experts more regularly as part of grant investigations, and tried a few new approaches to getting external feedback. While it’s tough to establish impact, we think this led to four smaller grants we might not have made otherwise (totalling ~$1 million) and led us to deprioritize a ~$10 million grant we might’ve made otherwise.
More detail on what we said we’d do to address this issue and what we found (text in italics is drawn from our original report):
- More regularly attend conferences with experts in areas in which we fund programs (malaria, vaccination, etc.).
- In 2024, our research team attended 16 conferences, or ~140 days, compared to ~40 days at conferences in 2023.35
- We think these conferences helped us build relationships with experts and identify new grant opportunities. Two examples:
- A conversation with another funder at a conference led us to re-evaluate our assumptions on HPV coverage and ultimately deprioritize a roughly $10 million grant we may have made otherwise.36
- We learned about Uduma, a for-profit rural water utility, at a conference and made a $480,000 grant to them in November 2024.37
- We also made more site visits. In 2023, we spent approximately 20 days on site visits. In 2024, the number was approximately 100 days.38
- Reach out to experts more regularly as part of grant investigations and intervention research. We’ve always consulted with program implementers, researchers, and others through the course of our work, but we think we should allocate more relative time to conversations over desk research in most cases.
- Our research team has allocated more time to expert conversations. A few examples:
- Our 2024 grants for VAS to Helen Keller International relied significantly on conversations with program experts. Excluding conversations with the grantee, we had 15 external conversations.
- We’ve set up longer-term contracts with individuals who provide us regular feedback. For example, our water and livelihoods team has engaged Daniele Lantagne and Paul Gunstensen for input on grant opportunities and external review of our research.
- We spoke with other implementers about programs we’re considering. For example, we discussed our 2024 grant to support PATH’s technical assistance to support the rollout of malaria vaccines with external stakeholders in the space.39
- This led to learning about some new grant opportunities. For example:
- The $150,000 grant to International Rescue Committee (IRC) for desk-based scoping of programs to increase vaccination coverage stemmed from a conversation with an expert.
- We connected with Zipline at a conference, and made a $55,000 grant to them for desk-based scoping of drones to increase vaccination coverage in December 2024.
- We are currently considering a $4 million grant that we learned about through an expert conversation.40
- Experiment with new approaches for getting feedback on our work.
- In addition to the above, we tried a few other approaches we hadn’t (or hadn’t extensively) used before. Three examples:
- Following our red teaming of GiveWell’s top charities, we decided to review our iron grantmaking to understand what were the top research questions we should address as we consider making additional grants in the near future. We had three experts review our work in parallel to internal red teaming, so we could get input and ask questions along the way.41 We did not do this during our top charities red teaming, in the report of which we wrote “we had limited back-and-forth with external experts during the red teaming process, and we think more engagement with individuals outside of GiveWell could improve the process.”
- We made a grant to Busara to collect qualitative information on our grants to Helen Keller International's vitamin A supplementation program in Nigeria.42
- We funded the Center for Global Development to understand why highly cost-effective GiveWell programs aren’t funded by other groups focused on saving lives. This evaluation was designed to get external scrutiny from an organization with expertise in global health and development, and by other funders and decision-makers in low- and middle-income countries.
Some quick reactions:
Ben Kuhn has a great essay about how
all my favorite people are great at a skill I’ve labeled in my head as “staring into the abyss.”1
Staring into the abyss means thinking reasonably about things that are uncomfortable to contemplate, like arguments against your religious beliefs, or in favor of breaking up with your partner. It’s common to procrastinate on thinking hard about these things because it might require you to acknowledge that you were very wrong about something in the past, and perhaps wasted a bunch of time based on that (e.g. dating the wrong person or praying to the wrong god). However, in most cases you have to either admit this eventually or, if you never admit it, lock yourself into a sub-optimal future life trajectory, so it’s best to be impatient and stare directly into the uncomfortable topic until you’ve figured out what to do. ...
I noticed that it wasn’t just Drew (cofounder and CEO of Wave) who is great at this, but many the people whose work I respect the most, or who have had the most impact on how I think. Conversely, I also noticed that for many of the people I know who have struggled to make good high-level life decisions, they were at least partly blocked by having an abyss that they needed to stare into, but flinched away from.
So I’ve come to believe that becoming more willing to stare into the abyss is one of the most important things you can do to become a better thinker and make better decisions about how to spend your life.
I agree, and I think there's an organisational analogue as well, which GiveWell exemplifies above.
CE/AIM-incubated orgs run lean. (Some past discussion here if you're interested.) I also don't live in a high CoL country, which helps.
Reread Patrick McKenzie (patio11)'s inspirational oral history of VaccinateCA and thought to pull out a few quotes for my own edification. (Patrick posted about this on the forum awhile back, that's worth reading too.)
The following is what it looks like to bake in triage into org decision-making from the top down:
I think in absolute terms plenty of orgs do this, Patrick just so happens to be a good writer. But in relative terms it's quite rare, and very meaningful to see, especially for folks like me with a bit of mission orientation. Also this:
On entrepreneurship:
This is somewhat reminiscent of what Scott Alexander wrote about a very different person, although what Patrick calls "carefully titrating the amount of truth to various parties" Scott outright labeled "blatant lies"; my takeaway is that it's possible to do a more ethical version of the description below:
But I digress. Relatedly:
On do-gooder precocity:
On ownership and accountability, a case study:
On how much of Patrick's job in the early days as CEO was bringing in funding:
On the advantages private individuals and organizations have over official initiatives:
(It's hard to convey how much I like and appreciate that last paragraph.)
On funder-nonprofit misalignment:
(I'll be upfront that despite having spent a couple of my formative years in California, my bias leans so far in Patrick's direction that it'd probably be useful for me to hear out the strongest counterargument, especially in the context of triage.)