There is a good and widely accepted approach to assessing testable projects - roughly what GiveWell does.  It is much less clear how EA research organisations should assess projects, interventions and organisations with very uncertain non-testable impact, such as policy work or academic research. There are some disparate materials on this question on blogs, Open Phil's website, on 80k's website, in the academic/grey literature etc. However, this information is not centralised; it's not clear what the points of agreement and disagreement are; lots of the organisations who will have thought about this question will have insights that have not been shared with the community (e.g. maybe CSER, FHI?); and the mechanisms for sharing relevant information in the future are unclear. 

Ultimately, it would be good to collate and curate all the best material on this, so that EA researchers at separate EA orgs would have easy access to it and would not have to approach this question on their own. As a first step, we invite people who have thought about this question to discuss their insights in the comments to this post. Topics could include:

  • How far should we use quantified models?
    • e.g. The Oxford Prioritisation Project used quantified models to assess really uncertain things like 80k and MIRI. 
    • Open Phil doesn't appear to do this (they don't mention that often in their public facing docs.)
  • What role should the Importance/Neglected/Tractable framework play?
    • Should it be used to choose between interventions and/or causes?
    • Should quantitative models be instead of ITN?
    • How quantified should the ITN framework be? As quantified as 80k's? More intuitive?
  • What are the key takeaways from the history of philanthropy, and the history of scientific research?
  • What's the best way to assess historical impact?
    • Process tracing or something like it?
    • What are the main biases at play in assessing historical impact?
    • Who do you ask ?
  • Is hits-based giving the right approach and what follows from it?
    • How relevant is track record, on this approach? Sometimes Open Phil takes account of track record, other times not. 
    • Should we favour choosing a cause area and then making lots of bets, or should we be more discerning?
  • What are the most important considerations for assessing charities doing uncertain-return stuff?
    • Strength of team
    • Current strategy
    • Potential to crowd in funding. 
  • What are the best theories of how to bring about political change?
  • How much weight should we put on short to medium-term tractability?
    • Given the nonlinear nature of e.g. political change, current tractability may not be the best guide. 
  • Are there any disciplines we could learn from?
    • Intelligence analysis.
    • Insurance (especially catastrophe insurance). 

 

Thanks, John and Marinella @ Founders Pledge. 

18

0
0

Reactions

0
0
Comments23


Sorted by Click to highlight new comments since:

In academic research, government and foundation grants are often awarded using criteria similar to ITN, except:

1) 'importance' is usually taken as short-term importance to the research field, and/or to one country's current human inhabitants (especially registered voters),

2) 'tractability' is interpreted as potential to yield several journal publications, rather than potential to solve real-world problems,

3) 'neglectedness' is interpreted as addressing a problem that's already been considered in only 5-20 previous journal papers, rather than one that's totally off the radar.

I would love to see academia in general adopt a more EA perspective on how to allocate scarce resources -- not just when addressing problems of human & animal welfare and X-risk, but in addressing any problem.

Another big area you didn't mention is Superforecasting, prediction markets and that kind of thing.

[anonymous]0
0
0

good shout - does anyone have any thoughts on this that aren't well-known or disagree with Tetlock?

This isn't a unique thought, but I just want to make sure the EA community knows about Gnosis and Augur, decentralized prediction markets built on Ethereum.

https://gnosis.pm/

https://augur.net/

Here's a link to the talk for those without Facebook: https://www.youtube.com/watch?v=67oL0ANDh5Y (go to 21:00)

Here's a written version by 80,000 Hours: https://80000hours.org/articles/problem-framework/

This is sort of a meta-comment, but there's loads of important stuff here, each of which could have it's own thread. Could I suggest someone (else), organises a (small) conference to discuss some of these things?

I've got quite a few things to add on the ITN framework but nothing I can say in a few words. Relatedly, I've also been working on a method for 'cause search' - a ways of finding all the big causes in a given domain - which is the step before cause prio, but that's not something I can write out succinctly either (yet, anyway).

I have organized retreat/conferency things before and would probably be up for organizing something like this if there was interest. I can contact some people and see if they think it would be worthwhile, I am not sure what I expect attendance to be like though (would 20 people be optimistic?)

I have a lot of thoughts on cause search, but possibly at a more granular level. One of the big challenges when you switch from an assessing to a generating perspective is finding the right problems to work on, and it's not easy at all.

Concur. If only we were meeting in the pub in, say, 1 hour to discuss this...

I think there's one happening in London in November that is discussing questions of this nature - it may be worth seeing if they will add it to the schedule if it's not already there.

[anonymous]0
0
0

I think splitting off these questions would balkanise things too much, making it harder for people interested in this general question to get relevant information.

Open Phil doesn't appear to do this (they don't mention that often in their public facing docs.)

They do, though it's usually not the first thing they mention e.g. here: https://docs.google.com/document/d/1DTl4TYaTPMAtwQTju9PZmxKhZTCh6nmi-Vh8cnSgYak/edit

This seems like the right approach. Quantify but don't hype the numbers.

I definitely agree that information on these topics is ripe for aggregation/curation.

My instinct is to look to the VC/startup community for some insight here, specifically around uncertainty (they're in the business of "predicting/quantifying/derisking uncertain futures/projects"). Two quick examples:

I would expect an "EA-focused uncertainty model" to include gates that map a specific project through time given models of macro future trends.

One difference between for-profits and non-profits is that the worst case scenario for a for-profit is going out of business. The worst case for a non-profit is unintentionally causing harm. Someone mentioned that there aren't a lot of historians in EA--I wonder if this is because the history of attempts to do good is discouraging.

If you haven't come across it yet, you might like to look at Back of the Envelope Guide to Philanthropy, which tries to estimate the value of some really uncertain stuff.

I started putting together a reading list, but it's going to take longer than I thought. To avoid making the perfect the enemy of the good, and to toot my own horn, I thought you might like to read How do EA Orgs Account for Uncertainty in their Analysis? by myself, Kathryn Mecrow, and Simon Beard.

Yep yep, happy to! A couple things come to mind:

  1. We could track the "stage" of a given problem/cause area, in a similar way that startups are tracked by Seed, Series A, etc. In other words, EA prioritization would be categorized w.r.t. stages/gates. I'm not sure if there's an agreed on "stage terminology" in the EA community yet. (I know GiveWell's Incubation Grants http://www.givewell.org/research/incubation-grants and EAGrants https://www.effectivealtruism.org/grants/ are examples of recent "early stage" investment.) Here would be some example stages:

Stage 1) Medium dive into the problem area to determine ITN. Stage 2) Experiment with MVP solutions to the problem. Stage 3) Move up the hierarchy of evidence for those solutions—RCTs, etc. Stage 4) For top solutions with robust cost-effectiveness data, begin to scale.

(You could create something like a "Lean Canvas for EA Impact" that could map the prioritized derisking of these stages.)

  1. From the "future macro trends" perspective, I feel like there could be more overlap between EA and VC models that are designed to predict the future. I'm imagining this like the current co-evolving work environment with "profit-focused AI" (DeepMind, etc.) and "EA-focused AI" (OpenAI, etc.). In this area, both groups are helping each other pursue their goals. We could imagine a similar system, but for any given macro trend. i.e. That macro trend is viewed from a profit perspective and an impact/EA perspective.

In other words, this is a way for the EA community to say "The VC world has [x technological trend] high on their prioritization list. How should we take part from an EA perspective?" (And vice versa.)

(fwiw, I see two main ways the EA community interacts in this space—pursuing projects that either a) leverage or b) counteract the negative externalities of new technologies. Using VR for animal empathy is an example of leverage. AI alignment is an example of counteracting a negative externality.)

Do those examples help give a bit of specificity for how the EA + VC communities could co-evolve in "future uncertainty prediction"?

I think ITN is good for scoping problems, but many think it shouldn't be used for interventions. So for interventions, I think quantitative models like this one are valuable.

What are the most important considerations for assessing charities doing uncertain-return stuff?

I think an important one is: How likely is the project to reduce the uncertainty of the return?

E.g. will it decide a crucial consideration

Edit to give more detail:

Resolving a crucial consideration increases the value of all your future research massively. Take for example the question of whether that will be a hard or slow take off. Hard take off favours AI safety now, whereas soft take off favours building political and social institutions that encourage cooperation and avoid wars. As they both have humanity's future on the line they are both equally massively important, conditioned on them being the scenario that might happen.

Resolving the question (or at least driving down the uncertainty) would allow the whole community to focus on the right scenario and get a lot better bang for their buck. Even if it doesn't directly address the problem.

[anonymous]0
0
0
  • How far should we use quantified models?

I think quantifying tends to be the right approach almost always when it's an in-depth exploration based on 80,000 hours' reasoning. That said, when it is used in a public summary of reasoning it can give a false pretense of certainty in cases where the estimate is very uncertain, leading to problems like those described here in animal advocacy. I think the best solution is to reason quantitatively when possible but to keep that in a public document linked to in any announcements and not to highlight the quantitative estimates in a way that often misleads people.

Another important step to take on this issue is probably to distinguish between problems that are unmeasurable and those that simply have not been measured yet. On those that have not been measured yet, we should try to measure them, and that might take some creativity by ingenious researchers.

thanks

Curated and popular this week
Paul Present
 ·  · 28m read
 · 
Note: I am not a malaria expert. This is my best-faith attempt at answering a question that was bothering me, but this field is a large and complex field, and I’ve almost certainly misunderstood something somewhere along the way. Summary While the world made incredible progress in reducing malaria cases from 2000 to 2015, the past 10 years have seen malaria cases stop declining and start rising. I investigated potential reasons behind this increase through reading the existing literature and looking at publicly available data, and I identified three key factors explaining the rise: 1. Population Growth: Africa's population has increased by approximately 75% since 2000. This alone explains most of the increase in absolute case numbers, while cases per capita have remained relatively flat since 2015. 2. Stagnant Funding: After rapid growth starting in 2000, funding for malaria prevention plateaued around 2010. 3. Insecticide Resistance: Mosquitoes have become increasingly resistant to the insecticides used in bednets over the past 20 years. This has made older models of bednets less effective, although they still have some effect. Newer models of bednets developed in response to insecticide resistance are more effective but still not widely deployed.  I very crudely estimate that without any of these factors, there would be 55% fewer malaria cases in the world than what we see today. I think all three of these factors are roughly equally important in explaining the difference.  Alternative explanations like removal of PFAS, climate change, or invasive mosquito species don't appear to be major contributors.  Overall this investigation made me more convinced that bednets are an effective global health intervention.  Introduction In 2015, malaria rates were down, and EAs were celebrating. Giving What We Can posted this incredible gif showing the decrease in malaria cases across Africa since 2000: Giving What We Can said that > The reduction in malaria has be
Ronen Bar
 ·  · 10m read
 · 
"Part one of our challenge is to solve the technical alignment problem, and that’s what everybody focuses on, but part two is: to whose values do you align the system once you’re capable of doing that, and that may turn out to be an even harder problem", Sam Altman, OpenAI CEO (Link).  In this post, I argue that: 1. "To whose values do you align the system" is a critically neglected space I termed “Moral Alignment.” Only a few organizations work for non-humans in this field, with a total budget of 4-5 million USD (not accounting for academic work). The scale of this space couldn’t be any bigger - the intersection between the most revolutionary technology ever and all sentient beings. While tractability remains uncertain, there is some promising positive evidence (See “The Tractability Open Question” section). 2. Given the first point, our movement must attract more resources, talent, and funding to address it. The goal is to value align AI with caring about all sentient beings: humans, animals, and potential future digital minds. In other words, I argue we should invest much more in promoting a sentient-centric AI. The problem What is Moral Alignment? AI alignment focuses on ensuring AI systems act according to human intentions, emphasizing controllability and corrigibility (adaptability to changing human preferences). However, traditional alignment often ignores the ethical implications for all sentient beings. Moral Alignment, as part of the broader AI alignment and AI safety spaces, is a field focused on the values we aim to instill in AI. I argue that our goal should be to ensure AI is a positive force for all sentient beings. Currently, as far as I know, no overarching organization, terms, or community unifies Moral Alignment (MA) as a field with a clear umbrella identity. While specific groups focus individually on animals, humans, or digital minds, such as AI for Animals, which does excellent community-building work around AI and animal welfare while
Max Taylor
 ·  · 9m read
 · 
Many thanks to Constance Li, Rachel Mason, Ronen Bar, Sam Tucker-Davis, and Yip Fai Tse for providing valuable feedback. This post does not necessarily reflect the views of my employer. Artificial General Intelligence (basically, ‘AI that is as good as, or better than, humans at most intellectual tasks’) seems increasingly likely to be developed in the next 5-10 years. As others have written, this has major implications for EA priorities, including animal advocacy, but it’s hard to know how this should shape our strategy. This post sets out a few starting points and I’m really interested in hearing others’ ideas, even if they’re very uncertain and half-baked. Is AGI coming in the next 5-10 years? This is very well covered elsewhere but basically it looks increasingly likely, e.g.: * The Metaculus and Manifold forecasting platforms predict we’ll see AGI in 2030 and 2031, respectively. * The heads of Anthropic and OpenAI think we’ll see it by 2027 and 2035, respectively. * A 2024 survey of AI researchers put a 50% chance of AGI by 2047, but this is 13 years earlier than predicted in the 2023 version of the survey. * These predictions seem feasible given the explosive rate of change we’ve been seeing in computing power available to models, algorithmic efficiencies, and actual model performance (e.g., look at how far Large Language Models and AI image generators have come just in the last three years). * Based on this, organisations (both new ones, like Forethought, and existing ones, like 80,000 Hours) are taking the prospect of near-term AGI increasingly seriously. What could AGI mean for animals? AGI’s implications for animals depend heavily on who controls the AGI models. For example: * AGI might be controlled by a handful of AI companies and/or governments, either in alliance or in competition. * For example, maybe two government-owned companies separately develop AGI then restrict others from developing it. * These actors’ use of AGI might be dr