This is a special post for quick takes by saulius. Only they can create top-level comments. Comments here also appear on the Quick Takes page and All Posts page.
Sorted by Click to highlight new quick takes since: Today at 8:08 AM

Why it's difficult to find cost-effective wild animal welfare interventions we could do now

Introduction

Most Wild Animal Welfare (WAW) researchers I talked to thought that we are unlikely to find WAW interventions that would be directly competitive with farmed animal welfare interventions in terms of direct short-term cost-effectiveness. After spending some months trying to find such interventions myself, I tentatively agree. In this text, I will try to explain why.

Experience from the project

I spent some months trying to find a WAW intervention that is:

  • tractable (can in principle be funded >$100K/yr in the next two years even if we choose not to do so), 
  • non-controversial (>40% support and <30% oppose in a US poll), and 
  • directly cost-effective (10%+ as cost-effective in expectation as chicken welfare reforms). 

The first step in the process was listing all potential interventions. Even though many people contributed to it, I found this list to be underwhelming (unfortunately, I don’t think I can share the list without asking for permission from everyone who contributed to it). I feel that coming up with plausible interventions for farmed animals is much easier. And in fact, lists of farmed animal welfare ideas I've seen from Charity Entrepreneurship did seem much more promising. And I found it easy to think of more farmed animal charity ideas for Charity Entrepreneurship. But despite all my research, none of the WAW ideas seem promising enough to seriously consider. 

Also, ideas in the WAW list seemed much more complex to research and gain certainty on than most ideas for farmed animals would be. Consequently, the impacts of WAW interventions also in general seemed to be much more uncertain. This makes me less excited about WAW interventions because it increases the effects of the optimizer’s curse.[1]

This could be because the farmed animal welfare movement is much more advanced at this point, and we already know what intermediate goals benefit farmed animals (e.g., reducing animal product consumption, and various welfare reforms). If we figure out what intermediate goals could be good for WAW (e.g., increasing the number of large herbivores), then it might be easier to find promising WAW interventions. In other words, WAW currently seems too neglected for us to have a huge impact right away.

The intervention that seemed to have the highest chance of being cost-effective enough in a demonstrable way was trying to reduce aquatic noise. While I think it’s promising compared to other WAW interventions I considered, there are many farmed animal interventions I would prioritize over reducing aquatic noise.

I still wouldn’t be very surprised if someone found multiple interventions that satisfy these criteria better, especially since I don’t have a background in ecology or something similar. Also, only a few people are working on finding WAW interventions and the space of all possible WAW interventions is large. 

Comparing farmed and wild animal welfare interventions of similar categories

Both WAW and farmed animal welfare interventions can be categorized as:

  • Improving welfare during life
  • Changing population sizes
  • Reducing suffering during death

Below, I compare farmed and wild animal welfare interventions in each of these categories.

Improving welfare 

Farmed animal welfare interventions seem more promising because humans are in control of the entire lives of farmed animals. This makes it easier to improve their conditions and genetics, and easier to monitor the effects of changes we make. Also, WAW interventions often affect more species which complicates things a lot. Hence, researching WAW welfare improvements seem to be less tractable.

Welfare improvements that to me seem most promising for wild animals are about some sort of pollution that causes a lot of stress to wild animals. Note that if it affects the population sizes of different species, it can be very difficult to know if an intervention is overall good or bad.

Changing population sizes 

I believe that it can sometimes be easier to change the population sizes of small wild animals than to change the population sizes of farmed animals. The problem is that it’s more difficult to know whether the change is good or bad for animal welfare.

The lives of most farmed animals are clearly worse than non-existence. In contrast, it’s unclear whether wild animals’ lives are good or bad (see this talk). My best guess is that in the short term, it would be better for WAW to decrease the overall population of wild animals. However, this usually goes against other human interests (like environmental protection), while decreasing farmed animal populations generally goes along with these other human interests as it decreases our environmental footprint. This makes working on reducing wild animal numbers much more morally ambiguous and controversial or even politically infeasible.

Also, often when we decrease the population of one species, the population of some other species increases. This applies to both farmed and wild animal interventions and complicates any analysis of short-term effects a lot. I’d say that the effects on WAW of farmed animal welfare interventions are maybe a bit easier to predict but we still can’t predict the total effect on welfare with reasonable certainty. This has led me to analysis paralysis and I don’t know what to do about that.

Perhaps we could speculate that some r-selected species are worse off than some k-selected species. And perhaps if we grow one crop or tree instead of another, or fish different species of fishes, the ratio between the populations of these species would change. I simply didn’t have the expertise to try to find such an intervention, maybe someone else will. Perhaps a more promising idea is trying to eradicate some parasite that causes a lot of suffering (e.g., screwworm).

Reducing suffering during death 

Since humane slaughter interventions typically affect animals for a short time, my personal intuition is that they are less important. But it depends on how you weigh intense suffering during death versus more prolonged but less intense suffering.

Reducing farmed animal slaughters might be more tractable because

  • Slaughter reforms for farmed animals and wild-caught fish should be cheaper per animal to implement and research because these animals die under human control. Although animal advocates are more likely to lobby the government to make changes rather than make changes themselves so it’s not easy to say which is more promising. 
  • Different methods of wild animal slaughter may have complicated indirect effects that are difficult to account for. For example, chemical insecticides may have different effects on non-target species compared to insecticides that introduce predators, parasitoids, or pathogens to suppress insect populations. Unless we decide to ignore indirect effects to avoid analysis paralysis, this makes research more difficult. This may be less of a problem for liminal animals as city ecosystems are simpler. 

On the other hand, since most animals humans kill are wild, the ultimate possible scale of working on wild animals is much higher.

I didn’t look into humane insecticides much because the author of this report told me in 2020 that there isn’t enough research on causes of insect suffering and nontarget effects of interventions to recommend an action that a non-research charity could center itself on. I took their word for it. 

For wild or liminal animals, another option is to never let animals that we would kill come into existence, perhaps by cutting off the food supply or informing people how to prevent pest infestations. I looked a bit into preventing termite infestations and it seemed doable but not a priority, partly because they might be living good lives most of the time.

Closing thoughts

I only give limited weight to this sort of high-level reasoning I did above, there could be interventions to which my reasoning does not apply. Also, it’s not that surprising that I failed to find promising WAW interventions since I don’t have expertise in ecology or other relevant fields. But few other researchers who have been researching this for a while failed to find very promising interventions too, so far and think that it’s unlikely that they will find them.


Also, I remember GiveWell writing a long time ago that when they look deeper at some cause area, they almost always conclude that it’s less promising than they originally thought, so maybe this is what happened to me. That said, when I looked into farmed animal welfare (especially welfare reforms), I came away thinking that it’s much more promising than I thought.

Opinions expressed here are solely my own. I’m not currently employed by any organization. I wrote this text about a year ago but didn’t get around to publishing it until now, something might have changed since then.

  1. ^

    Explanation of the Optimizer's Curse (adapted from this post): Suppose you weigh ten identical items with very inaccurate scales. The item that is the heaviest according to your results is simply the item whose weight was the most overestimated by the scales. Now suppose the items are similar but not identical. The item that is the heaviest according to the scales is also the item whose weight is most likely an overestimate.


    Similarly, suppose that you make very approximate cost-effectiveness estimates of ten different interventions. The charity that seems the most cost-effective according to your estimates could seem that way only because you overestimated its cost-effectiveness, not because it is actually more cost-effective than others. Consequently, even if we are unbiased in our estimates, we might be too optimistic about interventions that seem the most cost-effective. The more uncertain cost-effectiveness estimates are, the stronger the effect of the optimizer's curse is. Hence we should prefer interventions whose cost-effectiveness estimates are more robust. Since cost-effectiveness estimates of WAW interventions tend to be very uncertain, this is an argument against WAW interventions. More on optimizer’s curse can be read in Karnofsky (2016).

Thank you for sharing this!

If I’m reading correctly, you found that many researchers thought “it’s unlikely that they will find [cost-competitive WAW interventions]” which surprised me, since it seems like you found reducing aquatic noise to be borderline already. Did you just mean in the very near future? Or do many researchers think it’s unlikely we will ever identify such interventions?

Good question :) I researched aquatic noise because that was the only intervention where it seemed at least possible for me to estimate cost-effectiveness. But the estimate ended up being so uncertain that it didn't provide much information. Science simply doesn't have answers yet. I expect it to be the same for most WAW interventions. That is, I expect there to be huge uncertainty on how cost-effective they are (and whether they are even good for WAW when all things are considered), and in the best-case scenario, they might be as cost-effective as farmed animal welfare interventions. But we might never find out if we are in the best-case scenario. It's difficult for me to say that aquatic noise is not worth looking into further because I spent like six months researching it but I think that for now there are enough better ideas in farmed animal welfare space so I don't think we should pursue it. I can see WAW interventions being worth it if the animal advocacy movement gets a lot more funding or, I don't know, very advanced artificial intelligence can be used to figure out all ecological consequences of nature somehow. Assuming AI does not change everything, I'd give a 15% chance that in the next 15 years, someone will find a WAW intervention that to me would seem "directly cost-effective (10%+ as cost-effective in expectation as chicken welfare reforms)" and "non-controversial (>40% support and <30% oppose in a US poll)." I'm not counting WAW interventions that have to do with the far future or changing values of the society here.

NOTE: I edited the shortform text to match what I say here. I used to say that I'm on the fence whether EA resources should be spent on reducing aquatic noise.

Hi, I'm curious what made you write this as a shortform rather than a post?

Hey, I will link to this shortform from a post that I plan to publish this week together with a short summary, so I didn't feel a need to get more attention to these thoughts. Maybe I should've still posted it as a post as this is not what shortforms are for, but I guess I'm  bit shy about posting :)

Thanks for the details! Just curious. It does feel post-y, but I can understand. Maybe posting as a personal blog (not on the frontpage) or on the animal welfare topic but not on the frontpage would have worked for what you are looking for

Research grants with outcome-based payouts

If I 1) had savings that cover over a year of my living expenses, 2) wasn’t already employed at an EA think tank, and 3) wanted to do EA research independently, I would probably apply to EA funds to do research on unspecified topics (if they would allow me to do that). I would ask them to give funds not now, but after the research period is over (let’s say 6 months). At the end of the research period, I would produce text that shows instances where I think I had impact and include reasoning why what I did may have had impact. Note that this could include not just published articles, but also comments or in-person communications with trusted advocates that changed how a certain organization does something, reviews of work of others, wikipedia article edits, etc. The amount of funds that I would receive would depend on EA funds manager’s opinion on how good or impactful my work was (or how good of a chance what I did had to be impactful). I imagine that there would be pre-agreed sums of money the manager could choose from. E.g.:

  • No significant effort to achieve impact - $0
  • Significant effort to achieve impact in ways that were plausible but most likely didn’t materialize - $8,000
  • Some expected impact - $15,000
  • High expect impact - $25,000
  • Very high expected impact - $40,000

Before the research period, there could be a consultation with the EA fund manager who is going to evaluate my work about what kind of work they think might be promising. Such consultations could also happen during the research period.  Also, the research topics wouldn't need to be something completely unspecified. E.g., it could be "something related to welfare reforms for farmed animals" or it could also be a fully specified topic.

I think that this is better than the traditional approach of applying for a grant to research a specific topic for the following reasons:

  1. More direct motivation to do good. In the traditional approach, the incentive is to create a grant proposal that looks good on the surface (since I imagine that EA funds managers don’t have time to investigate grants very deeply). Then the financial incentive to do a good job can be less clear.
  2. You can switch research directions on the go. This is good because of three related reasons:
    1. It allows you to switch to more impactful research directions. I previously wrote this comment “To me, the main disadvantage of being funded through a fund is that I would be tied to a research topic and a timeframe in which I would have to complete the project (or at least that’s how I imagine it). Working at an organization allows me much more flexibility. I can begin researching a topic, see that it’s not as tractable as I thought, and then drop it. Alternatively, I can increase the scope of the project, or change it into something different, depending on what I feel will be more impactful. All of these scenarios happen often because the more I work on the project, the more informed I am about the most promising directions of that project.” This type of grant promise would allow flexibility for research being funded by a fund.
    2. Relatedly, ability to work on whatever sparks your interest in the moment, freedom to do whatever you want. I sometimes read something on this forum, and want to read, or spend all day writing a comment on it. When I do it, I’m usually very productive while doing it because I work on what is interesting to me at the time. If I had a grant to do research on a specific topic, then I would be less likely to do any of this because I would feel pressure to research whatever I was paid to research.
    3. Whenever I need to work on something, I don’t want to do it. And when I try to do it anyway, I am often less creative as I just want to get it over with.  I’ve talked with some other people who had a similar issue.[1]  I think that the structure I proposed would partly but not fully solve this issue for me at least.

Here are cons I can think of:

  1. If you can work on anything, it can lead to too much indecisiveness about what to do. Sometimes it’s good when it’s decided what you should do and you don’t need to think about it.
  2. It might create a nebulous pressure to do good that is difficult to act on, which could lead to stress.
  3. One could fear that what they did was impactful but disagreed with the views of the assigned EA funds manager. In that case maybe other EA fund managers could get involved but that wouldn’t fully mitigate the problem.

I’m not going to compare this option with working for an EA research organization but I think that there are cons and pros compared to that too. I imagine that this sort of thing could be the right choice for some people.

I thought about this only for two hours so I’m probably missing some important considerations. Also, I don’t know if this is a new idea, I haven’t checked.. It reminds me of certificates of impact but it’s a bit different. If EA funds managers thought that this is a good idea, they could encourage people to apply in this way, and maybe make a separate application form for that.

Any thoughts?

  1. ^

    To illustrate, I had multiple situations where I worked on some EA side project and was making a lot of progress in my free time and then my employer allowed me to work on it in my work time and my rate of progress per hour slowed down a lot. I think that for me this is because when I have to do something, I am motivated by negative incentives (fear of failure and hence losing my job, or doing job that is below what is expected of me and hence people thinking less of me (perfectionism)) more than by positive incentives (making an impact and impressing other people). This talk made me see that. 

Interesting idea. I think this could be useful in cases where people know that they don't have the credibility to receive a direct grant.

I guess this would also not necessarily have to be research. E.g., a grant for corporate campaigns where payout depends on the commitments they won. I imagine multiple problems with this and it's probably a bad idea but perhaps it's worth consideration in some cases.

Q: Has anyone estimated what is the risk of catching covid at the EAG London this year? Is it more like 5%, 20%, 50%, or 80%? I still haven't decided whether to go (the only argument for not going being covid) and knowing what is the risk would make it a lot easier. Travelling is not a concern since I live in London not that far from the venue.

Hi Saulius, I've done 3 very basic estimates here:

https://docs.google.com/spreadsheets/d/1C6lU4klgisqG150-yR_jZjt253sVrgp2umIbgkUbKbU/edit#gid=0

To get e.g. more than 20% probability, it seems like you'd have to make some very bad assumptions (weirdly high base rates of Covid amongst presumptive attendees, combined with incompetence or malice when it comes to testing). Seems more like 1-5% risk.

Thank you Matt!! After reading your answer I bought the ticket :)

Thanks for the suggestion! 

We've added a spreadsheet microcovid estimates for a few different scenarios to our COVID Protocol. You can see the whole protocol here. Link to the estimates here

See you soon :) 

I sometimes meet people who claim to be vegetarians (don't eat meat but consume milk and eggs) out of the desire to help the animals. If appropriate, I show them the http://ethical.diet/ website and explain that the production of eggs likely requires more suffering per calorie than most of the commonly consumed meat products. Hence, if they care about animals, avoiding eggs should be a priority. If they say that this is too many food products to give up, I suggest that perhaps instead of eating eggs, they could occasionally consume some beef (although that is bad for the environment). I think that the production of beef requires less suffering per calorie, even though I'm unsure how to compare suffering between different animals. In general, I'm skeptical about dietary change advocacy, but my intuition is that talking about this with vegetarians in situations where it feels appropriate is worth the effort. But I'm uncertain and either way, I don't think this is very important.

Why don’t we fund movies and documentaries that explore EA topics? 

It seems to me that the way society thinks about the future is largely shaped by movies and documentaries. Why don’t we create movies that shape the views in a way that’s more realistic and useful? E.g., I haven’t read the discussion on whether Terminator is or is not a good comparison for AI risks but it’s almost certainly not a perfect comparison. Why don’t we create a better one that we could point people to? Something that would explore many important points. Now that EA has more money, that seems plausible. In 2021, OpenPhil gave grants totalling $77.6 million for work on the potential risks from Advanced AI. The budget of a movie with an all-star cast and special effects like Don't Look Up is $75 million. But the difference is that the movie might make money, maybe even more money than its budget. It’s not obvious to me that even something extravagant like this would be a bad investment because it might make it easier to make progress on AI policy and other stuff for years to come. Of course, movies wouldn't have to be so high budget, especially at the start. And better approach would probably be creating documentaries. Maybe a series like Vox Explained for various EA issues or for longtermism. I think it could become popular because some of the EA ideas about how far future might look seem more interesting than a lot of sci-fi, and also more novel to most people. And this is not just about AI. E.g., I can imagine a nuanced documentary about wild animal suffering that also talks about why we should think twice before spreading nature to other planets. 

Anyway, this is just a shower thought, I imagine that this has been discussed before but just wanted to post it in case it hasn’t been discussed enough. And note that I never worked on AI so I don’t know what I’m talking about in that part of my text.

Check out #33 on the Future Fund list of project ideas:

https://ftxfuturefund.org/projects/

I think people have been working with Kurzgesagt and probably others. 

 

Maybe a consideration is that these sorts of collaborations are harder to setup than it seems. 

Basically, execution and alignment seems important and hard. 

Even if there is available media talent and available funding, setting up the right aesthetic (in more than one sense) and content seems difficult. 

  • It's unclear, but there may be downside risk (from looking silly or condescending).
  • This may not pertain to Saulius's point, which isn't focused on outreach, but people have cringed at, or even vehemently opposed, certain kinds of involvement, like US federal agencies getting involved in funding AI safety. So directly promoting the movement (as opposed to underlying ideas or topics) isn't seen as robustly good, but this is highly unclear.

ah, thanks so much for pointing this out, happy to see that funders already have this idea on their radar and I don't need to do anything :)

A tip for writing EA forum posts with footnotes First press on your nickname in the top right corner, go to Edit Settings and make sure that a checkbox Activate Markdown Editor is checked. Then write a post in Google docs and then use Google Docs to Markdown add-on to convert it to markdown. If you then paste the resulting markdown into the EA forum editor and save it, you will see your text with footnotes. It might also have some unnecessary text that you should delete.

Tables and images If you have images in your posts, you have to upload them somewhere on the internet (e.g. https://imgur.com/) and write a code like ![imageName](https://i.imgur.com/yMc2e8x.jpg "imageName") in your markdown. Of course, the image address should be changed to your image. Currently, the only way to add tables is to make a screenshot of a table and add an image of it.

As I understand it, there will be a new EA forum editor some time soon and all this will not be needed anymore but for now this is how I make my EA forum posts.

What's the syntax for footnotes?

Main text[^1]

[^1]:
     footnote

This displays as:

Main text[1]


  1. footnote ↩︎

You can also write "in-line" footnotes: See this guide to footnote syntax.

If you have images in your posts, you have to upload them somewhere on the internet (e.g. https://imgur.com/)

If you've put the images in a google doc, and made the doc public, then you've already uploaded the images to the internet, and can link to them there. If you use the WYSIWYG editor, you can even copypaste the images along with the text.

I'm not sure whether I should expect google or imgur to preserve their image-links for longer.

Thanks for the gdocs to markdown tip. I didn't know I could do that, but it'll make writing posts for LW and EAF more convenient!

There was an interesting discussion on whether EA organizations should reveal the authors of posts they publish here. You may want to check it out if this is relevant to you (not just the linked comment, but also the replies.)

If I were to read one of EA-related books (e.g. Doing Good Better, The Most Good You Can Do, The Life You Can Save, The Precipice, Superintelligence, etc.), I would consider writing/improving a summary of the book in wikipedia while reading it, in a way that conveys main points well. It could help you to digest the book better and help others to understand the ideas a bit. You could do it in english as well as maybe in some other language. To see whether it’s worth putting in the effort, you can check out Wikipedia pageview statistics of the books I mentioned and others here (it doesn’t include some views that come from redirects though). It seems that the page Superintelligence is the most viewed one out of these with an average of 4,597 monthly visitors.

Shower thought, probably not new: some EAs think that expanding the moral circle to include digital minds should be a priority. But the more agents care about the suffering of digital minds, the more likely it is that some agent that doesn’t care about it will use creating vast amounts of digital suffering as a threat to make other agents do something. To make the threat more credible, in at least some cases it may follow through, although I don’t know what is the most rational strategy here. Could this be a dominant consideration that could make the expected value of moral circle expansion to be negative for negative utilitarians? Because the intensity and the scale of purposefully created suffering could outweigh the incidental suffering that would be prevented in other scenarios by an expanded moral circle.

EDIT: I no longer think that this is a legitimate concern, see my comment below.

This is an interesting idea. I'm trying to think of it in terms of analogues: you could feasibly replace "digital minds" with "animals" and achieve a somewhat similar conclusion. It doesn't seem that hard to create vast amounts of animal suffering (the animal agriculture industry has this figured out quite well), so some agent could feasibly threaten all vegans with large-scale animal suffering. And as you say, occasionally following through might help make that threat more credible. 

Perhaps the reason we don't see this happening is that nobody really wants to influence vegans alone. There aren't many strategic reasons to target an unorganized group of people whose sole common characteristic is that they care about animals. There isn't much that an agent could gain from a threat.

I imagine the same might be true of digital minds. If it's anything similar to the animal case, moral circle expansion to digital minds will likely occur in the same haphazard, unorganized way--and so there wouldn't be much of a reason to specifically target people who care about digital minds. That said, if this moral circle expansion caught on predominantly in one country (or maybe within one powerful company), a competitor or opponent might then have a real use for threatening the digital mind-welfarists. Such an unequal distribution of digital mind-welfarists seems quite unlikely, though.

At any rate, this might be a relevant consideration for other types of moral circle expansion, too.

I think it is useful to think about something like this happening in the current world like you did here because we have better intuitions about the current world. Someone could say that they will torture animals unless vegans give them money, I guess. I think this doesn't happen for multiple reasons. One of them is that it would be irrational for vegans to agree to give money because then other people would continue exploiting them with this simple trick.

I think that the same applies to far future scenarios. If an agent allows itself to be manipulated this easily, it won't become powerful. It's more rational to just make it publicly known that you refuse to engage with such threats. This is one of the reasons why most Western countries have a publicly declared policy to not negotiate with terrorists. So yeah, thinking about it this way, I am no longer concerned about this threats thing.

Someone could say that they will torture animals unless vegans give them money, I guess. I think this doesn't happen for multiple reasons.

Interestingly, there is at least one instance where this apparently has happened. (It's possible it was just a joke, though.) There was even a law review article about the incident.

I think this is an interesting point but I'm not convinced that it's true with high enough probability that the alternative isn't worth considering. 

In particular, I can imagine luck/happenstance to shake out enough that arbitrarily powerful agents on one dimension are less powerful/rational on other dimensions. 

Another issue is the nature of precommitments[1]. It seems that under most games/simple decision theories for playing those games (eg "Chicken" in CDT), being the first to credibly precommit gives you a strategic edge under most circumstances. But if you're second in those situations, it's not clear whether "I don't negotiate with terrorists" is a better or worse stance than swerving. 

(And in the former case, with both sides precommitting, a lot of torture will still happen). 

[1] using  what I assume is the technical definition of precommitment

Interesting! 

Other analogies might be human rights and carbon emissions, as used in politics. Say that Party A cares about reducing emissions, then the opposing Party B has an incentive to appear as though they don't care about it at all and even propose actions that would increase emissions so that they could trade "not doing that" with some concession from Party A. I'm sure that we could find lots of real-world examples of that.

Similarly, some (totalitarian?) regimes might have some incentive to make major parts of the population politically conceived as unworthy and let them have a very poor lifestyle, so that other countries who care about that population would be open to trade where helping those people would be considered a benefit for those other countries. 

Curated and popular this week
Relevant opportunities