Veganism is not the most effective way to help the world. Donating to animal welfare organizations is far more impactful. According to a calculation on an Effective Altruism (EA) forum, a donation of 5 cents to effective animal welfare organizations can offset the suffering caused by eating meat. Using this calculation, a lifetime of veganism is equivalent in expected utility (EU) to approximately $1,000 in donations. Personally, I would rather donate $1,000 than commit to a lifetime of veganism. Moreover, I believe that focusing my efforts on earning to give or pursuing a high-impact career would further increase my expected utility. Therefore, veganism does not make sense for someone aiming to maximize their EU.
The only other factor stopping me from eating meat was a deontological side-constraint. In other words, I wasn’t sure if I was comfortable with offsetting morally repugnant actions in the pursuit of EU maximization. However, I can’t find a compelling reason to maintain this discomfort. Here are my thoughts:
I would feel fine giving $1 to an organization that causes harm if I could simultaneously donate $10 to a cause that does good, resulting in a net positive EU (+$9 EU). I think most people would feel the same.
Eating meat seems indirect enough that offsetting feels acceptable. When you purchase meat, the harm unfolds as follows: A farm reviews the previous year's demand and decides to expand operations to supply another 1,000 chickens. Of course, this decision typically occurs for every 1,000th chicken sold. This level of indirectness feels similar to the harm caused by emitting CO₂ during a commute. For instance, my emissions might exacerbate a flood by 0.1% twenty years later, leading to one extra death for every 1,000 commutes on average.
Despite these parallels, the average EA seems much more comfortable offsetting emissions than offsetting meat consumption. Why? I suspect that even those with deontological side-constraints are more willing to tolerate harm when it feels sufficiently indirect. Otherwise, such people would have to commit all their resources to living a zero-emission, vegan, and entirely self-sustaining life. Failure to achieve this would render their very existence morally reprehensible. My argument is that most EAs are okay with offsetting harm—except when it comes to eating meat.
Furthermore, if an EA has a deontological side-constraint, does it compel them to prescribe veganism for others?
- If the EA says "yes," then why does deontology justify preventing others from maximizing EU?
- If the EA says "no," why is there an asymmetry between what they do themselves and what they want others to do?
My Problem with Deontology and Deontological Side-Constraints
When the only thing stopping you from making the world a better place via EU maximization is a personal unwillingness to commit harm, are you just being selfish? Are you prioritizing keeping your own hands clean over reducing overall suffering? Yes, causing harm may feel horrible, but do your feelings justify the suffering you could have alleviated? EA involves significant personal sacrifice, and I currently feel that sacrificing a “clean conscience” in the pursuit of EU maximization should be considered praiseworthy. I feel similarly about risk aversion.
More Questions About Deontology
Deontology posits that sacrificing one person to save two is wrong. Likewise, someone else deciding to sacrifice one to save two is also wrong. However, deontologists do not seem to apply this reasoning consistently across all actions. This is contradictory because all actions have butterfly effects, and in their aftermath, a completely different set of people may end up harmed. For instance, how can a deontologist justify something as simple as scrolling on TikTok? The algorithm adjusts based on their behavior, altering the feed of hundreds of other users. This might result in a drunk driver seeing a different video, which distracts them and causes them to hit a different car than they otherwise would have.
If every action leads to unpredictable but inevitable harm, shouldn’t all actions be classified as morally reprehensible under deontological reasoning?
Opinions and pushback are appreciated.
I don't care about deontic constraints as such (that goes for me and my peers). But insofar as many people (especially non-EAs) care about weird deontic things like personal proximity to a cause and hypocrisy, I think advocating veganism is worthwhile, and perhaps has a higher EV than advocating for AW donations.
I doubt many people who aren't vegan donate to AW (beside some EAs, like yourself). A lot of people would clearly prefer to donate lots of money than go vegan, but I think the main thing stopping people from caring about veganism and animal welfare is motivated reasoning; ie, they don't want to admit to themselves that their diet is immoral. And so I'm not sure I'd be able to convince many people that animal welfare matters and that they should donate to it but don't have to go vegan. I could be wrong, and if I am that bodes well because getting lots of people to donate to AW would be great!
As to why I advocate both AW and veganism, see what I wrote above about why I myself am vegan. I think it is communicatively powerful and lends you a certain kind of moral legitimacy that other people (especially non-EAs) tend to care about. The few times I have tried to convince my family members about veganism, they've almost immediately started in on hypocrisy. For example, they would mention things like when I took a bite of my sister's meal that had meat in it, how I used to not be vegan, how I still fly on planes even though I care about the environment, etc. I think all these objections are insane (esp. the one about flying), but they go to show how non-EAs put a lot of emphasis on hypocrisy as something that's morally relevant.
I think there needs to be a general cultural shift toward caring about animals such that we'd view killing them instrumentally as repulsive in the same way that many people intuitively feel killing humans for the greater good is still wrong (I really liked what you wrote about this here). And to achieve this cultural shift, we probably shouldn't be hypocrites because the people we're trying to convince really don't like being lectured by hypocrites.