My Problem with Veganism and Deontology

Veganism is not the most effective way to help the world. Donating to animal welfare organizations is far more impactful. According to a calculation on an Effective Altruism (EA) forum, a donation of 5 cents to effective animal welfare organizations can offset the suffering caused by eating meat. Using this calculation, a lifetime of veganism is equivalent in expected utility (EU) to approximately $1,000 in donations. Personally, I would rather donate $1,000 than commit to a lifetime of veganism. Moreover, I believe that focusing my efforts on earning to give or pursuing a high-impact career would further increase my expected utility. Therefore, veganism does not make sense for someone aiming to maximize their EU.

The only other factor stopping me from eating meat was a deontological side-constraint. In other words, I wasn’t sure if I was comfortable with offsetting morally repugnant actions in the pursuit of EU maximization. However, I can’t find a compelling reason to maintain this discomfort. Here are my thoughts:

I would feel fine giving $1 to an organization that causes harm if I could simultaneously donate $10 to a cause that does good, resulting in a net positive EU (+$9 EU). I think most people would feel the same.

Eating meat seems indirect enough that offsetting feels acceptable. When you purchase meat, the harm unfolds as follows: A farm reviews the previous year’s demand and decides to expand operations to supply another 1,000 chickens. Of course, this decision typically occurs for every 1,000th chicken sold. This level of indirectness feels similar to the harm caused by emitting CO₂ during a commute. For instance, my emissions might exacerbate a flood by 0.1% twenty years later, leading to one extra death for every 1,000 commutes on average.

Despite these parallels, the average EA seems much more comfortable offsetting emissions than offsetting meat consumption. Why? I suspect that even those with deontological side-constraints are more willing to tolerate harm when it feels sufficiently indirect. Otherwise, such people would have to commit all their resources to living a zero-emission, vegan, and entirely self-sustaining life. Failure to achieve this would render their very existence morally reprehensible. My argument is that most EAs are okay with offsetting harm—except when it comes to eating meat.

Furthermore, if an EA has a deontological side-constraint, does it compel them to prescribe veganism for others?

  • If the EA says “yes,” then why does deontology justify preventing others from maximizing EU?

  • If the EA says “no,” why is there an asymmetry between what they do themselves and what they want others to do?

My Problem with Deontology and Deontological Side-Constraints

When the only thing stopping you from making the world a better place via EU maximization is a personal unwillingness to commit harm, are you just being selfish? Are you prioritizing keeping your own hands clean over reducing overall suffering? Yes, causing harm may feel horrible, but do your feelings justify the suffering you could have alleviated? EA involves significant personal sacrifice, and I currently feel that sacrificing a “clean conscience” in the pursuit of EU maximization should be considered praiseworthy. I feel similarly about risk aversion.

More Questions About Deontology

Deontology posits that sacrificing one person to save two is wrong. Likewise, someone else deciding to sacrifice one to save two is also wrong. However, deontologists do not seem to apply this reasoning consistently across all actions. This is contradictory because all actions have butterfly effects, and in their aftermath, a completely different set of people may end up harmed. For instance, how can a deontologist justify something as simple as scrolling on TikTok? The algorithm adjusts based on their behavior, altering the feed of hundreds of other users. This might result in a drunk driver seeing a different video, which distracts them and causes them to hit a different car than they otherwise would have.

If every action leads to unpredictable but inevitable harm, shouldn’t all actions be classified as morally reprehensible under deontological reasoning?

Opinions and pushback are appreciated.