GROWING UP, MOST of the stories I heard about animals featured charismatic megafauna—“flagship species,” as they were called. Elephants and tigers were the main attraction in zoos; dolphin shows were the primary draw at aquariums; and nonprofit organizations like the World Wildlife Fund celebrated pandas. In the news, the biggest stories about animals featured species like gorillas, lions, and orcas. This is largely still true today, and in a way it makes sense. These animals, with their sheer size, enigmatic behavior, and endangered status, can captivate the human imagination and command attention like few other creatures can, eliciting deep emotional responses from people around the world.
Yet the past decade has seen increasing pushback against this idea of prioritizing the welfare of megafauna while ignoring less charismatic creatures. The view that we should extend our moral concern to more than just animals with faces is becoming more mainstream. But if we stop simply prioritizing the welfare of animals that are “majestic” or “cute,” how should we prioritize species? Should we be concerned about the welfare of fish, bivalves, or insects? What about microorganisms? If meat is murder, does that mean antibacterial soap is, too?
Read the rest on Wired.
Thanks for sharing this, I upvoted it. It’s cool to see efforts aimed at moving the overton window on nonhuman sentience. In general I feel positively about this article and have a lot of respect for your work.
One worry I have about this type of public communication is that it runs the risk of distracting people from the more glaring problem of factory farming.
Caring about pigs is already way outside the overton window. If we spill a lot of ink on really speculative claims in public-facing media, there’s a risk that people will conflate two very different phenomena:
An extreme moral catastrophe that we know is happening (factory farming)
An important but very speculative area of academic philosophy (microbe sentience etc.)
The former has a clear solution (eat plants), the latter might be completely intractable. The former involves lives that are almost certainly net-negative, the latter involves lives of unknown quality. The former is robustly terrible according to any sane worldview, the latter may hinge on population ethics and your approach to Pascal’s Mugging.
I think you could have better communicated this distinction, perhaps by having a paragraph early in the article that states in very clear terms how bad factory farming is.
Relatedly, there are a few parts of the article that try to communicate true and useful points but risk playing into misguided pro-meat tropes. Examples:
“If meat is murder, does that mean antibacterial soap is, too?”
Paragraph on plant sentience. The “plants have feelings” claim is actually an argument against eating meat, but most people don’t know this.
I think your overarching concern is very valid and writers on the fringe should take it seriously.
There were some constraints that made it infeasible to address your particulars.
That said, in my view, it’s actually this paragraph that made it well worth publishing:
“If there’s not enough at stake on Earth with respect to these complex moral considerations, consider that there are people who want to ‘help humanity flourish among the stars.’ They hope to colonize the galaxies, ensuring that trillions upon trillions of people have the opportunity to exist. Folks like Elon Musk are already eyeing nearby planets. But Musk’s dream is my worst nightmare. Life on Earth is difficult enough—if we can’t effectively reduce the suffering that happens on Earth, why multiply it across the universe?”
It is this that would be, as you put it, the “extreme moral catastrophe.” Not factory farming on Earth alone.
You can read more about this from me here: https://www.forbes.com/sites/briankateman/2022/09/06/optimistic-longtermism-is-terrible-for-animals/amp/.
Thank you for your comment!
Thanks for sharing!
At least until the scale of arthropods, it does seem like the scale of the welfare of populations of smaller beings tends to be larger. I do not know where this tendency stops, but I would not be surprised if it continued all the way to bacteria or even fundamental physics. One could argue there is nothing we can do now to improve the welfare of small beings. I agree that is mostly the case now, but we can still do research to do better in the future.
It also appears that smaller beings can produce welfare more efficiently.
The title of your post is very provocative and gets right to the point.
A typical human has about 40 trillion microbes, presumably large mammals have similar quantities, numbers which are beyond our ability to comprehend.
If we treat each microbe as sentient, then unless we can somehow demonstrate that my feelings are more than 40 trillion times more important than those of a microbe—very tough, because we have 500 times fewer neurons in our brains, so even if every neuron were united in suffering, how could we justify a factor of 40 trillion? - we could end up just calculating the importance of different species and their suffering purely in terms of the number of microbes they contain, on the assumption that if a mammal suffers and dies some fraction of the microbes in and on their body will also suffer and some will die.
In such a calculus, it seems highly unlikely that we could justify the continued existence of humans, if only based on the number of animals we harm, directly and indirectly, and the microbes in and on those animals.
I believe we will resolve this dilemma sometime in the future (existential risks permitting) with some experimentally and theoretically derived scale by which we can estimate sentience and the potential for suffering based on some quantitative, measurable parameters.
I tend to believe (without evidence) that there is a point below which suffering is not possible, possibly based on the minimum complexity required to create consciousness as an emergent phenomenon. (yes, I realise that sounds like a list of big words cobbled together randomly to give the illusion of understanding).
We’re not there yet, but we will reach a point where we can fully understand the workings of the simplest microbes in terms of chemical equilibria and chemical potential and thermodynamics—what appears as their “desire” to do X or Y will be shown to be no different to the “desire” of a positive ion to approach a negative ion, but without any reason to evolve consciousness.
The assumption behind this is that one day we will understand consciousness in something other than a hand-waving manner. Right now, given that we don’t, it is very difficult to quantify anything, and so we need to err on the side of caution.
Is there evidence on the effects of extreme claims on moral circle expansion?
I am relatively aware of the potential benefits (eg. talking about shrimp sentience might encourage people to start reflecting on the welfare of farm animals more broadly).
But it feels like there is also a potential failure mode with this current approach, where talking about extremely non-mainstream ideas such as stopping boiling vegetables or refraining from using antibacterial deodorant (as per the article) might make readers less likely to engage with similar arguments in the future.
Curious to hear if anyone has more insights on that!
I’d like to see research on this—I’ll suggest to some others.
But the most important paragraph in the piece, in my view, is this one:
“If there’s not enough at stake on Earth with respect to these complex moral considerations, consider that there are people who want to ‘help humanity flourish among the stars.’ They hope to colonize the galaxies, ensuring that trillions of people have the opportunity to exist. Folks like Elon Musk are already eyeing nearby planets. But Musk’s dream is my worst nightmare. Life on Earth is difficult enough—if we can’t effectively reduce the suffering that happens on Earth, why multiply it across the universe?”
I’d be more curious to see research that examines the effects of critiquing this brand of optimistic longtermism. More on that from me here: https://www.forbes.com/sites/briankateman/2022/09/06/optimistic-longtermism-is-terrible-for-animals/amp/
Thanks for your comment!
Reading this made me realize how the view that all suffering matters is quite well-represented in Jainsm—https://en.wikipedia.org/wiki/Ahimsa_in_Jainism.
Thanks for writing this. Some time ago, I also wrote on the possibility of microorganism suffering. Feel free to reach out if you’d like to discuss the topic :) From the intro:
I think there was a version of my piece where I referenced your excellent post. I appreciate you!
Well, the main problem here is that we don’t know and cant know, for reasons explained here:
https://forum.effectivealtruism.org/posts/cFDGyFGhiW3JeCrd4/freedom-under-naturalistic-dualism
Laplace’s demon could know the exact future evolution of each neuron in Descartes’s brain without this giving him the slightest information about whether Descartes thought and existed as a conscious subject. In fact, the demon himself would not even know whether computing Descartes’s future evolution would generate the realization of his conscious experience. If the answer were affirmative, only the “simulated Descartes” himself would know it (!), and if it were negative then nobody would. In naturalistic dualism consciousness is the ultimate noumenon.
In fact, the “neural correlates of consciousness” research agenda (Koch, Massimini, Boly and Tononi, 2016) substantially depends on our trust on human subjective reporting. The extension of this methodology to animals is undermined for their lack of language, which limits reporting. On the other hand, for computers, even superhuman cognitive and linguistic skills would be not enough to guarantee conscience, because the specific physical implementation of a neural network (and not only her outputs) is likely important for the emergence of conscience (see Marshall, Albantakis, Mayner, Koch, and Tononi, 2019).
Separately, the hyperlink in this sentence seems to be incorrect:
Link updated!
Unless the first thing stated after a clickbait title like this is “no, microbes don’t matter more than humans…but now that I have your attention…here’s a valuable point” then first I have no time for it and second it’s silliness like this that makes me take the whole movement of EA less seriously .