Shall we all re-read Parfit’s ‘Overpopulation and The Quality of Life’ before taking certain assumptions for granted?
I am motivated to write this post in response to @Rafael Ruiz ’s recent publication, but these are thoughts I’ve been contemplating for some time. Particularly, I’ve noticed how various members of the EA community not only engage in deep discussions about the hypothetical value of life for different arthropods and invertebrates but also experience genuine distress over the need to remove ants or moths from their homes. Similarly, I’ve read here about the ‘anxious, dizzying existential crisis I’m facing recently’ which is at least partly triggered by adhering to classical utilitarianism as a primary moral framework. Sometimes, it seems like this perspective is presented as if utilitarianism as an ethical stance is both obvious and unavoidable.
As the author points out and we all know, utilitarianism, particularly in its classical form, can indeed lead us to several ‘crazy towns’, and prima facie, one shouldn’t abandon these ‘crazy trains’ if they consider themselves to be intellectually and morally honest. I somewhat agree in a trivial sense, but it often seems as though we completely gloss over the crucial initial step of adopting classical utilitarianism as a main moral view—an assumption that is frequently taken for granted in many of these forum discussions.
I just find it surprising to encounter people who assert so categorically and nonchalantly that they are adherents of classical utilitarianism. Given the caliber of this forum, I assume that the most obvious problems with utilitarianism have been discussed extensively, so I have no intention of reiterating those discussions. Personally, I find Derek Parfit’s analysis particularly illuminating; he clearly defines what is ultimately the central problem of utilitarianism with the repugnant conclusion. His famous essay ‘Overpopulation and The Quality of Life’ not only clarifies the issue but also suggests a possible solution with his concept of ‘Perfectionism’—a concept which, as far as I remember, he does not fully develop in ‘Reasons and Persons’. This is not the only objection against the moral view, but I find that it's a particularly robust one, making it sufficient as a primary reference here as an example.
‘Intellectual and moral honesty’ is just as important, if not more, when addressing the problems of Utilitarianism
I sincerely believe that addressing these aspects of utilitarianism is crucial for anyone striving to be ‘intellectually and morally honest’, as it feels unfair to imply that anyone who gets out of the train is just due to intellectual or moral dishonesty. Neglecting these issues is equally indefensible, considering the significant consequences if utilitarianism is, even partially, flawed. A realistic possibility from carelessly adhering to a utilitarian stance might be an EA member inadvertently influencing a future AI to transform the universe into ‘happy ants’ (as an example, continuing the theme of the referenced post). If the repugnant conclusion poses a real problem, and if Parfit’s notion of Perfectionism (or any similar concept addressing current challenges in utilitarianism) holds true, then EA could end up being responsible for an immensely catastrophic act of evil. This would render incidents like the FTX debacle irrelevant in comparison, and become a maximum example of how ‘the road to hell is paved with good intentions’.
Again, the author expresses confusion, saying, ‘I’m baffled, I feel somewhat anxious, uneasy. I feel bad about any time I waste that isn’t related to astronomical numbers. It is just plain weird that alien invertebrates might be by far the most important moral subjects in the universe.’ What I want to point out is how such anxiety may not stem from being ‘intellectually and morally honest’, but instead might arise from failing to confront the serious issues within utilitarianism with the same level of intellectual and moral scrutiny. You might fully understand the implications of the repugnant conclusion and other critiques against utilitarianism and choose to ‘bite the bullet anyway.’ But that would be your prerogative, don’t presume it to be the obvious course of action. This stance is highly personal and subjective, not universally held among other thoughtful individuals. Again I am continually surprised by how frequently this oversight occurs in many discussions, to the point that it seems truly forsaken in their minds.
A question about EA’s Approach to Metaethical Issues
At this point, in my view if longtermism is indeed compelling, its primary value lies in giving us time to resolve ethical dilemmas (a point that has been previously made in several discussions). But we should not jump the gun before we have truly arrived at solutions, as the potential consequences could be cosmically severe.
Reflecting on the previous discussion, I would like to pose a more specific question regarding the EA approach to this topic. I understand that within the context of EA, there should also be a dedicated area for ‘funding’ or research aimed at addressing these important questions—deepening and clarifying the challenges associated with utilitarianism. What are these initiatives? Or is EA currently just accepting utilitarianism as the default normative ethic and limiting its efforts to merely optimizing ‘concrete implementations’ from various institutions and research endeavors?
I found @Wei Dai ’s comment in the previous post particularly interesting, as it highlights a neglected area, which is surprising since I believe this should be a priority for clarification before we proceed. He notes,
‘As far as actionable points, I’ve been advocating for work on metaphilosophy or AI philosophical competence as a means to accelerate philosophical progress in general—ensuring it keeps pace with other forms of intellectual progress, such as scientific and technological advancements, which are likely to be accelerated by AI development. This would also improve the chances that human-descended civilizations will eventually arrive at correct conclusions on important moral and philosophical issues, and be motivated and guided by those conclusions.’
And I guess this is the kind of work I mean, I am keen to learn if anyone is aware of specific initiatives aimed at clarifying the metaethical landscape at this juncture (beyond simply employing longtermism to eventually ‘have time to resolve metaethics’). From a personal viewpoint, I believe the next significant step in our ethical and moral advancement will occur once we fully address the ‘easy problem of consciousness’ and better understand the translation from the physical to the phenomenal ( thou to reference another popular recent post in the forum, there is still no even consensus on whether phenomenal consciousness is indeed relevant for moral status here. Again, we are at such an early stage, and our focus needs to be on enhancing our moral understanding.).
Finishing with a thought experiment, The Countless Ants Trolley, or how I would sincerely save your lovely child from arthropods
Lastly, I want to present a personal thought experiment that I find useful for linking the repugnant conclusion with the invertebrate issue that Rafael mentions. Although there’s nothing fundamentally new here—it’s essentially a reiteration of the repugnant conclusion—I still find it illuminating. If anyone has a catchy name for this, I’m all ears; perhaps something like the ‘Countless Ants Trolley’ as I say in the title heh, thou I suspect similar descriptions might already exist.
The scenario involves a classic trolley problem: on one side of the track stands a bright and kind human child, perhaps your own son or daughter (I literally picture this while thinking about other people’s children, not my own) . This child is capable of deep love and compassion, curious about mathematics and philosophy, and any other traits you find meaningful and important. They have the potential to push the boundaries of human knowledge and exhibit the best virtues of the human condition—traits that are rare but do exist. This child not only brings unique and profound happiness to their parents but also shares in this deep joy. These moments, like tucking them in at night and embracing after a day filled with shared discoveries and growth, foster a mutual bond and contentment that material means alone could never achieve. Some parents here may know exactly what I’m talking about, yet, I fear that I am too clumsy with words to fully capture this experience. Perhaps words themselves might never truly suffice.
Anyways, on the opposite track, imagine we must decide at what point the number of small invertebrates—be it ants, cockroaches, moths, or any sentient being deemed to have the minimal positive value of experience—would justify diverting the trolley away from the child. Assume the trolley, initially headed towards this track, would instantly and painlessly eliminate these insects, causing no explicit suffering. So here is the thing: I firmly believe that if I were actually in such a situation, no finite number of ants could compel me to sacrifice your son or daughter. The value of that child’s life is immeasurable compared to the sum of the ants. They simply belong to different dimensions of valuation, they are from a different nature, one is not the result of any kind of aggregation of the other; they are just incommensurable. As I often summarize to myself, ‘You just can’t always do arithmetic with phenomenology,’ which is what I understand Parfit truly revealed with the repugnant conclusion.
In any case, this may all just be a meandering mess I’ve dabbled in as a gut reaction, compounded by the fact that English is not my native language, so please forgive my lack of precision and insight. I do understand that most people in these forums are already trying harder and with more sincerity than most. However, I can’t shake this nagging feeling that concepts like Parfit’s Perfectionism should be more thoroughly considered, or at least, more present before making certain assumptions about our starting moral stance. Love to everybody.
Thanks @Richard Y Chappell🔸 , I truly enjoyed that one (you’re right that all this leans more towards ethical theory or normative ethics than metaethics; my apologies for the slip. I particularly resonated with:
I completely agree that moving towards an objective list theory may not only be plausible but crucial, given the risks of overlooking the possibility that it may be closer to the truth.
In any case, this is precisely the type of topic and nuance that I find lacking in most EA discussions, I find it surprising that considering such important questions is often not even seen as a possibility.
Are posts like this, then, a rarity within the EA context? Are there any sub-communities, study groups, or institutions that focus seriously on these types of issues? (I assume there aren’t, as you likely would have mentioned them, but I remain surprised)
Additionally, if you have any other references to essays or articles that explore different types of perfectionism as a potential solution to some of the challenges posed by the repugnant conclusion, I would greatly appreciate it .
Thanks again!