Biting bullets (accepting a moral argument that's intuitively unpleasant because one feels logically compelled to) is central to many people's experiences in EA. In fact, bullet-biting often resembles a contest to demonstrate one's commitment to EA. I have even heard someone brag of having “teeth of steel for biting bullets”.
However, the compulsion to bite bullets is at odds with moral anti-realism, another popular belief among EAs. If there is no objective morality, then there is no reason to prefer a principled, fundamentalist morality to an ad hoc, intuitionist morality. In fact, for the following reasons, it would be expected that bullet-biting does not accord with most people's experiences of morality.
Is morality law-like?
The paper “Is Life Law-Like?” criticizes the quest of biologists to derive laws for their observations, pointing out that evolution leads to ad hoc solutions and that biological processes occur probabilistically. So is morality closer to physics or biology? An empirical approach to morality would view it as stemming from evolutionary psychology, social structures, and historical serendipity. Psychology, sociology, and history are some of the only fields even less law-like than biology. Attempts to derive laws for morality are largely confined to recent Western history.
Scott Alexander argues for “high-energy ethics”, the idea that only through extreme thought experiments and edge cases can we discern the true nature of morality. The allusion to physics is no accident: it is the only discipline that thrives in such extremes. Applying such an approach to psychology would be absurd. For example, some psychologists believe that tall men are viewed as more attractive. One psychologist tries to disprove this by engineering a nine-foot-tall man, who is not viewed as more attractive. Of course, she didn't actually refute the hypothesis, only demonstrated that observations in the social sciences are context-dependent.
Optimizing for multiple values
A popular bullet to bite is an argument of the form “X is theft/rape/murder”, where X is an act that is widely believed to be morally acceptable but that has superficial similarity to a serious crime. This has been called the worst argument in the world. The reason that it's naive and often rejected is that serious crimes are typically viewed as immoral for multiple reasons. Murder is immoral because, among other reasons, it causes fear and suffering through the act itself, the victim had a desire to continue living, the victim would have had future positive experiences, and the death brings grief to family and friends. X typically has one or several of these characteristics but not all, so it is commonly judged to be not as bad as murder.
Morality becomes even more complex when it involves competing values. There is no inconsistency in believing that airports should X-ray luggage to reduce security risks and simultaneously believing that widespread surveillance of citizens is unjustified. One can value both security and privacy and believe that in some cases one outweighs the other. This point is often lost in bullet-biting morality, which views “inconsistency” as a product of hypocrisy and cowardice.
Optimizing for a single value when one has multiple values will almost always lead to the sacrifice of some values. Paying workers by the hour leads to slow work; paying by the task leads to shoddy work. Similarly, optimizing for naive definitions of utility will lead to paperclipping. For example, if we believe in one definition of utility, we may end up with a universe tiled with thermostats. (Each thermostat is “happy” because it is programmed to “want” the temperature of the cosmic background radiation.)
Conclusion
Moral anti-realism doesn't like neat conclusions: though there's no reason to favor biting bullets, there's no reason to disfavor it either. However, I have two pragmatic observations. First, the pressure to bite bullets (and the implication of irrationality if one does not) can be be off-putting to some EAs. Second, it may be easier to maintain commitments to beliefs if they are sincere, rather than conclusions that you grudgingly accept because you see no other choice.
One question is what we want "morality" to refer to under anti-realism. For me, what seems important and action-guiding is what I want to do in life, so personally I think of normative ethics as "What is my goal?".
Under this interpretation, the difference between biting bullets or not is how much people care about their theories being elegant, simple, parsimonious, vs how much they care about tracking their intuitions as closely as possible. You mention two good reasons for favoring a more intuition-tracking approach.
Alternatively, why might some people still want to bite bullets? Firstly, no one wants to accept a view that seems unacceptable. Introspectively biting a bullet can feel "right", if I am convinced that the alternatives feel worse and if I realize that the aversion-generating intuitions are not intuitions that my rational self-image would endorse. For instance, I might feel quite uncomfortable with the thought to send all my money to people far away, while neglecting poor people in my community. I can accept this feeling as a sign that community matters intrinsically to me, i.e. that I care (somewhat) more strongly about the people close to me. Or I could bite the bullet and label "preference for in-group" as a “moral bias” – biased in relation to what I want my life-goals to be about. Perhaps, upon reflection, I decide that some moral intuitions matter more fundamentally to me, say for instance because I want to live for something that is “altruistic”/"universalizable" from a perspective like Harsanyi’s Veil of Ignorance. Given this fundamental assumption, I’ll be happy to ignore agent-relative moral intuitions. Of course, it isn’t wrong to end up with a mix of both ideas if the intuition “people in my community really matter more to me!” is just as strong strong as the intuition that you want your goal to work behind a veil of ignorance.
On Lesswrong, people often point out that human values are complex, and that those who bite too many bullets are making a mistake. I disagree. What is complex are human moral intuitions. Values, by which I mean "goals" or "terminal values", are chosen, not discovered. (Because consequentialists goals are new and weird and hard for humans to have, so why would they be discoverable in a straightforward manner from all the stuff we start out with?) And just because our intuitions are complex – and totally contradicting each other sometimes – doesn't mean that we're forced to choose goals that look the same. Likewise, I think people who think some form of utiltiarianism must be the thing are making a mistake as well.
If values are chosen, not discovered, then how is the choice of values made?
Do you think the choice of values is made, even partially, even implicitly, in a way that involves something that fits the loose definition of a value--like "I want my values to be elegant when described in english" or "I want my values to match my pre-theoretic intuitions about the kinds of cases that I am likely to encounter?" Or do you think that the choice of values is made in some other way?
I too think that values are chosen, but I think that the choice inv... (read more)