Epistemic status: Not an expert on moral uncertainty.
The Model
There have been many models proposed to resolve moral uncertainty, but I would like to introduce one more. Instead of acting in accordance to the moral theory we are most confident in (my favorite theory) or making complex electoral systems (MEC, the parliamentary model), we might want to pick a moral theory at random. Just assign to every moral theory you know the probability of how confident you are about this theory, put them in a row from least to most likely (or any sequence really) and pick a random real number between 0 and 100. E.g: Say you have 1% credence in 'Kantian ethics', 30.42% in 'Average utilitarianism' and 68.58% in 'Total utilitarianism' and you generate the random number 31, you will therefore pursue 'Average utilitarianism'. Whenever you update your probabilities you can reroll the dice (another version would be that you reroll at a fixed frequency of intervals, e.g every day). Here are some of the advantages and disadvantages of this model.
The Good
- It represents your probabilities
- It is fair, every theory gets an equal chance
- It is easy to understand
- It is fast
- It stops a moral theory from dominating even though it only has slightly more credence than the second largest theory (e.g 49%, 50%)
- It stops a moral theory from dominating even though it has a minority credence (e.g 20%, 20%, 20%, 40%)
- It stops the problem of theory-individuation
- It has no need for intertheoretic comparisons of value
- It makes you less fanatical
- It is cognitively easy (no need to do complex calculations in your head)
The Bad
- Humans need to use something other than their brain (dice/computers) to choose randomly (for an A.I this would not be a problem)
- You're not considering a lot of information about the moral theories. This could lead to you violating “moral dominance” e.g picking a theory that decides on an option that it doesn't have much stake in while another theory screams from the sideline (this problem could potentially be solved by making the 'stakes' an additional metric for deciding any given option, but that increases the complexity)
- It makes you more inconsistent and therefore harder to cooperate with
- Someone might get a wrong impression of you because they met you on a day of very low probability
Overall I'm not really convinced this is the path to a better model of moral uncertainty (or value uncertainty, since this model could also be applied there). I think some variation of MEC is probably the best route. The reason I posted this was because:
- Maybe someone with more expertise in moral uncertainty could expand upon this model to make it better
- Maybe sortition elements could be included in other theories to improve them
- Maybe sortition elements could be included in other theories to make them more useful in practice, since sortition is so easy without sacrificing fairness
Regarding stakes, I think OP's point is that it's not obvious that being sensitive to stakes is a virtue of a theory, since it can lead to low credence-high stakes theories "swamping" the others, and that seems, in some sense, unfair. Bit like if you're really pushy friend always decides where the your group of friends goes for dinner, perhaps. :)
I'm not sure your point about money pumping works, at least as stated: you're talking about a scenario where you lose money over successive choices. But what we're interested in is moral value, and the sortition model will simply deny their's a fixed amount of money in the envelope each time one 'rolls' to see what one's moral view is. It's more like there's $10 in the envelope at stage 1, $100 at stage 2, $1 at stage 3, etc. What this brings out is the practical inconsistency of the view. But again, one might think that's a theoretical cost worth paying to avoid other theories costs, e.g. fanaticism.
I rather like the sortition model - I don't know if I buy it, but it's at least interesting and one option we should have on the table - and I thank the OP for bringing it to my attention. I would flag the "worldview diversification" model of moral uncertainty has a similar flavour, where you divide your resources into different 'buckets' depending on the credence you have in each bucket. See all the bargaining-theoretic model, which treats moral uncertainty as a problem of intra-personal moral trade. This two models also avoid fanaticism and leave one open to practical inconsistency.