All of Johannes_Treutlein's Comments + Replies

One way I imagine dealing with this is that there is an oracle that tells us with certainty, for two algorithms and their decision situations, what the counterfactual possible joint outputs are. The smoothness then comes from our uncertainty about (i) the other agents' algorithms (ii) their decision situation (iii) potentially the outputs of the oracle. The correlations vary smoothly as we vary our probability distributions over these things, but for a fully specified algorithm, situation, etc., the algorithms are always either logically identical or not.

U... (read more)

Not to trigger you, but I think by now it's probably more than 4%. The reason being America's soft power :/

Thanks for this meta-advice! Will try to adhere to it when asking for advice the next time :)

Just as a side note, Harsanyi's result is not directly applicable to a formal setup involving subjective uncertainty, such as Savage's or the Jeffrey-Bolker framework underlying evidential and causal decision theory. Though there are results for the Savage setup too, e.g., https://www.jstor.org/stable/10.1086/421173, and Caspar Oesterheld and I are working on a similar result for the Jeffrey Bolker framework. In this setup, to get useful results, the indifference Axiom can only be applied to a restricted class of propositions where everyone agrees on beliefs.

I don't think Romeo even has to deny any of the assumptions. Harsanyi's result, derived from the three assumptions, is not enough to determine how to do intersubjective utility comparisons. It merely states that social welfare will be some linear combination of individual utilities. While this already greatly restricts the way in which utilities are aggregated, it does not specify which weights to use for this sum.

Moreover, arguing that weights should be equal based on the veil of ignorance, as I believe Harsanyi does, is not sufficient, since ut... (read more)

3
RomeoStevens
4y
> there seems to be no way to determine what equal weights should look like, without settling on a way to normalize utility functions, e.g., by range normalization or variance normalization. I think the debate about intersubjective utility comparisons comes in at the point where you ask how to normalize utility functions. yup, thanks. Also across time as well as across agents at a particular moment.
And it turns out that the utilitarian approach of adding up utilities is *not* a bargaining solution, because it violates Pareto-optimality in some cases. Does that "disprove" total utilitarianism?

I'm not sure this is right. As soon as you maximize a weighted sum with non-negative coefficients your solution will be weakly Pareto optimal. As soon as all coefficients are strictly positive, it will be strongly Pareto optimal. The axioms mentioned above don't imply non-negative coefficients, so theoretically they are also satisfied by "... (read more)

2
Tobias_Baumann
4y
You're right; I meant to refer to the violation of individual rationality. Thanks!

Your argument seems to combine SSA style anthropic reasoning with CDT. I believe this is a questionable combination as it gives different answers from an ex-ante rational policy or from updateless decision theory (see e.g. https://www.umsu.de/papers/driver-2011.pdf). The combination is probably also dutch-bookable.

Consider the different hingeynesses of times as the different possible worlds and your different real or simulated versions as your possible locations in that world. Say both worlds are equally likely a priori and there is one real version of you

... (read more)

EV stands for Expected Value. (I think I actually meant Expected Utility more precisely)

Thanks a lot for this article! I just wanted to link to Lukas Gloor's new paper on Fail-Safe AI, which discusses the reduction of "quality future-risks" in the context of AI safety. It turns out that there might be interventions that are less directed at achieving a perfect outcome, but instead try to avoid the worst outcomes. And those interventions might be more tractable (because they don't aim at such a tiny spot in value-space) and more neglected than other work on the control problem. https://foundational-research.org/wp-content/uploads/2016/08/Suffering-focused-AI-safety.pdf

(Edit: I no longer endorse negative utilitarianism or suffering-focused ethics.)

Thank you! Cross-posting my reply as well:

If we adopt more of a preference-utilitarian view, we end up producing contradictory conclusions in the same scenarios that I discussed in my original essay—you can't claim that AMF saves 35 DALYs without knowing AMF's population effects.

Shouldn't this be fixed by negative preference utilitarianism? There could be value in not violating the "preference-equivalent" of dying one year earlier, but no value in cre... (read more)

(Edit: I no longer endorse suffering-focused ethics.)

Regardless of your stance on population ethics, I think in general it makes sense to take DALYs as a heuristic for how much good you can do with your money. Clearly all population ethical views consider improving existing lives in quality (decreasing YLDs, years lived with disability) a good thing. Preventing deaths expressed through reducing YLLs (Years of Life Lost) is probably overall good as well, although different views will assign more or less value to it. I agree with Michael Dickens that if the ... (read more)

0
Philip_B
7y
Excuse me, what does EV stand for?
2
Jonas V
8y
One thing that seems noteworthy is the fact that the population effect actually brings people closer together than they were before: Ignoring population effects, AMF has high impact from a CU perspective but low impact from a suffering-focussed perspective; accounting for population effects, the difference almost vanishes. Another way of looking at it: In situations where the population remains constant, population ethics becomes irrelevant. So accounting for population effects mainly gives us these two updates: 1. Population-ethical views become less relevant for prioritisation between various GiveWell charities (and not more relevant, as some seemed to suggest (possibly with the exception of the negative preference view)). 2. AMF might be less effective than deworming charities according to most population-ethical views (but still more effective than cash transfers due to developmental effects of malaria prevention).
5
MichaelDickens
8y
Cross-posting my reply: I believe this is the most plausible attempt at a resolution I've heard so far. Thanks, Johannes. Like some other responses I've heard, if we accept your proposed view on population ethics, we'd still have to substantially update the common view on the value of AMF. Remember, I'm not saying that YLL's don't have value; I'm saying that it's controversial and probably incoherent to claim that the value of AMF's lives saved equal the (time-discounted) number of additional life-years lived. If the importance of YLL's comes from the suffering of parents, as you suggest, YLL's will look really different than just one DALY per year of life lost. If we adopt more of a preference-utilitarian view, we end up producing contradictory conclusions in the same scenarios that I discussed in my original essay—you can't claim that AMF saves 35 DALYs without knowing AMF's population effects. If you're inconsistent, you cannot coherently maximize EV. You can only maximize EV if you can apply a unique real-valued EV function over states or actions, and such a function only exists in a consistent system.