A couple of years ago a thought experiment occurred to me, following spending some time seeing how well Effective Altruism could be baked into a system aimed at ethics. A year ago I put that experiment to paper, and this year I finally decided to share it. See the full text available here: http://dx.doi.org/10.13140/RG.2.2.26522.62407
The paper discusses a few critical points for calculating ethical value, positive and negative, including an edge case that some members of this community have been unwittingly sitting in the middle of. No one has yet refuted even a single point made in the paper, though several have pointed to portions of it being emotionally unappealing. Some discomfort is to be expected, as reality offers no sugar-coating.
I’m sharing it now to see if the EA community fairs any better than the average person when it comes to ethics, or if it is perhaps driven more by emotional phenomena than strictly ethical motives as cognitive bias reseach could imply. One of the wealthiest individuals in this community has failed already, after investing repeatedly in AI frauds who preyed on this community. The question this post will answer for me is if that event was more likely random, or systemic.
Thank you in advance. I look forward to hearing any feedback.
*Note: Unlike the infamous “Roko’s Basilisk”, it doesn’t matter at all if someone reads it or not. In any scenario where humanity doesn’t go extinct the same principles apply. People remain accountable for their actions, proportionate to the responsibilities they carry, regardless of their beliefs or intentions.
This statement: "An ethical system with any practical value must be able to reward those who do good" is what I disagree about. An "ethical system" is mainly a mental construction, as Linear Algebra or Set Theory. Sometimes Linear Algebra "rewards" those who undertand it with some technological results, or a position as university professor, but its truth is not a result of enforcement.
In any case, AGI has not incentive for retrospective punishment. The past is irrevocable, so Eleazar would be likely spared of any useless punishment by an almigthy AGI, because when you are powerfull enougth you dont need to use deterrence, and punishment becames useless A Godlike benevolent being has not credibiliby for retrospective punishment: this is elementary game theory, isn't it?.
PD.- I see you arrive before me to this argument. I will leave comment in the post to recognize it:
https://forum.effectivealtruism.org/posts/6j6qgNa3uGmzJEMoN/artificial-intelligence-as-exit-strategy-from-the-age-of
Perhaps this could be of interest to you:
https://forum.effectivealtruism.org/posts/uHeeE5d96TKowTzjA/world-and-mind-in-artificial-intelligence-arguments-against
It is baked into a few different papers I’ve written since 2020. I think part of the disconnect here is that the fundamental purpose of the system of ethics I’m talking about isn’t an intention to alter future behavior. The purpose isn’t to optimize human behavior, but rather to measure it.
A system that can’t measure the ethical value of actions, positive or negative, can’t react appropriately to them. A system that can measure them and chooses not to react appropriately isn’t ethical, and if it reacts only to some fraction according to bias, then it... (read more)