Indirect normativity is an approach to the AI alignment problem that attempts to specify AI values indirectly, such as by reference to what a rational agent would value under idealized conditions, rather than via direct specification.

Bibliography

Bostrom, Nick (2014) Superintelligence: Paths, Dangers, Strategies, Oxford: Oxford University Press.

Christiano, Paul (2012) A formalization of indirect normativity, Ordinary Ideas, April 21....

(Read More)

Posts tagged Indirect normativity
Most Relevant