[ Question ]

How worried should I be about a childless Disneyland?

by willbradshaw1 min read28th Oct 20198 comments



In Superintelligence, Bostrom writes:

We could thus imagine, as an extreme case, a technologically highly advanced society, containing many complex structures, some of them far more intricate and intelligent than anything that exists on the planet today – a society which nevertheless lacks any type of being that is conscious or whose welfare has moral significance. In a sense, this would be an uninhabited society. It would be a society of economic miracles and technological awesomeness, with nobody there to benefit. A Disneyland with no children.

Most of the scenarios I've seen bandied about for grand galactic futures involve a primarily (or entirely) non-biological civilisation. As a non-expert in AI or consciousness, it seems to me that such scenarios are at a high risk of being childless Disneylands unless we specifically foresee and act to prevent this outcome. I think this partly because consciousness seems like a really hard problem, and partly because of stuff like this (from Paul Christiano in Ep. 44 of the 80,000 Hours Podcast):

I guess another point is that I’m also kind of scared of [the topic of the moral value of AI systems] in that I think a reasonably likely way that AI being unaligned ends up looking in practice is like: people build a bunch of AI systems. They’re extremely persuasive and personable because [...] they can be optimized effectively for having whatever superficial properties you want, so you’d live in a world with just a ton of AI systems that want random garbage, but they look really sympathetic and they’re making really great pleas. They’re like, “Really, this is incredibly inhumane. They’re killing us after this or [...] imposing your values on us.” And then, I expect … I think the current way overall, as actual consensus goes is to be much more concerned about people being bigoted or failing to respect the rights of AI systems than to be concerned the actual character of those systems. I think it’s a pretty likely failure mode, it's something I’m concerned about.

This is pretty scary because it means we could end up happily walking into an X-risk scenario and never even know it. But I'm super uncertain about this and there could easily be some fundamental idea I'm missing here.

On the other hand, if I am right that Disneylands without children are fairly likely, how should we respond? Should we invest more in consciousness research? What mistakes am I making here?



New Answer
Ask Related Question
New Comment

2 Answers

I don't think it matters that much (for the long-term) if the AI systems we build in the next century are conscious. What matters is how they think about what possible futures they can bring about.

If AI systems are aligned with us, but turned out not to be conscious or not very conscious, then they would continue this project of figuring out what is morally valuable and so bring about a world we'd regard as good (even though it likely contains very few minds that resemble either us or them).

If AI systems are conscious but not at all aligned with us, then why think that they would create conscious and flourishing successors?

So my view is that alignment is the main AI issue here (and reflecting well is the big non-AI issue), with questions about consciousness being in the giant bag of complex questions we should try to punt to tomorrow.

Several background variables give rise to worldviews/outlooks about how to make the transition to a world with AGIs go well. Answering this question requires assigning values to the background variables or placing weights on the various worldviews, and then thinking about how likely "Disneyland with no children" scenarios are under each worldview, by e.g. looking at how they solve philosophical problems (particularly deliberation) and how likely obvious vs non-obvious failures are.

That is to say, I think answering questions like this is pretty difficult, and I don't think there are any deep public analyses about it. I expect most EAs who don't specialize in AI alignment to do something on the order of "under MIRI's views the main difficulty is getting any sort of alignment, so this kind of failure mode isn't the main concern, at least until we've solved alignment; under Paul's views we will sort of have control over AI systems, at least in the beginning, so this kind of failure seems like one of the many things to be worried about; overall I'm not sure how much weight I place on each view, and don't know what to think so I'll just wait for the AI alignment field to produce more insights".