Summary: This post argues that three common objections to an approach in AI governance called AI ideal governance are misguided. Firstly, developing AI ideal governance theories is not necessarily an anti-pluralistic activity or one that promotes an irrational focus on the ends of our actions. Secondly, AI ideal governance theories are not necessarily de-motivating. Thirdly, AI ideal governance theories can be useful in developing plans for action.
Most of the discussion of advanced AI in the EA community has been focussed on what could go wrong, from the possibility of AI cementing authoritarianism to it leading to outright human extinction. However, if we manage to safely develop AI, there is also a fair amount that can go right.
The extent to which life in the future could be ‘profoundly good’ seems to be something that’s often neglected. The future may well include vast amounts of wealth, the expansion of the liveable galaxy, or even the development of the human form. Nonetheless, I sympathise with those who might initially feel somewhat reluctant about focussing on what could go right, when what could go wrong seems like the more pressing issue. With this in mind, this post will assess three possible arguments for scepticism about an approach in the field of AI governance that focuses on what could go right called AI ideal governance. I conclude that these arguments fall short, and that AI ideal governance is a valuable approach to thinking about AI.
As a field, AI governance focuses on the norms, institutions and policies that are required to ensure the safe development of AI. The sub-field of AI ideal governance was first identified by Allan Dafoe in his 2018 research agenda. It considers what the institutions and norms governing AI should look like in the future, rather than what they look like now. For example, imagine that we wanted to assess the role of private research labs in the safe development of AI. A traditional approach in AI governance may begin by evaluating the current regulation and incentive structures of these institutions. In contrast, a potential AI ideal governance approach might take a step back and ask questions such as: What would a private research lab committed to AI safety have in its constitution? Which norms or values would it be guided by?
The term ‘ideal’ may conjure an association with utopianism, which, for my purposes, can simply be understood as the approach of developing positive blueprints for institutions in the far future. Utopianism has a long and storied history, from the terms coinage by Thomas More through to socialist utopianism and even the technological aspirations of Silicon Valley. However, I don't think that all models developed through AI ideal governance need be utopian. This is because we can be more or less realistic about our ideal model. For example, in his recent post on ideal governance, Holden Karnofsky discusses the method of sortition as a possible means to improve the governance of certain political institutions. I think it's a stretch to suggest that this is a utopian suggestion; part of the reason we may favour sortition is to help address people’s imperfections or inadequacies. Nonetheless, it's pretty clearly grounded in a view of what such institutions should look like in the future, so it's fair to call it a discussion in ideal governance.
Objection 1. AI ideal governance theories are dangerous
The most stark and immediately pressing objection to AI ideal governance theories is that they are dangerous. Historically, this objection has an intellectual lineage to arguments against utopian thought by liberal thinkers during the Cold War, who commonly associated utopianism with socialist ideas and the Soviet Union. There are multiple reasons why someone could think that AI ideal governance theories are dangerous, but I’ll address two fairly common ones here. Both of these arguments focus on the more utopian AI ideal governance theories. I’m interested in arguments for why there’s something about (at least certain kinds of) ideal theories that makes them intrinsically dangerous, as to avoid a historical debate about whether particular uses of ideal theory (e.g., in the 20th century utopian political projects) have been dangerous. Although I ultimately conclude that AI ideal governance theories are not intrinsically dangerous, I do think there are important lessons to be learnt from each objection that can help us develop better plans.
1.1. AI ideal governance theories are anti-pluralistic
Wittgenstein, Elizabeth Taylor, Bertrand Russell, Thomas Merton, Yogi Berra, Allen Ginsburg, Harry Wolfson, Thoreau, Casey Stengel, The Lubavitcher Rebbe, Picasso, Moses, Einstein, Hugh Heffner, Socrates, Henry Ford, Lenny Bruce, Baba Ram Dass, Gandhi, Sir Edmund Hillary, Raymond Lubitz, Buddha, Frank Sinatra, Columbus, Freud, Norman Mailer, Ayn Rand, Baron Rothschild, Ted Williams, Thomas Edison, H.L. Mencken, Thomas Jefferson, Ralph Ellison, Bobby Fischer, Emma Goldman, Peter Kropotkin, you, and your parents. Is there really one kind of life which is best for each of these people?
The above quote, dated cultural figures and all, comes from the (often ignored) final section of Robert Nozick’s famous Anarchy, State and Utopia (1999, p. 310). It nicely illustrates the pluralism issue by highlighting the difficulty of drawing up ideal institutions that govern a large group of diverse individuals. To the extent that some AI ideal governance theories draw up such institutions, they might face the same problem. Previous utopians often haven’t grappled with this fact, contributing to the impression that utopias often don’t allow for people to live different kinds of lives. This is clearly a moral problem for those with liberal intuitions about the importance of respecting people’s choices to live their life as they please. Those sympathetic to Nick Bostrom’s views about population growth might also think that this pluralism issue is particularly pressing when thinking about the long-term future. 10^58 is a large number of sentient beings, and if a future governance structure required them to sacrifice their individuality, this could lead to a large amount of suffering.
Although there have been many utopias that don’t pay enough attention to people’s different lives, I struggle to see why this is a necessary problem of utopianism or the AI ideal governance approach more generally. Firstly, it’s not clear that this objection would immediately apply to many of the institutions involved in AI governance. For example, one could focus on drawing up plans about institutions without creating guidelines about people’s personal lives, perhaps by making use of a liberal distinction based on neutrality between different conceptions of the good. Secondly, there could be utopias that aim to maximise individual freedoms and that thereby allow people to live the lives that they choose. For example, in an earlier study on utopianism, Karnofsky drew up a number of possible utopias and performed a poll to find out which ones they’d most like to live in. The utopia that was ‘reminiscent of Nozick’s’ went as follows :
Everything is set up to give people freedom. If you aren’t interfering with someone else’s life, you can do whatever you want. People can sell anything, buy anything, choose their daily activities, and choose the education their children receive. Thanks to advanced technology and wealth, in this world everyone can afford whatever they want (education, food, housing, entertainment, etc.) Everyone feels happy, wealthy, and fulfilled, with strong friendships and daily activities that they enjoy.
Examples like this suggest that it is possible to create utopias that respect pluralism. Incidentally, the results of Karnofsky’s poll vindicates the view that utopias that don’t respect pluralism aren’t popular. The Nozickian utopia that respects freedom did much better than its less liberal alternatives. Though I think the anti-pluralism objection that I addressed in this section doesn’t give us reason to abandon AI ideal governance, I do think being mindful of the fact that anti-pluralistic utopias often aren’t popular can help us be better at planning for our future.
1.2. AI ideal governance theories promote a misguided focus on the ends of our actions
Another possible worry one might have about AI ideal governance theories is that aiming for certain fixed destinations might promote a misguided focus on the ends of our actions. Developing utopias could help justify the neglect (or even the sanctioning) of suffering in the nearer term, when no justification is really possible. This is the objection that comes closest to the concerns of many Cold War liberals. Famously, Popper calls utopianism an ‘all too attractive theory’, highlighting that it ‘leads to violence’. It’s beyond the scope of this paper to assess whether putting ends before means could ever be justified (for example through a form of utilitarianism), and instead I’m going to focus on how AI ideal governance theories might promote a form of irrationality.
One reason that AI ideal theories might promote irrationality is by painting an intoxicating vision of the future that is so appealing that it encourages us to forego traditional safeguards and sacrifice other concerns. More utopian ideal theories are often presented through fiction, or through a practice of worldbuilding. A good writer able to create a tangible and coherent world may encourage a feeling of longing from the reader.
I think there are a couple of counter-tendencies that should lead us to think that this concern is overstated. Firstly, there is no reason why we should consider utopias as a fixed plan rather than as a possible destination that could be updated if we think of something better. The misunderstanding that utopias have to function as an unchangeable final destination may result from the historical context of utopias being used as a legitimating ideology, but in a different context utopias need not have such rigidity.
Secondly, AI ideal governance theories could help tackle some existing irrationalities. This could be achieved by framing an issue in a new light, or illustrating the possibility that a particular set of actions opens up. One way in which ideal theories may actually boost rationality is by functioning similarly to philosophical thought experiments. Such experiments might allow us to further understand the institutional implications of maximising the presence of a certain virtue, or the complexity involved in balancing different desires for a future civilization.
The two tendencies I raise here cannot guarantee that a particular AI ideal governance theory might not inspire irrational action on the part of its reader, but give us reason to think such a concern may be overstated. They should highlight the importance of combining ideal governance theories with a healthy epistemic discourse, in which genuine examination of one’s views and the changing of one’s mind is possible.
Objection 2. AI ideal governance theories are de-motivating
Another possible objection one might have to AI ideal governance theories is that the practice of coming up with visions of how institutions should be is de-motivating. Even if this initially seems counter-intuitive, there are a number of reasons this might be the case. Firstly, without providing a roadmap, a vague blueprint for how an institution might be better may instead encourage lamentation at the state of the world, rather than helping to create something better. In particular, developing utopias may reveal the gulf between them and our current world, which might reveal the scale of the challenge and make it seem more difficult. Secondly, ideal theories can often look weird and may be difficult to relate to, which may simply turn people off. There has also been a relevant discussion to this topic within the EA community around branding, with one view being that EA should be at least somewhat mindful of how weird it comes across. Thirdly, it might be that other ways of talking about AI are simply more motivating. One recent suggestion in the EA community has been that the best way to motivate people to action may be just to talk about the possibility that inaction in the face of serious risks could lead to an existential catastrophe. This approach seemingly requires less agreement (as most people would agree that extinction is at least prima facie bad), and can be used to appeal to people's selfish sensibilities.
In this section, I’ll focus on offering a response to the third reason after briefly addressing the other two here. Firstly, as I explain in the next section, I think AI ideal governance theories can help to provide a roadmap. This can let us move beyond lamentation and towards concrete political action. In addition, I think there could be a useful division of labour here, and that it is rather uncharitable to expect all ideal governance theories to provide an accompanying policy programme. I think the second reason is a very important consideration in developing ideal governance theories, and I have no real response other than to say that weird ideal governance theories may well be ineffective.
As for the third reason, I think there is reason to think that ideal theories will probably be motivating to a certain kind of person. People's imaginations are captured by different things, and it doesn’t seem implausible to think that ideal theories might motivate some people to a greater extent than, say, a sole focus on a possible existential catastrophe might do. A common reason in favour of developing such theories is that they allow us to move beyond short-term concerns and to consider possibilities that had not been previously heard. Consideration of ideals can exercise imagination and offer hope. One way in which they can do this is by offering concrete suggestions about what the implementation of a certain policy or principle might look like.
Personally speaking, I have found Anders Sandberg’s recent discussions of grand futures to be particularly engaging in thinking through the possible futures that humanity might enjoy. Another popular example of motivational utopianism can be found in Toby Ord’s The Precipice (2020, p. 112). In his discussion of humanity’s ‘potential for flourishing’, Ord writes about the limits of our understanding of pleasure. He writes:
Most of these peaks [of pleasure] fade all too quickly. The world deadens; we settle into our routines; our memories dim. But we have seen enough to know that life can offer something far grander and more alive than the standard fare. If humanity can survive, we may one day learn to dwell more and more deeply in such vitality; to brush off more and more dust; to make a home amidst the beauty of the world. Sustaining such heights might not be easy, or simple. It could require changes in our psychology that we should approach with caution. But we know of nothing, in principle, that stands in our way, and much to recommend the exploration.
Although these motivational writings leave work to do to figure out how to progress to that point, their appealing nature may give us good inspiration to do so.
There are also historical examples that show the power of ideal visions. There has been much written about the role that utopianism has played in the development of Silicon Valley, in which founders are often motivated by the possibility of a grander future. A concrete example of this is Elon Musk's naming of the SpaceX rockets after various science fiction books. In addition, in a policy context, thinking about how the future might go well may allow a candidate to come across as engaging, progressive and forward-looking. One famous example illustrating the ability of ideal visions to motivate is Al Gore’s famous speech highlighting the capacity of the internet. The speech, often mis-referred to as ‘Al Gore invents the internet’, waxes lyrical about the ‘planetary information network that transmits messages and images with the speed of light’, allowing us to ‘transcend the barriers of time and distance’ and making possible ‘robust and sustainable economic progress, strong democracies, better solutions to global and local environmental challenges, improved health care’. These examples should give us reason to think that AI ideal theories could be similarly motivating.
Objection 3. AI ideal governance theories are not helpful for action
The third objection I’m going to address here is the claim that AI ideal governance theories are not helpful in guiding action. A potential sceptic may grant my case above that utopias have a degree of motivational value, but this might mean that only a small number of them ever need to be written. Instead, the vast majority of people should perhaps be working on more practical endeavours to influence policymakers. Here, I want to tackle this scepticism by illustrating the concrete role that ideal governance theories could play in developing AI strategy. I argue that AI ideal theories can help provide general orientations to guide policy making and can even provide actionable objectives.
AI ideal governance theories are useful because they allow us to set positive objectives. Having these positive objectives gives us something to aim for, which can in turn orient action. These plans wouldn’t necessarily have to be incredibly precise, and could instead be used to choose a broad positioning that will later be better defined. In previous mappings of the AI governance space, useful distinctions have been drawn between research at different levels of abstraction, from ‘macro strategy’ all the way down to ‘standards-setting’ and the evaluation of a particular regulation after it has been implemented. Setting broad objectives through AI ideal governance theories could form part of AI governance macro strategy, with further consideration about how to actually reach a given ideal or how to enact a certain strategy at a later stage in the division of labour. One method that has been particularly popular recently in setting such broad objectives has been worldbuilding, which has been recently promoted by the Future of Life Institute contest.
Alternatively, one could develop AI ideal governance theories that operate on shorter time-scales and that can provide more immediately actionable objectives. For example, in 2021, the World Economic Forum published a ‘Positive AI Economic Futures’ report that detailed six possible scenarios that might occur in the future given AI’s impact on the employment market. The impact of increased automation from AI might lead us to automate mundane tasks that humans don’t enjoy in favour of humans performing ‘fulfilling jobs’, such as those that involve using specialised skills (e.g., empathy or critical thinking). On the other hand, we might favour the automation of all work (assuming this is possible) so that humans are able to enjoy their leisure time. Sketching out these different positive visions allows us to decide which would be more appealing. This choice then has important policy consequences. If one favours the former option of only automating mundane jobs, this could lead us to try to develop AIs that can work with humans. In contrast, if the automation of all work is our aim, policies such as a universal basic income might better serve this.
However, an important objection to my argument that AI ideal governance theories are helpful in setting goals for action is that it’s very difficult to set positive objectives for AI development because AI pathways are so uncertain. This is a particularly challenging objection to the practice of developing utopian visions involving AI, as it seems unlikely we could follow a set model of how AI develops without it being a little incorrect.
Here, I suggest two possible responses. Firstly, developing macro-strategies for humanity’s future using AI ideal governance theories might not rely on particular controversial technical claims about AI development. A more convincing strategy might involve multiple different ways of reaching it, and thereby limit its contingency. Secondly, the process of developing plans might be useful even if those plans are not used. Plans can help clarify our motivations and illustrate what’s appealing about a particular policy. For example, it may not be possible to automate all work, even if we decided that it would be ideal to do so. This does limit how action guiding the plan is in this particular scenario, but the fact that we find the scenario in which work is no longer a feature of our lives an appealing one is still an important finding. This could then inform future action, for example in exploring whether any of the positive features of leisure could be better distributed by attempting to reduce inequality.
In conclusion, I think that the objections I’ve tried to raise fairly here do not give us reason to dismiss the approach of AI ideal governance. I argued that AI ideal governance theories are not necessarily dangerous, that they can be motivating and that they can be helpful in action. I hope to have another forum post appearing soon illustrating further directions in AI ideal governance and open questions that I think would be valuable for the sub-field to answer, which may be of particular interest to those that got this far in this post.
Acknowledgements: This post benefited greatly from the support of (amongst others) LP Bova, Miles Brundage, Joseph Carlsmith, Ross Gruetzemacher, Julia Karbing and Risto Uuk.
AI governance is not limited to a focus on government institutions, and can also include analysis of non-state actors like NGOs and intergovernmental organisations. For further discussion, see here. ↩︎
These questions are similar to questions of institutional design. A great piece in this field is here. Discussion of values governing AI creates an overlap with the (much more well-developed) field of AI ethics. I do not spend time here differentiating the fields, other than to say that the institutional focus of AI ideal governance is often not present in AI ethics. ↩︎
There is a long-standing debate about how to define utopianism which I do not enter into here. Compared to some authors, my definition is fairly narrow. Much of what I say will still be relevant to those who understand utopianism in a broader manner. ↩︎
For illustrative purposes, I draw upon literature in the utopian tradition to make my argument. Not all of these utopias are examples of AI ideal governance; not all of them deal with AI. Such examples should simply be understood as illustrative. ↩︎
This is Bostrom’s estimation including artificial beings found in Superintelligence (2017, p. 103). ↩︎
See ‘Utopia and Violence’ (1986) by Karl Popper, p. 5. ↩︎
Utopias and ‘ends justify the means’ worries are also discussed here. ↩︎
On the ability of philosophical thought experiments to re-frame issues and function as intuition pumps, see Daniel Dennett’s Intuition Pumps and other Tools for Thinking (2013). ↩︎
For more on political philosophy and lamentation, see here. ↩︎
On simplifying EA pitches to warnings about existential catastrophes, see here. ↩︎
For an example of Sandberg’s discussion of grand futures, see this talk. ↩︎
For a nice forum post that collects some of these more grandiose visions of possible futures, see here. ↩︎
On Silicon Valley and utopianism, see Howard Segal’s Technological Utopianism in American Culture (2005). ↩︎
In writings on this topic, there is sometimes a misguided suggestion that AI ideal governance theories are the only way to set objectives, which isn’t true. The objective to reduce the chance of an existential catastrophe isn’t an objective that stems from ideal theory but it clearly is an objective, for example. ↩︎
This also could be an objection to the motivational function of AI ideal governance theories. If we think that we have no idea if these utopias can come to be, this might make us less motivated by them. ↩︎
On the value of planning, see this piece. ↩︎
I really liked this post, thanks for writing it! I'm much more sympathetic to ideal governance now.
Two fairly off-the-cuff reactions:
First, I would guess that some (most?) of the appeal of utopias is conditional on lack of scarcity? I'm not sure how to interpret Karnofsky's results here: he notes that his freedom-focused utopia without mention of happiness, wealth or fulfilment was still the third most popular utopia amongst respondents. However, the other top five utopias highlight a lack of scarcity, and even the freedom-focused utopia without conclusion implies a lack of scarcity ("If you aren't interfering with someone else's life, you can do whatever you want"). Naively, I'd guess that at least some people value freedom so highly only if they can do what they like with that freedom.
I think this is a problem for creating utopias that respect pluralism - it's still possible, but I think it's harder than it appears. I would expect people to have strong opinions about resource allocation, which makes it hard to get to post-scarcity for everyone, which makes it hard to reach a utopia that many people like. The counterargument here is that if we had resource abundance, people would be much happier sharing resources, and this would make getting to a post-scarcity state for everyone much easier, but I'm a little sceptical about this. (Billionaires are quite close to resource abundance, and most of them seem quite keen to hold onto their personal wealth/don't seem very interested in redistributing their resources?)
It might still be motivating to construct pluralistic utopias which assume post-scarcity for everyone, even if this condition is unlikely to be met in practice, but I'm less confident that utopias which require a fairly unlikely condition will be action-guiding.
Second, I agree that AI ideal governance theories are useful for action. But shouldn't we care more about how useful for action they are? I'm not sure how useful it is to work on AI ideal governance vs. working on e.g. more applied AI policy, and it seems like you need a stronger claim than "AI ideal governance" is useful to motivate working on it. (Probably >0 people should work on ideal governance? But without a compelling argument that it's more valuable on the margin than other AI governance work, I'm not sure if many people should work on it.)
Thank you! A few quick thoughts on some great points:
Other than Karnofsky's piece, I didn't find too much empirical research trying to understand why people find utopias appealing, but I share the intuition that a lack of scarcity/great wealth is often a plausible reason. There's an interesting open empirical question about how that relates with people's views about freedom, and a related normative one about our intuitions regarding the importance of freedom more generally.
Agree that it's very hard to construct pluralism-respecting utopias and even harder to work out ways to get there. In the post, main aim was to question the idea that the former is impossible.
Definitely think the question of how useful AI ideal governance theories can be is next step of discussion after establishing ways in which they can be helpful. I don't have too many abstract thoughts on how many people should be working in this sub-field (this may depend on what problem is trying to be solved at a given time) - mainly wanted to establish that, e.g., people interested in policy, shouldn't see it as totally irrelevant approach.
Thanks for this Rory, I'm excited to see what else you have to say on this topic.
One thing I think this post is missing is a more detailed response to the 'ideal governance as weird' criticism. You write that 'weird ideal governance theories may well be ineffective', but I would suggest that almost all fleshed-out theories of ideal AI governance will be inescapably weird, because most plausible post-trasformative AI worlds are deeply unfamiliar by nature.
A good intuition pump for this is to consider how weird modern Western society would seem to people from 1,000 years ago. We currently live in secular market-based democratic states run by a multiracial, multigender coalition of individuals whose primary form of communication is the instantaneous exchange of text via glowing, beeping machines. If you went back in time and tried to explain this world to an inhabitant of a mediaeval European theocratic monarchy, even to a member of the educated elite, they would be utterly baffled. How could society maintain order if the head of state was not blue-blooded and divinely ordained? How could peasants (particularly female ones) even learn to read and write, let alone effectively perform intellectual jobs? How could a society so dependent on usury avoid punishment by God in the form of floods, plagues or famines?
Even on the most conservative assumptions about AI capabilities, we can expect advanced AI to transform society at least as much as it has changed in the last 1,000 years. At a minimum, it promises to eliminate most productive employment, significantly extend our lifetimes, allow us to intricately surveil each and every member of society, and to drastically increase the material resources available to each person. A world with these four changes alone seems radically different and unfamiliar to our own, meaning any theory about its governance is going to seem weird. Throw in ideas like digital people and space colonisation and you're jumping right off the weirdness deep end.
Of course, weirdness isn't per se a reason not to go ahead with investigation into this topic, but I think the Wildeford post you cited is on the right track when it comes to weirdness points. AI Safety and Governance already struggles for respectability, so if you're advocating for more EA resources to be dedicated to the area I think you need to give a more thorough justification for why it won't discredit the field.
Thanks for this comment John. Briefly, one related thought I have is that weirdness is an important concern but that not all AI ideal governance theories are necessarily so weird that they're self-defeating. I'm less concerned that theories that consider what institutions/norms should look like in the nearer future (such as around the future of work) are too weird, for example.
Broadly, I think your comment reinforces an important concern, and that further research on this topic would benefit from being mindful of the purpose it is trying to serve and its intended audience.
I have a couple thoughts on this.
First - if you're talking about nearer-term questions, like 'What's the right governance structure for a contemporary AI developer to ensure its board acts in the common interest?' or 'How can we help workers reskill after being displaced from the transport industry' then I agree that doesn't seem too strange. However, I don't see how this would differ from the work that folks at places LPP and Gov.AI are doing already.
Second - if you're talking about longer-term ideal governance questions, I reckon even relatively mundane topics are likely to seem pretty weird when studied in a longtermist context, because the bottom line for researchers will be how contemporary governance affects future generations.
To use your example of the future of work, an important question in that topic might be whether and when we should attribute legal personhood to digital labourers, with the bottom line concerning the effect of any such policy on the moral expansiveness of future societies. The very act of supposing that digital workers as smart as humans will one day exist is relatively weird, let alone considering their legal status, let further alone discussing the potential ethics of a digital civilisation.
This is of course a single, cherry-picked example, but I think that most papers justifying specific positive visions of the future will need to consider the impact of these intermediate positive worlds on the longterm future, which will appear weird and uncomfortably utopian. Meanwhile, I suspect that work with a negative focus ('How can we prevent an arms race with China?') or a more limited scope ('How can we use data protection regulations to prevent bad actors from accessing sensitive datasets?') doesn't require this sort of abstract speculation, suggesting that research into ideal AI governance carries reputational hazards that others forms of safety/governance work do not. I'm particularly concerned that this will open up AI governance to more hit-pieces of this variety, turning off potential collaborators whose first interaction with longtermism is bad faith critique.
Thanks again for your comment. Two quick related points:
People at the places you mention are definitely already doing interesting work relevant to ideal theory, e.g., regarding institutional design. Something distinctive about AI ideal governance research that I do think is less common is consideration of the normative components of AI governance issues.
On reflection, your comments and examples have convinced me that in the original post I didn't take the 'weirdness problem' seriously enough. Although I'd guess we might still have a slight disagreement about the scope (and possibly the implications) of the problem, I certainly see that it is particularly salient for longtermists at the moment given the debates around the publication of Will MacAskill's new book. As an example of ideal governance research that considers longer-term issues (including some 'weird' ones) in an analytical manner, Nick Bostrom, Allan Dafoe and Carrick Flynn's paper on 'Public Policy and Superintelligent AI' may be of interest.
Taking each of your points in turn: