AI is getting more powerful, and at some point could cause significant damage.
There's a certain amount of damage that an AI could do that would scare the whole world (similar effects on government and populace psychology as coronavirus -- both willing to make sacrifices). The AI that could cause this (I naively expect) could be well short of the sophistication / power needed to really "rule the world", be unstoppable by humans.
So it seems likely to me (but not certain) that we will get a rude (coronavirus-like) awakening as a species before it gets to the point that we are totally helpless to suppress AI. This awakening would give us the political will / sense of urgency to be willing and feel compelled to do something to limit AI.
(Limiting / suppressing AI: making it hard to make supercomputers, concentrate compute -- technologically, legally. Also, maybe, making the world less computer-legible so that AI that do get made have less data to work with / have less connection to the world. Making it so that the inevitable non-aligned AI are stoppable. Or anything else along those lines.)
It seems like if suppressing AI were easy / safe, that would have been the first choice of AI safety people, at least until such time as alignment is thoroughly solved. But it seems like most discussion is about alignment, not suppression, I would assume on the assumption that suppression is not actually a viable option. However, given the possibility that governments may all be scrambling to do something in the wake of a "coronavirus AI", what kind of AI suppression techniques would they be likely to try? What problems could come from them? (One obvious fear being that a government powerful enough to suppress AI could itself cause a persistent dystopia.) Is there a good, or at least better, way to deal with this situation, which EAs might work toward?
I at least approximately agree with that statement.
I think there'd still be some reasons to think there won't someday be significantly non-aligned AIs. For example, a general argument like: "People really really want to not get killed or subjugated or deprived of things they care about, and typically also want that for other people to some extent, so they'll work hard to prevent things that would cause those bad things. And they've often (though not always) succeeded in the past."
(Some discussions of this sort of argument can be found in the section on "Should we expect people to handle AI safety and governance issues adequately without longtermist intervention?" in Crucial questions.)
But I don't think those arguments make significantly non-aligned AIs implausible, let alone impossible. (Those are both vague words. I could maybe operationalise that as something like a 0.1-50% chance remaining.) And I think that that's all that's required (on this front) in order for the rest of your ideas in this post to be relevant.
In any case, both that quoted statement of yours and my tweaked version of it seem very different from the claim "if we don't currently know how to align/control AIs, it's inevitable there'll eventually be significantly non-aligned AIs someday"?