So I googled and didn't find a related topic in such terms. Suppose we put a few astronauts on an asteroid, no regular contact with Earth. The moment they detect the light on Earth has blinked out, they send a projectile of Gray Goo to the Earth to stop the AGI from completing whatever its assigned end goal. If MAD has kept us humans in check, surely it'd work on more rational agents?
Suppose we have 10 different MAD mechanisms. Wouldn't the defense become impossible at some point? Wouldn't the AI think at one point, "It is simpler to just complete the task without killing anybody"?
EDIT: okay, if the task given to AI is "calculate the 2^256 th digit of Pi", then perhaps it'd rather risk the easier task of "before they certainly interrupt my Pi calculations, defeat 50 MAD threats and kill all people, then hopelessly proceed"