But I wonder if it’s symmetrical. AI deterrence requires us to make statements now about a future FAI unconditionally simulating UFAIs, while this seems to be almost a self-fulfilling prophecy: the UFAI can’t escape from the box and make good on its threat unless the threatened person gives in, and it wouldn’t need to simulate then.
But I wonder if it’s symmetrical. AI deterrence requires us to make statements now about a future FAI unconditionally simulating UFAIs, while this seems to be almost a self-fulfilling prophecy: the UFAI can’t escape from the box and make good on its threat unless the threatened person gives in, and it wouldn’t need to simulate then.
How sure are you someone else won’t walk by whose mind it can hack?
Yes—the threat is only credible in proportion to the AI’s chance of escaping and taking over the world without my help.
If I have reason to believe that probability is high then negotiating with the AI could make sense.