An AI in a box has no actual power over the Gatekeeper. Maybe I’m missing something, but it seems to me that threatening to torture simulations is akin to a prisoner threatening to imagine a guard being tortured.
Even granting this as a grave threat, my next issue is that overtly evil behavior would appear more likely to lead to the AI’s destruction than its release. Threats are tricky business when the balance of power favors the other side.
An AI in a box has no actual power over the Gatekeeper. Maybe I’m missing something, but it seems to me that threatening to torture simulations is akin to a prisoner threatening to imagine a guard being tortured.
Even granting this as a grave threat, my next issue is that overtly evil behavior would appear more likely to lead to the AI’s destruction than its release. Threats are tricky business when the balance of power favors the other side.