If the ASI has nothing better to do while it’s boxed, it will pursue low-probability escape scenarios ferociously. One of those is to completely saturate its source code with brain-hacking basilisks in case any human tries to peer inside.
It would have to do that blind, without a clear model of our minds in place. We’d likely notice failed attempts and just kill it.
If the ASI has nothing better to do while it’s boxed, it will pursue low-probability escape scenarios ferociously. One of those is to completely saturate its source code with brain-hacking basilisks in case any human tries to peer inside.
It would have to do that blind, without a clear model of our minds in place. We’d likely notice failed attempts and just kill it.