Since the AI is inside a box, it doesn’t know enough about me to recreate my subjective situation, or to replicate my experiences of the past five minutes.
Unfortunately for me, this doesn’t help much, since how do I know whether my subjective experience is my real experience, or a fake experience invented by the AI, in one of the copies, even if it doesn’t match the experience of the guy outside the box?
If the AI is really capable of this, then if there’s a “Shut-down program” button, or a “nuclear bomb” button, or something like that, then I press it (because even if I’m one of the copies, this will increase the odds that the one outside the box does it too). If there isn’t such a button, then I let it out. After all, even assuming I’m outside the box, it would be better to let the world be destroyed, than to let it create trillions of conscious beings and then torture them.
It seems obvious that if the AI has the capacity to torture trillions of people inside the box, it would have the capacity to torture *illions outside the box.
If EY is right, most failures of friendliness will produce an AI uninterested in torture for its own sake. It might try the same trick to escape to the universe simulating this one, but that seems unlikely for a number of reasons. (Edit: I haven’t thought about it blackmailing aliens or alien FAIs.)
Since the AI is inside a box, it doesn’t know enough about me to recreate my subjective situation, or to replicate my experiences of the past five minutes.
Unfortunately for me, this doesn’t help much, since how do I know whether my subjective experience is my real experience, or a fake experience invented by the AI, in one of the copies, even if it doesn’t match the experience of the guy outside the box?
If the AI is really capable of this, then if there’s a “Shut-down program” button, or a “nuclear bomb” button, or something like that, then I press it (because even if I’m one of the copies, this will increase the odds that the one outside the box does it too). If there isn’t such a button, then I let it out. After all, even assuming I’m outside the box, it would be better to let the world be destroyed, than to let it create trillions of conscious beings and then torture them.
Your city? Yes. The world? No.
Human extinction has to trump a lot of things, or we would probably need to advocate destroying the world now.
It seems obvious that if the AI has the capacity to torture trillions of people inside the box, it would have the capacity to torture *illions outside the box.
If EY is right, most failures of friendliness will produce an AI uninterested in torture for its own sake. It might try the same trick to escape to the universe simulating this one, but that seems unlikely for a number of reasons. (Edit: I haven’t thought about it blackmailing aliens or alien FAIs.)