That was my understanding, but I think that any world in which there is an AGI that isn’t Friendly probably won’t be very stable. If that happens, I think there’s a lot more chances that humanity will be destroyed quickly and you won’t be woken up than that a stable but “worse than death” world will form and decide to wake you up.
But maybe I’m missing something that makes such “worse than death” worlds plausible.
That was my understanding, but I think that any world in which there is an AGI that isn’t Friendly probably won’t be very stable. If that happens, I think there’s a lot more chances that humanity will be destroyed quickly and you won’t be woken up than that a stable but “worse than death” world will form and decide to wake you up.
But maybe I’m missing something that makes such “worse than death” worlds plausible.
I think you’re right. The main risk would be Friendly to Someone Else AI.