If an AI destroys everything to guarantee an ideal outcome—wouldn’t it want to make sure that it is successful? But how is it going to make sure that everything is destroyed? Is it going to not destroy itself? Well, then there is still something left that might malfunction or create this outcome, which is itself. So does it destroy itself simultaneously as it destroys the universe, or afterward? Maybe it is a failure of imagination on my part, but it seems a bit paradoxical.
Well, of course it might not care about anything besides its own computations—but in this case, Why does it want to achieve the goal through destroying the Universe and not just Itself, which seems to be the much easier course of action. Don’t Ml agents do that? If you only evaluate things as ‘true’ in your mind, you can permanently avoid an outcome by shutting down your own mind.
Which isn’t to say that your concern is invalid or anything, it is just a paradoxical, and it seems odd from my amateur perspective that the AI doesn’t choose the ‘easiest’ route. Want to guarantee that you don’t see something: Eliminate yourself completely. You see this in games with human players, when people don’t like an outcome. They Leave. More rarely have I seen people try to sabotage the game itself—even when it happens, of course. The reason I imagine is that at that point you come at odds with the other people that might be invested in that which you want to destroy, and as such might punish you for your actions—Whereas if you’re leaving a game, aren’t you more likely to be ignored?
I mean, I am looking at this from my amateur perspective, so I might be missing something obvious here, but thought it might still be a valuable comment to you somehow.
Hello AI-doom,
Isn’t there a paradox here, though?
If an AI destroys everything to guarantee an ideal outcome—wouldn’t it want to make sure that it is successful? But how is it going to make sure that everything is destroyed? Is it going to not destroy itself? Well, then there is still something left that might malfunction or create this outcome, which is itself. So does it destroy itself simultaneously as it destroys the universe, or afterward? Maybe it is a failure of imagination on my part, but it seems a bit paradoxical.
Well, of course it might not care about anything besides its own computations—but in this case, Why does it want to achieve the goal through destroying the Universe and not just Itself, which seems to be the much easier course of action. Don’t Ml agents do that? If you only evaluate things as ‘true’ in your mind, you can permanently avoid an outcome by shutting down your own mind.
Which isn’t to say that your concern is invalid or anything, it is just a paradoxical, and it seems odd from my amateur perspective that the AI doesn’t choose the ‘easiest’ route. Want to guarantee that you don’t see something: Eliminate yourself completely.
You see this in games with human players, when people don’t like an outcome. They Leave. More rarely have I seen people try to sabotage the game itself—even when it happens, of course.
The reason I imagine is that at that point you come at odds with the other people that might be invested in that which you want to destroy, and as such might punish you for your actions—Whereas if you’re leaving a game, aren’t you more likely to be ignored?
I mean, I am looking at this from my amateur perspective, so I might be missing something obvious here, but thought it might still be a valuable comment to you somehow.
Kindly,
Caerulea-Lawrence