Although the AI could threaten to simulate a large number of people who are very similar to you in most respects but who do not in fact press the reset button. This doesn’t put you in a box with significant probability and it’s a VERY good reason not to let the AI out of the box, of course,but it could still get ugly. I almost want to recommend not being a person very like Eliezer but inclined to let AGIs out of boxes, but that’s silly of me.
I’m not sure I understand the point of this argument… since I always push the “Reset” button in that situation too, an AI who knows me well enough to simulate me knows that there’s no point in making the threat or carrying it out.
It’s conceivable that an AI could know enough to simulate a brain, but not enough to predict that brain’s high-level decision-making. The world is still safe in that case, but you’d get the full treatment.
Although the AI could threaten to simulate a large number of people who are very similar to you in most respects but who do not in fact press the reset button. This doesn’t put you in a box with significant probability and it’s a VERY good reason not to let the AI out of the box, of course,but it could still get ugly. I almost want to recommend not being a person very like Eliezer but inclined to let AGIs out of boxes, but that’s silly of me.
I’m not sure I understand the point of this argument… since I always push the “Reset” button in that situation too, an AI who knows me well enough to simulate me knows that there’s no point in making the threat or carrying it out.
It’s conceivable that an AI could know enough to simulate a brain, but not enough to predict that brain’s high-level decision-making. The world is still safe in that case, but you’d get the full treatment.