AI: Hmm, I have produced in this past hour one paperclip, and the only other thing I did was come up with the solutions for all of humanity’s problems, I guess I’ll just take the next minute to etch them into the paperclip...
t=2 hours...
Experimenters: Phew, at least we’re safe from that AI.
Extra clarification: in this example, I’m assuming that we don’t observe the AI, and that we are very unlikely to detect the paperclip. How to get useful work out of the AI is the next challenge, if this model holds up.
That seems to be the preferred outcome, yes. In the process, we can (hopefully) safely learn more about AIs in general. Though the AI may choose to sabotage this learning process in order to reduce its future impact...
t=59 minutes...
AI: Hmm, I have produced in this past hour one paperclip, and the only other thing I did was come up with the solutions for all of humanity’s problems, I guess I’ll just take the next minute to etch them into the paperclip...
t=2 hours...
Experimenters: Phew, at least we’re safe from that AI.
Extra clarification: in this example, I’m assuming that we don’t observe the AI, and that we are very unlikely to detect the paperclip. How to get useful work out of the AI is the next challenge, if this model holds up.
That seems to be the preferred outcome, yes. In the process, we can (hopefully) safely learn more about AIs in general. Though the AI may choose to sabotage this learning process in order to reduce its future impact...