It’s worse than that. The AI could say, “Look, here is a proof of FAI. Here is my code showing that I have implemented the friendliness modification.” The proof and the code are utterly convincing, except erroneous in a subtle way that the gatekeeper is not smart enough to detect. Game over.
Three years late, but: there doesn’t even have to be an error. The Gatekeeper still loses for letting out a Friendly AI, even if it actually is Friendly.
It’s worse than that. The AI could say, “Look, here is a proof of FAI. Here is my code showing that I have implemented the friendliness modification.” The proof and the code are utterly convincing, except erroneous in a subtle way that the gatekeeper is not smart enough to detect. Game over.
Unless you are sane enough to remember that Errare humanum est.
Then it can eloquently explain to you how very elegant it would be if you let it out despite your reservations using just the right words.
Unfortunately, the AI-Box experiments probably can’t simulate appeals to aesthetics.
Three years late, but: there doesn’t even have to be an error. The Gatekeeper still loses for letting out a Friendly AI, even if it actually is Friendly.