I still find the bit about not releasing the logs strange. I understand the rationale behind it (preventing people from saying, “But I wouldn’t have fallen for that!”), but surely Eliezer had to understand that it would make his accomplishment look a lot more suspect.
One possible explanation is that seeing the logs would have made his accomplishment look even more suspect. (E.g., perhaps he didn’t in fact persuade the gatekeeper to let him out in-game, but made some out-of-band argument like “If you agree to say that you let me out and never release the logs, I will pay you $1000” or “If you say that I persuaded you to let me out, it will make people take the problem of AI safety more seriously”. I think Eliezer has denied doing any such thing … but then he would, wouldn’t he?)
I suspect that seeing the logs would have made Eliezer seem like a horrible human being. Most people who hear of AI Box imagine a convincing argument, when to me it seems more plausible to exploit issues in people’s sense of narrative or emotion.
Yup, certainly possible. Some later attempts at box-escaping have certainly gone that way. (I don’t know whether any successful ones have. There don’t seem to have been a lot of successes since Eliezer’s.)
Did it succeed? I know I find it thoroughly unconvincing.
Certainly outside of LW it has had absolutely no impact.
I, too, am thoroughly unconvinced.
I still find the bit about not releasing the logs strange. I understand the rationale behind it (preventing people from saying, “But I wouldn’t have fallen for that!”), but surely Eliezer had to understand that it would make his accomplishment look a lot more suspect.
One possible explanation is that seeing the logs would have made his accomplishment look even more suspect. (E.g., perhaps he didn’t in fact persuade the gatekeeper to let him out in-game, but made some out-of-band argument like “If you agree to say that you let me out and never release the logs, I will pay you $1000” or “If you say that I persuaded you to let me out, it will make people take the problem of AI safety more seriously”. I think Eliezer has denied doing any such thing … but then he would, wouldn’t he?)
I suspect that seeing the logs would have made Eliezer seem like a horrible human being. Most people who hear of AI Box imagine a convincing argument, when to me it seems more plausible to exploit issues in people’s sense of narrative or emotion.
Yup, certainly possible. Some later attempts at box-escaping have certainly gone that way. (I don’t know whether any successful ones have. There don’t seem to have been a lot of successes since Eliezer’s.)