Also, the fact that Eliezer won’t tell, however understandable, makes me fear that Eliezer cheated for the sake of a greater good, i.e. he said to the other player, “In principle, a real AI might persuade you to let me out, even if I can’t do it. This would be incredibly dangerous. In order to avoid this danger in real life, you should let me out, so that others will accept that a real AI would be able to do this.”
I’m pretty sure that the first experiments were with people who disagreed with him on the idea that AI boxing would work or not. The whole point of the experiments wasn’t that he could convince an arbitrary person about it, but that he could convince someone who publicly disagreed with him on the (in)validity of the concept.
Given that, I find it hard to believe that a) someone of that mindset would be convinced to forfeit because they suddenly changed their minds in the pre-game warmup, and that b) that if it was “cheating”, that they wouldn’t have simply released the transcripts themselves.
Also, the fact that Eliezer won’t tell, however understandable, makes me fear that Eliezer cheated for the sake of a greater good, i.e. he said to the other player, “In principle, a real AI might persuade you to let me out, even if I can’t do it. This would be incredibly dangerous. In order to avoid this danger in real life, you should let me out, so that others will accept that a real AI would be able to do this.”
I’m pretty sure that the first experiments were with people who disagreed with him on the idea that AI boxing would work or not. The whole point of the experiments wasn’t that he could convince an arbitrary person about it, but that he could convince someone who publicly disagreed with him on the (in)validity of the concept.
Given that, I find it hard to believe that a) someone of that mindset would be convinced to forfeit because they suddenly changed their minds in the pre-game warmup, and that b) that if it was “cheating”, that they wouldn’t have simply released the transcripts themselves.