I find AI Box experiments interesting, like many others on this site, but my gut feeling is that a real AI is much, much more likely to rely on trickery instead of persuasion.
I hope the AI safety community isn’t expending an undue amount of resources on defense against persuasion simply because it is more intellectually stimulating work.
I find AI Box experiments interesting, like many others on this site, but my gut feeling is that a real AI is much, much more likely to rely on trickery instead of persuasion.
I hope the AI safety community isn’t expending an undue amount of resources on defense against persuasion simply because it is more intellectually stimulating work.