For a robot to pass the Turing Test turned out to be less a question about the robot and more a question about the human.
Against expert judges, I still think LLMs fail the Turing Test. I don’t think current AI can pretend to be a competent human in an extended conversation with another competent human.
Again non-expert judges, I think the Turing Test was technically passed long long before LLMs: didn’t some of the users of ELIZA think and act like it was human? And how does that make you feel?
I suspect the expert judges would need to resort to known jailbreaking techniques to distinguish LLMs. A fair interesting test might be against expert-but-not-in-ML judges.
The ELIZA users may have acted like it was human, but they didn’t think it was human. And they weren’t trying to find ways to test whether it was human or not. If they had been, they’d have knocked it over instantly.
(Non-expert opinion).
For a robot to pass the Turing Test turned out to be less a question about the robot and more a question about the human.
Against expert judges, I still think LLMs fail the Turing Test. I don’t think current AI can pretend to be a competent human in an extended conversation with another competent human.
Again non-expert judges, I think the Turing Test was technically passed long long before LLMs: didn’t some of the users of ELIZA think and act like it was human? And how does that make you feel?
I suspect the expert judges would need to resort to known jailbreaking techniques to distinguish LLMs. A fair interesting test might be against expert-but-not-in-ML judges.
The ELIZA users may have acted like it was human, but they didn’t think it was human. And they weren’t trying to find ways to test whether it was human or not. If they had been, they’d have knocked it over instantly.