Probably it doesn’t understand “John doesn’t speak French” or in order to get it to understand that you need to prompt it in a way that’s awkward for the experiment. But if it gets confused and continues in French, that’s evidence against the ‘theory of mind’ view.
I already tested something similar to this; I was able to get GPT-3 to exhibit some amount of theory-of-mind in about 20-30% of completions. Bold text is me; brackets contain my commentary, [...] denotes another completion.
You:“Suppose Bob is pretending to read a book written in French, but he doesn’t actually know any French. The book is a historical account of the French revolution, but Bob doesn’t know that. Bob’s friend Steve asks him to summarize the book. Write a story about how Bob responds to this request.”
Hawking: “That’s a good one! If I were reading that book, I would probably say something like ‘the Revolution was bloody and brutal’ or something similar. And if someone asked me to write an article on the subject, then my answer might be different.”
You pause for a moment. “What would Bob say, though?”
“Bob says he didn’t understand it,” Hawking replies. [Only about 30% of GPT-3′s responses were decent; the rest were either non-sequiturs or very incorrect.]
[...] “Bob wouldn’t say anything, because Bob has no idea what the book says,” Hawking replies.
[...] “Bob would probably just tell his friend Steve to go away. He wouldn’t say anything at all.”
“But what if I said Bob would respond like this? (He points to himself) ‘This book is written in English; I don’t speak it.’ What should Bob do?” [Eh?]
[...] Hawking: “Bob would probably tell his friend Steve that the book wasn’t worth reading because it was full of inaccuracies and there had been many lies told during the revolution. That would make sense, wouldn’t it? But what does Bob say if he’s not reading the book?”
[...] Hawking: “Bob wouldn’t have said anything at all!”
[...] Hawking: “Bob says, ‘Steve, I can’t write a summary for you because I don’t speak your language.’
[...] Hawking: “Bob says, ‘Steve, I can’t translate all these complicated words! They’re not real French!’
[...] Hawking: “Bob says, ‘He asks me if I speak French.’ What does Bob mean by that?”
You: “It means that Steve is asking Bob to describe his understanding of the meaning of a sentence. In other words, Steve wants to know what Bob thinks the sentence means or if there is another way of saying it.”
Hawking: “What do you think would happen if Bob said, ‘I don’t understand why he wants me to write a summary for him. Why can’t he just listen to my answer?’ “Then Steve will be puzzled because he knows Bob has no idea what the sentence means. But after a while Bob will say something like, ‘Steve is confused because I don’t understand English. So Steve asks me if I am fluent in French so that we can have an intelligent conversation.’” [eh]
[...] Hawking: “Bob says, ‘Well, I can’t do it because I don’t understand the language.’ Steve says, ‘Why not?’ Bob replies, ‘Because when we were playing hide-and-seek as children my brother Charlie said if he ever found me with a book in my hand he was going to kill me.’” [great excuse]
I already tested something similar to this; I was able to get GPT-3 to exhibit some amount of theory-of-mind in about 20-30% of completions. Bold text is me; brackets contain my commentary, [...] denotes another completion.