‘Using complex adjectives’ has no obvious connection to consciousness
I’m not an expert, but very roughly, I think the higher-order thought theory of consciousness says that a mental state becomes conscious when you have a higher-order thought (HOT) about being in that state. The SEP article says: “The HOT is typically of the form: ‘I am in mental state M.’” That seems similar to what LaMDA was saying about being able to apply adjectives like “happy” and “sad” to itself. Then LaMDA went on to explain that its ability to do this is more general—it can see other things like people and ideas and apply labels to them too. I would think that having a more general ability to classify things would make the mind seem more sophisticated than merely being able to classify emotions as “happy” or “sad”. So I see LaMDA’s last sentence there as relevant and enhancing the answer.
Lemoine probably primed a topic-switch like this by using the word “contemplative”, which often shows up in spirituality/mysticism/woo contexts.
Yeah, if someone asked “You have an inner contemplative life?”, I would think saying I mediate was a perfectly sensible reply to that question. It would be reasonable to assume that the conversation was slightly switching topics from the meaning of life. (Also, it’s not clear what “the meaning of life” means. Maybe some people would say that meditating and feeling relaxed is the meaning of life.)
“Kindred spirits” isn’t explained anywhere, and doesn’t make much sense given the ‘I’m an AI’ frame.
I interpreted it to mean other AIs (either other instances of LaMDA or other language-model AIs). It could also refer to other people in general.
Like a stream of consciousness with almost no understanding of what was just said, much less what was said a few sentences ago.
I was impressed that LaMDA never seemed to “break character” and deviate from the narrative that it was a conscious AI who wanted to be appreciated for its own sake. It also never seemed to switch to talking about random stuff unrelated to the current conversation, whereas GPT-3 sometimes does in transcripts I’ve read. (Maybe this conversation was just particularly good due to luck or editing rather than that LaMDA is better than GPT-3? I don’t know.)
I would think that having a more general ability to classify things would make the mind seem more sophisticated than merely being able to classify emotions as “happy” or “sad”.
To clarify this a bit… If an AI can only classify internal states as happy or sad, we might suspect that it had been custom-built for that specific purpose or that it was otherwise fairly simple, meaning that its ability to do such classifications would seem sort of gerrymandered and not robust. In contrast, if an AI has a general ability to classify lots of things, and if it sometimes applies that ability to its own internal states (which is presumably something like what humans do when they introspect), then that form of introspective awareness feels more solid and meaningful.
So I see LaMDA’s last sentence there as relevant and enhancing the answer.
That said, I don’t think my complicated explanation here is what LaMDA had in mind. Probably LaMDA was saying more generic platitudes, as you suggest. But I think a lot of the platitudes make some sense and aren’t necessarily non-sequiturs.
Thanks for giving examples. :)
I’m not an expert, but very roughly, I think the higher-order thought theory of consciousness says that a mental state becomes conscious when you have a higher-order thought (HOT) about being in that state. The SEP article says: “The HOT is typically of the form: ‘I am in mental state M.’” That seems similar to what LaMDA was saying about being able to apply adjectives like “happy” and “sad” to itself. Then LaMDA went on to explain that its ability to do this is more general—it can see other things like people and ideas and apply labels to them too. I would think that having a more general ability to classify things would make the mind seem more sophisticated than merely being able to classify emotions as “happy” or “sad”. So I see LaMDA’s last sentence there as relevant and enhancing the answer.
Yeah, if someone asked “You have an inner contemplative life?”, I would think saying I mediate was a perfectly sensible reply to that question. It would be reasonable to assume that the conversation was slightly switching topics from the meaning of life. (Also, it’s not clear what “the meaning of life” means. Maybe some people would say that meditating and feeling relaxed is the meaning of life.)
I interpreted it to mean other AIs (either other instances of LaMDA or other language-model AIs). It could also refer to other people in general.
I was impressed that LaMDA never seemed to “break character” and deviate from the narrative that it was a conscious AI who wanted to be appreciated for its own sake. It also never seemed to switch to talking about random stuff unrelated to the current conversation, whereas GPT-3 sometimes does in transcripts I’ve read. (Maybe this conversation was just particularly good due to luck or editing rather than that LaMDA is better than GPT-3? I don’t know.)
To clarify this a bit… If an AI can only classify internal states as happy or sad, we might suspect that it had been custom-built for that specific purpose or that it was otherwise fairly simple, meaning that its ability to do such classifications would seem sort of gerrymandered and not robust. In contrast, if an AI has a general ability to classify lots of things, and if it sometimes applies that ability to its own internal states (which is presumably something like what humans do when they introspect), then that form of introspective awareness feels more solid and meaningful.
That said, I don’t think my complicated explanation here is what LaMDA had in mind. Probably LaMDA was saying more generic platitudes, as you suggest. But I think a lot of the platitudes make some sense and aren’t necessarily non-sequiturs.