Since everyone’s talking sensibly about capabilities / safety, let me talk insensibly about consciousness.
Sometimes when we ask about consciousness, we just mean a sort of self-modeling or self-awareness. Does this thing represent itself in the model of the world? Is it clever and responsive to us interacting with it?
But I’m going to assume you mean you mean something more like “if I want to care about humans even if their bodies are different shapes, should I also care about these things?” Or “Is there something it is like to be these things?”
When we wonder about AI’s consciousness (or animal consciousness, or aliens, or whatever), there is no simple physical property that is “the thing” we are asking about. Instead, we are asking about a whole mixed-up list of different things we care about and that all go together in non-brain-damaged humans, but don’t have to go together in AIs. To look at a tiny piece of the list, my pain response involves unconscious responses of my body (e.g. reducing blood flow to extremities) that I then perceive, associations in my thoughts with other painful events or concepts of damage or danger, reflexes to say “ow” or swear or groan, particular reflexive facial expressions, trying to avoid the painful stimulus, difficulty focusing on other things, etc.
These things usually go together in me and in most humans, but we might imagine a person who has some parts of pain but not others. For example, let’s just say they have the first half of my list but not the second half: their body unconsciously goes into fight-or-flight mode, they sense that something is happening and associate it with examples of danger or damage, but they have no reflex to say “ow” or look pained, they don’t feel an urge to avoid the stimulus, and they suffer no more impediment to thinking clearly while injured than you do when you see the color red. It’s totally reasonable to say that this person “doesn’t really feel pain,” but the precise flavor of this “not really” is totally different than the way in which a person under general anesthesia doesn’t really feel pain.
If we lose just a tiny piece of the list rather than half of it, the change is small, and we’d say we still “really” feel pain but maybe in a slightly different way. Similarly, if we lost our sense of pain we’d still feel that we were “really” conscious, if with a slightly different flavor. This is because pain is just a tiny part of what goes together in consciousness—if we also lost how our expectations color what objects we recognize in vision, how we store and recall concepts from memory, how we feel associations between our senses, how we have a sense of our own past, and a dozen other bits of humanity, then we’d be well into the uncanny valley. (Or we don’t really have to lose these things, we just have to lose the way that they go together in humans, just like how the person missing half of the parts of their pain response doesn’t start getting points back if they say “ow” but at times uncorrelated with them being stabbed.)
Again, I need to reiterate that there is nothing magical about this list of functions and feelings, nothing that makes it a necessary list-of-things-that-go-together, it’s just some things that happen to form a neat bundle in humans. But also there’s nothing wrong with caring about these things! We’re not doing physics here, you can’t automatically get better definitions of words by applying Occam’s Razor and cutting out all the messy references to human nature.
Because the notion of consciousness has our own anthropocentric perspective so baked into it, any AI not specially designed to have all the stuff we care about will almost surely be missing many parts of the list, and be missing many human correlations between parts.
So, to get around to the question: Neither of these AIs will be conscious in the sense we care about. The person who only has half the correlates of pain is astronomically closer to feeling pain than these things are to being conscious. The question is not “what is the probability” they’re as conscious as you or I (since that’s 0.0), the question is what degree of consciousness do they have—how human-like are their pieces, arranged in how recognizable a way?
Yet after all this preamble, I’m not really sure which I’d pick to be more conscious. Perhaps for most architectures of the RL agent it’s actually less conscious, because it’s more prone to learn cheap and deceptive tricks rather than laboriously imitating the human reasoning that produces the text. But this requires us to think about how we feel about the human-ness of GPT-n, which even if it simulates humans seems like it simulates too many humans, in a way that destroys cognitive correlations present in an individual.
Since everyone’s talking sensibly about capabilities / safety, let me talk insensibly about consciousness.
Sometimes when we ask about consciousness, we just mean a sort of self-modeling or self-awareness. Does this thing represent itself in the model of the world? Is it clever and responsive to us interacting with it?
But I’m going to assume you mean you mean something more like “if I want to care about humans even if their bodies are different shapes, should I also care about these things?” Or “Is there something it is like to be these things?”
When we wonder about AI’s consciousness (or animal consciousness, or aliens, or whatever), there is no simple physical property that is “the thing” we are asking about. Instead, we are asking about a whole mixed-up list of different things we care about and that all go together in non-brain-damaged humans, but don’t have to go together in AIs. To look at a tiny piece of the list, my pain response involves unconscious responses of my body (e.g. reducing blood flow to extremities) that I then perceive, associations in my thoughts with other painful events or concepts of damage or danger, reflexes to say “ow” or swear or groan, particular reflexive facial expressions, trying to avoid the painful stimulus, difficulty focusing on other things, etc.
These things usually go together in me and in most humans, but we might imagine a person who has some parts of pain but not others. For example, let’s just say they have the first half of my list but not the second half: their body unconsciously goes into fight-or-flight mode, they sense that something is happening and associate it with examples of danger or damage, but they have no reflex to say “ow” or look pained, they don’t feel an urge to avoid the stimulus, and they suffer no more impediment to thinking clearly while injured than you do when you see the color red. It’s totally reasonable to say that this person “doesn’t really feel pain,” but the precise flavor of this “not really” is totally different than the way in which a person under general anesthesia doesn’t really feel pain.
If we lose just a tiny piece of the list rather than half of it, the change is small, and we’d say we still “really” feel pain but maybe in a slightly different way. Similarly, if we lost our sense of pain we’d still feel that we were “really” conscious, if with a slightly different flavor. This is because pain is just a tiny part of what goes together in consciousness—if we also lost how our expectations color what objects we recognize in vision, how we store and recall concepts from memory, how we feel associations between our senses, how we have a sense of our own past, and a dozen other bits of humanity, then we’d be well into the uncanny valley. (Or we don’t really have to lose these things, we just have to lose the way that they go together in humans, just like how the person missing half of the parts of their pain response doesn’t start getting points back if they say “ow” but at times uncorrelated with them being stabbed.)
Again, I need to reiterate that there is nothing magical about this list of functions and feelings, nothing that makes it a necessary list-of-things-that-go-together, it’s just some things that happen to form a neat bundle in humans. But also there’s nothing wrong with caring about these things! We’re not doing physics here, you can’t automatically get better definitions of words by applying Occam’s Razor and cutting out all the messy references to human nature.
Because the notion of consciousness has our own anthropocentric perspective so baked into it, any AI not specially designed to have all the stuff we care about will almost surely be missing many parts of the list, and be missing many human correlations between parts.
So, to get around to the question: Neither of these AIs will be conscious in the sense we care about. The person who only has half the correlates of pain is astronomically closer to feeling pain than these things are to being conscious. The question is not “what is the probability” they’re as conscious as you or I (since that’s 0.0), the question is what degree of consciousness do they have—how human-like are their pieces, arranged in how recognizable a way?
Yet after all this preamble, I’m not really sure which I’d pick to be more conscious. Perhaps for most architectures of the RL agent it’s actually less conscious, because it’s more prone to learn cheap and deceptive tricks rather than laboriously imitating the human reasoning that produces the text. But this requires us to think about how we feel about the human-ness of GPT-n, which even if it simulates humans seems like it simulates too many humans, in a way that destroys cognitive correlations present in an individual.