Just wanted to share a new paper on AI consciousness with Simon Goldstein that members of this community might be interested in. Here’s the abstract:
It is generally assumed that existing artificial systems are not phenomenally conscious, and that the construction of phenomenally conscious artificial systems would require significant technological progress if it is possible at all. We challenge this assumption by arguing that if Global Workspace Theory (GWT) — a leading scientific theory of phenomenal consciousness — is correct, then instances of one widely implemented AI architecture, the artificial language agent, might easily be made phenomenally conscious if they are not already. Along the way, we articulate an explicit methodology for thinking about how to apply scientific theories of consciousness to artificial systems and employ this methodology to arrive at a set of necessary and sufficient conditions for phenomenal consciousness according to GWT.
This is a mix of nitpicks and more in-depth comments. I just finished reading the paper and these are my notes. I liked that it was quite technical, suggesting specific changes to existing systems and precise enough to make testeable predictions.
Some discussion about the type of consciousness applied in the paper: Dehaene’s Phenomenal Consciousness And The Brain reviewed on ACX
In the introduction the verb “to token” is used two times without it being clear what that means:
The authors analyze multiple criteria for conscious Global Workspace systems and based on plausibility checks synthesize the following criteria:
In the introduction, the authors write:
Access conscious means the ability to report on perceptions. We can apply this statement in reverse as the condition that a system that isn’t access conscious can’t meet all of the above conditions.
To do so, we apply the thought experiment from the paper to test this: Young babies do not yet represent their desires and while they do show some awareness, many people would agree that they are not conscious (yet) and are likely not aware of their own existence, specifically, indepenent of their caretakers. Also, late stage dementia patients lose the ability to recognize themselves, in this model resulting from the loss of the ability to represent the self-concept in GWS. This indicates that something is missing.
Indeed, in section 7, the authors discuss the consequence that their four criteria could be fulfilled by very simple systems:
The authors get quite close to an additional criteria in their discussion:
But refrain from offering one:
As in the thought experiements above, there are readily available examples where too simple or impaired neuronal nets fail to appear conscious and to me this suggests the following criteria:
(5) To be phenomenally conscious, the system needs to have sufficient structure or learning capabilities to represent (specific or general) observations or perceptions as concepts and determine that the concept applies to the system itself.
In terms of the Smallville system discussed, this criteria may already be fulfilled by the modeling strength of GTP-3.5 and would likely be fulfilled by later LLM versions. And as required, it is not fulfilled by simple neuronal networks that can’t represent self-representation. This doesn’t rule out system with far fewer neurons than humans, e.g., by avoiding all the complexities of sense processing and interacting purely textually with simulated worlds.
Am I the only one who, upon reading the title, wondered “do they mean arguments that conscious AIs would be better than unconscious AIs, or do they mean arguments that existing AIs are conscious?”