Yeah, this theory definitely needs far better methodologies for testing this theory, and while I wouldn’t be surprised by at least part of the answer/solution to the Hard Problem or problems of Consciousness being that we have unnecessarily conflated various properties that occur in various humans in the word consciousness because of political/moral reasons, whereas AIs don’t automatically have all the properties of humans here, so we should create new concepts for AIs, it’s still methodologically bad.
But yes, this post at the very least relies on a theory that hasn’t been tested, and while I suspect it’s at least partially correct, the evidence in the conflationary alliances post is basically 0 evidence for the proposition.
Yeah, this theory definitely needs far better methodologies for testing this theory, and while I wouldn’t be surprised by at least part of the answer/solution to the Hard Problem or problems of Consciousness being that we have unnecessarily conflated various properties that occur in various humans in the word consciousness because of political/moral reasons, whereas AIs don’t automatically have all the properties of humans here, so we should create new concepts for AIs, it’s still methodologically bad.
But yes, this post at the very least relies on a theory that hasn’t been tested, and while I suspect it’s at least partially correct, the evidence in the conflationary alliances post is basically 0 evidence for the proposition.