I think this problem goes both ways. So even if we could get some kind of AI to translate the knowledge into verbal statements for us, it would be impossible, or very difficult, for anything resembling a normal human to gain “experiential knowledge” just by reading the verbal statements.
Mary isn’t a normal human. The point of the story is to explore the limites of explanation. That being the case, Mary is granted unlimited intelligence, so that whatever limits he encountes are limits of explanation, and not her own limits.
I think the most likely reason that qualia seem irreducible is because of some kind of software problem in the brain that makes it extremely difficult, if not impossible, for us to translate the sort of “experiential knowledge” found in the unconscious “black box” parts of the brain into the sort of verbal, propositional knowledge that we can communicate to other people by language. The high complexity of our minds probably compounds the difficulty even further.
Whatever is stopping Mary from understanding qualia, if you grant that she does not, is not their difficulty in relation to her abilities, as explained above. We might not be able to understand oiur qualia because we are too stupid, but Mary does notnhave that problem.
If you’re asserting that Mary does not have the software problem that makes it impossible to derive “experential knowledge” from verbal data, then the answer to the puzzle is “Yes, Mary does know what red looks like, and won’t be at all surprised. BTW the reason our intuition tells us the opposite is because our normal simulate-other-humans procedures aren’t capable of imagining that kind of architecture.”
Otherwise, simply postulating that she has unlimited intelligence is a bit of a red herring. All that means is she has a lot of verbal processing power, it doesn’t mean all bugs in her mental architecture are fixed. To follow the kernel object analogy: I can run a program on any speed of CPU, it will never be able to get a handle to a kernel redness object if it doesn’t have access to the OS API. “Intelligence” of the program isn’t a factor (this is how we’re able to run high-speed javascript in browsers without every JS program being a severe security risk).
Mary isn’t a normal human. The point of the story is to explore the limites of explanation. That being the case, Mary is granted unlimited intelligence, so that whatever limits he encountes are limits of explanation, and not her own limits.
Whatever is stopping Mary from understanding qualia, if you grant that she does not, is not their difficulty in relation to her abilities, as explained above. We might not be able to understand oiur qualia because we are too stupid, but Mary does notnhave that problem.
If you’re asserting that Mary does not have the software problem that makes it impossible to derive “experential knowledge” from verbal data, then the answer to the puzzle is “Yes, Mary does know what red looks like, and won’t be at all surprised. BTW the reason our intuition tells us the opposite is because our normal simulate-other-humans procedures aren’t capable of imagining that kind of architecture.”
Otherwise, simply postulating that she has unlimited intelligence is a bit of a red herring. All that means is she has a lot of verbal processing power, it doesn’t mean all bugs in her mental architecture are fixed. To follow the kernel object analogy: I can run a program on any speed of CPU, it will never be able to get a handle to a kernel redness object if it doesn’t have access to the OS API. “Intelligence” of the program isn’t a factor (this is how we’re able to run high-speed javascript in browsers without every JS program being a severe security risk).
If this is the case then, as I said before, my intuition that she would not understand qualia disappears.
For any value of abnormal? SHe is only quantitatively superior: she does not have brain-rewiring abilities.