Reality is weirdly normal
Related to: When Anthropomorphism Became Stupid, Reductionism, How to Convince Me That 2 + 2 = 3
“Reality is normal.” That is: Surprise, confusion, and mystery are features of maps, not of territories. If you would think like reality, cultivate outrage at yourself for failing to intuit the data, not resentment at the data for being counter-intuitive.
“Not one unusual thing has ever happened.” That is: Ours is a tight-knit and monochrome country. The cosmos is simple, tidy, lawful. “[T]here is no surprise from a causal viewpoint — no disruption of the physical order of the universe.”
“It all adds up to normality.” That is: Whatever is true of fundamental reality does not exist in a separate universe from our everyday activities. It composes those activities. The perfected description of our universe must in principle allow us to reproduce the appearances we started with.
These maxims are remedies to magical mereology, anthropocentrism, and all manner of philosophical panic. But reading too much (or too little) into them can lead seekers from the Path. For instance, they may be wrongly taken to mean that the world is obliged to validate our initial impressions or our untrained intuitions. As a further corrective, I suggest: Reality is weirdly normal. It’s “normal” in odd ways, by strange means, in surprising senses.
At the risk of vivisecting poetry, and maybe of stating the obvious, I’ll point out that the maxims mean different things by “normal”. In the first two, what’s “normal” or “usual” is the universe taken on its own terms — the cosmos as it sees itself, or as an ideally calibrated demon would see it. In the third maxim, what’s “normal” is the universe humanity perceives — though this still doesn’t identify normality with what’s believed or expected. Actually, it will take some philosophical work to articulate just what Egan’s “normality” should amount to. I’ll start with Copernicanism and reductionism, and then I’ll revisit that question.
Everything is usual. Very nearly nothing is familiar.
Spell 1: Since the beginning, not one unusual thing has ever happened. And intelligence explosion sounds… well, ‘unusual’ is the mildest word that comes to mind. So it won’t happen.
Counterspell: Reality is weirdly normal. It’s not the kind of normal that sounds normal.
Relative to the universe’s laws, everything is par for the course. But relative to human standards of normality — what I’ll call the familiar — barely anything that actually exists is par for the course. Nearly all events are senseless, bizarre, inhuman. But that’s because our mapping hardware is adapted to a very specific environment; if anything’s objectively weird, it’s we, not the other denizens of Everythingland. That’s much of why absurdity is a poor guide to probability.
Egan’s Law reminds us that beneath our human surface weirdness lies a deeper regularity, a deeper unity with the rest of Nature. But to call this alien order ‘regular’ already assumes a shift in perspective from what a person off the street would initially think of as ‘regular’, to the hidden patterns of the very large and very small. We should prize the ability to shift between these points of view, while carefully avoiding conflating them.
Spell 2: Reality is normal. So it shouldn’t be difficult for me to think like it.
Counterspell: What’s normal from a God’s-eye view is wont to be weird from a human’s. And vice versa.
Thinking like reality requires understanding reality. You can’t just will yourself into becoming a high-fidelity map. If nothing else, you’ll be too vulnerable to knowledge gaps. Pretending you think like quantum physics is even worse than rebelling against the physics for being too confusing. At least in the latter case you’ve noticed the disparity between your map and the territory.
To take pride in one’s confusion at the quantum, rather than striving mightily to understand, is an epistemic sin. But to deny one’s confusion at the quantum, to push it from one’s mind and play-act at wisdom, is a graver sin. The goal isn’t to do away with one’s confusion; it’s to do away with one’s best reasons to be confused.
What adds up to the familiar needn’t be familiar.
Spell 3: It all adds up to normality. But relativity sure sounds abnormal! So relativity will be replaced by something normaler.
Counterspell: Physics is the weird addenda, not the normal sum.
x + y + z + … = Normality. But x ≠ Normality. Nor does y. In fact, even the complete quantum-mechanical description of the human world would not look normal, from a human perspective. But it would be a description of exactly the world we live in.
This is another reminder that surprise is a feature of maps, not of territories. Very different maps — a wave function scribbled in silicon, a belief, a gesture, a child’s drawing — can represent the same territory. The drawing is normal (familiar), and the wave function isn’t. But the referent of the two maps may well be the same. In asserting “quantum mechanics adds up to normality”, we’re really asserting that our maps of familiar objects, to the extent they’re accurate, would co-refer with portions of the maps of an ideal Finished Science. They’re two different languages, but faithful translation is possible.
Spell 4: It all adds up to normality. So I should be able to readily intuit how QM yields the familiar.
Counterspell: Reality forms what’s normal, but by weird means.
The relationship between the quantum and the familiar can be thoroughly unfamiliar. The way in which the unintuitive fundamental truths yield an intuitive Middle World is not itself intuitive.
We could call this “emergence”, if we wished to craft a territory predicate out of mapstuff. Otherwise, I suggest calling this predicate”wow-sometimes-I’m-not-very-good-at-keeping-track-of-lots-of-small-things-ence”. Understanding the logical or mathematical implications of simple patterns is not humanity’s specialty. Causal and part-whole relations are no exception to this rule.
It all adds up to the phenomenon.
Spell 5: It all adds up to normality. It’s common sense that time ‘flows’. So science will ultimately show time is not at all like space.
Counterspell: We’d find even the familiar alien, if we but understood it.
Normality isn’t common sense. Normality isn’t our beliefs, expectations, assumptions, or strongest convictions. Normality, in the sense relevant to a true and binding “It all adds up to normality”, is the phenomenon, the human world as it appears. (Not the human world as it is believed to appear. The human world as we actually encounter it.)
What’s “the phenomenon”?
I’m going to be willfully cagey about that. What I really mean by “the phenomenon” is “whatever’s going on”. Or whatever’s going on that we’re cognitively accessing or representing. It needn’t all add up to a world that makes my map true; but it will add up to a world that explains why my map looks the way it does.
Even saying that much risks overly restricting the shape explanations are allowed to take. Egan’s Law can be used as a general constraint on successful explanations — they must account for the explanandum — but only if we treat the explanandum loosely enough to permit dissolutions and eliminations alongside run-of-the-mill reductions. It all adds up to an explanation of how our beliefs arose, though not necessarily a validation of them.
Spell 6: It all adds up to normality. I have an immediate epistemic acquaintance with the irreducible phenomenal character of my experience. So, whatever be the final theory, it will certainly include qualia.
Counterspell: Every map is in a territory. But no meta-map is beyond suspicion.
It’s obvious that appearances can deceive us about underlying reality. But a variety of perceptual illusions demonstrate that experiences can also consistently mislead us about themselves. It all adds up to normality, but normality lies.
In some cases, we can’t accurately describe our surface impressions until we’ve understood their underlying mechanism. Since so much of science is revisionary, we mustn’t interpret Egan’s Law in a way that unduly privileges first-pass descriptions. When data and theory conflict, it’s sometimes more likely that the data has been misrepresented than that the theory is false.
How far does this openness to map revision extend? As far as the dynamics of one’s brain allows. From the sidelines, it may appear to me that in principle a thinker should be able to have certainty about some propositions — for instance, ‘an experience is occurring’. But when I actually find myself living through such a thought, I don’t in fact experience infinite confidence. It remains physically possible for me to be persuaded otherwise.
This point ought to be a bit controversial, and more defense of it is needed. But I do insist on treating ‘phenomenal experience’ and ‘phenomenon’ as prima facie independent concepts. Egan’s Law really is the law, whereas claims about some inerrant mode of reasoning or perception will at best qualify as well-supported hypotheses.
Egan’s Law is not about saving conscious experience, or our theories, or our axioms, or our interpretations or descriptions of the phenomenon. It’s about saving the phenomenon itself — the piece of the world in which we are, in fact, submerged. Egan’s Law can be restated: The part of reality that (under its familiar description) puzzled or surprised us is identical to (or otherwise lawfully derivable from) the part of reality that (under its more fundamental description) explains it.
The human piece of the universe is of a piece with everything else. And the Everything Else gets explanatory priority. Of all our science’s findings to date, that may well be the most startling.
- Bridge Collapse: Reductionism as Engineering Problem by 18 Feb 2014 22:03 UTC; 81 points) (
- 11 Jun 2020 4:51 UTC; 4 points) 's comment on Public Static: What is Abstraction? by (
- 25 Oct 2020 20:47 UTC; 3 points) 's comment on Public Static: What is Abstraction? by (
Short version: the issue with “reality is normal” is that the English statement looks bidirectional when the concept is actually unidirectional assignment. That is, a CS way of stating it would be:
This is, as you say, the solution to a lot of philosophical panic. The reverse statement
leads to a lot of confusion, and it’s unfortunate that “reality is normal” doesn’t exclude that interpretation, because this can lead to the double illusion of transparency. It looks to me like this post is valuable because it highlights that risk and tries to counteract it.
I’m not sure that your attempt to counteract it is very effective, though. When I first read the phrase “reality is weirdly normal,” I thought you had totally missed the point. To quote from Think Like Reality:
Perhaps “reality is unintuitively normal” is closer to what you were trying to say- but again the issue is that you want reality to be intuitively normal. That’s what it means to understand that part of reality. For humans, it takes work to align your intuition with reality but it can be done, and so something like:
seems like it actively disallows that possibility (and tries to have the map influence the territory).
I’m not sure I agree with this interpretation. To go with the wiki’s explanation:
If by “as humanity sees it” you mean “observed reality,” okay, but “as humanity sees it” seems very prone to the other interpretations you want to exclude. “Observed reality” seems to exclude them, and also to concord nicely with the demonic view.
I think the rest of the post does a good job of highlighting and dispelling various confusions one could have about this concept, and I think banishing confusions is a very important part of education and communication.
I like your analogy!
There are multiple independent insights I’m trying to communicate in this post, so I don’t expect to be able to express it all in the title. I do want a title that’s relevant and memorable (to help chunk the information here), eye-catching (so people read it), and simple (so people who like non-technical stuff read it), and that doesn’t worsen the problem. I went with ‘Reality is weirdly normal’ because I figured the paradoxical appearance would make it harder to come up with a harmful pet interpretation, forcing readers to use my detailed explanations to generate reasonable semantic values for the title. Feel free to suggest other title options; I may well change it.
Yeah, that phrasing is definitely less confusing. I prefer it, but in context it doesn’t clearly contrast with the divine or demonic observation I use for the other sense of ‘normal’. How about “as humanity perceives it” in place of “as humanity sees it”? It’s still not 100% transparent, but it at least is less idiomatically tied to belief.
Very nice! You should move this to Main.
OK!
I’d like to point out the parallel to the existentialist concept of “absurdity”. From Wikipedia:
To draw out the connection: rather than saying “reality is weirdly normal”, I would say: Reality is absurd in human terms. Physics does not care about you. ‘Every action has an equal and opposite reaction’ is not a principle of justice; it has only to do with kinetic motion. Conservation of energy does not entail survival after death. The universe is allowed to kill you. Physics will happily twist everything you find meaningful beyond recognition (including you yourself).
This has a depressing aspect which is not present in the post. Perhaps that is the largest flaw of existentialism. However, I think this concept of absurdity is valuable. “Absurd” takes things one step further than “weird”, in a way I feel clarifies things.
Reality is absurdly normal.
=Edit=
The ensuing discussin has made me change my mind. The concept of the absurd is not an improvement here. Insisting that reality is absurd in human terms (strongly violating our intuitions and also our values) is not helpful; it’s relevantly similar to insisting that life is a mystery or that quantum mechanics is impossible to understand.
Sartre is my prototype for ‘existentialist’ (and he was the first person to adopt the term), so I’m usually talking about his philosophy when I say ‘existentialism’. Camus’ philosophy is the thing I’d label ‘absurdism’. The two thought they had different philosophies, though I think they mainly disagreed on emphasis and tone rather than substance. I’ve seen it put best:
“Sartre: life is meaningless, SO HUMANS CREATE MEANING.
“Camus: LIFE IS MEANINGLESS, so humans create meaning.”
Camus is also frequently mopey and mournful about the fact that we live in a godless, purposeless universe. Sartre is positively gleeful and exultant about it. (Though, admittedly, glee and exultation can look pretty scholastic and grim when you funnel it through Sartre’s authorial demeanor.) So I think of absurdism as the brand of existentialism that emphasizes our tragic helplessness to escape our world’s absurdity (the Droopy Dog Method for overcoming nihilism), while Sartrean existentialism emphasizes the mad power and freedom that comes with atheism (the Calvinball Method for overcoming nihilism). I believe Sartre avoids the flaw Eliezer’s pointing at. (Though he suffers from plenty of other flaws.)
Perhaps it is a slight digression, but I don’t see Sartre and Camus as two sides of the existential coin. I believe Camus concludes the Myth of Sisyphus with the hero happy, and concluding all is well, despite the absurdity of reality. I don’t get “Droopy Dog” from Camus.
...
I really like your first sentence...
In my head (and in the context of Camus’s absurdism) it reads like this:
“Reality is normal, though it may appear at times to be perfectly absurd. That is, absurdity, and all that it entails, is a feature of how we may come to view the map, not a characteristic of the territory itself. To win, spend your life experiencing & studying the territory, as well as working on how to draw a better (more accurate) map; don’t waste resources pouting about the ‘way the territory is’...just keeping living & studying & improving that map.”
Anyhow, interesting post. Much to consider.
This is a key reason I see Sartre and Camus as substantively in agreement, and disagreeing just in tone and emphasis; Camus’ insistence that in spite of all he’s said Sisyphus must somehow be happy, aligns him with Sartre’s view that human value can emerge even in nightmarish circumstances, via individuals’ attitudes toward their circumstances. Sisyphus’ sudden and inexplicable turn-around is the moment in the Droopy Dog cartoons when the narrative arc is suddenly broken and Droopy smashes everyone in sight, bringing things back into order by deus ex animale. (In contrast, Sartre tries to give explicit arguments for why Sisyphus’ position is a superior one.)
I would indeed call this the largest flaw of existentialism. For one thing, reality would probably seem a lot less existential!absurd in a happy intergalactic civilization, also permitted by physics.
Robby’s totally right with respect to Sartre. And a statement that doesn’t hold for Sartre doesn’t firmly stand as a statement about existentialism, I think; he’s too central. Virtually everything I’ve read of Sartre’s engenders a triumphant resolution to create the world I want to live in, and a sense that doing so is my personal responsibility. His philosophy is not all ponies and rainbows, because it’s sometimes extremely frightening, but it’s the opposite of depressing.
I’m not sure exactly when existentialism’s image switched over from Sartre’s ‘fuck yeah, life is meaningless! let’s do all the things!’ to ‘alas, life is meaningless! we are all sad and alone in the cosmos and stuff.’ I guess Camus happened. Plus Sartre is really terrible at PR.
Nope. If the Big Rip doesn’t do you in, the Heat Death of the universe will. Enjoy billions of years contemplating your inevitable demise.
Or, enjoy several weeks of Cognitive Behavioural Therapy as you train yourself out of pointless rumination followed by billions of years enjoying utopia. The universe having a trait that is not preferred does not thereby obligate you to be overwhelmed by angst.
Can I do it now and enjoy my remaining lifespan? And if so, do I really need intergalactic civilization?
Yes. By all means.
Cognitive Behavioural Therapy isn’t a cure for death. Yes, the intergalactic civilisation and the accompanying technology sounds rather important.
Intergalactic civilization isn’t a cure for death, either.
The simplistic view is that 50 years of life is, all else being equal, less desirable than 50 billion years of life. As it happens 50 billion years of life is actually better than 50 years of similar quality life. This isn’t a complicated or deep issue.
This isn’t what I’m arguing against here. Of course 50 billion years is better, ceteris paribus.
50 billion years still remain absurdly small compared to eternity.
What I am arguing against are these particular comments.
If the clever rhetorical questions you ask are rhetorically wrong then they remain wrong even if you have some good point that you want to ultimately support. I know this is a peculiar cultural norm to go by but it’s a valuable one.
Quite a lot smaller, yes.
The interesting question to ask here is what chance of eternal life we would be willing to exchange for a given 1% chance of 50 billion years of life in terms of hypothetical decision making. For instance someone who valued years of life linearly would, given the chance, dedicate the cosmic commons to computronium dedicated to optimistic matrix hacking even if almost completely certain there is no escape possible. I don’t value life years linearly and I’m quite sure that there are finite numbers of life years that I would prefer over chances of eternal life but I don’t know quite what the equivalence function looks like.
I’d like to take a million years or two to fully study the issue before I accept that my demise is inevitable.
I would still be curious how much I can get out of life in billions of years.
You are curious now. You might become less curious after the first billion years. :-)
PS. I’m not advocating deathism. Just saying that any period of time is insignificant, when compared to a much larger period of time. The billion-year happy intergalactic civilization is still absurdly tiny compared to the eternity of Boltzmann-brain-infested Darkness that follows...
PPS. Or we could just taboo the word “absurd”.
As a function of how long the universe will exist? ETA: a short period of time might be significantly located.
This reminds of a website where you can check if the Earth has been destroyed.
International Earth-Destruction Advisory Board.
See also: http://hasthelargehadroncolliderdestroyedtheworldyet.com/
Thanks. Now we need one to check the existence of the entire universe. Of course, the site will have to be hosted off-universe.
The absurd claim is “there is nothing you ought to do or ought to not do”. The claim “life is tough” is not absurd. ETA: existentialism in the absurdist flavor (as opposed to for example the Christian flavor) is a form of value anti-realism which is not nihilism. It denies that there are values that could guide choices, but puts intrinsic value into making choices.
Dude, this post is amazing (though I could probably articulate (or resolve in the process) a few quibbles I have with it, if I had time for that)! You should seriously consider rewriting some of Eliezer’s stuff.
I noticed that “reality is normal” and “reality is weirdly normal” both commit the mind projection fallacy. However, this is OK (though probably not optimal) if the reader has a technical understanding of what you mean (which is more likely to be true for someone who is reading this than for someone who is not).
That being said, I have a challenge for you. Since I think that this exposition is better than Eliezer’s, I challenge you to rewrite this post without any direct dependencies on his work (of course, you will have to formally cite any pages of his whose ideas you consult). Write it so that a competent person who has never heard of Less Wrong could understand it in as close to a technical way as is possible with as abstract a concept as this (and you’ve inspired me to try the same!).
I can feel the void.
Thanks, notsonewuser! Do tell me some of your quibbles if you have time.
Perhaps I’ll add a cautionary link to the Mind Projection Fallacy where I explicitly allude to this—“the cosmos as it sees itself”.
I like your challenge a lot! I’ll rewrite this for a more general audience and post it to my blog. Possibly as two separate posts, the first introducing ‘Reality is normal’ and the unusualness koan, the second discussing Egan’s Law. I’ll try to remove ‘dependencies’ in the sense of prerequisites for comprehension, but I don’t think I’ll try to remove the large number of intellectual connections to what Eliezer’s written. What Eliezer’s written is too useful, and part of my reason to make this more accessible is so new people will check out the Sequence posts I link.
That sounds to me like, “Some events appear improbable to us, but nothing that actually occurs was improbable.” That might be correct in some multiverse universes, but is trivially and dramatically false in non-multiverse universes. Some people argue that the impact of highly-improbable events dominates over the impact of common, highly-probable events.
You need to distinguish between surprise about events, and surprise about models. Egan’s law should only cover surprise about your models. I can be surprised that you won the lottery without needing to change my models.
I’m glad you raised this issue, because I don’t think it’s a simple task to unpack ‘not one unusual thing has ever happened’ in a way that is neither trivial nor false. (It’s also quite difficult to do this with Egan’s Law.)
A trivial reading of the usualness koan: ‘Since the beginning, nothing that never happens has ever happened.’ (Here ‘unusual’ means ‘violating a Law of Nature, in the sense of violating a true generalization about the universe’.)
A false reading of the usualness koan: ‘Since the beginning, nothing that infrequently happens has ever happened.’
A non-trivial and true reading will be somewhat sophisticated, and will (I think) narrow down what sort of universe we live in a lot more than ‘reality is normal’ or Egan’s Law do. What the koan means is that we live in a cosmos whose structure and dynamics are determined by
a short list of
simple
exceptionless rules
that are uniform across space and time
and deterministic. (Or, more strongly, locally deterministic, allowing one to use spacetime regions to predict the properties of their neighbors.)
A universe in which (cosmically) “unusual” things sometimes happen would be one where the rules vary substantially across spacetime regions, where things sometimes happen for no reason (indeterminism), or where the universal rules are extremely convoluted and gerrymandered.
Interestingly, there is one candidate event for a cosmically unusual thing that happened in our own universe. To our knowledge, only one unusual thing has ever happened — at t=0, a state of low entropy occurred. But since then, every spacetime region has followed the same rules as every other spacetime region. Knowing that our universe is lawful, and knowing everything about any two contiguous spacetime regions, would together (I think) allow one to immediately infer the causal dynamics of all other spacetime regions in the observable universe.
How do you know the list is short and the rules are simple?
What do the words “short” and “simple” mean here?
I don’t think that most high-complexity algorithms for building a life-permitting observable universe would allow a theory as simple as human physics to approximate the algorithm as well as our observable universe does.
Do you think the observable universe is a lot more complicated than it appears?
This is trivially false. Imagine, for the sake of argument, that there is a short, simple set of rules for building a life permitting observable universe. Now add an arbitrary, small, highly complex perturbation to that set of rules. Voila, infinitely many high complexity algorithms which can be well-approximated by low complexity algorithms.
How does demonstrating ‘infinitely many algorithms have property X’ help falsify ‘most algorithms lack property X’? Infinitely many integers end with the string …30811, but that does nothing to suggest that most integers do.
Maybe most random life-permitting algorithms beyond a certain level of complexity have lawful regions where all one’s immediate observations are predictable by simple rules. But in that case I’d want to know the proportion of observers in such universes that are lucky enough to end up in an island of simplicity. (As opposed to being, say, Boltzmann brains.)
The observable universe is enormously complicated, not in its rules but in its configuration (“indexical” complexity = complexity).
I have no idea what these algorithms might be are and neither do you. Accordingly I don’t see any basis for speculating what will they allow.
I think the observable universe appears to be very complicated.
I am still interested in what do you mean by “short” and “simple”. The default rule is that “man is the measure of all things” so presumably you are using these words in the context of what is short and simple for the human brain.
Requiring the universe to be constructed in a way that is short and simple for the brains of a single species on a planet in a provincial star system in some galaxy seems to be carrying the anthropic principle a bit too far.
Well, let’s think about whether we have a proof of concept. What’s an example of a generalization about high-complexity algorithms that might show most of them to be easily usefully compressed, for an observer living inside one? At this point it’s OK if we don’t know that the generalization holds; I just want to know what it could even look like to discover that a universe that looks like ours (as opposed to, say, one that looks like a patchwork or a Boltzmann Braintopia) is the norm for high-complexity sapience-permitting worlds.
ETA: Since most conceivable universes are very very complicated, I’d agree that we probably live in a very very complicated universe, if it could be shown that our empirical data doesn’t strongly support nomic simplicity.
No, I’m saying it’s short and simple relative to the number of ways a universe could be, and short and simple relative to the number of ways a life-bearing universe could be. There’s no upper bound on how complicated a universe could in principle be, but there is a lower bound, and our physics is, even in human terms, not far off from that lower bound.
Humans have a preference for simple laws because those are the ones we can understand and reason about. The history of physics is a history of coming up with gradually more complex laws that are better approximations to reality.
Why not expect this trend to continue with our best model of reality becoming more and more complex?
In this case I can apply the “short & simple” descriptor to anything at all in the observable universe. That makes it not very useful.
Uh, most of these are demonstrably false, so the koan does not seem overly useful.
The list of rules has been growing awfully quick, and there is no guarantee that it is finite. Here is the latest list of the basic rules, already quite compressed. And it does not even describe any macroscopic phenomena, which have their own rules, some more ad hoc than others. Thus there is no indication that “the territory” can be fully described by a map with “a short list of simple rules”, though some subsets of it certainly can.
If you think that the above is simple, I shudder to think what you consider complex.
“exceptionless rules” is either vacuous or false. We observe plenty of phenomena we don’t know the rules for. If the koan means that “but the rules are still there even for these exceptions, we just don’t have them on the map yet”, this meta-model seems to contradict 1.
Uniformity across space and time does seem to apply, roughly, to the observed universe, but plenty of models in the modern high-energy physics suggest that the laws and values we observe and deduce might be an accident of some local false vacuum state, or just one of many options in the chaotically inflating multiverse.
Local determinism is only saved in QM by postulating that “everything possible happens, even if we can never observe it”, and even this locality is severely challenged by the EPR/Bell. Even if “the cosmos” were deterministic, it is still not necessarily predictable, both due to chaotic effects and due to potential inherent Knightean uncertainty related to the uninteracted parts of the Big Bang.
Anyway, I agree that the “weird rules” are there first to explain the (weird and non-weird) observations, but I disagree with the narrow interpretation in the wiki
The purpose of the rules is actually to change reality, at least as we perceive it. While we are not (yet) able to change some “fundamental” laws, we certainly affect reality by learning (and changing) some of the others.
Nearly all possible lists of rules would be too lengthy and complex to be encoded in a space the size of the observable universe. By comparison, our universe’s rules seem likely to be able to be written on a t-shirt or two, especially when you consider how many of the rules are structurally similar. So, yes, I consider that simple. Simple to build into a universe simulator, if not simple for a human to inutit. (I’m not sure what you mean by “macroscopic phenomena… have their own rules”.)
I shudder too! There may be simpler possible rules, but the number of sapience-permitting rules much simpler than ours is probably very small. (And it’s certainly vanishingly small relative to the number of sapience-permitting rules more complex than ours.)
For now, we can combine ‘exceptionless’ with ‘uniform across space and time’, unless someone has a thought about how to distinguish the two.
Yes. My expectation is that we live in a bubble of simplicity within a more complex structure. I think the koan is meant to be a generalization about our own observable universe (hence its temporal character), not speculation about our world’s metaphyical substrate. Though it’s obviously at least a clue.
I agree there are costs to saving local determinism (and serious unsolved questions in the neighborhood), but it’s still an extremely plausible model. And when you combine it with non-local determinism, we’ll have accounted for all the plausible hypotheses. Determinism rather than locality is the point I want to emphasize, since it only requires that we deny Collapse interpretations.
EPR doesn’t challenge MWI-style local determinism. (Though it does limit the usefulness of that knowledge, since we don’t know which part of the wave function we’re in.)
I’m not seeing why that disagrees with the wiki. One goal is just more proximate than the other, and more specific to the case at hand. The purpose of a hammer is to improve human life; but the purpose of a hammer is also to put nails in stuff.
The lagrangian in that PDF is about three transformations away from the most compact specification of the SM, which would be “the most general renormalizable field theory with these local symmetries and with fermions transforming in these representations”. If you then wrote down the lagrangian immediately implied by that specification, then changed field variables to incorporate the effects of the Higgs, and finally chose a gauge and included “ghost” fields, then you could get that long expression.
That doesn’t appear to be a list, or a rule, or anything meaningful, since it has no equality or inequality symbols.
Your main point seem basically correct. I think RobbBB is trying to get at something meaningful, but is heading in the wrong direction with his demand for definitive, exceptionless, deterministic rules. It’s all about information, and information accommodates exceptions and non-determinism.
Yeah, sorry, this is just the Lagrangian of the Standard Model of particle physics, it’s used to calculate probability amplitudes. That’s where you get the equal sign.
As far as I can tell, you are talking about Qualia, the subjective experience. Don’t be shy—we are allowed to talk about it.
Qualia are a candidate for being part (or all) of the phenomenon. Egan’s Law doesn’t rule out the possibility that eliminativism is true, and qualia are an illusion. But it does insist that if they’re an illusion (the sort of illusion zombie Chalmers suffers from in zombie world), the physical basis for the illusion is then the normality that our finished theory has to reproduce and explain.
I’m not setting qualia to one side because I think they’re a bad topic for discussion. I’m setting them to one side because they’re less metaphysically innocent than the ‘normality’ of a (true and binding) ‘It all adds up to normality’, less metaphysically innocent than the ‘phenomenon’ of a (true and binding) ‘Save the phenomenon’.
I wasn’t aware of eliminativism. After reading the wikipedia page, eliminativism seems to be nothing more than reductionism applied to the philosophy of mind, but I don’t see what problems reductionism poses for qualia. I don’t think that I’m missing a logical step, nor do I feel confused on this genre of philosophical issues. So if many people perceive that qualia and reductionism are incompatible, my current hypothesis is that “qualia’ has some sort of definitional connotation attached to it that I’m not aware of, which somehow interferes with reductionism. I’d like to be informed about these connotations.
I guess “experience” is the most innocent word? Honestly, “Qualia”, “(Subjective) experience”, and “(save the) Phenomenon” all seem precisely identical to me, and I only use “qualia” because it’s short and doesn’t have any other definitions. If it’s picked up additional connotations, then I’ll have to find a new label for what I will temporarily call “Property Q, which separates counterfactual mathematical structures from reality”.
Reductionism says there is some thing existing X which is composed of, undestandable in terms of, and ultimately identical to some other existing thing Y. ELiminativism says X doesn’t exist. Heat has been reduced, phlogiston has been eliminated.
Qualia are the specific qualitative properties of our experience, as accessed by first-person introspection. The raw redness of the red you notice in your field of vision, for instance, as contrasted with the functional state of visually detecting light of wavelength 620–740 nm that an alien building a cognitive model of your behavior might initially construct. (Which is not to say that those two properties are non-identical. But if the two are identical, this must be discovered, not just stipulated.)
There are several popular arguments for the irreducibility of qualia (and, more generally, against the reliability of introspection as a method for directly reading off part of our world’s ontology), which has made them controversial posits.
In a conversation about mindstuff, what Eliezer calls ‘Reductionism’ is what I’d call ‘physicalism’ — the doctrine that the mental (unlike physicsy stuff) is non-fundamental, that it can in principle be fully explained in non-mental terms. The ‘reductionism’ I think about (which I’ll distinguish by making it minuscule) is the more specific doctrine that mental stuff — in this case, phenomenal properties, qualia — exists and is reducible. So a physicalist has to either eliminate or reduce every mental posit.
Eliminativists might insist that we reject qualia because the evidence for them is strictly first-person and introspective (phenomenological) rather than sensory and in-the-world (scientific). Or an eliminativist might think that we have strong prior grounds for accepting physicalism, but that reductionism is a doomed project, say, because of the Mary’s Room argument.
...what exactly is the difference between these two things? Observation, evidence, etc… are qualia.
If you are saying that problem of qualia is in the realm of neuroscience, I think this is the wrong way to go about it. Neuroscience might make some things more obvious—as it has done with free will—but the answer has been available to us all along. We need not look at an actual brain...we don’t even need to look at an actual universe. These sorts of answers hold true no matter what sort of universe you are in, so the word “exists” is inappropriate—you wouldn’t say that Bayes theorem “exists”, would you?
Qualia (or property Q) is the property that makes things Real in the first place. Qualia is the property that separates the actual universe from the range of all possible universes. All of science and epistemic rationality is an attempt to create models that describe qualia.
Introspection is a perception-like, reflective ‘inner sense’. It’s modeling your mental states as mental states, by a relatively non-inferential process. Sensory perception is vision, taste, etc.
But there are parts of the universe no one is experiencing. What does it mean to say that there are qualia (properties of experience) there, if there aren’t any subjects of experience to be found?
Do you have in mind something like Russell’s The relation of sense-data to physics?
Don’t all properties distinguish actual universes from non-actual ones? If something isn’t actually a bowling ball, then it isn’t a bowling ball. I don’t see what work qualia is doing here, or what problem it’s solving.
Yeah, that’s what the words mean, but really, what’s the difference between the two? There’s no sharp distinction between sensory and introspective perception.
1) Well firstly, qualia are only for you, not for others—I’m pseudo-solopsist that way. But to answer the question: qualia are the Known parts of Reality, and everything else is the Unknown. You can guess at the Unknown using Bayes theorem, and stuff.
2) Maybe? He’s being a bit long winded and I’m having trouble ascertaining the main point without spending too much time.
Yes, this is what I mean by qualia being knowns and everything else being unknowns
I’m not sure where he is going with that one as there are many possible interpretations as to what he means. I do think that models are only candidates for being True if they output the correct qualia—and all such models that fit this criteria are then ranked from lowest to highest complexity.
3) No. The counterfactual universe y=3x+7 has properties: the slope is 3, it contains the point (0, 7) and so on. But our reality is not merely “y=3x+7”. You know this because you are experiencing things that are decidedly not in the universe “y=3x+7″. In the same vein, the counterfactual universe where the sky is green is not real either, and you know this for the same reason.
If your qualia consisted of [(-1, 4), (0, 7), (1, 10)] and so on, with a new point appearing every second, you might consider “y=3x+7 with x=x+1 every second” as a practical model for your universe (although not a complete model—you still need to explain the existence of your thoughts). But that’s not our observation...hence, that’s not a candidate model of what our universe looks like.
To put it in different terms: the thing that distinguishes reality from theoretical alternatives is observation.
I agree with most of this, although I am not sure that the way strawberries taste to me is a posit.
I’ve never gotten this either. It has always seemed to me that qualia exist, and that they can fully be explained by reductionism and physicalism (presumably as some sort of function of our nervous system interacting with stimuli). There are apparently some people who have a strong intuition that they can’t be explained in such a fashion, but I do not share this intuition.
It seems to me that attempting to eliminate qualia is a repeat of the comedy of behaviorism. “All these mystical people claim that qualia can’t be explained by physics, so I’ll say qualia don’t exist at all! That’ll show ’em!”
(At his blog Eric S. Raymond wrote an article arguing that qualia are probably the sensation one feels when one’s stimuli processing systems light up, and that attempting to eliminate them is silly).
I think I may write a sequence about this. I’ve noticed that there are a lot more LW posts trying to solve the Hard Problem (or insisting that it’s a pseudo-problem) than trying to explain what ‘Hard Problem’ means in the first place, or trying to state it precisely.
Thus I see a lot of people insisting that the Hard Problem either isn’t a problem, or isn’t hard, without investing any time into steel-manning (or even reading) the Other Side. Eliezer, actually, is one of the few LWers I’ve seen who generally grants that it’s both hard and a problem.
A sequence that spent more time trying to figure out what the problem is, and what methodology is appropriate for such a strange topic, might also be more domain-generally useful than one that leaps straight into picking the best solutions (or mocking the worst),
Do you understand exactly why they have the intuition, and what their intuition amounts to?
That may be true for eliminativists who are behaviorists, like perhaps Dennett. But it’s not true for eliminativists who acknowledge that introspective evidence is admissible evidence, and just deny that the evidence for qualia outweighs the evidence for the conjunction ‘physicalism is true, and phenomenal reductionism is false’.
If you can’t regenerate the reasons people disagree with you—if you’re still at the stage where the opposing side purely sounds like a silly caricature, with no coherent supporting arguments—then you should have low confidence that you know their positions’ strong and weak points.
Can you point me to such an explanation??
There’s actually one in that essay I linked to at the end of my post. Here is the most relevant paragraph (discussing the Mary’s Room problem):
Reading Wikipedia’s entry on qualia, it seems to me that most of the arguments that qualia can’t be explained by reductionism are powered by the same intuition that makes us think that you can give someone superpowers without changing them in any other way. Anyone with a basic knowledge of physiology knows the idea you can give someone the powers of Spider-Man or Aquaman without changing their physical appearance or internal anatomy is silly. Modern superhero writers have actually been forced to acknowledge this by occasionally referencing ways that such characters are physically different from humans (in ways that don’t cosmetically affect them, of course).
But because qualia are a property of our brain’s interaction with external stimuli, rather than a property of our bodies, the idea that you could change someone’s qualia without changing their brain or the external world fails to pass our nonsense detector. If I wake up and the spectrum is inverted, something is wrong with my brain, or something is wrong with the world.
That isn’t a reductive explanaiton, becuase no attempt is made to show how Mary;s red quale breaks down into smaller component parts. In fact, it doens;t do much more than say subjectivity exists, and occurs in sync with brain states. As such, it is compatible with dualism.
You mean p-zombie arguments?
Whatever,tThat doesn;t actuall provide an explanation of qualia.
I presume that would be “Mary’s qualia are caused by the feeling the color-processing pathways of her brain light up. The color processing parts are made of neurons, which are made of molecules, which are made of atoms. Those parts of the brain are then connected to another part of the brain by more neurons, which are similarly composed. When those color processing parts fire this causes the connecting neurons to fire in a certain pattern. These patterns of firings are what her feelings are made of. Feelings are made out of firing neurons, which are in turn made out of atoms.”
I don’t get the appeal of dualism. Qualia can’t run on machines made out of atoms and quarks, but there is some other mysterious substance that composes our mind, and qualia can run on machines made out of this substance? Why the extra step? Why not assume that atoms and quarks are the substrate that qualia run on? What hypothetical special properties does this substance have that let qualia run on it, but not on atoms?
I’m sure that if we ever did discover some sort of disembodied soul made out of a weird previously unknown substance that was attached to the brain and appeared to contain our consciousness, Dave Chalmers would argue that qualia couldn’t possibly be reduced down to something as basic as [newly discovered substance], and that obviously this disembodied soul couldn’t possibly contain consciousness, that has to be contained somewhere else. There is no possible substance, no possible anything, that could ever satisfy the dualist’s intuitions.
Yes, plus the inverted spectrum argument, and all the other “conceivability arguments.” I can conceive of myself walking on walls, bench-pressing semi-trucks, and flying without making any modifications to my body or changing the external world. But that’s because my brain is bad at conceiving stuff and fudges using shortcuts. If I actually start thinking in extremely detailed terms of my muscle tissues and the laws of physics, it becomes obvious that you can’t conceive of such a thing.
If anyone argued “I can imagine an anorexic person with almost no muscles lifting a truck, therefore strength cannot be caused by one’s muscles,” they would be laughed at. P-zombies and inverted spectrums deserve similar ridicule.
A claim that some X is made of some Y is not showing how X’s are made of Y’s. Can you explain why red is produced and not soemthing other.
I wasn’t selling dualism, was noting that ESR’s account is not particualrly phsycialist as well as being not particularly explanatory,
I find the Mary argument more convincing.
There are many different neuron firing patterns. Some produce various shades of red, other produce other stuff.
The intuition that Mary’s Room activates is that no amount of book-learning can substitute for firsthand experience. This is because we can’t always use knowledge we obtain from reading about experiences to activate the same neurons that having those experiences would activate. The only way to activate them and experience those feeling is to have the activating experience.
Now, in Dennett’s RoboMary variation of the experience, RoboMary would probably not say “Wow!” That is because she is capable of constructing a brain emulator of herself seeing red inside her own head, and then transferring the knowledge of what those neurons (or circuits in this case) felt when activated. She already knows what seeing red feels like, even though she’s never seen it.
The dualist says: ‘I imagine Mary the color-blind learning all the scientific facts about color vision, including the fine neurological details, and correctly drawing any relevant inferences from these facts. Yet when I imagine Mary seeing red for herself for the first time, it seems to me that she would think that further epistemically open possibilities have been ruled out, that were previously open. There seemed to be more than one candidate subjective character red-detecting brain states could add up to, and learning “oh, that’s what red feels like!” narrowed down the model further.’
Since Mary is color-omniscient, some explanation then is needed for why she would harbor this false belief, or for why she wouldn’t really think that the first-hand experience had further narrowed down the experiential possibilities for her.
Saying ‘she hadn’t instantiated the property X’ doesn’t explain why anyone has this intuition, because in nearly all cases it’s possible to understand and expect properties without instantiating them oneself. If Mary were a volcanologist, there wouldn’t be some factual information she’s missing by virtue of not having her brain instantiate all the properties of a volcano. What is it about certain mental properties that makes them relevantly different?
Mary isn’t really color omniscient. The thought experiment has the hidden false assumption that human beings can learn all types of knowledge by study alone. Since we all know this isn’t true, when we hear “Mary knows everything about color” our brain translate that into “Mary knows everything about color that one can learn by studying.” Our intuitions about whether she’ll say “Wow” or not are based on this translation.
jimrandoh makes a similar point:
In other words, Mary can’t figure out what qualia feel like because she is using the “linguistic” program and needs to use the “visual processing one.” It’s like trying to do a slideshow using Notepad instead of Powerpoint.
Because human brains are much more complicated than volcanoes. Humans are only capable of assimilating so much prepositional knowledge and we are severely limited in our ability to convert it into other types of knowledge. orthonormal makes this point when explaining qualia.
Now, you could make Mary a superhuman creature that can assimilate vast amounts of knowledge and control and restructure her brain any way she wants. But if this assumption is made explicit my intuition that she would say “Wow!” when she goes outside disappears. A superhuman creature like that probably could figure out how seeing red felt without ever seeing it.
That’s another major problem with Mary’s Room. It posits a superhuman creature, capable of feats of learning and knowledge no human can achieve, but downplays that fact so that our intuitions are still conditioned to act like Mary is human.
Should probably be “propositional”.
Ex hypothesi, Mary knows all the relevant third-person specifiable color facts. Our inability to simulate her well doesn’t change that fact. If you’re saying there are some physical facts it’s physically possible for any agent to be able to figure out scientifically in principle, then you’ll need to explain why.
But the intuition isn’t that Mary would acquire the ability to recognize red objects for the first time. It’s that she’d learn new facts about what redness feels like. Consider the Marianna variant:
“Like Mary, Marianna first (at t1) lives in a black and white environment. Contrary to Mary (at a later moment t2) she gets acquainted with colors by seeing arbitrarily colored objects (abstract paintings, red chairs, blue tables, etc. but no yellow bananas, no pictures of landscapes with a blue sky etc.). Marianna is therefore unable to relate the kinds of color experiences she now is acquainted with to what she already knew about them at t1. At t2, Marianna may wonder which of four slides (a red, a blue, a green and a yellow slide) appears to her in the color normal people experience when looking at the cloudless sky. At t2 Marianna knows, in a sense, what it is like to have experiences of red, blue, etc. But she still lacks the relevant items of knowledge about what other people experience: there is a clear sense in which she still may not know that the sky appears blue to normal perceivers, she may even have the false believe that it appears to normal perceivers like the red slide appears to her and thus believe, in a sense, that the sky appears red to normal perceivers. Only at t3, when Marianna is finally released and sees the sky, does she gain this item of knowledge.”
Everything about volcanoes can be translated into a linguistic program, without information loss. Why can’t everything about visual processing by translated into a linguistic program without loss? If it’s merely a matter of qualia being complicated, then shouldn’t all other complicated systems yield relevantly identical Hard Problem intuitions? E.g., shouldn’t the planet Mars appear irreducible and ineffable and unphysical?
My intuition is that making Mary superhuman doesn’t change that experiencing red seems to narrow down the possibilities for her. Analogously, a superhuman wouldn’t be able to scientifically narrow down the possibilities for what it’s like to be a bat to a single model, without generating bat-experiences of its own. Can you explain why this intuition persists for me, when (as far as I can tell) it doesn’t for any other complex system?
Maybe, but in that case the challenge is to explain, at least schematically, what superhuman power Mary obtains that lets her solve the Hard Problem. Mere increased processing power alone doesn’t seem to dissolve the problem.
It does if our inability to simulate her well messes with our intuitions. If, as I conjectured, we tend to translate “omniscient person” with “scholar with lots of book-learning” then our intuitions will reflect that, and will hence be wrong.
Is Marianna omniscient about light and neuroscience like Mary? If she is, she’d be able to figure out which color is which fairly easily.
It’s not just a matter of qualia being complicated, it’s a matter of the human brain being bad at communicating certain things, of which qualia are only one thing of many. And this isn’t just an issue of processing power and the complexity of something being processed, it’s an issue of software problems. There are certain problems we have trouble processing regardless of what level of power we have, because of our mind’s internal architecture. Wei Dei puts it well when he says:
Furthermore, there are in fact other things that humans have a lot of difficulty communicating besides qualia. For instance, it’s common knowledge that people with a few days of job experience are much better at doing jobs than people who have spent months reading about the job.
I disagree. If Mary was a superhuman she could study what functions of the brain cause us to experience “qualia,” and then study the memories these processes generated. She could then generate such memories in her own brain, giving her the knowledge of what qualia feel like without ever experiencing them. She would see red and not be surprised at all.
If qualia were not a physical part of the brain, duplicating the memories of someone who had experienced them would not have this effect. However, I think it very likely that doing so would have this effect.
Because, as I said before, our emotions are “black boxes” that humans are very bad at understanding and explaining. Their Kolmogorov complexity is extraordinarily high, but we feel like they are simple because of our familiarity with them.
I think the ability to study and modify her own source code and memory, as well as the source code and memory of others is probably all she’d need, but I could be wrong.
You… disagree? Do you mean your own intuition is different, or do you mean you have some special insight into my psychology that tells you that I’m misunderstanding or misrepresenting my own intuitions?
I’m reporting on psychological data about what my intuitions are indicating to me. I’m not a dualist, so I’m not (yet) making any assertions about what Mary would actually do or say or know. I’m explaining what output my simulator is giving me when I run the thought experiment.
You’re assuming that all superhumans intelligent enough to understand the biophysics of color vision will also necessarily have a module that allows them to self-modify in a way that they have whatever first-person subjective experience they wish. There’s no reason to assume that. As long as a Mary without this capacity (but with the third-person biophysics-comprehending capacity) is possible, the argument goes through. The fact that a Mary that can spontaneously generate its own experience of redness is also possible doesn’t make any progress toward refuting or dissolving the Mary hunch.
It sounds to me like you’ve been reading too much Dennett. Dennett is not a careful or patient dissector of the Hard Problem. The entire RoboMary paper, for instance, is a non sequitur in relation to the arguments it’s meant to refute. It’s fun and interesting, but it’s talking about a different subject matter.
That’s not true at all. Most forms of dualism allow Mary to generate the relevant mental states by manipulating the physical states they are causally tied to.
This doesn’t look to me like an explanation yet, even an outline of one. In fact, it looks like an appeal to the Black Box black box: ‘Black box’ is being used as a special word meant to pick out some uniquely important and effective category of Unknown Thingie. But just saying ‘we don’t understand emotions yet’ doesn’t tell me anything about why emotions appear irreducible to me, while other things I don’t understand do seem reducible to me.
I don’t feel that mental states are simple! Yet the Mary hunch persists. You seem to be hopping back and forth between the explanations ‘qualia seem irreducible because we don’t know enough about them yet’ and ‘qualia seem irreducible because we don’t realize how complicated they are’. But neither of these explanations makes me any less confused, and they’re both incredibly vague. I think this is a legitimate place to insist that we say not “complexity”.
Why, specifically, would any of those four abilities help? Are all four needed? Are some more important than others? Why, for instance, wouldn’t just studying my own source code and memory (without being able to do radical surgery on it) suffice for knowing the phenomenal character of redness, or the phenomenal character of a bat’s echolocation...?
I mean my intuition is different.
Alright, I’ll try to stop hopping and nail down what I’m saying:
I think the most likely reason that qualia seem irreducible is because of some kind of software problem in the brain that makes it extremely difficult, if not impossible, for us to translate the sort of “experiential knowledge” found in the unconscious “black box” parts of the brain into the sort of verbal, propositional knowledge that we can communicate to other people by language. The high complexity of our minds probably compounds the difficulty even further.
I think this problem goes both ways. So even if we could get some kind of AI to translate the knowledge into verbal statements for us, it would be impossible, or very difficult, for anything resembling a normal human to gain “experiential knowledge” just by reading the verbal statements.
In addition to making qualia seem irreducible, this phenomenon explains other things, such as the fact that many activities are easier to learn to do by experience.
I’ve never actually read any Denett, except for short summaries of some of his criticisms written by other people. One person who has influenced me a lot is Thomas Sowell, who frequently argues that the most important knowledge is implicit and extremely difficult, if not impossible, to articulate into verbal form. He does this in terms of economics, but when I started reading about the ineffability of qualia I immediately began to think “This probably has a similar explanation.”
Mary isn’t a normal human. The point of the story is to explore the limites of explanation. That being the case, Mary is granted unlimited intelligence, so that whatever limits he encountes are limits of explanation, and not her own limits.
Whatever is stopping Mary from understanding qualia, if you grant that she does not, is not their difficulty in relation to her abilities, as explained above. We might not be able to understand oiur qualia because we are too stupid, but Mary does notnhave that problem.
If you’re asserting that Mary does not have the software problem that makes it impossible to derive “experential knowledge” from verbal data, then the answer to the puzzle is “Yes, Mary does know what red looks like, and won’t be at all surprised. BTW the reason our intuition tells us the opposite is because our normal simulate-other-humans procedures aren’t capable of imagining that kind of architecture.”
Otherwise, simply postulating that she has unlimited intelligence is a bit of a red herring. All that means is she has a lot of verbal processing power, it doesn’t mean all bugs in her mental architecture are fixed. To follow the kernel object analogy: I can run a program on any speed of CPU, it will never be able to get a handle to a kernel redness object if it doesn’t have access to the OS API. “Intelligence” of the program isn’t a factor (this is how we’re able to run high-speed javascript in browsers without every JS program being a severe security risk).
If this is the case then, as I said before, my intuition that she would not understand qualia disappears.
For any value of abnormal? SHe is only quantitatively superior: she does not have brain-rewiring abilities.
‘Ex hypothesi’ here means ‘by stipulation’ or ‘by the terms of the conditional argument’. The assumption is ‘Mary is a color scientist who knows all the relevant facts about color vision, but has never experienced color in her own visual field’. You aren’t denying that this is a consistent, coherent hypothetical. All you’re suggesting is that a being that satisfied this hypothetical would have transhuman or posthuman capacities for data storage and manipulation. So far so good.
You then insist that such a being, if it acquired color vision, would be completely unsurprised by the particular shade of red it now (for the first time) encounters; whereas the dualist insists in that situation the transhuman would learn a new fact, would acquire new, possibilities-ruling-out information. (A sentient supercomputer without the capacity to experience color would run into the exact same trouble.)
Up to that point, the two of you remain in a stalemate. (Or worse than a stalemate, from your perspective, since you find it baffling that anyone could share the dualist’s intuitions or reasoning, whereas the dualist perfectly well understands the intuitive force of your argument, and just doesn’t think it’s strong enough.)
So you assert. The goal here isn’t to just repeatedly assert, in various permutations, that dualists are wrong. The goal is to figure out why they think as they do, so we can dissolve the question. Swap out ‘free will’ for ‘irreducible qualia’ in Eliezer’s recommendation:
“It is a fact about human psychology that people think they have free will. Finding a more defensible philosophical position doesn’t change, or explain, that psychological fact. Philosophy may lead you to reject the concept, but rejecting a concept is not the same as understanding the cognitive algorithms behind it. [...]
“The key idea of the heuristics and biases program is that the mistakes we make, often reveal far more about our underlying cognitive algorithms than our correct answers [...] But once you understand in detail how your brain generates the feeling of the question [...] then you’re done. Then there’s no lingering feeling of confusion, no vague sense of dissatisfaction.
“If there is any lingering feeling of a remaining unanswered question, or of having been fast-talked into something, then this is a sign that you have not dissolved the question. A vague dissatisfaction should be as much warning as a shout. Really dissolving the question doesn’t leave anything behind.
“A triumphant thundering refutation of free will, an absolutely unarguable proof that free will cannot exist, feels very satisfying—a grand cheer for the home team. And so you may not notice that—as a point of cognitive science—you do not have a full and satisfactory descriptive explanation of how each intuitive sensation arises, point by point.
“You may not even want to admit your ignorance, of this point of cognitive science, because that would feel like a score against Your Team. In the midst of smashing all foolish beliefs of free will, it would seem like a concession to the opposing side to concede that you’ve left anything unexplained.
’And so, perhaps, you’ll come up with a just-so evolutionary-psychological argument that hunter-gatherers who believed in free will, were more likely to take a positive outlook on life, and so outreproduce other hunter-gatherers—to give one example of a completely bogus explanation. If you say this, you are arguing that the brain generates an illusion of free will—but you are not explaining how. You are trying to dismiss the opposition by deconstructing its motives—but in the story you tell, the illusion of free will is a brute fact. You have not taken the illusion apart to see the wheels and gears.”
If you keep rushing again and again to swiftly solve the problem—or, worse, rushing again to affirm that the problem is solved—then it will be harder to notice the points that cause you confusion. My appeal to the Marianna example is a key example of a place that should have made you stop, furrow your brow, and notice that the explanation you gave before to dispell Mary-intuitions doesn’t work for Marianna-intuitions, even though the two seem to be of the same kind. It would be surprising indeed if ‘Mary lacked the ability to visualize redness’ were a big part of the explanation in the former case, yet not in the least bit a part of the latter case, given their obvious parallelism. This suggests that the explanation you first gave is off-base in the Mary case too. Retreating to just asserting that dualism is wrong is missing the important tidal shift that just happened.
OK. But ‘Qualia seem irreducible because something about how our brains work makes them seem irreducible’ isn’t the most satisfying of explanations. Could you give a little more detail?
OK. But couldn’t all of the same be said of ordinary macroscopic objects in our environment, too? When I see a table (a physical table in my environment—not a table-shaped quale in my visual field), I can’t directly access the underlying fine-grained quantum description of the table. Nor can I make tables spontaneously appear in my environment by acquiring superhuman knowledge of the physics of tables. Yet tables don’t seem to pose any problem at all for reductionism.
If tables and qualia have all these things in common, then where does the actual difference lie, the difference that explains why there seems to be a Hard Problem in one case and not in the other?
But is that because people who only learn about jobs indirectly are lacking certain key pieces of factual knowledge? The problem raised by Mary’s Room isn’t ‘Explain why Mary intuitively seems to get better at completing various tasks’; it’s ‘Explain why Mary intuitively seems to learn new factual knowledge’. This is made clearer by the Marianna example. Your analogy only helps us give a physicalistic explanation of the former, not the latter.
Mary is a super-sceintist in tersm of intelligence and memory, but doesn’t have special abilities to rewire her own cortex. Internally gerneating Red is a cheat, like pricking her thumb to observe the blood.
She isn’t generating Red, she’s generating a memory of the feeling Red generates without generating Red. She now knows what emotional state Red would make her feel, but hasn’t actually made herself see red. So when she goes outside she doesn’t say “Wow” she says “Oh, those feelings again, just as I suspected.”
Why is she generating a memory? How is she generatign a memory?
So she’s bound and gagged, with no ability to use her knowledge? Seems implausible, but OK. (Did she get this knowledge by dictation, or by magically reaching out to the Aristotelian essences of neurons?)
In any case, at least two of us have linked to orthonormal’s mini-sequence on the matter. Those three posts seem much better than ESR’s attempt at the quest.
If by “using her knowledge” you mean performing neurosurgery in herself, I have to repeat that that is a cheat.Otherwise, I ha e to point put that knowledge of, eg. phontosynthesis, doesn’t cause photosynthesis.
Sure, that would be this mini-sequence by orthonormal.