The I-Less Eye
or: How I Learned to Stop Worrying and Love the Anthropic Trilemma
Imagine you live in a future society where the law allows up to a hundred instances of a person to exist at any one time, but insists that your property belongs to the original you, not to the copies. (Does this sound illogical? I may ask my readers to believe in the potential existence of uploading technology, but I would not insult your intelligence by asking you to believe in the existence of a society where all the laws were logical.)
So you decide to create your full allowance of 99 copies, and a customer service representative explains how the procedure works: the first copy is made, and informed he is copy number one; then the second copy is made, and informed he is copy number two, etc. That sounds fine until you start thinking about it, whereupon the native hue of resolution is sicklied o’er with the pale cast of thought. The problem lies in your anticipated subjective experience.
After step one, you have a 50% chance of finding yourself the original; there is nothing controversial about this much. If you are the original, you have a 50% chance of finding yourself still so after step two, and so on. That means after step 99, your subjective probability of still being the original is 0.5^99, in other words as close to zero as makes no difference.
Assume you prefer existing as a dependent copy to not existing at all, but preferable still would be existing as the original (in the eyes of the law) and therefore still owning your estate. You might reasonably have hoped for a 1% chance of the subjectively best outcome. 0.5^99 sounds entirely unreasonable!
You explain your concerns to the customer service representative, who in turn explains that regulations prohibit making copies from copies (the otherwise obvious solution) due to concerns about accumulated errors (the technical glitches in the early versions of the technology that created occasional errors have long been fixed, but the regulations haven’t caught up yet). However, they do have a prototype machine that can make all 99 copies simultaneously, thereby giving you your 1% chance.
It seems strange that such a minor change in the path leading to the exact same end result could make such a huge difference to what you anticipate, but the philosophical reasoning seems unassailable, and philosophy has a superb track record of predictive accuracy… er, well the reasoning seems unassailable. So you go ahead and authorize the extra payment to use the prototype system, and… your 1% chance comes up! You’re still the original.
“Simultaneous?” a friend shakes his head afterwards when you tell the story. “No such thing. The Planck time is the shortest physically possible interval. Well if their new machine was that precise, it’d be worth the money, but obviously it isn’t. I looked up the specs: it takes nearly three milliseconds per copy. That’s into the range of timescales in which the human mind operates. Sorry, but your chance of ending up the original was actually 0.5^99, same as mine, and I got the cheap rate.”
“But,” you reply, “it’s a fuzzy scale. If it was three seconds per copy, that would be one thing. But three milliseconds, that’s really too short to perceive, even the entire procedure was down near the lower limit. My probability of ending up the original couldn’t have been 0.5^99, that’s effectively impossible, less than the probability of hallucinating this whole conversation. Maybe it was some intermediate value, like one in a thousand or one in a million. Also, you don’t know the exact data paths in the machine by which the copies are made. Perhaps that makes a difference.”
Are you convinced yet there is something wrong with this whole business of subjective anticipation?
Well in a sense there is nothing wrong with it, it works fine in the kind of situations for which it evolved. I’m not suggesting throwing it out, merely that it is not ontologically fundamental.
We’ve been down this road before. Life isn’t ontologically fundamental, so we should not expect there to be a unique answer to questions like “is a virus alive” or “is a beehive a single organism or a group”. Mind isn’t ontologically fundamental, so we should not expect there to be a unique answer to questions like “at what point in development does a human become conscious”. Particles aren’t ontologically fundamental, so we should not expect there to be a unique answer to questions like “which slit did the photon go through”. Yet it still seems that I am alive and conscious whereas a rock is not, and the reason it seems that way is because it actually is that way.
Similarly, subjective experience is not ontologically fundamental, so we should not expect there to be unique answer to questions involving subjective probabilities of outcomes in situations involving things like copying minds (which our intuition was not evolved to handle). That’s not a paradox, and it shouldn’t give us headaches, any more than we (nowadays) get a headache pondering whether a virus is alive. It’s just a consequence of using concepts that are not ontologically fundamental, in situations where they are not well defined. It all has to boil down to normality—but only in normal situations. In abnormal situations, we just have to accept that our intuitions don’t apply.
How palatable is the bullet I’m biting? Well, the way to answer that is to check whether there are any well-defined questions we still can’t answer. Let’s have a look at some of the questions we were trying to answer with subjective/anthropic reasoning.
Can I be sure I will not wake up as Britney Spears tomorrow?
Yes. For me to wake up as Britney Spears, would mean the atoms in her brain were rearranged to encode my memories and personality. The probability of this occurring is negligible.
If that isn’t what we mean, then we are presumably referring to a counterfactual world in which every atom is in exactly the same location as in the actual world. That means it is the same world. To claim there is or could be any difference is equivalent to claiming the existence of p-zombies.
Can you win the lottery by methods such as “Program your computational environment to, if you win, make a trillion copies of yourself, and wake them up for ten seconds, long enough to experience winning the lottery. Then suspend the programs, merge them again, and start the result”?
No. The end result will still be that you are not the winner in more than one out of several million Everett branches. That is what we mean by ‘winning the lottery’, to the extent that we mean anything well-defined by it. If we mean something else by it, we are asking a question that is not well-defined, so we are free to make up whatever answer we please.
In the Sleeping Beauty problem, is 1⁄3 the correct answer?
Yes. 2⁄3 of Sleeping Beauty’s waking moments during the experiment are located in the branch in which she was woken twice. That is what the question means, if it means anything.
Can I be sure I am probably not a Boltzmann brain?
Yes. I am the set of all subpatterns in the Tegmark multiverse that match a certain description. The vast majority of these are embedded in surrounding patterns that gave rise to them by lawful processes. That is what ‘probably not a Boltzmann brain’ means, if it means anything.
What we want from a solution to confusing problems like the essence of life, quantum collapse or the anthropic trilemma is for the paradoxes to dissolve, leaving a situation where all well-defined questions have well-defined answers. That’s how it worked out for the other problems, and that’s how it works out for the anthropic trilemma.
- An attempt to dissolve subjective expectation and personal identity by 22 Feb 2013 20:44 UTC; 53 points) (
- Anthropics: A Short Note on the Fission Riddle by 28 Jul 2018 4:14 UTC; 11 points) (
- Revisiting the Anthropic Trilemma II: axioms and assumptions by 16 Feb 2011 9:42 UTC; 9 points) (
- 7 May 2010 16:39 UTC; 9 points) 's comment on Beauty quips, “I’d shut up and multiply!” by (
- A simple-minded theory of “observer fluid” by 16 Feb 2011 16:40 UTC; 6 points) (
- 1 Feb 2012 12:35 UTC; 6 points) 's comment on Open Thread, February 1-14, 2012 by (
- Revisiting the anthropic trilemma I: intuitions and contradictions by 15 Feb 2011 11:18 UTC; 3 points) (
- 11 Oct 2010 15:47 UTC; 3 points) 's comment on References & Resources for LessWrong by (
- 20 Dec 2010 15:09 UTC; 3 points) 's comment on Copying and Subjective Experience by (
- 1 Feb 2011 10:39 UTC; 3 points) 's comment on Reply to Stuart on anthropics by (
- 26 May 2010 7:39 UTC; 1 point) 's comment on Shock Level 5: Big Worlds and Modal Realism by (
- 7 May 2010 19:46 UTC; 1 point) 's comment on Beauty quips, “I’d shut up and multiply!” by (
- 28 Mar 2010 18:52 UTC; 1 point) 's comment on Addresses in the Multiverse by (
- 23 Feb 2013 15:48 UTC; 0 points) 's comment on An attempt to dissolve subjective expectation and personal identity by (
- 26 Jun 2010 9:57 UTC; 0 points) 's comment on MWI, copies and probability by (
- 6 Jan 2016 20:05 UTC; 0 points) 's comment on Your transhuman copy is of questionable value to your meat self. by (
That’s not the way my subjective anticipation works, so the assertion of uncontroversialness is premature. I anticipate that after step one I have a 100% chance of being the copy, and a 100% chance of being the original. (Which is to say, both of those individuals will remember my anticipation.)
Right, I’m getting the feeling I was too focused on the section of the audience that subscribes to the theory of subjective anticipation against which I was arguing, and forgetting about the section that already doesn’t :-)
Cyan, I gave another argument against subjective anticipation, which does cover the way your subjective anticipation works. Please take a look. (I’m replying to you here in case you miss it.)
Thanks for the link. When you write that it’s an argument against subjective anticipation, I’m not sure what you are specifically arguing against. If you’re just saying that my kind of subjective anticipation will lead to time-inconsistent decisions (and hence is irrational), I agree.
I think this guy disagrees: Weatherson, Brian. Should We Respond to Evil with Indifference? Philosophy and Phenomenological Research 70 (2005): 613-35. Link: http://brian.weatherson.org/papers.shtml
I would prefer if, before I click on the link, the comment tells me something more than someone disagrees with Cyan on the internet.
Good information to include would be the nature of the disagreement (what competing claim is made) and a summary of the reasoning that backs up that competing claim.
I further note that your link points to a list of articles, none of which have the name you cited. This is not helpful.
It’s hidden in “Older Work”—you have to click on “Major Published Papers” to see it.
But agreed on all other points.
Here’s another, possibly more general, argument against subjective anticipation.
Consider the following thought experiment. You’re told that you will be copied once and then the two copies will be randomly labeled A and B. Copy A will be given a button with a choice: either push the button, in which case A will be tortured, or don’t push it, in which case copy B will be tortured instead, but for a longer period of time.
From your current perspective (before you’ve been copied), you would prefer that copy A push the button. But if A anticipates any subjective experiences, clearly it must anticipate that it would experience being tortured if and only if it were to push the button. Human nature is such that a copy of you would probably not push the button regardless of any arguments given here, but let’s put that aside and consider what ideal rationality says. I think it says that A should push the button, because to do otherwise would be to violate time consistency.
If we agree that the correct decision is to push the button, then to reach that decision A must (dis)value any copy of you being tortured the same as any other copy, and its subjective anticipation of experiencing torture ends up playing no role in the decision.
Eliezer wrote that we should make beliefs pay rent in anticipated experiences, and I think we should also make anticipation pay rent in correct decisions. It seems that is only possible under a limited set of circumstances (namely, with no mind copying).
Personal identity/anticipated experience is a mechanism through which a huge chunk of preference is encoded in human minds, on an intuitive level. A lot of preference is expressed in terms of “future experience”, which breaks down once there is no unique referent for that concept in the future. Whenever you copy human minds, you also copy this mechanism, which virtually guarantees lack of reflective consistency in preference in humans.
Thought experiments with mind-copying effectively involve dramatically changing the agent’s values, but don’t emphasize this point, as if it’s a minor consideration. Getting around this particular implementation, directly to preference represented by it, and so being rational in situations of mind-copying, is not something humans are wired to be able to do.
Morendil’s comment made me realize that my example is directly analogous to your Counterfactual Mugging: in that thought experiment, Omega’s coin flip splits you into two copies (in two different possible worlds), and like in my example, the rational thing to do, in human terms, is to sacrifice your own interests to help your copy. To me, this analogy indicates that it’s not mind-copying that’s causing the apparent value changes, but rather Bayesian updating.
I tend to agree with you, but I note that Eliezer disagrees.
Locating future personal experience is possible when we are talking about possible futures, and not possible when we are talking about the future containing multiple copies at the same time. Only in the second case does the mechanism for representing preference breaks down. The problem is not (primarily) in failure to assign preference to the right person, it’s in failure to assign it at all. Humans just get confused, don’t know what the correct preference is, and it’s not a question of not being able to shut up and calculate, as it’s not clear what the answer should be, and how to find it. More or less the same problem as with assigning value to groups of other people: should we care more when there are a lot people at stake, or the same about them all, but less about each of them? (“Shut up and divide”.)
In counterfactual mugging, there is a clear point of view (before the mugging/coin flip) from where preference is clearly represented, via intermediary of future personal experience, as seen from that time, so we can at least shut up and calculate. That’s not the issue I’m talking about.
While for some approaches to decision-making, it might not matter whether we are talking about multiplicative indexical uncertainty, or additive counterfactuals, the issue here is the concept of personal identity through which a chunk of preference is represented in human mind. Decision theories can handle situations where personal identity doesn’t make sense, but we’d still need to get preference about those situations from somewhere, and there is no clear assignment of it.
Some questions about fine distinctions in preference aren’t ever going to be answered by humans, we don’t have the capacity to see the whole picture.
Which brings up the question: suppose that your values are defined in terms of an ontology which is not merely false but actually logically inconsistent, though in a way that is too subtle for you to currently grasp. Is it rational to try to learn the logical truth, and thereby lose most or all of what you value? Should we try to hedge against such a possibility when designing a friendly AI? If so, how?
Do you want to lose what you value upon learning that you were confused? More realistically, the correct preference is to adapt the past preference to something that does make sense. More generally, if you should lose that aspect of preference, it means you prefer to do so; if you shouldn’t, it means you don’t prefer to do so. Whatever the case, doing what you prefer to do upon receiving new information is in accordance with what you prefer.
This is all tautologous, but you are seeing a conflict of interest somewhere, so I don’t think you’ve made the concepts involved in the situation explicit enough to recognize the tautologies.
Preference talks about what you should do, and what you do is usually real (until you pass to a next level).
Perhaps an example will illustrate. The theist plans his life around doing God’s will: when he is presented with a persuasive argument from scripture that God’s will is for him to do X rather than Y, he will do X. Perhaps he has frequently adjusted his strategies when considering scripture, revelations (which are, in fact, hallucinations his subconscious generates), and Papal decree.
It seems that he loses a lot upon learning that God does not exist. As a matter of pure psychological fact, he will be depressed (probably). Moreover, suppose that he holds beliefs that are mutually contradictory, but only in subtle ways; perhaps he thinks that God is in complete control of all things in the world, and that God is all-loving (all good), but the world which he thinks he lives in manifestly contains a lot of suffering. (The Theodicy Problem).
It seems that the best thing for him is to remain ignorant of the paradox, and of his false, inconsistent and confused beliefs, and for events to transpire in a lucky way so that he never suffers serious material losses from his pathological decision-making.
Consider the claim that what a Friendly AI should do for such a person is the following: keep them unaware of the facts, and optimize within their framework of reality.
This seems to confuse stuff that happens to a human with decision theory. What happens with a human (in human’s thoughts, etc.) can’t be “contradictory” apart from a specific interpretation that names some things “contradictory”. This interpretation isn’t fundamentally interesting for the purposes of optimizing the stuff. The ontology problem is asked about the FAI, not about a person that is optimized by FAI. For FAI, a person is just a pattern in the environment, just like any other object, with stars and people and paperclips all fundamentally alike; the only thing that distinguishes them for FAI is what preference tells should be done in each case.
When we are talking about decision theory for FAI, especially while boxing the ontology inside the FAI, it’s not obvious how to connect that with particular interpretations of what happens in environment, nor should we try, really.
Now, speaking of people in environment, we might say that the theist is going to feel frustrated for some time upon realizing that they were confused for a long time. However I can’t imagine the whole process of deconverting to be actually not preferable, as compared to remaining confused (especially given that in the long run, the person will need to grow up). Even the optimal strategy is going to have identifiable negative aspects, but it may only make the strategy suboptimal if there is a better way. Also, for a lot of obvious negative aspects, such as negative emotions accompanying an otherwise desirable transition, FAI is going to invent a way of avoiding that aspect, if that’s desirable.
And that the person might be the source of preference. This is fairly important. But, in any case, FAI theory is only here as an intuition pump for evaluating “what would the best thing be, according to this person’s implicit preferences?”
If it is possible to have preference-like things within a fundamentally contradictory belief system, and that’s all the human in question has, then knowing about the inconsistency might be bad.
This is actually wrong. Whatever the AI starts with is its formal preference, it never changes, it never depends on anything. That this formal preference was actually intended to copycat an existing pattern in environment is a statement about what sorts of formal preference it is, but it is enacted the same way, in accordance with what should be done in that particular case based on what formal preference tells. Thus, what you’ve highlighted in the quote is a special case, not an additional feature. Also, I doubt it can work this way.
True, but implicit preference is not something that person realizes to be preferable, and not something expressed in terms of confused “ontology” believed by that person. The implicit preference is a formal object that isn’t built from fuzzy patterns interpreted in the person’s thoughts. When you speak of “contradictions” in person’t beliefs, you are speaking on a wrong level of abstraction, like if you were discussing parameters in a clustering algorithm as being relevant to reliable performance of hardware on which that algorithm runs.
A belief system can’t be “fundamentally contradictory” because it’s not “fundamental” to begin with. What do you mean by “bad”? Bad according to what? It doesn’t follow from confused thoughts that preference is somehow brittle.
A Friendly AI might also resolve the situation by presenting itself as god, eliminating suffering in the world, and then giving out genuine revelations with adequately good advice.
Eliminating the appearance of suffering in the world would probably be bad for such a theist. He spends much of his time running Church Bazaars to raise money for charity. Like many especially dedicated charity workers, he is somewhat emotionally and axiologically dependent upon the existence of the problem he is working against.
In that case, eliminate actual suffering as fast as possible, then rapidly reduce the appearance of suffering in ways calculated to make it seem like the theist’s own actions are a significant factor, and eventually substitute some other productive activity.
To get back at this point: This depends on how we understand “values”. Let’s not conceptualize values is being defined in terms of an “ontology”.
You do not lose any options by gaining more knowledge. If the optimal response to have when your values are defined in terms of an inconsistent ontology is to go ahead and act as if the ontology is consistent then you can still choose to do so even once you find out the dark secret. You can only gain from knowing more.
If your values are such that they do not even allow a mechanism for creating an best effort approximation of values in the case of ontological enlightenment then you are out of luck no matter what you do. Even if you explicitly value ignorance of the fact that nothing you value can have coherent value, the incoherency of your value system makes the ignorance value meaningless too.
Make the most basic parts of the value system in an ontology that has as little chance as possible of being inconsistent. Reference to actual humans can ensure that a superintelligent FAI’s value system will be logically consistent if it is in fact possible for a human to have a value system defined in a consistent ontology. If that is not possible then humans are in a hopeless position. But at least I (by definition) wouldn’t care.
If preference is expressed in terms of what you should do, not what’s true about the world, new observations never influence preference, so we can fix it at the start and never revise it (which is an important feature for constructing FAI, since you only ever have a hand in its initial construction).
(To whoever downvoted this without comment—it’s not as stupid an idea as it might sound; what’s true about the world doesn’t matter for preference, but it does matter for decision-making, as decisions are made depending on what’s observed. By isolating preference from influence of observations, we fix it at the start, but since it determines what should be done depending on all possible observations, we are not ignoring reality.)
In the situation described by Roko the agent has doubt about its understanding of the very ontology that its values are expressed in. If it were an AI that would effectively mean that we designed it using mathematics that we thought was consistent but turns out to have a flaw. The FAI has self improved to a level where it has a suspicion that the ontology that is used to represent its value system is internally inconsistent and must decide whether to examine the problem further. (So we should have been able to fix it at the start but couldn’t because we just weren’t smart enough.)
If its values are not represented in terms of an “ontology”, this won’t happen.
See the example of the theist (above). Do you really think that the best possible outcome for him involves knowing more?
How could it be otherwise? His confusion doesn’t define his preference, and his preference doesn’t set this particular form of confusion as being desirable. Maybe Wei Dai’s post is a better way to communicate the distinction I’m making: A Master-Slave Model of Human Preferences (though it’s different, the distinction is there as well).
No, I think his values are defined in terms of a consistent ontology in which ignorance may result in a higher value outcome. If his values could not in fact be expresesd consistently then I do hold that (by definition) he doesn’t lose by knowing more.
You might be able to get a scenario like this without mind-copying by using a variety of Newcomb’s Problem.
You wake up without any memories of the previous day. You then see Omega in front of you, holding two boxes. Omega explains that if you pick the first box, you will be tortured briefly now. If you pick the second box, you won’t be.
However, Omega informs you that he anticipated which box you would choose. If he predicted you’d pick the first box, the day before yesterday he drugged you so you’d sleep through the whole day. If he predicted you’d pick the second box he tortured you for a very long period of time the previous day and erased your memory of it afterward. He acknowledges that torture one doesn’t remember afterwards isn’t as bad as torture one does, and assures you that he knows this and extended the length of the previous day’s torture to compensate.
It seem to me like there’d be a strong temptation to pick the second box. However, your self from a few days ago would likely pay to be able to stop you from doing this.
Is that an area in which a TDT would describe the appropriate response using different words to a UDT, even if they suggest the same action? I’m still trying to clarify the difference between UDT, TDT and my own understanding of DT. I would not describe the-updating-that-causes-the-value-changes as ‘bayesian updating’, rather ‘naive updating’. (But this is a terminology preference.)
My understanding is that TDT would not press the button, just like it wouldn’t give $100 to the counterfactual mugger.
Thanks. So they actually do lead to different decisions? That is good to know… but puts me one step further away from confidence!
I wish I could upvote twice as this is extremely important.
ISTM that someone who would one-box on Newcomb for the reasons given by Gary Drescher (act for the sake of what would be the case, even in the absence of causality) would press the button here; if you’re the kind of person who wouldn’t press the button, then prior to copying you would anticipate more pain than if you’re the other kind.
Getting the button is like getting the empty large box in the transparent boxes version of Newcomb’s problem.
This problem is effectively equivalent to counterfactual mugging. I don’t know whether you should pay/press in this problem, but you should certainly pre-commit to doing this beforehand.
Anyway, it doesn’t prove that you value these copies intrinsically, just that you’ve agreed to a trade where you take into account their interests in return for them taking into account yours.
I’m not sure what this “whole business of … anticipation” has to do with subjective experience.
Suppose that, a la Jaynes, we programmed a robot with the rules of probability, the flexibility of recognizing various predicates about reality, and the means to apply the rules of probability when choosing between courses of action to maximize a utility function. Let’s assume this utility function is implemented as an internal register H which is incremented or decremented according to whether the various predicates are satisfied.
This robot could conceivably be equipped with predicates that allow for the contingency of having copies made of itself, copies which we’ll assume to include complete records of the robot’s internal state up to the moment of copying, including register H.
The question then becomes one of specifying what, precisely, is meant by maximizing the expected value of H, given the possibility of copying.
Suppose we want to know what the robot would decide given a copy-and-torture scenario as suggested by Wei Dai. The question of “what the robot would do” surely does not depend on whether the robot thinks of itself as rational, whether it can be said to have subjective anticipation, whether time consistency is important to it, and so on. These considerations are irrelevant to predicting the robot’s behaviour.
The question of “what the robot would do” depends solely on what it formally means to have the robot maximize the expected value of H, since “the value of H” becomes an ambiguous specification from the moment we allow for copying.
(On the other hand, was that specification ever unambiguous to begin with?)
If the robot is programmed to construe “the value of H” as meaning what we might call the “indexical value” of H, that is, the value-held-by-the-present-copy, then it (or rather its A copy) would presumably act in the torture scenario as Wei Day claims most humans would act, and refuse to press the button. But since the “indexical value of H” is ill defined from the perspective of the pre-copying robot, with respect to the situation after the copy, the robot would err when making this decision prior to copying, and would therefore predictably exhibit what we’d call a time inconsistency.
If the robot is programmed to construe “the value of H” as the sum (or the average) of the indexical values of H for all copies of its state which are descendants of the state which is making the decision, then—regardless of when it makes the choice and regardless of which copy is A or B—it would decide as I have claimed a one-boxer would decide.(Though, working out these implications of the “choice machine” frame, I’m less sure than before of the relation between Wei Dai’s scenario and Newcomb’s problem.)
While writing the above, I realized—this is where I’m driving at with the parenthetical comment about ambiguity—that even in a world without copying you get to make plenty of non-trivial decisions about what it means, formally, to maximize the value of H. In particular, you could be faced with decisions you must make now but which will have an effect in the future and whose effect on H may depend on the value of H at that time. (There are plenty of real-life examples, which I’ll leave as exercise for the reader.) Just how you program the robot to deal with those seems (as far as I can tell) underspecified by the laws of probability alone.
A shorter way of saying all the above is that if we taboo “anticipation”, when predicting what a certain class of agent will do, we don’t necessarily find that there is anything particularly strange about a predicate saying “the present state of the robot is copy N of M”. What we find is that we might want to program the robot differently if we want it to deal in a certain way with the contingency of copying; that is unsurprising. We also find that subjective experience needn’t enter the picture at all.
I know p-zombies are unpopular around here, so maybe by ‘equivalent’ you merely meant ‘equivalently wacky’, but it’s worth noting that the view you rightly dismiss here is quite different (and arguably more radical) than Chalmersian property dualism. After all, phenomenal ‘qualia’ are still qualitative (descriptive) features of a world, whereas you’re imagining someone who thinks that there can be differences in numerical identity even between qualitatively identical worlds. (In academic circles this view is called ‘haecceitism’.)
Anyway, as a property dualist who shares your anti-haecceitism, I just found it a bit strange to see you describe haecceitism as ‘equivalent’ to property dualism, and so wanted to clarify that the two views are actually quite independent.
I don’t really understand this distinction. If property dualism is to explain subjective experience at all, the word ‘me’ must refer to a bundle of phenomenological properties associated to e.g. Richard Chappell’s brain. Saying that ‘I am now Britney Spears’ would just mean that the same identifier now referred to a different bundle of phenomenal qualia. True, the physical and even mental features of the world would be unchanged, but it seems easy to model haeccetism just by adding a layer of indirection between your subjective experience and the actual bundle of qualia. And given that the possiblity of ‘waking up as someone else’ is somewhat intuitive, this might be worthwhile.
Upvoted, but the Boltzmann problem is that it casually looks like the vast majority of subpatterns that match a given description ARE Boltzmann Brains. After all, maxentropy is forever.
But so is eternal inflation, so we are comparing infinities of the same cardinality. The solution seems to be that the Kolmogorov complexity of a typical Boltzmann brain is high, because its space-time coordinates have a length in bits exceeding the length of the description of the brain itself; by Solomonoff induction, we can therefore assign them a very low measure, even in total.
The way subjective probability is usually modeled, there is this huge space of possibilities. And there is a measure defined over it. (I’m not a mathematician, so I may be using the wrong terminology, but what I mean is that every ‘sufficiently nice’ subset of this set of possibilities has a number attached which behaves something like an area for that subset of the space.)
And then, in this model, the probability of some proposition is the measure of the subset where the proposition is true divided by the measure of the whole set. Numerator and denominator. And then each time you learn something, you throw away all of the points in that space that are no longer possible. So, you have typically decreased (never increased) both numerator and denominator. Do the division again and get the new updated probabilities. The space of all possibilities only loses points and measure, it never gains.
But I am not so sure this rule still applies when copying is involved. I think that each time you copy, you need to duplicate the subjective space of possibilities. The original space covered the measures of possibilities from one subjective viewpoint. At the point of copying, that space is duplicated because you now have two viewpoints. Initially, both original and copy are unsure which half of the space is theirs. But when they find out, they each throw out half of the doubled space. And then, as they learn more, possibilities are thrown away from one or the other of the spaces and each one updates to his own subjective probabilities.
So how does this apply to the copying scenario above? Start with one universe. Copy it when you copy the person. Produce a second copy when you produce the second copy of the person. Produce the 99th copy of subjective reality when you produce the 99th copy of the person. If at any stage, one of these persons learns for sure which copy is his, then he can prune his own subjective universe back to the original size.
So, if the protocol is that after each copying, the copy is told that he is a copy and the original is told that he is the original, then before any copying, the person should anticipate being told “You are original” N times, where N is between 0 and 99 inclusive. And he should attach equal probability to each of those events. That is, he should be 99 to 1 sure he will be the original the first time, 98 to one sure the second time, etc.
Forgive me if this is already known as one of the standard approaches to the problem.
Interesting! So you propose to model mind copying by using probabilities greater than 1. I wonder how far we can push this idea and what difficulties may arise...
This is really well written. I hope you post more.
Again, probability doesn’t work with indexical uncertainty.
Could you explain how the post you linked to relates to your comment?
His comments point towards expanding the gp to something like
I wrote a response to this post here:
I would much rather make a billion clones of myself whenever I experience great sex with a highly desirable partner. Point: making the clones to experiencing the lottery is about the experience and not the lottery. I’m not sure I particularly want to have orgasmic-clone-mes but if for whatever arbitrary reason I value having wireheaded simulations of myself experiencing the positive experiences I have had then I may as well do it well. I’m tempted to make a word play here on ‘getting lucky’ but the point is it is not about increasing your luck at all, in either case. It’s about experience cloning.
My impression was that the Anthropic Trilemma was Eliezer uncharacteristically confusing himself when reality itself didn’t need to be.
Joe is just confused about the math. Multiplying the 0.5s together like that is the wrong thing to do. The subjective expectation I would have after being copied for the 99th time is that I have 50% chance of being told I am a copy.
Perhaps Joe’s confusion is not fully specifying what his expectation is before the process starts. He expects that after he is cloned:
p(there is an original) = 1
p(there is a clone) = 1
p(after one clone is made the original will expect to be told that he is a clone) = 0.5
p(after one clone is made that clone will expect to be told that he is a clone) = 0.5
p(after two clones are made there will be an original) = 1
p(after two clones are made there will be a first clone) = 1
p(after two clones are made there will be a second clone) = 1
p(after two clones have been made the original clone will expect to be told that he is the second clone) = 0.5
p(after two clones have been made the first clone will expect to be told that he is the second clone) = 0 (The first clone doesn’t expect to be told anything more.)
p(after two clones have been made the second clone will expect to be told that he is the second clone) = 0.5
Now, consider multiplying
p(after two clones have been made the second clone will expect to be told that he is the second clone) * p(after one clone is made that clone will expect to be told that he is a clone).
Why would I do that? That makes no sense. That would give me 0.25 but so what? It certainly doesn’t mean that p(after two clones have been made the second clone will expect to be told that he is the second clone | after one clone is made that clone will expect to be told that he is a clone) = 0.25. If Joe thinks you can do that he is just wrong and so will end up making a decision that doesn’t optimally reflect his preferences.
It seems like you’re discussing two types of copying procedure, or could be. The Ebborian copying seems to strongly imply the possibility of waking up as the copy or original (I have no statement on the probabilities), but a “teleporting Kirk” style of copying doesn’t seem to imply this. You’re presumably not Ebborian copied into Britney Spears, but if you’re talking about Joe making 99 copies through Kirk copying, it seems like he’d have an equal chance to be any of them at any time, assuming he has any chance not to be the original.
I suppose this may not be the case if you view the subjective experience as being equally likely to be in either Joe until the first moment of differing input, here being the vocal information on what copy he is.
If you aren’t discussing Ebborian copying, why does the creation of a physically identical version of oneself imply a possibility of being them? This seems like a widespread view, and I’ve long been confused by it, so I’d enjoy illumination by anyone on this point.
Does anyone have a sense of what the lower limit is on meaningful probability estimates for individual anticipation? Right, like there should be some probability p(E) where, upon experiencing E, even a relatively sane and well-balanced person ought to predict that the actual state of the world is ~E, because p(I’m Crazy or I’ve Misunderstood) >> p(E).
More to the point, p(E) should be roughly constant across apparently sane people; I would guess that the probability of hallucination doesn’t vary by much more than a factor of 10 among people who have no reason to expect that they are hallucinating. 10^1 might be small relative to whatever the Minimum Probability turns out to be.
Careful; your question contains the implied assumption that P(hallucinate X) doesn’t vary with X. For example, suppose I look at a string of 100 digits produced by a random number generator. Whatever that string is, my prior probability of it being that particular string was 10^-100, but no matter how long that string is, it doesn’t mean I hallucinated it. What really matters is the ratio of how likely an event is to how likely it is that your brain would’ve hallucinated it, and that that depends more on your mental represenations than reality.
I respectfully disagree.
Suppose I bet that a 30-digit random number generator will deliver the number 938726493810487327500934872645. And, lo and behold, the generator comes up with 938726493810487327500934872645 on the first try. If I am magically certain that I am actually dealing with a random number generator, I ought to conclude that I am hallucinating, because p(me hallucinating) > p(guessing a 30-digit string correctly).
Note that this is true even though p(me hallucinating the number 938726493810487327500934872645) is quite low. I am certainly more likely, for example, to hallucinate the number 123456789012345678901234567890 than I am to hallucinate the number 938726493810487327500934872645. But since I am trying to find the minimum meaningful probability, I don’t care too much about the upper bounds on the odds that I’m hallucinating—I want the lower bound on the odds that I’m hallucinating, and the lower bound would correspond to a mentally arbitrary number like 938726493810487327500934872645.
In other words, if you agree with me that p(I correctly guess the number 938726493810487327500934872645) < p(I’m hallucinating the number 938726493810487327500934872645), then you should certainly agree with me that p(I correctly guess the number 123456789012345678901234567890) < p(I’m hallucinating the number 123456789012345678901234567890). The probability of guessing the correct number is always 10^-30; the probability of hallucinating varies, but I suspect that the probability of hallucinating is more than 10^-30 for either number.
Choosing a number and betting that you will see it increases the probability that you will wrongly believe that you have seen that number in the future to a value that does not depend on how long that number is. P(hallucinate number N|placed a bet on N) >> P(hallucinate number N).
Yes, I completely agree. To show that I understand your point, I will suggest possible numbers for each of these variables. I would guess, with very low confidence, that on a daily basis, P(hallucinate a number) might be something like 10^-7, that P(hallucinate a 30-digit number N) might be something like 10^-37, and that P(hallucinate a 30-digit number N | placed a bet on N) might be something like 10^-9. Obviously, p(correctly guess a 30-digit number) is still 10^-30.
Even given all of these values, I still claim that we should be interested in P(hallucinate a 30-digit number N | placed a bet on N). This number is probably roughly constant across ostensibly sane people, and I claim that it marks a lower bound below which we should not care about the difference in probabilities for a non-replicable event.
I am not certain of these claims, and I would greatly appreciate your analysis of them.
Note that there are explanations other than “I correctly guessed”, and “I’m hallucinating”. “This generator is broken and always comes up 938726493810487327500934872645, but I’ve forgotten that it’s broken consciously, but remember that number”, or “The generator is really remotely controlled, and it has a microphone that heard me guess, and transmitted that to the controller, who wants to mess with my head.”
Oh, I completely agree. I’m using “hallucinating” as shorthand for all kinds of conspiracy theories, and assuming away the chance that the generator is broken.
Obviously the first thing you should do if you appear to guess right is check the generator.
By the way: Welcome to Less Wrong! If you want to post an introduction, you can do so in that thread.
This seems related to the post I submitted in January, The Prediction Hierarchy. I think I’d have to know what you’re using it for to know what to do with any given lower bound.
rwallace, nice reductio ad adsurdum of what I will call the Subjective Probability Anticipation Fallacy (SPAF). It is somewhat important because the SPAF seems much like, and may be the cause of, the Quantum Immortality Fallacy (QIF).
You are on the right track. What you are missing though is an account of how to deal properly with anthropic reasoning, probability, and decisions. For that see my paper on the ‘Quantum Immortality’ fallacy. I also explain it concisely on on my blog on Meaning of Probability in an MWI.
Basically, personal identity is not fundamental. For practical purposes, there are various kinds of effective probabilities. There is no actual randomness involved.
It is a mistake to work with ‘probabilities’ directly. Because the sum is always normalized to 1, ‘probabilities’ deal (in part) with global information, but people easily forget that and think of them as local. The proper quantity to use is measure, which is the amount of consciousness that each type of observer has, such that effective probability is proportional to measure (by summing over the branches and normalizing). It is important to remember that total measure need not be conserved as a function of time.
As for the bottom line: If there are 100 copies, they all have equal measure, and for all practical purposes have equal effective probability.
Interesting one. 100 hundred Joes, one ‘original’, some ‘copies’.
If we copy Joe once, and let him be, he’s 50% certainly original. If we copy the copy, Joe remains 50% certainly original, while status of copies does not change.
After the first copying process, we ended up with the original and a copy. 50% of the resulting sentient beings were original. If we do that again, again, we have two sentient beings, original and a new copy. Again, 50% chance for a random sentient byproduct of this copying process to be the original.
But there’s something you didn’t take into account. You can’t just multiply the 50% of the first and second process, as your Joe wasn’t chosen randomly after every copying process. After every single process, you took the original and copied only him again. Math should reflect this.
New take. The problem can be described as a branching tree, where each copy-branch is cut off, leaving only 1 copy.
So, at step 2, we would’ve had 4 possibilities, 1 original and three copies, but branches of the copy were cut away, so we are left with three Joes, 1 original, 1 equally likely copy, and… 1 copy that’s twice as likely?
So let’s look what happens in this process.
t=1: You know that you are the original t=2: We create a clone in such a way that you don’t know whether you are a clone or not. At this time you have a subjective probability of 50% of being a clone. t=3: We tell clone 1 that they are a clone. Your subjective probability of being a clone is now 0% since you were not informed that you were a clone. t=4: We create another clone that provides you with a subjective probability of being a clone of 50% t=5: Clone 2 finds out that they are a clone. Since you weren’t told you were a clone, you know you aren’t a clone, so your subjective probability of being you goes back up to 100%.
Let’s now imagine that we want no-one to know if they are clones or not. We will imagine that people initially know that they are not the new clone, but this information is erased.
t=1: We copy the original person so that we have two clones. We erase any information that would indicate who is original. t=2: We create a third clone, but allow the first two people to know they aren’t the third clone t=3: We erase information from the first two people about whether or not they are the third clone.
At t=1, you have a 50% chance of being a clone and a 50% chance of being the original. At t=2, you still have a 50% chance as you know you aren’t the third clone At t=3, you have lost information about whether you are the third clone. You can now be any of the clones and there is no distinguishing information, so the probability becomes 1⁄3 of being the original. Probability mass isn’t just redistributed from the chance of you being the original but also from the chance of you being the first clone.
When we have created n clones, your odds of being the original will be 1/n.
It makes no difference whether the steps at t=2 and t=3 occur separately or together, I simply separately them to show that it was the loss of information about identity, not the cloning that changed the probability.
So if the clones weren’t informed about their number after cloning, we would get the same result whether we produced 99 clones at once or one at a time.
Lastly, let’s suppose that the clone is told that they are a clone, but the original doesn’t know they won’t be told. This won’t affect the subjective probabilities of the original, only that of the clones, so again there isn’t a paradox.
This paradox is based upon a misunderstanding of how cloning actually works. Once this is modelled as information loss, the solution is straightforward.
I don’t really get what you’re saying.
The normal way of looking at it is that you are only going to be you in the future. The better way of looking at it is that an unknown person is equally likely any person during any period of a given length.
The results of the former don’t work well. They lead to people preferentially doing things to help their future selves, rather than helping others. This is rather silly. Future you isn’t you either.
I don’t believe in the Tegmark multiverse. ;)
I’m sorry, I didn’t read the rest of your post after seeing the 0.5^99 estimate of the probability of being the original because the math looked very wrong to me, but I didn’t know why. While I agree there is nothing controversial about saying that after one step you have a 50% chance of being the original, I’m pretty sure it is not true that you only have a 25% chance after two steps. Yes, if you are the original after step one, you have a 50% chance probability of still being the original after step two. So, I Oi is the probability of being the original after step i, then P(O2|O1) = 0.5, P(01) = 0.5, P(O2|~O1) = 0, so P(O2) = P(O2|O1) * P(O1). BUT, after step 2, O1 is not 0.5, since there are 3 copies of you in existence, and 2 of those WERE the original after step 1, so P(O1) after step 2 is 2⁄3, so P(O2) = 2⁄6 = 1⁄3. Don’t think it affects the substance of anything else in your argument, though, but the math bothered me since intuitively the expected value of originals after any number of steps should be 1, which should be equal to the number of copies in existence times the number of copies. If I’m missing something I’d welcome an explanation of what.
Hmm. Perhaps I should’ve put in a note to the effect of if you don’t subscribe to the theory of subjective anticipation which would give that estimate, great, just skip to the summary break and read on from there.
If P(O1) is the probability of an event, then it doesn’t change.
I believe in continuity of substance, not similarity of pattern, as the basis of identity. If you are the original, that is what you are for all time. You cannot wake up as the copy. At best, a new mind can be created with false beliefs (such as false memories, of experiences which did not happen to it). Do I still face a problem of “subjective anticipation”?
ETA: Eliezer said of the original problem, “If you can’t do the merge without killing people, then the trilemma is dissolved.” Under a criterion of physical continuity, you cannot go from two objects to one object without at least one of them ceasing to be. So the original problem also appears to be a non-problem for me.
So Scotty killed Kirk and then created a zombie-Kirk back on the Enterprise? It would seem that the whole Star Trek is a fantasy story about a space faring necromancer who repeatedly kills his crew then uses his evil contraption to reanimate new beings out of base matter while rampaging through space seeking new and exotic beings to join his never ending orgy of death.
Yes, yes he did, time and again (substituting “copy” for “zombie”, as MP points out below). That’s the Star Trek paradox.
Imagine that there is a glitch in the system, so that the “original” Kirk fails to dematerialise when the “new” one appears, so we find ourselves with two copies of Kirk. Now Scotty says “Sowwy Captain” and zaps the “old” Kirk into a cloud of atoms. How in the world does that not constitute murder?
That was not the paradox. The “paradox” is this: the only difference between “innocuous” teleportation, and the murder scenario described above, is a small time-shift of a few seconds. If Kirk1 disappears a few seconds before Kirk2 appears, we have no problem with that. We even show it repeatedly in programmes aimed at children. But when Kirk1 disappears a few seconds after Kirk2 appears, all of a sudden we see the act for what it is, namely murder.
How is it that a mere shift of a few seconds causes such a great difference in our perception? How is it that we can immediately see the murder in the second case, but that the first case seems so innocent to us? This stark contrast between our intuitive perceptions of the two cases, despite their apparent underlying similarity, constitutes the paradox.
And yes, it seems likely that the above also holds when a single person is made absolutely unconscious (flat EEG) and then awakened. Intuitively, we feel that the same person, the same identity, has persisted throughout this interruption; but when we think of the Star Trek paradox, and if we assume (as good materialists) that consciousness is the outcome of physical brain activity, we realise that this situation is not very different from that of Kirk1 and Kirk2. More generally, it illustrates the problems associated with assuming that you “are” the same person that you were just one minute ago (for some concepts of “are”).
I was thinking of writing a post about this, but apparently all of the above seems to be ridiculously obvious to most LWers, so I guess there’s not much of a point. I still find it pretty fascinating. What can I say, I’m easily impressed.
I’m not comfortable with ‘for what it is, namely’. I would be comfortable with ‘see the act as murder’. I don’t play ‘moral reference class tennis’. Killing a foetus before it is born is killing a foetus before it is born (or abortion). Creating a copy then removing the original is creating a copy and then removing the original (or teleportation). Killing someone who wants to die is killing someone who wants to die (or euthanasia). Calling any of these things murder is not necessarily wrong but it is not a factual judgement it is a moral judgement. The speaker wants people to have the same kind of reaction that they have to other acts that are called ‘murder’.
‘Murder’ is just more complex than that. So is ‘killing’ and so is ‘identity’. You can simplify the concepts arbitrarily so that ‘identity’ is a property of a specific combination of matter if you want to but that just means you need to make up a new word to describe “that thing that looks, talks and acts like the same Kirk every episode and doesn’t care at all that he gets de-materialised all the time”. If you don’t keep a separate word to describe that concept then you will end up making some extremely silly life choices when you come to be exposed to choices outside of the cultural norm. Like Cryonics. Depending on how many exceptions you want to allow yourself you’ll also need to be careful about blood transfusions, shedding skin, neurogenesis over time, definitely organ transplants.
For my part I feel comfortable obliterating the atomic structure of my body and having it recreated perfectly on the surface of a planet. I wouldn’t be comfortable with being conscious in the ‘obliteration’ process but I don’t care at all whether the copy is created before or after the original is destroyed. No, scratch that, I do care. I want the original destroyed after the copy is created so it can double check it has it right first! I care about my ‘pattern’ rather a lot more than my ‘originality’ so if “identity” means ‘the original substance’ then I jolly well want a new word that means ‘me’!
It doesn’t and I don’t. See above. (ie. Replace ‘our’ with ‘my’?) That just isn’t how my intuition is wired. I actually feel a little nervous about there not being a copy of me outside of the Enterprise’s RAM. Although I suppose I would adjust to that once people explain “it’s safer than driving in your car, no really!”.
Any plans Kirk had prior to his “original” being dematerialized are still equally likely to be carried out by the “copy” Kirk, any preferences he had will still be defended, and so on. Nothing of consequence seems to have been lost; an observer unaware of this little drama will notice nothing different from what he would have predicted, had Kirk traveled by more conventional means.
To say that a murder has been committed seems like a strained interpretation of the facts. There’s a difference between burning of the Library of Alexandria and destroying your hard drive when you have a backup.
Currently, murder and information-theoretic murder coincide, for the same reasons that death and information-theoretic death coincide. When that is no longer the case, the distinction will become more salient.
And here is something that bugs me in Sci. Fi. shows. It’s worse than ’Sound in space? Dammit!” Take Carter from Stargate. She has Asgard beaming technology and the Asgard core (computer). She can use this to create food, a Chelo for herself and Tritonin for Teal’c. The core function of the device is to take humanoid creatures and re-materialise them somewhere else. Why oh why do they not leave the originals behind and create a 50-Carter strong research team, a million strong Teal’c army and an entire wizard’s circle of Daniel Jacksons with whatever his mind-power of the episode happens to be? There are dozens of ways to clone SG1. The robot-SGI is the mundane example. The Stargates themselves have the capability and so do Wraith darts. The same applies to Kirk and his crew. But no. let’s just ignore the most obvious use of the core technology.
If Kirk1 disappears a few seconds before Kirk2 appears, we assume that no subjective experience was lost; a branch of length 0 was terminated. If the transporter had predictive algorithms good enough to put Kirk2 into the exact same state that Kirk1 would be in a few seconds later, then painlessly dematerialized Kirk1, I would have no more problem with it than I do with the original Star Trek transporter.
A copy, not a zombie.
It is a shame that the term was reserved for ‘philosophical zombies’. I mean, philosophical zombies haven’t even been killed. Kirk was killed then reanimated. That’s real necromancy for you.
Not possible, according to Eliezer.
And what do you think? I disagree with Eliezer, and I can talk about my position, but I want to hear your opinion first.
I find Eliezer’s argument convincing.
OK. Well, here’s a different perspective.
Suppose we start with quantum mechanics. What is the argument that particles don’t have identity? If you start with particles in positions A and B, and end with particles in positions C and D, and you want to calculate the probability amplitude for this transition, you count histories where A goes to C and B goes to D, and histories where A goes to D and B goes to C. Furthermore, these histories can interfere destructively (e.g. this happens with fermions), which implies that the two endpoints really are the same place in configuration space, and not just outcomes that look the same.
From this it is concluded that the particles have no identity across time. According to this view, if you end up in the situation with particles at C and D, and ask if the particle at C started at A or started at B, there is simply no answer, because both types of history will have contributed to the outcome.
However, it is a curious fact that although the evolving superposition contains histories of both types, within any individual history, there is identity across time! Within an individual history in the sum over histories, A does go to strictly one of C or D.
Now I’m going to examine whether the idea of persistent particle-identity makes sense, first in single-world interpretations, then in many-world interpretations.
What do physicists actually think is the reality of a quantum particle? If we put aside the systematic attempts to think about the problem, and just ask what attitudes are implicitly at work from day to day, I see three attitudes. One is the positivistic attitude that it is pointless to talk or think about things you can’t observe. Another is the ignorance interpretation of quantum uncertainty; the particle always has definite properties, just like a classical particle, but it moves around randomly, in a way that adds up to quantum statistics. Finally, you have wavefunction realism: particles really are spread out in space or in superpositions. (The thinking of an individual physicist may combine several of these attitudes.)
The positivistic attitude is likely to dismiss the question of ‘which path the electron took’ or even ‘did the electron take a definite path’ as metaphysics and unanswerable, so it’s irrelevant to the present discussion. Wavefunction realism, pursued systematically, usually becomes a many-worlds philosophy, so I’ll save that option for the second part. So if we are asking whether electrons persist over time and follow definite paths in a single-world interpretation, we are really asking whether that is the case under an ignorance interpretation of quantum uncertainty.
I think it is obviously so. This way of thinking says that particles are just like classical particles—they always have a definite location, they always execute definite motions—except that they act randomly. If we have two particles apparently just sitting there, and we want to know whether they changed places or not, the real answer will be yes or no, even if we can never know which is right.
(A remark on the legitimacy of this way of thinking. Bell’s theorem evidently rattled a lot of people because it showed that a naive conception of how these random motions worked could not give rise to quantum mechanics—it could not produce sufficiently strong correlations at a distance. Nonetheless, it is possible to derive quantum probabilities from local random behavior, just as you can get a diffusion probability distribution from Brownian motion. The punchline is that it has to be local random motion in configuration space. In configuration space you treat the whole classical configuration as a single point in an infinite-dimensional abstract space, so “motion” in that abstract space will involve simultaneous changes to physical properties all across real space. This may sound like cheating; it means that when you go back to thinking in terms of real space, if your random motions are going to produce quantum statistics, then the randomness has to be correlated at a distance, without further cause. But some people are prepared to bite that bullet; that’s just how reality is, they’ll tell you.)
Now to many worlds. Here we are saying that superpositions are real; so the history where the particles stay where they are, and the history where they swap places, are both real, and they flow into the same world at the end. Now, surely, we cannot speak of a particle’s identity persisting over time. We started out with a world containing a particle at A and a particle at B; it evolved into a world that was a superposition (or was it a superposition of worlds?), each element of the superposition still containing two particles, but now in other positions; and it terminated in a world with a particle at C and a particle at D. Each final particle inherited a bit of amplitude from multiple predecessors, and for each there are paths heading back to A and to B. So we simply can’t say that the particle at C is the sole heir of either original particle.
However, perhaps we can say that these two particles were entangled, and that this entangled duo had a persistent identity across time! Certainly, as described, there were only ever two particles in the picture. You might object that in the real world, there would be other particles, and they would also interact with the duo, and even trade places with them in some histories, and so this notion of a locally encapsulated entanglement is false. Everything is entangled with everything else, indirectly if not directly, and so all I could say is that the universe as a whole has identity across time.
My response to that is that developing a coherent many-worlds interpretation is a lot more difficult than you might think. Many worlds has been presented here as the economical, no-collapse alternative to theories arbitrarily postulating a collapse process; but to actually find individual worlds in a universal wavefunction, you have to break it up somehow (break it up conceptually), and that is a project with a lot of hidden difficulties (significant example). The arbitrariness of the collapse postulate has its counterpart in the arbitrariness of how the worlds are defined. If a natural, non-arbitrary definition exists, it is going to have to find natural structures, such as temporarily localized entanglements; and I note Eliezer’s comment in the original article, “I’m calling you a factored subspace”. If that is so—if the idea can even make sense—then it will be that subspace which has continuity of identity across time.
So, whether you adopt a single-world or a many-world perspective, a nonpatternist theory of physical identity is viable.
We are actually talking about personal identity here, not physical identity, and that raises further issues. But if physical identity is a viable concept after all, then so too may be a concept of personal identity grounded in temporal persistence of physical identity.
I’ll grant that by being sufficiently clever, you can probably reconcile quantum mechanics with whatever ontology you like. But the real question is: why bother? Why not take the Schroedinger equation literally? Physics has faced this kind of issue before—think of the old episode about epicycles, for instance—and the lesson seems clear enough to me. What’s the difference here?
For what it’s worth, I don’t see the arbitrariness of collapse postulates and the arbitrariness of world-selection as symmetrical. It’s not even clear to me that we need to worry about extracting “worlds” from blobs of amplitude, but to the extent we do, it seems basically like an issue of anthropic selection; whereas collapse postulates seem like invoking magic.
But in any case you don’t really address the objection that
Instead, you merely raise the issue of finding “individual worlds”, and argue that if you can find manage to find an individual world, then you can say that that world has an identity that persists over time. Fair enough, but how does this help you rescue the idea that personal identity resides in “continuity of substance”, when the latter may still be meaningless at the level of individual particles?
The Schroedinger equation is an assertion about a thing called Psi. “Taking it literally” usually means “believe in many worlds”. Now even if I decide to try this out, I face a multitude of questions. Am I to think of Psi as a wavefunction on a configuration space, or as a vector in a Hilbert space? Which part of Psi corresponds to the particular universe that I see? Am I to think of myself as a configuration of particles, a configuration of particles with an amplitude attached, a superposition of configurations each with its own amplitude, or maybe some other thing, like an object in Hilbert space (but what sort of object?) not preferentially associated with any particular basis? And then there’s that little issue of deriving the Born probabilities!
Once you decide to treat the wavefunction itself as ultimate physical reality, you must specify exactly which part of it corresponds to what we see, and you must explain where the probabilities come from. Otherwise you’re not doing physics, you’re just daydreaming. And when people do address these issues, they do so in divergent ways. And in my experience, when you do get down to specifics, problems arise, and the nature of the problems depends very much on which of those divergent implementations of many-worlds has been followed.
It is hard to go any further unless you tell me more about what many-worlds means to you, and how you think it works. “Take the equation literally” is just a slogan and doesn’t provide any details.
By “world”, do you mean a universe-sized configuration, or just an element of a more localized superposition? It is another of the exasperating ambiguities of many-worlds discourse. Some people do make it clear that their worlds-in-the-wavefunction are of cosmic size, while others apparently prefer to think of the multiplicity of realities as a local and even relative thing—I think this is what “many minds” is about: the observer is in a superposition and we acknowledge that there are many distinct observers or distinct instances of the observer, but the rest of the universe is to be regarded as still in its transcendent pristine many-in-one multiverse unity… I speak sarcastically, but I do see among some many-worlders a sort of veneration of the wavefunction and a dislike for any attempt to break it up into worlds in a definite way, even though you absolutely need to do this to make contact with empirical reality.
So, anyway, I was talking about localized entanglements, or (equivalently) small factors of the total quantum state, as providing a basis for “continuity of substance” even if individual particles cannot. The relevance to personal identity is as follows. We are assuming that a person has something to do with the material world. The argument I dispute is the one that says personal identity cannot depend on the persistence through time of the person’s material parts, because there is no such thing as persistence through time of particles, because differently-braided particle histories all convey amplitude to the same configuration. And my proposition was that if you look at superpositions of these braidings and sub-braidings, you get localized entities which have ontological boundaries and persistence in time until they enter into a larger braiding; and this means you can after all talk about material parts of a person persisting in time.
What it means is that you let your ontology be dictated by the mathematical structure of the equation. So for instance:
It’s both—even when regarded purely as a mathematical object. The set of wavefunctions on a configuration space is (the unit sphere of) a Hilbert space. Specifically, as I understand it, configuration space is a measure space of some sort, and the set of wavefunctions is (the unit sphere in) L^2 of that measure space.
It seems to me that you’re a region of configuration space. There’s a subset of the measure space that consists of configurations that represent things like “you’re in this state”, “you’re in that state”, etc. We can call this subset the “you”-region. (Of course, these states also contain information about the rest of the universe, but the information they contain about you is the reason we’re singling them out as a subset.)
To repeat a point made before (possibly by Eliezer himself), this isn’t an issue that distinguishes between many-worlds and collapse postulates. With many-worlds, you have to explain the Born probabilities; with collapse interpretations, you have to explain the mysterious collapse process. It seems to me far preferable, all else being equal, to be stuck with the former problem rather than the latter—because it turns the mystery into an indexical issue (“Why are we in this branch rather than another?”) rather than writing it into the laws of the universe.
Why is this?
Okay, it now occurs to me that I may have been confusing “continuity of substance” (your criterion) with “identity of substance” (which is what Eliezer’s argument rules out). That’s still more problematic, in my opinion, than a view that allows for uploading and teleportation, but in any event I withdraw the claim that it is challenged by Eliezer’s quantum-mechanical argument about particle identity.
There are two issues here: many worlds, and the alleged desirability or necessity of abandoning continuity of physical existence as a criterion of identity, whether physical or personal.
Regarding many worlds, I will put it this way. There are several specific proposals out there claiming to derive the Born probabilities. Pick one, and I will tell you what’s wrong with it. Without the probabilities, you are simply saying “all worlds exist, this is one of them, details to come”.
Regarding “continuity of substance” versus “identity of substance”… If I was seriously going to maintain the view I suggested—that encapsulated local entanglements permit a notion of persistence in time—then I would try to reconceptualize the physics so that identity of substance applied. What was formerly described as three entangled particles, I would want to describe as one thing with a big and evolving state.
All this begs the question: Is personal identity made up of the same stuff as ‘blue’?
Are there any actual predictions that would be different with “continuity of substance” as the standard for identity rather than “similarity of composition”?
What does continuity of substance even mean with respect to Fock spaces or the path-integral formulation? All electrons (or protons, etc) are literally descended from all possible others.
These “subjective anticipations” are different, because I don’t try to think of my copies as me.
Discussed here.
This 0.5^99 figure only appears if each copy bifurcates iteratively.
Rather than
We’d have
No, as described, you have probability (1/2)^n of becoming copy #n, and #99 and the original share (1/2)^99.
The original is copied once—giving 50% #0 and 50% #1. Then #0 is copied again, giving 25% #0, and 25% #2. Then #0 is copied again, giving 12.5% #0, and 12.5% #3, and so forth.
This seems like a useful reductio ad absurdum of this means of calculating subjective expectation.
Hmmm.
Yes, I see it now. The dead-end copies function as traps, since they stop your participation in the game. As long as you can consciously differentiate your state as a copy or original, this works.