It looks like it turned awful since I’ve read it the last time:
This essay, while entertaining and useful, can be seen as Yudkowsky trying to reinvent the sense of awe associated with religious experience in the name of rationalism. It’s even available in tract format.
The most fatal mistake of the entry in its current form seems to be that it does lump together all of Less Wrong and therefore does stereotype its members. So far this still seems to be a community blog with differing opinions. I got a Karma score of over 1700 and I have been criticizing the SIAI and Yudkowsky (in a fairly poor way).
I hope you people are reading this. I don’t see why you draw a line between you and Less Wrong. This place is not an invite-only party.
LessWrong is dominated by Eliezer Yudkowsky, a research fellow for the Singularity Institute for Artificial Intelligence.
I don’t think this is the case anymore. You can easily get Karma by criticizing him and the SIAI. Most of all new posts are not written by him anymore either.
Members of the Less Wrong community are expected to be on board with the singularitarian/transhumanist/cryonics bundle.
Nah!
If you indicate your disagreement with the local belief clusters without at least using their jargon, someone may helpfully suggest that “you should try reading the sequences” before you attempt to talk to them.
I don’t think this is asked too much. As the FAQ states:
Why do you all agree on so much? Am I joining a cult?
We have a general community policy of not pretending to be open-minded on long-settled issues for the sake of not offending people. If we spent our time debating the basics, we would never get to the advanced stuff at all.
It’s unclear whether Descartes, Spinoza or Leibniz would have lasted a day without being voted down into oblivion.
So? I don’t see what this is supposed to prove.
Indeed, if anyone even hints at trying to claim to be a “rationalist” but doesn’t write exactly what is expected, they’re likely to be treated with contempt.
Provide some references here.
Some members of this “rationalist” movement literally believe in what amounts to a Hell that they will go to if they get artificial intelligence wrong in a particularly disastrous way.
I’ve been criticizing the subject matter and got upvoted for it, as you obviously know since you linked to my comments as reference. Further I never claimed that the topic is unproblematic or irrational but that I was fearing unreasonable consequences and that I have been in disagreement about how the content was handled. Yet I do not agree with your portrayal insofar that it is not something that fits a Wiki entry about Less Wrong. Because something sounds extreme and absurd it is not wrong. In theory there is nothing that makes the subject matter fallacious.
Yudkowsky has declared the many worlds interpretation of quantum physics is correct, despite the lack of testable predictions differing from the Copenhagen interpretation, and despite admittedly not being a physicist.
I haven’t read the quantum physics sequence but by what I have glimpsed this is not the crucial point that distinguishes MWI from other interpretations. That’s why people suggest one should read the material before criticizing it.
P.S.
I’m curious if you know of a more intelligent and rational community than Less Wrong? I don’t! Proclaiming that Less Wrong is more rational than most other communities isn’t necessarily factually wrong.
Edit: “[...] by what I have glimpsed this is just wrong.” now reads “[...] by what I have glimpsed this is not the crucial point that distinguishes MWI from other interpretations.”
Descartes is maybe the single best example of motivated cognition in the history of Western thought. Though interestingly, there are some theories that he was secretly an atheist.
I assume their point has something to do with those three being rationalists in the traditional sense… but I don’t think Rational Wiki is using the word in the traditional sense either. Would Descartes have been allowed to edit an entry on souls?
Do you mean Spinoza or Leibniz given their knowledge base and upbringing or the same person with a modern environment? I know everything Leibniz knew and a lot more besides. But I suspect that if the same individual grew up in a modern family environment similar to my own he would have accomplished a lot more than I have at the same age.
Sorry, I thought the notion was clear that one would be talking about same genetics but different environment. Illusion of transparency and all that. Explicit formulation: if one took a fertilized egg with Leibniz’s genetic material and raised in an American middle class family with high emphasis on intellectual success, I’m pretty sure he would have by the time he got to my age have accomplished more than I have. Does that make the meaning clear?
You think the average person on LessWrong ranks with Spinoza and Leibniz? I disagree.
Wedrifid_2010 was not assigning a status ranking or even an evaluation of overall intellectual merit or potential. For that matter predicting expected voting patterns is a far different thing than assigning a ranking. People with excessive confidence in habitual thinking patterns that are wrong or obsolete will be downvoted into oblivion where the average person is not, even if the former is more intelligent or more intellectually impressive overall.
I also have little doubt that any of those three would be capable of recovering from their initial day or three of spiraling downvotes assuming they were willing to ignore their egos, do some heavy reading of the sequences and generally spend some time catching up on modern thought. But for as long as those individuals were writing similar material to that which identifies them they would be downvoted by lesswrong_2010. Possibly even by lesswrong_now too.
Not exactly the point—I think the claim is that they would be downvoted even if they were providing modern, original content … which I would question, even then. We’ve had quite successful theist posters before, for example.
I think the claim is that they would be downvoted even if they were providing modern, original content
What would this even mean? Like, if they were transported forward in time and formed new beliefs on the basis of modern science? If they were cloned from DNA surviving in their bone marrow and then adopted by modern, secular families, took AP Calculus and learned to program?
Not exactly the point—I think the claim is that they would be downvoted even if they were providing modern, original content … which I would question, even then.
I would downvote Descartes based on the quality of his thinking and argument even if it was modern bad thinking. At least I would if he persisted with the line after the first time or two he was corrected. I suppose this is roughly equivalent to what you are saying.
Yudkowsky has declared the many worlds interpretation of quantum physics is correct, despite the lack of testable predictions differing from the Copenhagen interpretation, and despite admittedly not being a physicist.
I think there is a fair chance the many world’s interpretation is wrong but anyone who criticizes it by defending the Copenhagen ‘interpretation’ has no idea what they’re talking about.
I haven’t read the quantum physics sequence but by what I have glimpsed this is just wrong. That’s why people suggest one should read the material before criticizing it.
Irony.
Xixidu, you should also read the material before trying to defend it.
What single-world interpretation basically say to fit MWI: All but one world are eliminated by a magic faster than light non-local time-asymmetric acausal collapser-device.
So far MWI sounds like the most reasonable interpretation to me. And from what I have read I can tell that the sentence—“despite the lack of testable predictions differing from the Copenhagen interpretation”—is not crucial in favoring MWI over other interpretations.
Of course I am not able to judge that MWI is the correct interpretation but, given my current epistemic state, of all interpretations it is the most likely to be correct. For one it sounds reasonable, secondly Yudkowsky’s judgement has a considerable weight here. I have no reason to suspect that it would benefit him to favor MWI over other interpretations. Yet there is much evidence that suggests that he is highly intelligent and that he is able to judge what is the correct interpretation given all evidence a non-physicists can take into account.
Edit: “[...] is not correct, or at least not crucial.” now reads “[...] is not crucial in favoring MWI over other interpretations.”
And from what I have read I can tell that the sentence—“despite the lack of testable predictions differing from the Copenhagen interpretation”—is not correct, or at least not crucial.
It is correct, and it is crucial in the sense that most philosophy of science would insist that differing testable predictions is all that would favor one theory over another.
But other concerns (the Bayesian interpretation of Occam’s Razor (or any interpretation, probably)) make MWI preferred.
An interpretation of Occam’s Razor that placed all emphasis on space complexity would clearly favor the Copenhagen interpretation over the MW interpretation. Of course, it would also favor “you’re living in a holodeck” over “there’s an actual universe out there”, so it’s a poor formulation in it’s simplest form… but it’s not obvious (to me, anyway) that space complexity should count for nothing at all, and if it counts for “enough” (whatever that is, for the particular rival interpretation) MWI loses.
What particular gold-standard “Occam’s razor” are you adhering to, then? It seems to fit well with “entities must not be multiplied beyond necessity” and “pluralities must never be posited without necessity”.
Note that I’m not saying there is no gold-standard “Occam’s razor” to which we should be adhering (in terms of denotation of the term or more generally); I’m just unaware of an interpretaton that clearly lays out how “entities” or “assumptions” are counted, or how the complexity of a hypothesis is otherwise measured, which is clearly “the canonical Occam’s razor” as opposed to having some other name. If there is one, by all means please make me aware!
Is there anything in particular that leads you to claim Minimum Description Length is the only legitimate claimaint to the title “Occam’s razor”? It was introduced much later, and the wikipedia article claims it is “a forumlation of Occam’s razor”.
Certainly, William of Occam wasn’t dealing in terms of information compression.
The answer seems circular: because it works. The experience of people using Occam’s razor (e.g. scientists) find MDL to be more likely to lead to correct answers than any other formulation.
I don’t see that that makes other formulations “not Occam’s razor”, it just makes them less useful attempts at formalizing Occam’s razor. If an alternative formalization was found to work better, it would not be MDL—would MDL cease to be “Occam’s razor”? Or would the new, better formalization “not be Occam’s razor”? Of the latter, by what metric, since the new one “works better”?
For the record, I certainly agree that “space complexity alone” is a poor metric. I just don’t see that it should clearly be excluded entirely. I’m generally happy to exclude it on the grounds of parsimony, but this whole subthread was “How could MWI not be the most reasonable choice...?”
There’s an intent behind Occam’s razor. When Einstein improved on Newton’s gravity, gravity itself didn’t change. Rather, our understanding of gravity was improved by a better model. We could say though that Newton’s model is not gravity because we have found instances where gravity does not behave the way Newton predicted.
Underlying Occam’s razor is the simple idea that we should prefer simple ideas. Over time we have found ways to formalize this statement in ways that are universally applicable. These formalizations are getting closer and closer to what Occam’s razor is.
I see, I went too far in asserting something about MWI, as I am not able to discuss this in more detail. I’ll edit my orginal comments.
Edit—First comment: “[...] by what I have glimpsed this is just wrong.” now reads “[...] by what I have glimpsed this is not the crucial point that distinguishes MWI from other interpretations.”
Edit—Second comment: “[...] is not correct, or at least not crucial.” now reads “[...] is not crucial in favoring MWI over other interpretations.”
The problem isn’t that you asserted something about MWI—I’m not discussing the MWI itself here.
It’s rather that you defended something before you knew what it was that you were defending, and attacked people on their knowledge of the facts before you knew what the facts actually were.
Then once you got more informed about it, you immediately changed the form of the defense while maintaining the same judgment. (Previously it was “Bad critics who falsely claim Eliezer has judged MWI to be correct” now it’s “Bad critics who correctly claim Eliezer has judged MWI to be correct, but they badly don’t share that conclusion”)
This all is evidence (not proof, mind you) of strong bias.
Ofcourse you may have legitimately changed your mind about MWI, and legimitately moved from a wrongful criticism of the critics on their knowledge of facts to a rightful criticism of their judgment.
I’m also commenting on the blog of Neal Asher, a science fiction author I read. I have no problem making fun of his climate change skepticism although I doubt that any amateur, even on Less Wrong, would have the time to conclude that it is obviously correct. Yet I do not doubt it for the same reasons I do not doubt MWI:
There is no benefit in proclaiming the correctness of MWI (at least for Yudkowsky).
The argument used against MWI fails the argument used in favor of MWI on Less Wrong.
The person who proclaims the correctness of MWI is an expert when it comes to beliefs.
It’s the same with climate change. People saying—“look how cold it is in Europe again, that’s supposed to be global warming?!”—are, given my current state of knowledge, not even wrong. Not only will there be low-temperature records even given global warming (outliers), but global warming will also cause Europe to get colder on average. Do I know that this is correct? Nope, but I do trust the experts as I do not see that a global conspiracy is feasible and would make sense. It doesn’t benefit anyone either.
You are correct that I should stay away from calling people wrong on details when I’m not ready to get into the details. Maybe those people who wrote that entry are doing research on foundational physics, I doubt it though (writing style etc.).
I’m not sure about the details of your comment. I just changed my comment regarding the claim that there are testable predictions regarding MWI (although there are people on LW and elsewhere who claim this to be the case). As people started challenging me on that point I just retreated to not get into a discussion I can’t possible participate in. I did not change my mind about MWI in general. I just shortened my argument from MWI making testable predictions and being correct irrespective of testable predictions to the latter. That is, MWI is an implication of a theory that is more precise in its predictions, yet simpler, as the one necessary to conclude other interpretations.
My mistake was that I went to far. I read the Wiki entry and thought I’d write down my thoughts on every point. That point was behind my expertise indeed.
I haven’t seen any proof (stronger than “it seems like it”) that MWI is strictly simpler to describe. One good reason to prefer it is that it is nice and continuous, and all our other scientific theories are nice and continuous—sort of a meta-science argument.
In layman’s terms (to the best of my understanding), the proof is:
Copenhagen interpretation is “there is wave propagation and then collapse” and thus requires a description of how collapse happens. MWI is “there is wave propagation”, and thus has fewer rules, and thus is simpler (in that sense).
I agree denotatively—I don’t think the Copenhagen interpretation provides this description.
I am not sure what I do connotatively, as I am not sure what the connotations are meant to be. It would mean that quantum theory is less complete than it is if MWI is correct, but I’m not sure whether that’s a correct objection or not (and have less idea whether you intended to express it as such).
RationalWiki was the only place I saw this mistake made, so the i.e. seemed deserved to me.
It looks like it turned awful since I’ve read it the last time:
The most fatal mistake of the entry in its current form seems to be that it does lump together all of Less Wrong and therefore does stereotype its members. So far this still seems to be a community blog with differing opinions. I got a Karma score of over 1700 and I have been criticizing the SIAI and Yudkowsky (in a fairly poor way).
I hope you people are reading this. I don’t see why you draw a line between you and Less Wrong. This place is not an invite-only party.
I don’t think this is the case anymore. You can easily get Karma by criticizing him and the SIAI. Most of all new posts are not written by him anymore either.
Nah!
I don’t think this is asked too much. As the FAQ states:
Why do you all agree on so much? Am I joining a cult?
We have a general community policy of not pretending to be open-minded on long-settled issues for the sake of not offending people. If we spent our time debating the basics, we would never get to the advanced stuff at all.
So? I don’t see what this is supposed to prove.
Provide some references here.
I’ve been criticizing the subject matter and got upvoted for it, as you obviously know since you linked to my comments as reference. Further I never claimed that the topic is unproblematic or irrational but that I was fearing unreasonable consequences and that I have been in disagreement about how the content was handled. Yet I do not agree with your portrayal insofar that it is not something that fits a Wiki entry about Less Wrong. Because something sounds extreme and absurd it is not wrong. In theory there is nothing that makes the subject matter fallacious.
I haven’t read the quantum physics sequence but by what I have glimpsed this is not the crucial point that distinguishes MWI from other interpretations. That’s why people suggest one should read the material before criticizing it.
P.S. I’m curious if you know of a more intelligent and rational community than Less Wrong? I don’t! Proclaiming that Less Wrong is more rational than most other communities isn’t necessarily factually wrong.
Edit: “[...] by what I have glimpsed this is just wrong.” now reads “[...] by what I have glimpsed this is not the crucial point that distinguishes MWI from other interpretations.”
I know, I loved that quote. I just couldn’t work out why it was presented as a bad thing.
Descartes is maybe the single best example of motivated cognition in the history of Western thought. Though interestingly, there are some theories that he was secretly an atheist.
I assume their point has something to do with those three being rationalists in the traditional sense… but I don’t think Rational Wiki is using the word in the traditional sense either. Would Descartes have been allowed to edit an entry on souls?
You think the average person on LessWrong ranks with Spinoza and Leibniz? I disagree.
Do you mean Spinoza or Leibniz given their knowledge base and upbringing or the same person with a modern environment? I know everything Leibniz knew and a lot more besides. But I suspect that if the same individual grew up in a modern family environment similar to my own he would have accomplished a lot more than I have at the same age.
They wouldn’t be the same person. Which is to say, the the whole matter is nonsense as the other replies in this thread made clear.
Sorry, I thought the notion was clear that one would be talking about same genetics but different environment. Illusion of transparency and all that. Explicit formulation: if one took a fertilized egg with Leibniz’s genetic material and raised in an American middle class family with high emphasis on intellectual success, I’m pretty sure he would have by the time he got to my age have accomplished more than I have. Does that make the meaning clear?
Right, I understand. But Rational Wiki didn’t have clones of Leibniz and Spinoza in mind. I’m just saying, the whole thing is goofy.
Wedrifid_2010 was not assigning a status ranking or even an evaluation of overall intellectual merit or potential. For that matter predicting expected voting patterns is a far different thing than assigning a ranking. People with excessive confidence in habitual thinking patterns that are wrong or obsolete will be downvoted into oblivion where the average person is not, even if the former is more intelligent or more intellectually impressive overall.
I also have little doubt that any of those three would be capable of recovering from their initial day or three of spiraling downvotes assuming they were willing to ignore their egos, do some heavy reading of the sequences and generally spend some time catching up on modern thought. But for as long as those individuals were writing similar material to that which identifies them they would be downvoted by lesswrong_2010. Possibly even by lesswrong_now too.
Yes. Upvotes come from original, insightful contributions. Descartes’, Spinoza’s, and Liebnitz’s ideas are hundreds of years old and dated.
Not exactly the point—I think the claim is that they would be downvoted even if they were providing modern, original content … which I would question, even then. We’ve had quite successful theist posters before, for example.
What would this even mean? Like, if they were transported forward in time and formed new beliefs on the basis of modern science? If they were cloned from DNA surviving in their bone marrow and then adopted by modern, secular families, took AP Calculus and learned to program?
What a goofy thing to even be talking about.
Goofier than a universe in which humans work but matches don’t? Such ideas may be ill-formed, but that doesn’t make them obviously ill-formed.
Sure...
ETA: I’m making fun of that wiki, not you.
I would downvote Descartes based on the quality of his thinking and argument even if it was modern bad thinking. At least I would if he persisted with the line after the first time or two he was corrected. I suppose this is roughly equivalent to what you are saying.
I think there is a fair chance the many world’s interpretation is wrong but anyone who criticizes it by defending the Copenhagen ‘interpretation’ has no idea what they’re talking about.
Irony.
Xixidu, you should also read the material before trying to defend it.
Correct. Yet I have read some subsequent discussions about that topic (MWI) and also watched this talk:
I also read Decoherence is Simple and Decoherence is Falsifiable and Testable.
So far MWI sounds like the most reasonable interpretation to me. And from what I have read I can tell that the sentence—“despite the lack of testable predictions differing from the Copenhagen interpretation”—is not crucial in favoring MWI over other interpretations.
Of course I am not able to judge that MWI is the correct interpretation but, given my current epistemic state, of all interpretations it is the most likely to be correct. For one it sounds reasonable, secondly Yudkowsky’s judgement has a considerable weight here. I have no reason to suspect that it would benefit him to favor MWI over other interpretations. Yet there is much evidence that suggests that he is highly intelligent and that he is able to judge what is the correct interpretation given all evidence a non-physicists can take into account.
Edit: “[...] is not correct, or at least not crucial.” now reads “[...] is not crucial in favoring MWI over other interpretations.”
It is correct, and it is crucial in the sense that most philosophy of science would insist that differing testable predictions is all that would favor one theory over another.
But other concerns (the Bayesian interpretation of Occam’s Razor (or any interpretation, probably)) make MWI preferred.
An interpretation of Occam’s Razor that placed all emphasis on space complexity would clearly favor the Copenhagen interpretation over the MW interpretation. Of course, it would also favor “you’re living in a holodeck” over “there’s an actual universe out there”, so it’s a poor formulation in it’s simplest form… but it’s not obvious (to me, anyway) that space complexity should count for nothing at all, and if it counts for “enough” (whatever that is, for the particular rival interpretation) MWI loses.
That would not be Occam’s razor...
What particular gold-standard “Occam’s razor” are you adhering to, then? It seems to fit well with “entities must not be multiplied beyond necessity” and “pluralities must never be posited without necessity”.
Note that I’m not saying there is no gold-standard “Occam’s razor” to which we should be adhering (in terms of denotation of the term or more generally); I’m just unaware of an interpretaton that clearly lays out how “entities” or “assumptions” are counted, or how the complexity of a hypothesis is otherwise measured, which is clearly “the canonical Occam’s razor” as opposed to having some other name. If there is one, by all means please make me aware!
Minimum description length.
The MWI requires fewer rules than Copenhagen, and therefore its description is smaller, and therefore it is the strictly simpler theory.
Is there anything in particular that leads you to claim Minimum Description Length is the only legitimate claimaint to the title “Occam’s razor”? It was introduced much later, and the wikipedia article claims it is “a forumlation of Occam’s razor”.
Certainly, William of Occam wasn’t dealing in terms of information compression.
The answer seems circular: because it works. The experience of people using Occam’s razor (e.g. scientists) find MDL to be more likely to lead to correct answers than any other formulation.
I don’t see that that makes other formulations “not Occam’s razor”, it just makes them less useful attempts at formalizing Occam’s razor. If an alternative formalization was found to work better, it would not be MDL—would MDL cease to be “Occam’s razor”? Or would the new, better formalization “not be Occam’s razor”? Of the latter, by what metric, since the new one “works better”?
For the record, I certainly agree that “space complexity alone” is a poor metric. I just don’t see that it should clearly be excluded entirely. I’m generally happy to exclude it on the grounds of parsimony, but this whole subthread was “How could MWI not be the most reasonable choice...?”
There’s an intent behind Occam’s razor. When Einstein improved on Newton’s gravity, gravity itself didn’t change. Rather, our understanding of gravity was improved by a better model. We could say though that Newton’s model is not gravity because we have found instances where gravity does not behave the way Newton predicted.
Underlying Occam’s razor is the simple idea that we should prefer simple ideas. Over time we have found ways to formalize this statement in ways that are universally applicable. These formalizations are getting closer and closer to what Occam’s razor is.
I’ll accept that.
I see, I went too far in asserting something about MWI, as I am not able to discuss this in more detail. I’ll edit my orginal comments.
Edit—First comment: “[...] by what I have glimpsed this is just wrong.” now reads “[...] by what I have glimpsed this is not the crucial point that distinguishes MWI from other interpretations.”
Edit—Second comment: “[...] is not correct, or at least not crucial.” now reads “[...] is not crucial in favoring MWI over other interpretations.”
The problem isn’t that you asserted something about MWI—I’m not discussing the MWI itself here.
It’s rather that you defended something before you knew what it was that you were defending, and attacked people on their knowledge of the facts before you knew what the facts actually were.
Then once you got more informed about it, you immediately changed the form of the defense while maintaining the same judgment. (Previously it was “Bad critics who falsely claim Eliezer has judged MWI to be correct” now it’s “Bad critics who correctly claim Eliezer has judged MWI to be correct, but they badly don’t share that conclusion”)
This all is evidence (not proof, mind you) of strong bias.
Ofcourse you may have legitimately changed your mind about MWI, and legimitately moved from a wrongful criticism of the critics on their knowledge of facts to a rightful criticism of their judgment.
I’m also commenting on the blog of Neal Asher, a science fiction author I read. I have no problem making fun of his climate change skepticism although I doubt that any amateur, even on Less Wrong, would have the time to conclude that it is obviously correct. Yet I do not doubt it for the same reasons I do not doubt MWI:
There is no benefit in proclaiming the correctness of MWI (at least for Yudkowsky).
The argument used against MWI fails the argument used in favor of MWI on Less Wrong.
The person who proclaims the correctness of MWI is an expert when it comes to beliefs.
It’s the same with climate change. People saying—“look how cold it is in Europe again, that’s supposed to be global warming?!”—are, given my current state of knowledge, not even wrong. Not only will there be low-temperature records even given global warming (outliers), but global warming will also cause Europe to get colder on average. Do I know that this is correct? Nope, but I do trust the experts as I do not see that a global conspiracy is feasible and would make sense. It doesn’t benefit anyone either.
You are correct that I should stay away from calling people wrong on details when I’m not ready to get into the details. Maybe those people who wrote that entry are doing research on foundational physics, I doubt it though (writing style etc.).
I’m not sure about the details of your comment. I just changed my comment regarding the claim that there are testable predictions regarding MWI (although there are people on LW and elsewhere who claim this to be the case). As people started challenging me on that point I just retreated to not get into a discussion I can’t possible participate in. I did not change my mind about MWI in general. I just shortened my argument from MWI making testable predictions and being correct irrespective of testable predictions to the latter. That is, MWI is an implication of a theory that is more precise in its predictions, yet simpler, as the one necessary to conclude other interpretations.
My mistake was that I went to far. I read the Wiki entry and thought I’d write down my thoughts on every point. That point was behind my expertise indeed.
I haven’t seen any proof (stronger than “it seems like it”) that MWI is strictly simpler to describe. One good reason to prefer it is that it is nice and continuous, and all our other scientific theories are nice and continuous—sort of a meta-science argument.
In layman’s terms (to the best of my understanding), the proof is:
Copenhagen interpretation is “there is wave propagation and then collapse” and thus requires a description of how collapse happens. MWI is “there is wave propagation”, and thus has fewer rules, and thus is simpler (in that sense).
… which it doesn’t provide.
I agree denotatively—I don’t think the Copenhagen interpretation provides this description.
I am not sure what I do connotatively, as I am not sure what the connotations are meant to be. It would mean that quantum theory is less complete than it is if MWI is correct, but I’m not sure whether that’s a correct objection or not (and have less idea whether you intended to express it as such).