Now, do you have any actual argument as to why the ‘badness’ function computed over a box containing two persons with a dust speck, is exactly twice the badness of a box containing one person with a dust speck, all the way up to very large numbers (when you may even have exhausted the number of possible distinct people) ?
I don’t think you do. This is why this stuff strikes me as pseudomath. You don’t even state your premises let alone justify them.
You’re right, I don’t. And I do not really need it in this case.
What I need is a cost function C(e,n) - e is some event and n is the number of people being subjected to said event, i.e. everyone gets their own—where for ε > 0: C(e,n+m) > C(e,n) + ε for some m. I guess we can limit e to “torture for 50 years” and “dust specks” so this generally makes sense at all.
The reason why I would want to have such a cost function is because I believe that it should be more than infinitesimally worse for 3^^^^3 people to suffer than for 3^^^3 people to suffer. I don’t think there should ever be a point where you can go “Meh, not much of a big deal, no matter how many more people suffer.”
If however the number of possible distinct people should be finite—even after taking into account level II and level III multiverses—due to discreteness of space and discreteness of permitted physical constants, then yes, this is all null and void. But I currently have no particular reason to believe that there should be such a bound, while I do have reason to believe that permitted physical constants should be from a non-discrete set.
Well, within the 3^^^3 people you have every single possible brain replicated a gazillion times already (there’s only that many ways you can arrange the atoms in the volume of human head, sufficiently distinct as to be computing something subjectively different, after all, and the number of such arrangements is unimaginably smaller than 3^^^3 ).
I don’t think that e.g. I must massively prioritize the happiness of a brain upload of me running on multiple redundant hardware (which subjectively feels the same as if it was running in one instance; it doesn’t feel any stronger because there’s more ‘copies’ of it running in perfect unison, it can’t even tell the difference. It won’t affect the subjective experience if the CPUs running the same computation are slightly physically different).
edit: also again, pseudomath, because you could have C(dustspeck, n) = 1-1/(n+1) , your property holds but it is bounded, so if the c(torture, 1)=2 then you’ll never exceed it with dust specks.
Seriously, you people (LW crowd in general) need to take more calculus or something before your mathematical intuitions become in any way relevant to anything whatsoever. It does feel intuitively that with your epsilon it’s going to keep growing without a limit, but that’s simply not true.
I consider entities in computationally distinct universes to also be distinct entities, even if the arrangements of their neurons are the same. If I have an infinite (or sufficiently large) set of physical constants such that in those universes human beings could emerge, I will also have enough human beings.
edit: also again, pseudomath, because you could have C(dustspeck, n) = 1-1/(n+1) , your property holds but it is bounded, so if the c(torture, 1)=2 then you’ll never exceed it with dust specks.
No. I will always find a larger number which is at least ε greater. I fixed ε before I talked about n,m. So I find numbers m_1,m_2,… such that C(dustspeck,m_j) > jε.
Besides which, even if I had somehow messed up, you’re not here (I hope) to score easy points because my mathematical formalization is flawed when it is perfectly obvious where I want to go.
Well, in my view, some details of implementation of a computation are totally indiscernible ‘from the inside’ and thus make no difference to the subjective experiences, qualia, and the like.
I definitely don’t care if there’s 1 me, 3^^^3 copies of me, or 3^^^^3, or 3^^^^^^3 , or the actual infinity (as the physics of our universe would suggest), where the copies are what thinks and perceives everything exactly the same over the lifetime. I’m not sure how counting copies as distinct would cope with an infinity of copies anyway. You have a torture of inf persons vs dust specks in inf*3^^^3 persons, then what?
Albeit it would be quite hilarious to see if someone here picks up the idea and starts arguing that because they’re ‘important’, there must be a lot of copies of them in the future, and thus they are rightfully an utility monster.
Now, do you have any actual argument as to why the ‘badness’ function computed over a box containing two persons with a dust speck, is exactly twice the badness of a box containing one person with a dust speck, all the way up to very large numbers (when you may even have exhausted the number of possible distinct people) ?
I don’t think you do. This is why this stuff strikes me as pseudomath. You don’t even state your premises let alone justify them.
You’re right, I don’t. And I do not really need it in this case.
What I need is a cost function C(e,n) - e is some event and n is the number of people being subjected to said event, i.e. everyone gets their own—where for ε > 0: C(e,n+m) > C(e,n) + ε for some m. I guess we can limit e to “torture for 50 years” and “dust specks” so this generally makes sense at all.
The reason why I would want to have such a cost function is because I believe that it should be more than infinitesimally worse for 3^^^^3 people to suffer than for 3^^^3 people to suffer. I don’t think there should ever be a point where you can go “Meh, not much of a big deal, no matter how many more people suffer.”
If however the number of possible distinct people should be finite—even after taking into account level II and level III multiverses—due to discreteness of space and discreteness of permitted physical constants, then yes, this is all null and void. But I currently have no particular reason to believe that there should be such a bound, while I do have reason to believe that permitted physical constants should be from a non-discrete set.
Well, within the 3^^^3 people you have every single possible brain replicated a gazillion times already (there’s only that many ways you can arrange the atoms in the volume of human head, sufficiently distinct as to be computing something subjectively different, after all, and the number of such arrangements is unimaginably smaller than 3^^^3 ).
I don’t think that e.g. I must massively prioritize the happiness of a brain upload of me running on multiple redundant hardware (which subjectively feels the same as if it was running in one instance; it doesn’t feel any stronger because there’s more ‘copies’ of it running in perfect unison, it can’t even tell the difference. It won’t affect the subjective experience if the CPUs running the same computation are slightly physically different).
edit: also again, pseudomath, because you could have C(dustspeck, n) = 1-1/(n+1) , your property holds but it is bounded, so if the c(torture, 1)=2 then you’ll never exceed it with dust specks.
Seriously, you people (LW crowd in general) need to take more calculus or something before your mathematical intuitions become in any way relevant to anything whatsoever. It does feel intuitively that with your epsilon it’s going to keep growing without a limit, but that’s simply not true.
I consider entities in computationally distinct universes to also be distinct entities, even if the arrangements of their neurons are the same. If I have an infinite (or sufficiently large) set of physical constants such that in those universes human beings could emerge, I will also have enough human beings.
No. I will always find a larger number which is at least ε greater. I fixed ε before I talked about n,m. So I find numbers m_1,m_2,… such that C(dustspeck,m_j) > jε.
Besides which, even if I had somehow messed up, you’re not here (I hope) to score easy points because my mathematical formalization is flawed when it is perfectly obvious where I want to go.
Well, in my view, some details of implementation of a computation are totally indiscernible ‘from the inside’ and thus make no difference to the subjective experiences, qualia, and the like.
I definitely don’t care if there’s 1 me, 3^^^3 copies of me, or 3^^^^3, or 3^^^^^^3 , or the actual infinity (as the physics of our universe would suggest), where the copies are what thinks and perceives everything exactly the same over the lifetime. I’m not sure how counting copies as distinct would cope with an infinity of copies anyway. You have a torture of inf persons vs dust specks in inf*3^^^3 persons, then what?
Albeit it would be quite hilarious to see if someone here picks up the idea and starts arguing that because they’re ‘important’, there must be a lot of copies of them in the future, and thus they are rightfully an utility monster.