But maybe I misunderstand how he arrives at the belief that “wrong” can be a one-place predicate.
Yeah. While I’m reasonably confident that he holds the belief, I have no confidence in any theories how he arrives at that belief.
What I have gotten from his writing on the subject is a combination of “Well, it sure seems that way to me,” and “Well, if that isn’t true, then I don’t see any way to build a superintelligence that does the right thing, and there has to be a way to build a superintelligence that does the right thing.” Neither of which I find compelling.
But there’s a lot of the metaethics sequence that doesn’t make much sense to me at all, so I have little confidence that what I’ve gotten out of it is a good representation of what’s there.
It’s also possible that I’m completely mistaken and he simply insists on “right” as a one-place predicate as a rhetorical trick; a way of drawing the reader’s attention away from the speaker’s role in that computation.
If that is the case I don’t see how different agents could arrive at the same perception of right and wrong, if their preferences are fundamentally opposing, given additional computation
I am fairly sure EY would say (and I agree) that there’s no reason to expect them to. Different agents with different preferences will have different beliefs about right and wrong, possibly incorrigibly different.
Humans and Babykillers as defined will simply never agree about how the universe would best be ordered, even if they come to agree (as a political exercise) on how to order the universe, without the exercise of force (as the SHFP purpose to do, for example).
(if right and wrong designate future world states).
Um.
Certainly, this model says that you can order world-states in terms of their rightness and wrongness, and there might therefore be a single possible world-state that’s most right within the set of possible world-states (though there might instead be several possible world-states that are equally right and better than all other possibilities).
If there’s only one such state, then I guess “right” could designate a future world state; if there are several, it could designate a set of world states.
But this depends on interpreting “right” to mean maximally right, in the same sense that “cold” could be understood to designate absolute zero. These aren’t the ways we actually use these words, though.
If you just argue that we don’t have free will because what is right is logically implied by cause and effect,
I don’t see what the concept of free will contributes to this discussion.
I’m fairly certain that EY would reject the idea that what’s right is logically implied by cause and effect, if by that you mean that an intelligence that started out without the right values could somehow infer, by analyzing causality in the world, what the right values were.
My own jury is to some degree still out on this one. I’m enough of a consequentialist to believe that an adequate understanding of cause and effect lets you express all judgments about right and wrong action in terms of more and less preferable world-states, but I cannot imagine how you could derive “preferable” from such an understanding. That said, my failure of imagination does not constitute a fact about the world.
Humans and Babykillers are not talking about the same subject matter when they debate what-to-do-next, and their doing different things does not constitute disagreement.
There’s a baby in front of me, and I say “Humans and Babykillers disagree about what to do next with this baby.”
The one replies: “No, they don’t. They aren’t talking about the same subject when they debate what to do next; this is not a disagreement.”
“Let me rephrase,” I say. “Babykillers prefer that this baby be killed. Humans prefer that this baby have fun. Fun and babykilling can’t both be implemented on the same baby: if it’s killed, it’s not having fun; if it’s having fun, it hasn’t been killed.”
Have I left out anything of value in my restatement? If so, what have I left out?
More generally: given all the above, why should I care whether or not what humans and Babykillers have with respect to this baby is a disagreement? What difference does that make?
More generally: given all the above, why should I care whether or not what humans and Babykillers have with respect to this baby is a disagreement? What difference does that make?
If you disagree with someone, and you’re both sufficiently rational, then you can expect to have a good shot at resolving your disagreement by arguing. That doesn’t work if you just have fundamentally different motivational frameworks.
I don’t know if I agree that a disagreement is necessarily resolvable by argument, but I certainly agree that many disagreements are so resolvable, whereas a complete difference of motivational framework is not.
If that’s what EY meant to convey by bringing up the question of whether Humans and Babykillers disagree, I agree completely.
As I said initially: “Humans and Babykillers as defined will simply never agree about how the universe would best be ordered.”
To understand the other side of the argument, I think it helps to look at this:
all disagreements are about facts. What else would you be talking about?
One side has redefined “disagreement” to mean “a difference of opinion over facts”!
I think that explains much of the sound and fury surrounding the issue.
A “difference of opinion over goals” is not a “difference of opinion over facts”.
However, note that different goals led to the cigarette companies denying the link between cigarettes and cancer—and also led to oil company AGW denialism—which caused many real-world disagreements.
All of which leaves me with the same question I started with. If I know what questions you and I give different answers to—be they questions about facts, values, goals, or whatever else—what is added to my understanding of the situation by asserting that we disagree, or don’t disagree?
ata’s reply was that “we disagree” additionally indicates that we can potentially converge on a common answer by arguing. That also seems to be what EY was getting at about hot air and rocks.
That makes sense to me, and sure, it’s additionally worth clarifying whether you and I can potentially converge on a common answer by arguing.
Anything else?
Because all of this dueling-definitions stuff strikes me as a pointless distraction. I use words to communicate concepts; if a word no longer clearly communicates concepts it’s no longer worth anything to me.
ata’s reply was that “we disagree” additionally indicates that we can potentially converge on a common answer by arguing
That doesn’t seem to be what the dictionary says “disagreement” means.
Maybe if both sides realise that the argument is pointless, they would not waste their time—but what if they don’t know what will happen? - or what if their disagreement is intended to sway not their debating partner, but a watching audience?
I agree with you about what the dictionary says, and that people might not know whether they can converge on a common answer, and that people might go through the motions of a disagreement for the benefit of observers.
We talk about what is good, and Babykillers talk about what is eat-babies, but both good and eat-babies perform analogous functions. For building a Friendly-AI we may not give a damn about how to categorize such analogous functions, but I’ve got a feeling that simply hijacking the word “moral” to suddenly not apply to such similar things, as I think it is usually used, you’ve successfully increased my confusion through the last year. Either this, or I’m back at square one. Probably the latter.
Thanks for this, very enlightening! A very good framing and analysis of my beliefs.
Yeah. While I’m reasonably confident that he holds the belief, I have no confidence in any theories how he arrives at that belief.
What I have gotten from his writing on the subject is a combination of “Well, it sure seems that way to me,” and “Well, if that isn’t true, then I don’t see any way to build a superintelligence that does the right thing, and there has to be a way to build a superintelligence that does the right thing.” Neither of which I find compelling.
But there’s a lot of the metaethics sequence that doesn’t make much sense to me at all, so I have little confidence that what I’ve gotten out of it is a good representation of what’s there.
It’s also possible that I’m completely mistaken and he simply insists on “right” as a one-place predicate as a rhetorical trick; a way of drawing the reader’s attention away from the speaker’s role in that computation.
I am fairly sure EY would say (and I agree) that there’s no reason to expect them to. Different agents with different preferences will have different beliefs about right and wrong, possibly incorrigibly different.
Humans and Babykillers as defined will simply never agree about how the universe would best be ordered, even if they come to agree (as a political exercise) on how to order the universe, without the exercise of force (as the SHFP purpose to do, for example).
Um.
Certainly, this model says that you can order world-states in terms of their rightness and wrongness, and there might therefore be a single possible world-state that’s most right within the set of possible world-states (though there might instead be several possible world-states that are equally right and better than all other possibilities).
If there’s only one such state, then I guess “right” could designate a future world state; if there are several, it could designate a set of world states.
But this depends on interpreting “right” to mean maximally right, in the same sense that “cold” could be understood to designate absolute zero. These aren’t the ways we actually use these words, though.
I don’t see what the concept of free will contributes to this discussion.
I’m fairly certain that EY would reject the idea that what’s right is logically implied by cause and effect, if by that you mean that an intelligence that started out without the right values could somehow infer, by analyzing causality in the world, what the right values were.
My own jury is to some degree still out on this one. I’m enough of a consequentialist to believe that an adequate understanding of cause and effect lets you express all judgments about right and wrong action in terms of more and less preferable world-states, but I cannot imagine how you could derive “preferable” from such an understanding. That said, my failure of imagination does not constitute a fact about the world.
Humans and Babykillers are not talking about the same subject matter when they debate what-to-do-next, and their doing different things does not constitute disagreement.
There’s a baby in front of me, and I say “Humans and Babykillers disagree about what to do next with this baby.”
The one replies: “No, they don’t. They aren’t talking about the same subject when they debate what to do next; this is not a disagreement.”
“Let me rephrase,” I say. “Babykillers prefer that this baby be killed. Humans prefer that this baby have fun. Fun and babykilling can’t both be implemented on the same baby: if it’s killed, it’s not having fun; if it’s having fun, it hasn’t been killed.”
Have I left out anything of value in my restatement? If so, what have I left out?
More generally: given all the above, why should I care whether or not what humans and Babykillers have with respect to this baby is a disagreement? What difference does that make?
If you disagree with someone, and you’re both sufficiently rational, then you can expect to have a good shot at resolving your disagreement by arguing. That doesn’t work if you just have fundamentally different motivational frameworks.
I don’t know if I agree that a disagreement is necessarily resolvable by argument, but I certainly agree that many disagreements are so resolvable, whereas a complete difference of motivational framework is not.
If that’s what EY meant to convey by bringing up the question of whether Humans and Babykillers disagree, I agree completely.
As I said initially: “Humans and Babykillers as defined will simply never agree about how the universe would best be ordered.”
We previously debated the disagreements between those with different values here.
The dictionary apparently supports the idea that any conflict is a disagreement.
To understand the other side of the argument, I think it helps to look at this:
One side has redefined “disagreement” to mean “a difference of opinion over facts”!
I think that explains much of the sound and fury surrounding the issue.
A “difference of opinion over goals” is not a “difference of opinion over facts”.
However, note that different goals led to the cigarette companies denying the link between cigarettes and cancer—and also led to oil company AGW denialism—which caused many real-world disagreements.
All of which leaves me with the same question I started with. If I know what questions you and I give different answers to—be they questions about facts, values, goals, or whatever else—what is added to my understanding of the situation by asserting that we disagree, or don’t disagree?
ata’s reply was that “we disagree” additionally indicates that we can potentially converge on a common answer by arguing. That also seems to be what EY was getting at about hot air and rocks.
That makes sense to me, and sure, it’s additionally worth clarifying whether you and I can potentially converge on a common answer by arguing.
Anything else?
Because all of this dueling-definitions stuff strikes me as a pointless distraction. I use words to communicate concepts; if a word no longer clearly communicates concepts it’s no longer worth anything to me.
That doesn’t seem to be what the dictionary says “disagreement” means.
Maybe if both sides realise that the argument is pointless, they would not waste their time—but what if they don’t know what will happen? - or what if their disagreement is intended to sway not their debating partner, but a watching audience?
I agree with you about what the dictionary says, and that people might not know whether they can converge on a common answer, and that people might go through the motions of a disagreement for the benefit of observers.
We talk about what is good, and Babykillers talk about what is eat-babies, but both good and eat-babies perform analogous functions. For building a Friendly-AI we may not give a damn about how to categorize such analogous functions, but I’ve got a feeling that simply hijacking the word “moral” to suddenly not apply to such similar things, as I think it is usually used, you’ve successfully increased my confusion through the last year. Either this, or I’m back at square one. Probably the latter.