This is all true. But humans do not have utility functions… Humans are not the coherent, consistent agents you make them out to be.
If you think that’s relevant, you should also go write the same comment on Eliezer’s post on utilons and fuzzies. Having two coherent, consistent utility functions is no more realistic than having one.
If you want to be rational, you need to try to figure out what your values are, and what your utility function is. Humans don’t act consistently. Whether their preferences can be described by a utility function is a more subtle question whose answer is unknown. But in either case, in order to be more rational, you need to be able to approximate your preferences with a utility function.
Fighting for willpower is basically your far-self trying to wrest control of your behavior from your near-self.
You can alternately describe this as the place where the part of your utility function that you call your far self, and the part of your utility function that you call your near self, sum to zero and provide no net information on what to do. You can choose to describe the resultant emotional confusion as “fighting for willpower”. But this leads to the erroneous conclusions I described under the “ethics as willpower” section.
Just to clarify I am not, not, not defending the willpower model you described—I just don’t think willpower, properly understood as a conflict between near and far modes can be left out of an account of human decision making processes. I think the situation is both more complicated and more troubling than both models and don’t think it is rational to force the square peg that is human values into the round hole that is ‘the utility function’.
I’ll agree that willpower may be a useful concept. I’m not providing a full model, though—mostly I want to dismiss the folk-psychology close tie between willpower and morals.
Having two coherent, consistent utility functions is no more realistic than having one.
He never said these “utility functions” are coherent. In fact a large part of the problem is that the “fuzzies” utility function is extremely incoherent.
You keep using that word. I do not think it means what you think it means. A utility function that is incoherent is not a utility function.
If it is acceptable for Eliezer to talk about having two utility functions, one that measures utilons and one that measures fuzzies, then it is equally acceptable to talk about having a single utility function, with respect to the question of whether humans are capable of having utility functions.
A utility function that is incoherent is not a utility function.
I was using the same not-quite strict definition of “utility function” that you seemed to be using in your post. In any case, I don’t believe Eliezer ever called fuzzies a utility function.
If you want to be rational, you need to try to figure out what your values are, and what your utility function is. Humans don’t act consistently. Whether their preferences can be described by a utility function is a more subtle question whose answer is unknown. But in either case, in order to be more rational, you need to be able to approximate your preferences with a utility function.
This is neither here nor there. I have no doubt it can help to approximate your preferences with a utility function. But simply erasing complication by reducing all your preference-like stuff to a utility function decreases the accuracy of your model. You’re ignoring what is really going on inside. So yes, if you try to model humans as holders of single utility functions… morality has nothing to do with willpower! Congrats! But my point is that such a model is far too simple.
You can alternately describe this as the place where the part of your utility function that you call your far self, and the part of your utility function that you call your near self, sum to zero and provide no net information on what to do. You can choose to describe the resultant emotional confusion as “fighting for willpower”.
Well you can do that—it doesn’t seem at all representative of the way choices are made, though.
But this leads to the erroneous conclusions I described under the “ethics as willpower” section.
What erroneous conclusions? What does it predict that is not so?
If you think that’s relevant, you should also go write the same comment on Eliezer’s post on utilons and fuzzies. Having two coherent, consistent utility functions is no more realistic than having one.
If you want to be rational, you need to try to figure out what your values are, and what your utility function is. Humans don’t act consistently. Whether their preferences can be described by a utility function is a more subtle question whose answer is unknown. But in either case, in order to be more rational, you need to be able to approximate your preferences with a utility function.
You can alternately describe this as the place where the part of your utility function that you call your far self, and the part of your utility function that you call your near self, sum to zero and provide no net information on what to do. You can choose to describe the resultant emotional confusion as “fighting for willpower”. But this leads to the erroneous conclusions I described under the “ethics as willpower” section.
Just to clarify I am not, not, not defending the willpower model you described—I just don’t think willpower, properly understood as a conflict between near and far modes can be left out of an account of human decision making processes. I think the situation is both more complicated and more troubling than both models and don’t think it is rational to force the square peg that is human values into the round hole that is ‘the utility function’.
I’ll agree that willpower may be a useful concept. I’m not providing a full model, though—mostly I want to dismiss the folk-psychology close tie between willpower and morals.
He never said these “utility functions” are coherent. In fact a large part of the problem is that the “fuzzies” utility function is extremely incoherent.
You keep using that word. I do not think it means what you think it means. A utility function that is incoherent is not a utility function.
If it is acceptable for Eliezer to talk about having two utility functions, one that measures utilons and one that measures fuzzies, then it is equally acceptable to talk about having a single utility function, with respect to the question of whether humans are capable of having utility functions.
I was using the same not-quite strict definition of “utility function” that you seemed to be using in your post. In any case, I don’t believe Eliezer ever called fuzzies a utility function.
This is neither here nor there. I have no doubt it can help to approximate your preferences with a utility function. But simply erasing complication by reducing all your preference-like stuff to a utility function decreases the accuracy of your model. You’re ignoring what is really going on inside. So yes, if you try to model humans as holders of single utility functions… morality has nothing to do with willpower! Congrats! But my point is that such a model is far too simple.
Well you can do that—it doesn’t seem at all representative of the way choices are made, though.
What erroneous conclusions? What does it predict that is not so?