What you label “implicit utility function” sounds like instrumental goals to me. Some of that is also covered under Basic AI Drives.
I’m not familiar with the pig that wants to be eaten, but I’m not sure I would describe that as a conflicted utility function. If one has a utility function that places maximum utility on an outcome that requires their death, then there is no conflict, that is the optimal choice. Though I think human’s who think they have such a utility function are usually mistaken, but that is a much more involved discussion.
Not sure what the point of a dynamic utility function is. Your values really shouldn’t change. I feel like you may be focused on instrumental goals that can and should change and thinking those are part of the utility function when they are not.
What you label “implicit utility function” sounds like instrumental goals to me. Some of that is also covered under Basic AI Drives.
I’m not familiar with the pig that wants to be eaten, but I’m not sure I would describe that as a conflicted utility function. If one has a utility function that places maximum utility on an outcome that requires their death, then there is no conflict, that is the optimal choice. Though I think human’s who think they have such a utility function are usually mistaken, but that is a much more involved discussion.
Not sure what the point of a dynamic utility function is. Your values really shouldn’t change. I feel like you may be focused on instrumental goals that can and should change and thinking those are part of the utility function when they are not.