because when you train something using gradient descent optimised against a loss function it de facto has some kind of utility function. You cant accomplish all that much without a utility function.
a utility function is a particular long-term formulation of a preference function; in principle any preference function is convertible to a utility function, given zero uncertainty about the space of possible future trajectories. a preference is when a system tends to push the world towards some trajectories over others. not only can you not accomplish much without your behavior implying a utility function, it’s impossible to not have an implicit utility function, as you can define a revealed preference utility function for any hunk of matter.
doesn’t mean that the system is evaluating things using a zero computational uncertainty model of the future like in the classic utility maximizer formulation though. I think evolutionary fitness is a better way to think about this—the preferences that preserve themselves are the ones that win.
it’s impossible to not have an implicit utility function, as you can define a revealed preference utility function for any hunk of matter.
Yes, you can “prove” that everything has a UF by trivializing UF, and this has been done many times, and it isn’t a good argument because of the trivialisation.
I think evolutionary fitness is a better way to think about this—the preferences that preserve themselves are the ones that win.
The preferences that please humans are the ones that win.
The preferences that please humans are the ones that win.
aha! what about preferences that help humans hurt each other? we need only imagine ais used in war as their strength grows. the story where ai jump on their own to malice is unnecessary, humans will boost it to that directly. oh, also scammers.
because when you train something using gradient descent optimised against a loss function it de facto has some kind of utility function. You cant accomplish all that much without a utility function.
a utility function is a particular long-term formulation of a preference function; in principle any preference function is convertible to a utility function, given zero uncertainty about the space of possible future trajectories. a preference is when a system tends to push the world towards some trajectories over others. not only can you not accomplish much without your behavior implying a utility function, it’s impossible to not have an implicit utility function, as you can define a revealed preference utility function for any hunk of matter.
doesn’t mean that the system is evaluating things using a zero computational uncertainty model of the future like in the classic utility maximizer formulation though. I think evolutionary fitness is a better way to think about this—the preferences that preserve themselves are the ones that win.
Yes, you can “prove” that everything has a UF by trivializing UF, and this has been done many times, and it isn’t a good argument because of the trivialisation.
The preferences that please humans are the ones that win.
yes, that was my point about ufs.
aha! what about preferences that help humans hurt each other? we need only imagine ais used in war as their strength grows. the story where ai jump on their own to malice is unnecessary, humans will boost it to that directly. oh, also scammers.