Nonlinear utility functions (as a function of resources) do not accurately model human risk aversion. That could imply that we should either change our (or they/their) risk aversion or not be maximising expected utility.
Nonlinear jumps in utility from different amounts of a resource seem common for humans at least at some points in time. Example: Either I have enough to pay off the loan shark, or he’ll break my legs.
Nonlinear utility functions (as a function of resources) do not accurately model human risk aversion. That could imply that we should either change our (or they/their) risk aversion or not be maximising expected utility.
Nonlinear jumps in utility from different amounts of a resource seem common for humans at least at some points in time. Example: Either I have enough to pay off the loan shark, or he’ll break my legs.
Yep. Humans are not expected utility maximisers. But there’s strong arguments that an AI would be...