What would “incoherent decisions” look like for an agent that has a utility function defined on action-observation histories?
What would “incoherent decisions” look like for an agent that has a utility function defined on action-observation histories?