I wanted to write about my opinion that human values can’t be divided into final values and instrumental values, the way discussion of FAI presumes they can. This is an idea that comes from mathematics, symbolic logic, and classical AI. A symbolic approach would probably make proving safety easier. But human brains don’t work that way. You can and do change your values over time, because you don’t really have terminal values.
You may have wanted to—but AFAICS, you didn’t—apart from this paragraph. It seems to me that it fails to make its case. The split applies to any goal-directed agent, irrespective of implemetation details.
You may have wanted to—but AFAICS, you didn’t—apart from this paragraph. It seems to me that it fails to make its case. The split applies to any goal-directed agent, irrespective of implemetation details.