Given the extent to which the proper use of general categories of reason depends upon the ends you wish to use the concept for, and the extent to which goals and values are entangled, I’m wondering if it’s even possible to create an intelligent but non omniscient agent which uses these categories but that does not have some kind of implicit value preference structure.
I don’t think it is possible, which makes FAI even harder to achieve.
Given the extent to which the proper use of general categories of reason depends upon the ends you wish to use the concept for, and the extent to which goals and values are entangled, I’m wondering if it’s even possible to create an intelligent but non omniscient agent which uses these categories but that does not have some kind of implicit value preference structure.
I don’t think it is possible, which makes FAI even harder to achieve.