The AI reasons with its map, the map of the world. The map depicts events that happen in the world outside of AI, and it also depicts the events that happen to the AI, or to AI’s map of the world. In AI’s map, an event in the world and AI map’s picture of that event are different elements, just as they are different elements of the world itself. The goal that guides AI’s choice of action can then distinguish between an event in the world and AI map’s representation of that event, because these two events are separately depicted in its map.
Can it however distinguish between two different events in the world that result in same map state?
edit: here, example for you. For you, some person you care about, has same place in map even though the atoms get replaced etc. If that person gets ill, you may want to mind upload that person, into an indistinguishable robot body, right? You’ll probably argue that it is a valid solution to escaping death. A lot of people have different map, and they will argue that you’re just making a substitute for your own sake, as the person will be dead, gone forever. Some other people got really bizarre map where they are mapping ‘souls’ and have the person alive in the ‘heaven’, which is on the map. Bottom line is, everyone’s just trying to resolve the problem in the map. In the territory, everyone is gone every second.
edit: and yes, you can make a map which will distinguish between sending a glider that hits the computer, and making a ton of paperclips. You still have a zillion world states, including those not filled with paperclips, mapping to the same point in map as the world filled with paperclips. Your best bet is just making the AI narrow enough that it can only find the solutions where the world is filled with paperclips.
I don’t know, the above reads to me as “Everything is confusing. Anyway, my bottom line is .” I don’t know how to parse this as an argument, how to use it to make any inferences about .
The purpose of the grandparent was to show that it’s not in principle problematic to distinguish between a goal state and that goal state’s image in the map, so there is no reason for wireheading to be consequentialistically appealing, so long as an agent is implemented carefully enough.
The AI reasons with its map, the map of the world. The map depicts events that happen in the world outside of AI, and it also depicts the events that happen to the AI, or to AI’s map of the world. In AI’s map, an event in the world and AI map’s picture of that event are different elements, just as they are different elements of the world itself. The goal that guides AI’s choice of action can then distinguish between an event in the world and AI map’s representation of that event, because these two events are separately depicted in its map.
Can it however distinguish between two different events in the world that result in same map state?
edit: here, example for you. For you, some person you care about, has same place in map even though the atoms get replaced etc. If that person gets ill, you may want to mind upload that person, into an indistinguishable robot body, right? You’ll probably argue that it is a valid solution to escaping death. A lot of people have different map, and they will argue that you’re just making a substitute for your own sake, as the person will be dead, gone forever. Some other people got really bizarre map where they are mapping ‘souls’ and have the person alive in the ‘heaven’, which is on the map. Bottom line is, everyone’s just trying to resolve the problem in the map. In the territory, everyone is gone every second.
edit: and yes, you can make a map which will distinguish between sending a glider that hits the computer, and making a ton of paperclips. You still have a zillion world states, including those not filled with paperclips, mapping to the same point in map as the world filled with paperclips. Your best bet is just making the AI narrow enough that it can only find the solutions where the world is filled with paperclips.
I don’t know, the above reads to me as “Everything is confusing. Anyway, my bottom line is .” I don’t know how to parse this as an argument, how to use it to make any inferences about .
The purpose of the grandparent was to show that it’s not in principle problematic to distinguish between a goal state and that goal state’s image in the map, so there is no reason for wireheading to be consequentialistically appealing, so long as an agent is implemented carefully enough.