One small bit of such interaction could be rephrasing of commands.
Human: “I want apple.”
Robot: “Do you want a computer or a fruit?”
Another way of interaction is presenting of the plan of actions, may be drawing it as a visual image:
Robot: “To give you an apple, I have to go to the shop, which will take at least one hour.”
Human: “No, just find the apple in the refrigerator.”
The third way is to confirm that human still want X after reasonable amounts of time, say, every one day:
Robot: “Yesterday you asked for an apple. Do you still want it?”
The forth is sending reports after every timestamp, which describes how the the project is going and which new subgoals appeared:
Robot: “There is no apples in the shop. I am going to another village, but it will take two more hours.”
Human: “No, it is too long, buy me an orange.”
The general intuition pomp for such interactions is relations between a human and ideal human secretary, and such pomp could be even used to train the robot. Again, this type of learning is possible only after the biggest part of AI safety is solved, or the robot will go foom after the first question.
I agree that all of these seem like good aspects of human-AI interaction to have, especially for narrow AI systems. For superhuman AI systems, there’s a question of how much of this should the AI infer for itself vs. make sure to ask the human.
There is a problem of “moral unemployment”—that is, if superintelligent AI will do all hard work of analysing “what I should want”, it will strip from me the last pleasant duty I may have.
E.g: Robot: “I know that the your deepest desire, which you may be not fully aware of, but after a lot of suffering, you will learn it for sure – is to write a novel. And I already wrote this novel for you—the best one which you could possibly write.”
One small bit of such interaction could be rephrasing of commands.
Human: “I want apple.”
Robot: “Do you want a computer or a fruit?”
Another way of interaction is presenting of the plan of actions, may be drawing it as a visual image:
Robot: “To give you an apple, I have to go to the shop, which will take at least one hour.”
Human: “No, just find the apple in the refrigerator.”
The third way is to confirm that human still want X after reasonable amounts of time, say, every one day:
Robot: “Yesterday you asked for an apple. Do you still want it?”
The forth is sending reports after every timestamp, which describes how the the project is going and which new subgoals appeared:
Robot: “There is no apples in the shop. I am going to another village, but it will take two more hours.”
Human: “No, it is too long, buy me an orange.”
The general intuition pomp for such interactions is relations between a human and ideal human secretary, and such pomp could be even used to train the robot. Again, this type of learning is possible only after the biggest part of AI safety is solved, or the robot will go foom after the first question.
I agree that all of these seem like good aspects of human-AI interaction to have, especially for narrow AI systems. For superhuman AI systems, there’s a question of how much of this should the AI infer for itself vs. make sure to ask the human.
There is a problem of “moral unemployment”—that is, if superintelligent AI will do all hard work of analysing “what I should want”, it will strip from me the last pleasant duty I may have.
E.g: Robot: “I know that the your deepest desire, which you may be not fully aware of, but after a lot of suffering, you will learn it for sure – is to write a novel. And I already wrote this novel for you—the best one which you could possibly write.”