Every idea that comes to my mind is faced by the big question “if we were able to program a nice AI for that situation, why would we not program it to be nice in every situation”. I mean, it seems to me that in that scenario we would have both a solid definition of niceness and the ability to make the AI stick to it. Could you elaborate a little on that? Maybe an example?
This is basically in the line of my attempt to get high impact from reduced impact AIs. These are trying to extend part of “reduced impact” from a conditional situation, to a more general situation; see http://lesswrong.com/lw/m25/high_impact_from_low_impact/
Every idea that comes to my mind is faced by the big question “if we were able to program a nice AI for that situation, why would we not program it to be nice in every situation”. I mean, it seems to me that in that scenario we would have both a solid definition of niceness and the ability to make the AI stick to it. Could you elaborate a little on that? Maybe an example?
This is basically in the line of my attempt to get high impact from reduced impact AIs. These are trying to extend part of “reduced impact” from a conditional situation, to a more general situation; see http://lesswrong.com/lw/m25/high_impact_from_low_impact/
Nevermind this comment, I read some more of your posts on the subject and I think I got the point now ;)