While I don’t think AI should have objective functions at all (see Eric Drexler’s proposals here), I do like one thing about this one.
Curiosity is implementable. Is a tractable, numerical value.
Specifically, the machine can use it’s available sensors to get a compressed map of the current frame it’s on, M. It can make a future frame prediction, P., conditional on the machine’s actions, An.
Curiosity is satisfied with P has low confidence—it is not confidence on what the results of some actions will be.
So you could add to the machine’s reward counter positive reward for taking actions that gain information, if the gain is expected to be less than the cost.
With that said, it’s still a bad idea of general AIs free to roam in ‘our world’ to have this function.
While I don’t think AI should have objective functions at all (see Eric Drexler’s proposals here), I do like one thing about this one.
Curiosity is implementable. Is a tractable, numerical value.
Specifically, the machine can use it’s available sensors to get a compressed map of the current frame it’s on, M. It can make a future frame prediction, P., conditional on the machine’s actions, An.
Curiosity is satisfied with P has low confidence—it is not confidence on what the results of some actions will be.
So you could add to the machine’s reward counter positive reward for taking actions that gain information, if the gain is expected to be less than the cost.
With that said, it’s still a bad idea of general AIs free to roam in ‘our world’ to have this function.