Humans seem pretty good at making correct predictions even if they have made incorrect predictions in the past. More generally, any agent for whom a single wrong prediction throws everything into disarray will probably not continue to function for very long.
That’s basically my point. A human has to predict the answer to questions of the type “what would I do in situation X”, and their overall behavior is the sum of their actions over all situations, so they can still get the overall result roughly correct as long as they are correct on average. An AI that’s capable of self-modification also has to predict the answer to questions of the type “how would my behavior be affected if I modified my decision-making algorithm in this way”, where the answer doesn’t just influence the behavior in one situation but all the ones that follow. The effects of individual decisions become global rather than local. It needs to be able to make much more reliable predictions if it wants to have a chance of even remaining basically operational over the long term.
Fair enough. This is an admirable habit that is all too rare, so have an upvote :).
That’s basically my point. A human has to predict the answer to questions of the type “what would I do in situation X”, and their overall behavior is the sum of their actions over all situations, so they can still get the overall result roughly correct as long as they are correct on average. An AI that’s capable of self-modification also has to predict the answer to questions of the type “how would my behavior be affected if I modified my decision-making algorithm in this way”, where the answer doesn’t just influence the behavior in one situation but all the ones that follow. The effects of individual decisions become global rather than local. It needs to be able to make much more reliable predictions if it wants to have a chance of even remaining basically operational over the long term.
Thanks. :)
And more important, its creators want to be sure that it will be very reliable before they switch it on.