Isn’t the worst case one in which the AI optimizes exactly against human values?
I don’t know what it means, can you give a few examples?
An example is an AI making the world as awful as possible, e.g. by creating dolorium. There is a separate question about how likely this is, hopefully very unlikely.
Yeah, I would not worry about sadistic AI being super likely, unless specifically designed.
I think so, by definition, nothing can be worse than that.
(Assuming a perfect optimizer.)
Isn’t the worst case one in which the AI optimizes exactly against human values?
I don’t know what it means, can you give a few examples?
An example is an AI making the world as awful as possible, e.g. by creating dolorium. There is a separate question about how likely this is, hopefully very unlikely.
Yeah, I would not worry about sadistic AI being super likely, unless specifically designed.
I think so, by definition, nothing can be worse than that.
(Assuming a perfect optimizer.)