Hmm, maybe I’m interpreting the statement to mean something weaker and more handwavy than you are. I agree with claims like “with current technology, it can be hard to make an AI pursue some goals as competently as other goals” and “if a goal is hard to specify given available training data, then it’s harder to make an AI pursue it”.
However, I think how competently an AI pursues a goal is somewhat different than whether an AI tries to pursues a goal at all.(Which is what I think the strong version of the thesis is still getting at.) I was trying to get at the “hard to specify” thing with the simplicity caveat. There are also many other caveats because goals and other concepts are quite handwavy.
Doesn’t seem important to discuss further.
I think I agree with everything you said. (Except for the psychologising about Eliezer on which I have no particular opinion.)
Hmm, maybe I’m interpreting the statement to mean something weaker and more handwavy than you are. I agree with claims like “with current technology, it can be hard to make an AI pursue some goals as competently as other goals” and “if a goal is hard to specify given available training data, then it’s harder to make an AI pursue it”.
However, I think how competently an AI pursues a goal is somewhat different than whether an AI tries to pursues a goal at all.(Which is what I think the strong version of the thesis is still getting at.) I was trying to get at the “hard to specify” thing with the simplicity caveat. There are also many other caveats because goals and other concepts are quite handwavy.
Doesn’t seem important to discuss further.
I think I agree with everything you said. (Except for the psychologising about Eliezer on which I have no particular opinion.)