This part doesn’t seem to pass the ideological Turing test:
At the moment, AIs are not powerful enough to cause us harm, and we hardly know anything about the structures and uses of future AIs that might cause bigger problems. But instead of waiting to deal with such problems when we understand them better and can envision them more concretely, AI “doomers” want stronger guarantees now.
To clarify explicitly, people like Stuart Russell would point out that if future AIs are still built according to the “standard model” (a phrase I borrow from Russell) like the systems of today, then they will continue to be predictably misaligned.
Yudkowsky and others might give different reasons why waiting until later to gain more information about the future systems doesn’t make sense, including pointing out that that may lead us to missing our first “critical try.”
Robin, I know you must have heard these points before—I believe you are more familiar with e.g. Eliezer’s views than I am. But if that’s the case I don’t understand why you would write a sentence like last one in the quotation above. It sounds like a cheap rhetorical trick to say “but instead of waiting to deal with such problems when we understand them better and can envision them more concretely” especially without saying why people who don’t think we should wait don’t think that’s a good enough reason to wait / think there are pressing reasons to work on the problems now despite our relative state of ignorance compared to future AI researchers.
This part doesn’t seem to pass the ideological Turing test:
To clarify explicitly, people like Stuart Russell would point out that if future AIs are still built according to the “standard model” (a phrase I borrow from Russell) like the systems of today, then they will continue to be predictably misaligned.
Yudkowsky and others might give different reasons why waiting until later to gain more information about the future systems doesn’t make sense, including pointing out that that may lead us to missing our first “critical try.”
Robin, I know you must have heard these points before—I believe you are more familiar with e.g. Eliezer’s views than I am. But if that’s the case I don’t understand why you would write a sentence like last one in the quotation above. It sounds like a cheap rhetorical trick to say “but instead of waiting to deal with such problems when we understand them better and can envision them more concretely” especially without saying why people who don’t think we should wait don’t think that’s a good enough reason to wait / think there are pressing reasons to work on the problems now despite our relative state of ignorance compared to future AI researchers.