Yudkowsky and others might give different reasons why waiting until later to gain more information about the future systems doesn’t make sense, including pointing out that that may lead us to missing our first “critical try.”
Robin, I know you must have heard these points before—I believe you are more familiar with e.g. Eliezer’s views than I am. But if that’s the case I don’t understand why you would write a sentence like last one in the quotation above. It sounds like a cheap rhetorical trick to say “but instead of waiting to deal with such problems when we understand them better and can envision them more concretely” especially without saying why people who don’t think we should wait don’t think that’s a good enough reason to wait / think there are pressing reasons to work on the problems now despite our relative state of ignorance compared to future AI researchers.
Yudkowsky and others might give different reasons why waiting until later to gain more information about the future systems doesn’t make sense, including pointing out that that may lead us to missing our first “critical try.”
Robin, I know you must have heard these points before—I believe you are more familiar with e.g. Eliezer’s views than I am. But if that’s the case I don’t understand why you would write a sentence like last one in the quotation above. It sounds like a cheap rhetorical trick to say “but instead of waiting to deal with such problems when we understand them better and can envision them more concretely” especially without saying why people who don’t think we should wait don’t think that’s a good enough reason to wait / think there are pressing reasons to work on the problems now despite our relative state of ignorance compared to future AI researchers.