there are strong arguments that control of strongly superhuman AI systems will not be amenable to prosaic alignment
In which section of the linked paper is the strong argument for this conclusion to be found? I had a quick read of it but could not see it—I skipped the long sections of quotes, as the few I read were claims rather than arguments.
I’m not going to try to summarize the arguments here, but it’s been discussed on this site for a decade. And those quoted bits of the paper were citing the extensive discussions about this point—that’s why there were several hundred citations, many of which were to Lesswrong posts.
In which section of the linked paper is the strong argument for this conclusion to be found? I had a quick read of it but could not see it—I skipped the long sections of quotes, as the few I read were claims rather than arguments.
I’m not going to try to summarize the arguments here, but it’s been discussed on this site for a decade. And those quoted bits of the paper were citing the extensive discussions about this point—that’s why there were several hundred citations, many of which were to Lesswrong posts.