I had a long back-and-forth about that topic here. Among other things, I disagree that “more or less orthogonal to CEV” is the default in the absence of alignment research, because people will presumably be trying to align their AIs, and I think there are will be obvious techniques which will work well enough to get out of the “random goal” regime, but not well enough for reliability.
I had a long back-and-forth about that topic here. Among other things, I disagree that “more or less orthogonal to CEV” is the default in the absence of alignment research, because people will presumably be trying to align their AIs, and I think there are will be obvious techniques which will work well enough to get out of the “random goal” regime, but not well enough for reliability.
people trying to align their AIs == alignment research