A crux here is that I basically don’t think Coherent Extrapolated Volition of humanity type alignment strategies work, and I also think that it is irrelevant that we can’t align an AI to the CEV of humanity.
A crux here is that I basically don’t think Coherent Extrapolated Volition of humanity type alignment strategies work, and I also think that it is irrelevant that we can’t align an AI to the CEV of humanity.