AI Alignment is, effectively, a security problem. How do you secure against an adversary that is much smarter than you?
I think Eliezer would say that this is basically impossible to do in a way that leaves the AI useful—I think he would frame the alignment problem as the problem of making the AI not an adversary in the first place.
I think Eliezer would say that this is basically impossible to do in a way that leaves the AI useful—I think he would frame the alignment problem as the problem of making the AI not an adversary in the first place.