While this view may be correct, its optic is bad, as “alignment” become synonymous to “taking over the world”, and people will start seeing this before it is actually implemented.
They will see something like: “When they say “alignment”, they mean that AI should ignore anything I say and start taking the world, so it is not “AI alignment”, but “world alignment”.
They will see that AI alignment is opposite to AI safety, as Aligned AI must start taking very risky and ambitious actions to perform Pivotal act.
While this view may be correct, its optic is bad, as “alignment” become synonymous to “taking over the world”, and people will start seeing this before it is actually implemented.
They will see something like: “When they say “alignment”, they mean that AI should ignore anything I say and start taking the world, so it is not “AI alignment”, but “world alignment”.
They will see that AI alignment is opposite to AI safety, as Aligned AI must start taking very risky and ambitious actions to perform Pivotal act.