I agree that corrigibility, task AGI, etc. is a better thing for the field to focus on than value learning.
This seems like a real cost of the term “AI alignment”, especially insofar as researchers like Stuart Russell have introduced the term “value alignment” and used “alignment” as a shorthand for that.
I prefer “AI Safety” over “AI Alignment” because I associate the first more with Corrigibility, and the second more with Value-alignment.
It is the term “Safe AI” that implies 0% risk, while “AI Safety” seems more similar to “Aircraft Safety” in acknowledging a non-zero risk.
I agree that corrigibility, task AGI, etc. is a better thing for the field to focus on than value learning.
This seems like a real cost of the term “AI alignment”, especially insofar as researchers like Stuart Russell have introduced the term “value alignment” and used “alignment” as a shorthand for that.