I agree that corrigibility, task AGI, etc. is a better thing for the field to focus on than value learning.
This seems like a real cost of the term “AI alignment”, especially insofar as researchers like Stuart Russell have introduced the term “value alignment” and used “alignment” as a shorthand for that.
I agree that corrigibility, task AGI, etc. is a better thing for the field to focus on than value learning.
This seems like a real cost of the term “AI alignment”, especially insofar as researchers like Stuart Russell have introduced the term “value alignment” and used “alignment” as a shorthand for that.