I agree, one can conceive of AGI safety without alignment (e.g. if boxing worked), and one can conceive of alignment without safety (e.g. if the AI is “trying to do the right thing” but is careless or incompetent or whatever). I usually use the term “AGI Safety” when describing my job, but the major part of it is thinking about the alignment problem.
I agree, one can conceive of AGI safety without alignment (e.g. if boxing worked), and one can conceive of alignment without safety (e.g. if the AI is “trying to do the right thing” but is careless or incompetent or whatever). I usually use the term “AGI Safety” when describing my job, but the major part of it is thinking about the alignment problem.