FWIW I find using the word “alignment” when specifically what you mean is “safety” to be confusing. I consider AI alignment to be “getting actively positive things that you want because of AI”. I consider AI safety to be “getting no actively negative things to happen because of AI”. And it seems that Davidad’s approach is much more about safety than it is alignment.
FWIW I find using the word “alignment” when specifically what you mean is “safety” to be confusing. I consider AI alignment to be “getting actively positive things that you want because of AI”. I consider AI safety to be “getting no actively negative things to happen because of AI”. And it seems that Davidad’s approach is much more about safety than it is alignment.