Uncertainty about metaethics seems a serious source of risk in AI safety, and especially AI alignment. I’ve written a paper detailing how we might approach such fundamental uncertainty such that we can perform analysis to find positions which minimize risk such that we don’t unnecessarily expose ourselves to risk by unnecessarily making assumptions we need not make.
Uncertainty about metaethics seems a serious source of risk in AI safety, and especially AI alignment. I’ve written a paper detailing how we might approach such fundamental uncertainty such that we can perform analysis to find positions which minimize risk such that we don’t unnecessarily expose ourselves to risk by unnecessarily making assumptions we need not make.