Would AI safety be easy if all researchers agreed that the pleasure-pain axis is the world’s objective metric of value?
No. It would make a difference but it wouldn’t solve the problem. The clearest reason is that it doesn’t help with Inner Alignment at all.
Would AI safety be easy if all researchers agreed that the pleasure-pain axis is the world’s objective metric of value?
No. It would make a difference but it wouldn’t solve the problem. The clearest reason is that it doesn’t help with Inner Alignment at all.