p(UFAI) > p(Imminent, undetected catastrophe that only a FAI can stop)
Given UFAI results in “human extinction”, and my CEV assigns effectively infinite DISutility to that outcome, it would have to FIRST provide sufficient evidence for me to update to the catastrophe being more likely.
I’ve already demonstrated that an AI which can do exactly that will get more leniency from me :)
p(UFAI) > p(Imminent, undetected catastrophe that only a FAI can stop)
Given UFAI results in “human extinction”, and my CEV assigns effectively infinite DISutility to that outcome, it would have to FIRST provide sufficient evidence for me to update to the catastrophe being more likely.
I’ve already demonstrated that an AI which can do exactly that will get more leniency from me :)