Oracle AI has to be a last resort when the shit hits the fan.
If we had infinite time, I’d agree with you. But I’m feeling that we have little chance of solving FAI before the shit indeed does hit the fan and us. The route safe Oracle → Oracle asisted FAI design seems more plausible to me. Especially as we are so much better at correcting errors than preventing them, so a prediction Oracle (if safe) would play to our strengths.
But I’m feeling that we have little chance of solving FAI before the shit indeed does hit the fan and us.
If I assume a high probability of risks from AI and a short planning horizon then I agree. But it is impossible to say. I take the same stance as Holden Karnofsky from GiveWell regarding the value of FAI research at this point:
I think that if you’re aiming to develop knowledge that won’t be useful until very very far in the future, you’re probably wasting your time, if for no other reason than this: by the time your knowledge is relevant, someone will probably have developed a tool (such as a narrow AI) so much more efficient in generating this knowledge that it renders your work moot.
I think the same applies for fail-safe mechanisms and Oracle AI, although to a lesser extent.
The route safe Oracle → Oracle asisted FAI design...
What is your agenda for developing such a safe Oracle? Are you going to do AGI research first and along the way try to come up with solutions on how to make it safe? I think that would be a promising approach. But if you are trying to come up with ways on how to ensure the safety of a fictive Oracle, whose nature is a mystery to you, then the argument mentioned above counts again.
If we had infinite time, I’d agree with you. But I’m feeling that we have little chance of solving FAI before the shit indeed does hit the fan and us. The route safe Oracle → Oracle asisted FAI design seems more plausible to me. Especially as we are so much better at correcting errors than preventing them, so a prediction Oracle (if safe) would play to our strengths.
If I assume a high probability of risks from AI and a short planning horizon then I agree. But it is impossible to say. I take the same stance as Holden Karnofsky from GiveWell regarding the value of FAI research at this point:
I think the same applies for fail-safe mechanisms and Oracle AI, although to a lesser extent.
What is your agenda for developing such a safe Oracle? Are you going to do AGI research first and along the way try to come up with solutions on how to make it safe? I think that would be a promising approach. But if you are trying to come up with ways on how to ensure the safety of a fictive Oracle, whose nature is a mystery to you, then the argument mentioned above counts again.