The most obvious way to escape “false positives” is not to create AI at all, and in any other reasonable field it will be the correct solution. For example, if some organisation wants to create a reactor which has unmeasurable probability of catastrophic failure, no commission will ever allow it to be built.
But, as we all here understand, if aligned AI is not created and not used to stop AI race, another non-safe AI will eventually appear. Thus the only thing for which we really need the aligned AI is to stop other AI projects. In other words, we need aligned AI only as a weapon (or, to put in more socially acceptable terms, to make “pivotal acts”).
If we need aligned AI only as a weapon to stop other AI projects, the good question is: may be we have different safer instruments to stop other AI projects. One of such hypothetical instruments is the use of Narrow AI for global domination and policing, which I explored in my previous post. There could be other possible instruments, non of them is perfect, but they may have high probability to work.
The most obvious way to escape “false positives” is not to create AI at all, and in any other reasonable field it will be the correct solution. For example, if some organisation wants to create a reactor which has unmeasurable probability of catastrophic failure, no commission will ever allow it to be built.
But, as we all here understand, if aligned AI is not created and not used to stop AI race, another non-safe AI will eventually appear. Thus the only thing for which we really need the aligned AI is to stop other AI projects. In other words, we need aligned AI only as a weapon (or, to put in more socially acceptable terms, to make “pivotal acts”).
If we need aligned AI only as a weapon to stop other AI projects, the good question is: may be we have different safer instruments to stop other AI projects. One of such hypothetical instruments is the use of Narrow AI for global domination and policing, which I explored in my previous post. There could be other possible instruments, non of them is perfect, but they may have high probability to work.