If the AI is Unfriendly, then regardless of limitations imposed it will be harmful.
That’s not true. Unfriendly doesn’t mean that the AI necessarily tries to destroy the human race.
If you tell the paperclip AI: Produce 10000 paperclips, it might produce no harm. If you tell it to give you as many paperclips as possible it does harm.
When it comes to powerful entities you want checks&balances. The programmers of the AI can do a better job at checks&balances when the AI is completely truthful.
Sure, if the scale is lower it’s less likely to produce large-scale harm, but it is still likely to produce small-scale harm. And satisficing doesn’t actually protect against large-scale harm; that’s been argued pretty extensively previously, so the example you provided is still going to have large-scale harm.
Ultimately, though, checks & balances are also just rules for the genie. It’s not going to render an Unfriendly AI Friendly, and it won’t actually limit a superintelligent AI regardless, since they can game you to render the balances irrelevant. (Unless you think that AI-boxing would actually work. It’s the same principle.)
I’m really not seeing anything that distinguishes this from Failed Utopia 4-2. This even one of that genie’s rules!
I’m not sure how you could even specify ‘don’t game me’. That’s much more complicated than ‘don’t manipulate me’, which is itself pretty difficult to specify.
This clearly isn’t going anywhere and if there’s an inferential gap I can’t see what it is, so unless there’s some premise of yours you want to explain or think there’s something I should explain, I’m done with this debate.
How do you build a superintelligent AI in the first place? I think there are plenty of ways of allowing the programmers direct access to internal deliberations of the AI and see anything that looks like the AI even thinking about manipulating the programmers as a thread.
That’s not true. Unfriendly doesn’t mean that the AI necessarily tries to destroy the human race. If you tell the paperclip AI: Produce 10000 paperclips, it might produce no harm. If you tell it to give you as many paperclips as possible it does harm.
When it comes to powerful entities you want checks&balances. The programmers of the AI can do a better job at checks&balances when the AI is completely truthful.
Sure, if the scale is lower it’s less likely to produce large-scale harm, but it is still likely to produce small-scale harm. And satisficing doesn’t actually protect against large-scale harm; that’s been argued pretty extensively previously, so the example you provided is still going to have large-scale harm.
Ultimately, though, checks & balances are also just rules for the genie. It’s not going to render an Unfriendly AI Friendly, and it won’t actually limit a superintelligent AI regardless, since they can game you to render the balances irrelevant. (Unless you think that AI-boxing would actually work. It’s the same principle.)
I’m really not seeing anything that distinguishes this from Failed Utopia 4-2. This even one of that genie’s rules!
The fact that they could game you theoretically is why it’s important to give it a precommitment to not game you. To not even think about gaming you.
I’m not sure how you could even specify ‘don’t game me’. That’s much more complicated than ‘don’t manipulate me’, which is itself pretty difficult to specify.
This clearly isn’t going anywhere and if there’s an inferential gap I can’t see what it is, so unless there’s some premise of yours you want to explain or think there’s something I should explain, I’m done with this debate.
How do you give a superintelligent AI a precommitment?
How do you build a superintelligent AI in the first place? I think there are plenty of ways of allowing the programmers direct access to internal deliberations of the AI and see anything that looks like the AI even thinking about manipulating the programmers as a thread.