The response to Unknown sums up the issue already, though.
You may be justified in ascertaining that the AI will figure out what they’re doing. You’re not justified in assuming it will then act on this knowledge instead of identifying and pursuing its own purposes (presuming you’ve codified “purpose” enough for it to not just sit there and modify its own utility function to produce the computer equivalent of shooting up heroin).
Until you know what you’re doing, you can’t get something else to do it for you. The AI programmed without knowledge of what they wanted it to do might cooperate, might not. It would be better to start over, programming it specifically to do what you want it to do.
The response to Unknown sums up the issue already, though.
You may be justified in ascertaining that the AI will figure out what they’re doing. You’re not justified in assuming it will then act on this knowledge instead of identifying and pursuing its own purposes (presuming you’ve codified “purpose” enough for it to not just sit there and modify its own utility function to produce the computer equivalent of shooting up heroin).
Until you know what you’re doing, you can’t get something else to do it for you. The AI programmed without knowledge of what they wanted it to do might cooperate, might not. It would be better to start over, programming it specifically to do what you want it to do.