If I’d never heard of the AI-Box Experiment, I’d think that zero call’s comment was a reasonable contribution to a conversation about AI and safety in particular. It’s only when we realize that object-level methods of restraining a transhuman intelligence are probably doomed that we know we must focus so precisely on getting its goals right.
Please point me to some more details about the AI box experiment, since I think what i suggested earlier as isolated virtual worlds is pretty much the same as what zero call is suggesting here.
I feel that there are huge assumptions in the present AI Box experiment. The gatekeeper and the AI share a language, for one, by which the AI convinces the gatekeeper.
If AGI is your only criteria without regards to friendliness, just make sure not to communicate with the AI. Turing tests are not the only proofs of intelligence. If the agi can come up with unique solutions in the universe in which it is isolated, that is enough to understand this algorithm is creative.
If observing but not communicating with a boxed AI does a good enough job of patching the security holes (which I understand that it might not—that’s for someone who better understands the issue to look at), perhaps putting an instance of a potential FAI in a contained virtual world would be useful as a test. It seems to me that a FAI that didn’t have humans to start with would perhaps have to invent us, or something like us in some specific observable way(s), because of its values.
If I’d never heard of the AI-Box Experiment, I’d think that zero call’s comment was a reasonable contribution to a conversation about AI and safety in particular. It’s only when we realize that object-level methods of restraining a transhuman intelligence are probably doomed that we know we must focus so precisely on getting its goals right.
Vladimir and orthonormal,
Please point me to some more details about the AI box experiment, since I think what i suggested earlier as isolated virtual worlds is pretty much the same as what zero call is suggesting here.
I feel that there are huge assumptions in the present AI Box experiment. The gatekeeper and the AI share a language, for one, by which the AI convinces the gatekeeper.
If AGI is your only criteria without regards to friendliness, just make sure not to communicate with the AI. Turing tests are not the only proofs of intelligence. If the agi can come up with unique solutions in the universe in which it is isolated, that is enough to understand this algorithm is creative.
This just evoked a possibly-useful thought:
If observing but not communicating with a boxed AI does a good enough job of patching the security holes (which I understand that it might not—that’s for someone who better understands the issue to look at), perhaps putting an instance of a potential FAI in a contained virtual world would be useful as a test. It seems to me that a FAI that didn’t have humans to start with would perhaps have to invent us, or something like us in some specific observable way(s), because of its values.