I don’t understand why this comment was downvoted.
Yes, zero call asks a question many of us feel has been adequately answered in the past; but they are asking politely, and it would have taken extensive archive-reading for them to have already known about the AI-Box experiment.
Think before you downvote, especially with new users!
EDIT: As AdeleneDawner points out, zero call isn’t that new. Even so, the downvotes (at −2 when I first made my comment) looked more like signaling disagreement than anything else.
I downvoted the comment not because of AI box unsafety (which I don’t find convincing at the certainty level with which it’s usually asserted—disutility may well give weight to the worry, but not to the probability), but because it gives advice on the paint color for a spaceship in the time when Earth is still standing on a giant Turtle in the center of the world. It’s not a sane kind of advice.
If I’d never heard of the AI-Box Experiment, I’d think that zero call’s comment was a reasonable contribution to a conversation about AI and safety in particular. It’s only when we realize that object-level methods of restraining a transhuman intelligence are probably doomed that we know we must focus so precisely on getting its goals right.
Please point me to some more details about the AI box experiment, since I think what i suggested earlier as isolated virtual worlds is pretty much the same as what zero call is suggesting here.
I feel that there are huge assumptions in the present AI Box experiment. The gatekeeper and the AI share a language, for one, by which the AI convinces the gatekeeper.
If AGI is your only criteria without regards to friendliness, just make sure not to communicate with the AI. Turing tests are not the only proofs of intelligence. If the agi can come up with unique solutions in the universe in which it is isolated, that is enough to understand this algorithm is creative.
If observing but not communicating with a boxed AI does a good enough job of patching the security holes (which I understand that it might not—that’s for someone who better understands the issue to look at), perhaps putting an instance of a potential FAI in a contained virtual world would be useful as a test. It seems to me that a FAI that didn’t have humans to start with would perhaps have to invent us, or something like us in some specific observable way(s), because of its values.
Good thought, but on further examination it turns out that zero isn’t all that new—xe’s been commenting since November; xyr karma is low because xe has been downvoted almost as often as upvoted.
I don’t understand why this comment was downvoted.
Yes, zero call asks a question many of us feel has been adequately answered in the past; but they are asking politely, and it would have taken extensive archive-reading for them to have already known about the AI-Box experiment.
Think before you downvote, especially with new users!
EDIT: As AdeleneDawner points out, zero call isn’t that new. Even so, the downvotes (at −2 when I first made my comment) looked more like signaling disagreement than anything else.
I downvoted the comment not because of AI box unsafety (which I don’t find convincing at the certainty level with which it’s usually asserted—disutility may well give weight to the worry, but not to the probability), but because it gives advice on the paint color for a spaceship in the time when Earth is still standing on a giant Turtle in the center of the world. It’s not a sane kind of advice.
If I’d never heard of the AI-Box Experiment, I’d think that zero call’s comment was a reasonable contribution to a conversation about AI and safety in particular. It’s only when we realize that object-level methods of restraining a transhuman intelligence are probably doomed that we know we must focus so precisely on getting its goals right.
Vladimir and orthonormal,
Please point me to some more details about the AI box experiment, since I think what i suggested earlier as isolated virtual worlds is pretty much the same as what zero call is suggesting here.
I feel that there are huge assumptions in the present AI Box experiment. The gatekeeper and the AI share a language, for one, by which the AI convinces the gatekeeper.
If AGI is your only criteria without regards to friendliness, just make sure not to communicate with the AI. Turing tests are not the only proofs of intelligence. If the agi can come up with unique solutions in the universe in which it is isolated, that is enough to understand this algorithm is creative.
This just evoked a possibly-useful thought:
If observing but not communicating with a boxed AI does a good enough job of patching the security holes (which I understand that it might not—that’s for someone who better understands the issue to look at), perhaps putting an instance of a potential FAI in a contained virtual world would be useful as a test. It seems to me that a FAI that didn’t have humans to start with would perhaps have to invent us, or something like us in some specific observable way(s), because of its values.
Good thought, but on further examination it turns out that zero isn’t all that new—xe’s been commenting since November; xyr karma is low because xe has been downvoted almost as often as upvoted.