This is essentially the AI box experiment. Check out the link to see how even an AI that can only communicate with its handler(s) might be lethal without guaranteed Friendliness.
I don’t think the publicly available details establish “how”, merely “that”.
Sure, though the mechanism I was referring to is “it can convince its handler(s) to let it out of the box through some transhuman method(s).”
Wait, since when is Eliezer transhuman?
Who said he was? If Eliezer can convince somebody to let him out of the box—for a financial loss no less—then certainly a transhuman AI can, right?
Certainly they can; what I am emphasizing is that “transhuman” is an overly strong criterion.
Definitely. Eliezer reflects perhaps a maximum lower bound on the amount of intelligence necessary to pull that off.
This is essentially the AI box experiment. Check out the link to see how even an AI that can only communicate with its handler(s) might be lethal without guaranteed Friendliness.
I don’t think the publicly available details establish “how”, merely “that”.
Sure, though the mechanism I was referring to is “it can convince its handler(s) to let it out of the box through some transhuman method(s).”
Wait, since when is Eliezer transhuman?
Who said he was? If Eliezer can convince somebody to let him out of the box—for a financial loss no less—then certainly a transhuman AI can, right?
Certainly they can; what I am emphasizing is that “transhuman” is an overly strong criterion.
Definitely. Eliezer reflects perhaps a maximum lower bound on the amount of intelligence necessary to pull that off.