I don’t think Joe is proposing we find an AI design that is impossible to abuse even by malicious humans. The point so far seems to be making sure your own AI is not going to do some specific bad stuff.
If you solve the latter, you have not solved the former at all; if you solve the former, someone will solve the latter.
Insofar as this is true in your extended analogy, I think that’s a reflection of “completely proliferation-proof reactor” being a bad thing to just assume you can solve.
I don’t think Joe is proposing we find an AI design that is impossible to abuse even by malicious humans. The point so far seems to be making sure your own AI is not going to do some specific bad stuff.
Insofar as this is true in your extended analogy, I think that’s a reflection of “completely proliferation-proof reactor” being a bad thing to just assume you can solve.