a) AGI is easier than FAI b) Verification of “proof of friendliness” is easier than its production c) AI Boxing is possible
As far as I can tell, you agree with a) and b).
The comment of Eliezer’s does not seem to be mentioning the obvious difficulties with c) at all. In fact in the very part you choose to quote...
The probability I assign to achieving a capability state where it is (1) possible to prove a mind Friendly even if it has been constructed by a hostile superintelligence, (2) possible to build a hostile superintelligence, and (3) not possible to build a Friendly AI directly, is very low.
… it is b) that is implicitly the weakest link, with some potential deprecation of a) as well. c) is outright excluded from hypothetical consideration.
The comment of Eliezer’s does not seem to be mentioning the obvious difficulties with c) at all. In fact in the very part you choose to quote...
… it is b) that is implicitly the weakest link, with some potential deprecation of a) as well. c) is outright excluded from hypothetical consideration.