If the orthogonality thesis is incorrect and the reason (some) humans are friendly is because they are not intelligent enough (as Nick Land argue here), then friendly AGIs would be impssible. I think the arguments for this position are really badly argued, but this is still a good reductio of gworley’s philosophical conservatism.
If that’s the case we’re no longer addressing alignment and are forced to fall back on weaker safety mechanism. People are working in this direction, but alignment remains the best path until we see evidence it’s not possible.
Isn’t the least convenient world a world where FAI is outright impossible and all AGIs are Unfriendly ?
Given there are friendly human intelligences, what would have to be true about the universe in order for friendly AGI’s to be impossible?
If the orthogonality thesis is incorrect and the reason (some) humans are friendly is because they are not intelligent enough (as Nick Land argue here), then friendly AGIs would be impssible. I think the arguments for this position are really badly argued, but this is still a good reductio of gworley’s philosophical conservatism.
If that’s the case we’re no longer addressing alignment and are forced to fall back on weaker safety mechanism. People are working in this direction, but alignment remains the best path until we see evidence it’s not possible.