I agree that the conflation between maintaining human control and alignment and safety is a problem, and to be clear I’m not saying that the outcome of human-controlled AI taking over because someone ordered to do that is an objectively safe outcome.
I agree at present the AI safety field is poorly equipped to avoid catastrophic outcomes that don’t involve extinction from uncontrolled AIs.
I agree that the conflation between maintaining human control and alignment and safety is a problem, and to be clear I’m not saying that the outcome of human-controlled AI taking over because someone ordered to do that is an objectively safe outcome.
I agree at present the AI safety field is poorly equipped to avoid catastrophic outcomes that don’t involve extinction from uncontrolled AIs.