Q4 Time scale
In order to claim that we need to worry about AGI Alignment today, you need to prove that the time scale of development will be short. Common sense tells us that humans will be able to deal with whatever software we can create. 1) We create some software (eg self driving cars, nuclear power plant sofrtware) 2) People accidentally die (or have other “bad outcomes”) 3) Humans, governments, people in general will “course correct”.
So you have to prove (or convince) that an AGI will develop, gain control of it’s own resources and then be able to act on the world in a very short period of time. I haven’t seen a convincing argument for that.
Q5 Scenarios
I have different thoughts about different doomsday scenarios. I can think of two general categories, but maybe there are more.
A) “Build us a better bomb.”—The AGI is locked in service to a human organization who uses it’s superpowers to dominate and control the rest of the world. In this scenario the AGI is essentially a munitions that may appear in the piucture without warning (which takes us back to the time scale concern). This doesn’t require the AGI to become self-sufficient. Presumably lesser AIs would also be capable of building better bombs.
B) “Evil Overlord”—An AGI comes into being fast enough that nobody can stop it, and somehow it gains control of power and mechanical resources needed to preserve it’s own existence. I’m having a hard time visualizing how this happens with nobody noticing until it’s too late. Individual humans and even groups of humans have a hard enough time “preserving their existence” in the face to world troubles. If a physically centralized AGI threatens the world, it will get bombed out of existence. If it’s distributed it will be tracable and security measures will be put in place to prevent it from invading unauthorized computers.