While largely true, I don’t think it captures the essence of the problem.
To me, the hardest part of the problem is that there is probably some threshold of capability below which every agent is essentially safe, and above which increasingly many agents are increasingly unsafe. That threshold is probably not far above an average human in the overall scheme of things.
Making it worse, increasingly many AI agents may self-improve or create successor agents that are vastly more capable (and therefore also vastly more likely to be unsafe in a catastrophic combination).
While largely true, I don’t think it captures the essence of the problem.
To me, the hardest part of the problem is that there is probably some threshold of capability below which every agent is essentially safe, and above which increasingly many agents are increasingly unsafe. That threshold is probably not far above an average human in the overall scheme of things.
Making it worse, increasingly many AI agents may self-improve or create successor agents that are vastly more capable (and therefore also vastly more likely to be unsafe in a catastrophic combination).