Article itself claims it is not a good idea (because humanity would not survive the stampede of two AIs fighting off). But comments offer pretty good reasons of why it can work, if done right.
Author agrees with some points and clarifies his: “I am mostly objecting to strategies which posit one AI saving us from another as the primary mechanism of alignment”
Afaik it’s called the “Godzilla strategy” https://www.lesswrong.com/posts/DwqgLXn5qYC7GqExF/godzilla-strategies
Article itself claims it is not a good idea (because humanity would not survive the stampede of two AIs fighting off). But comments offer pretty good reasons of why it can work, if done right.
Author agrees with some points and clarifies his: “I am mostly objecting to strategies which posit one AI saving us from another as the primary mechanism of alignment”