The opposites of those four requirements sound also pretty good.
Exclusivity—Corrigibility
Humans that are being harmed should be able to effectively steer the AI to cease hurting them.
Benevolence—Servitude
The AI should serve humans and not put its own goals ahead of others.
Stability—Responsitivity
The AI should stay relevant and answer challenges to its existence. It should keep up with the world and not become out of distribution by turning into a relic.
Success at aligment—Fallibility
A minor mistake should not spell doom to the world. The setup should fail gracefully and accept fixes.
I’m not sure I understand what you mean. As I understand it, this comment seems a bit non sequitur to the post. First, I don’t agree with any of the four pairs you mentioned at all as being opposites. Second, it seems to me like you’re talking about an specific AI system, and not a TAI regime like I am.
The opposites of those four requirements sound also pretty good.
Exclusivity—Corrigibility
Humans that are being harmed should be able to effectively steer the AI to cease hurting them.
Benevolence—Servitude
The AI should serve humans and not put its own goals ahead of others.
Stability—Responsitivity
The AI should stay relevant and answer challenges to its existence. It should keep up with the world and not become out of distribution by turning into a relic.
Success at aligment—Fallibility
A minor mistake should not spell doom to the world. The setup should fail gracefully and accept fixes.
I’m not sure I understand what you mean. As I understand it, this comment seems a bit non sequitur to the post. First, I don’t agree with any of the four pairs you mentioned at all as being opposites. Second, it seems to me like you’re talking about an specific AI system, and not a TAI regime like I am.