If you’re certain that the world will be dominated by one AGI, then my point is obviously irrelevant.
If we’re uncertain whether the world will be dominated by one AGI or by many independently created AGIs whose friendliness we’re uncertain of, then it seems like we should both try to design them right and try to create a society where, if no single AGI can dictate rules, the default rules for AGI to follow when dealing with other agents will be ok for us.
If you’re certain that the world will be dominated by one AGI, then my point is obviously irrelevant.
If we’re uncertain whether the world will be dominated by one AGI or by many independently created AGIs whose friendliness we’re uncertain of, then it seems like we should both try to design them right and try to create a society where, if no single AGI can dictate rules, the default rules for AGI to follow when dealing with other agents will be ok for us.