I think it could be way easier to achieve robust Alignment in a system with a lot of individially weak agents. Because then to do something, they would need to cooperate, both implicitly and explicitly. So, they would need some common values and common culture code. Which, with some nudging from people, could converge into Algnment.
I think it could be way easier to achieve robust Alignment in a system with a lot of individially weak agents. Because then to do something, they would need to cooperate, both implicitly and explicitly. So, they would need some common values and common culture code. Which, with some nudging from people, could converge into Algnment.