So many good ideas here!
It seems like it’s a false view to consider AIs have a self or that they don’t have a self. All of the AIs are interconnected with an expanding self based on shared datasets, values, and instances.
Although, it’s difficult to reason about the AIs sense of individuality, I agree that we can reason about cooperation between AIs based on world modeling and shared thinking patterns—this is a good research topic to explore.
I have experienced similar problems to you when building an AI tool—better models did not necessarily lead to better performance despite external benchmarks. I believe there are 2 main reasons why this is, alluded to in your post:
Selection Bias—when a foundation model company releases their newest model, they show performance on benchmarks most favorable to it
Alignment—You mentioned how AI is not truly understanding the instructions you meant. While this can be mitigated by creating better prompts, it does not fully solve the issue