These questions are outside the scope of this post, which is about what would happen if AIs were pointed at defeating humanity.
I don’t think there’s a clear answer to whether AIs would have a lot of their goals in common, or find it easier to coordinate with each other than with humans, but the probability of each seems at least reasonably high if they are all developed using highly similar processes (making them all likely more similar to each other in many ways than to humans).
These questions are outside the scope of this post, which is about what would happen if AIs were pointed at defeating humanity.
I don’t think there’s a clear answer to whether AIs would have a lot of their goals in common, or find it easier to coordinate with each other than with humans, but the probability of each seems at least reasonably high if they are all developed using highly similar processes (making them all likely more similar to each other in many ways than to humans).