The closest thing that comes to mind is Critch’s work on multi-user alignment, e.g. What Multipolar Failure Looks Like, and Robust Agent-Agnostic Processes (RAAPs).
Here are a couple of other links that come to mind:
https://arxiv.org/abs/2008.02275
https://www.brookings.edu/research/aligned-with-whom-direct-and-social-goals-for-ai-systems/
The closest thing that comes to mind is Critch’s work on multi-user alignment, e.g. What Multipolar Failure Looks Like, and Robust Agent-Agnostic Processes (RAAPs).
Here are a couple of other links that come to mind:
https://arxiv.org/abs/2008.02275
https://www.brookings.edu/research/aligned-with-whom-direct-and-social-goals-for-ai-systems/