There’s a fair amount of discussion in AI alignment about what outer alignment requires, and how it’s not just pursuing goals of a single person who is supposed to be in control.As a few examples, you might be interested in some of these:https://www.alignmentforum.org/posts/Cty2rSMut483QgBQ2/what-should-ai-owe-to-us-accountable-and-aligned-ai-systems
https://www.cs.utexas.edu/~pstone/Papers/bib2html-links/ASIMOV2021-REUTH.pdf
https://www.lesswrong.com/posts/Tmvvvx3buP4Gj3nZK/learning-societal-values-from-law-as-part-of-an-agi
I appreciate you sharing these! I’ve already started to read them
There’s a fair amount of discussion in AI alignment about what outer alignment requires, and how it’s not just pursuing goals of a single person who is supposed to be in control.
As a few examples, you might be interested in some of these:
https://www.alignmentforum.org/posts/Cty2rSMut483QgBQ2/what-should-ai-owe-to-us-accountable-and-aligned-ai-systems
https://www.cs.utexas.edu/~pstone/Papers/bib2html-links/ASIMOV2021-REUTH.pdf
https://www.lesswrong.com/posts/Tmvvvx3buP4Gj3nZK/learning-societal-values-from-law-as-part-of-an-agi
I appreciate you sharing these! I’ve already started to read them