Thanks for sharing! I’m curious if any of these readings were most helpful around forming “theories of change toward achieving a limited form of global governance in specific key domains where it might be most important” or “viable mechanisms for partial global governance in those domains.”
As someone exploring alternative ways to govern near-global powerful technology organizations that interact closely with nation-states and fund a significant proportion of AI research, this is what I would be most curious about (and which also seems e.g. particularly relevant re. x-risks). In the linked doc, I focus on sortition-based systems as one potential approach, but there are additional routes (e.g. ML-augmented) that I am also exploring using ~this framework, and I’d be interested any I have not considered.
Assuming misaligned AI is a risk, is technical AI alignment enough, or do you need joint AI/Societal alignment?
My work has involved trying to support risk awareness and coordination similar to what has been suggested for AI alignment. For example, for mitigating harms around synthetic media / “deepfakes” (now rebranded to generative AI) and it worked for a few years with all the major orgs and most relevant research groups.
But then new orgs jumped in to fill the capability gap! (e.g. eleuther, stability, etc.)
Due to demand and for potentially good reasons: those capabilities which can harm people can also help people. The ultimate result is the proliferation/access/democratization of AI capabilities in the face of risks.
I’m currently skeptical that this sort of coordination is possible without some addressing deeper societal incentives (AKA reward functions; e.g. around profit/power/attention maximization, self-dealing, etc.) and related multi-principal-agent challenges. This joint/ai societal alignment or holistic alignment would seem to be a prerequisite to the actual implementation of technical alignment.[2]
This is assuming you can even get the major players on board, which isn’t true for e.g. misaligned recommender systems that I’ve also worked on (on the societal side).
This would also be generally good for the world! E.g. to address externalities, political dysfunction, corruption, etc.