Deep Democracy as a promising target for positive AGI futures

If you want the long-term future to go well by the lights of a certain value function, you might be tempted to try to align AGI(s) to your own values (broadly construed, including your deliberative values and intellectual temperaments).[1]

Why Shouldnt I Keep It Its Mine GIF - Why Shouldnt I Keep It Its Mine I  Found It - Discover & Share GIFs

Suppose that you’re not going to do that, for one of three reasons:

  1. You can’t. People more powerful than you are going to build AGIs and you don’t have a say over that.

  2. You object to aligning AGI(s) to your own values for principled reasons. It would be highly uncooperative, undemocratic, coercive, and basically cartoon supervillain evil.

  3. You recognize that this behaviour would, when pursued by lots of people, lead to a race to the bottom where everyone fights to build AGI aligned to their values as fast as possible and destroys a ton of value in the process, so you want to strongly reject this kind of norm.

Then a good next-best option is Deep Democracy. What I mean by this is aligning AGI(s) to a process that is arbitrarily sensitive to every person’s entire value function. Not democracy in the sense of the current Western electoral system, but in the idealistic theoretical sense of deeply capturing and being responsive to every single person’s values. (Think about the ideal that democratic mechanisms like quadratic voting and bargaining theory are trying to capture, where democracy is basically equivalent to enlightened preference utilitarianism.)

This is basically just the first class of Political Philosophy 101: it sure would be nice if you could install your favorite benevolent dictator, wouldn’t it? Well it turns out you can’t, and even if you could that’s evil and a very dangerous policy — what if someone else does this and you get a dictator you don’t like? As civilized people, let’s agree to give everyone a seat at the table to decide what happens.

Deep Democracy has a lot of nice properties:

  • It avoids the ascendence of an arbitrary dictator who decides the future.

  • Suitably deep kinds of democracy avoid the tyranny of the majority, where if 51% of people say they want something, it happens. Instead decisions are sensitive to everyone’s values. This means that if you personally value something really weird, that doesn’t get stamped out by majority values, it still gets a place in the future.

    • As a corollary, it makes outcomes sensitive to the number of people who care about something and how much they care about something.

    • And it means that what you specifically care about will have some place in the long-term future, no matter what it is.

  • It facilitates “moral hedging” — if everyone has a say, then everyone’s moral theories get a seat at the table in a real life moral parliament, hedging against both moral uncertainty and the possibility that a wrong moral theory wins and controls everything, destroying all value in the process.

    • If value is a power law or similarly distributed, then you have a high chance of at least capturing some of the stuff that is astronomically more valuable than everything else, rather than losing out on this stuff entirely.

  • It avoids races to the bottom because no one is advantaged by getting there first; no matter who wins, the outcome is the same.

  • It has a lot of resonance with enlightened political discourse around the world, making it much more achievable than some more narrow vision of the future.

  • Whereas Deep Democracy is fairly computationally intractable today[2] (we have to vote on representatives who act on our interests because preference utilitarianism is computationally intractable in the real world), advances in AI will make deeper and deeper versions of democracy more and more tractable, as we have more computational power to explore our own values, find positive-sum compromises, aggregate people’s preferences and local information, and the like.

    • And whereas Deep Democracy is politically intractable today because it upsets existing interested powers, if AGI shifts power structures a lot it will be more tractable to implement in the future.

  • And it’s of course extremely cooperative with everyone who values something.

There are reasons you could object to Deep Democracy, such as because there are other procedures that do better by the lights of your own values, or because you think evil people shouldn’t have a say on how the world goes. But I think by the lights of most value systems, in the real world, Deep Democracy is an extremely promising goal in the presence of value pluralism.

I don’t know what the best way is to practically achieve Deep Democracy. It might look like a top-down political mechanism of voting, or a rhizomatic mechanism of bargaining between AGIs who represent principals with different interests, or just really good market mechanisms facilitating trade between people under conditions of roughly equal power and wealth. It’s not obvious at all that it looks like something like aligning AI model specs to a democratic procedure or nationalizing and democratizing AGI companies, but these are possibilities as well. And there are plenty of more exotic mechanisms to consider like appointing a random lottery dictator. (It’s still Deep Democracy ex ante!) I think a lot of work needs to be done forecasting what is going to happen with AGI values and articulating and evaluating mechanisms for making the outcome one that is Deeply Democratic.

But as an ideal, Deep Democracy has a lot going for it, and it’s my current best candidate for an excellent north star that EAs and many other people could get behind and push for.

  1. ^

    I don’t mean to commit to any particular vision of the future here. By “align AGI(s) to your own values” I mean something like: whether there’s an ASI singleton, a bunch of competing AGIs with different values, or a bunch of personal AGIs aligned directly to each user’s intent, the overall way that decisions are made by the world’s AGI(s) (and beyond) is aligned with your values.

  2. ^

    But more tractable than many people assume. We would do much better as a society by implementing better voting procedures like approval voting and quadratic voting.