What actually is “value-alignment”?

Preamble

This is an extract from a post called “Doing EA Better”, which argued that EA’s new-found power and influence obligates us to solve our movement’s significant problems with respect to epistemics, rigour, expertise, governance, and power.

We are splitting DEAB up into a sequence to facilitate object-level discussion.

Each post will include the relevant parts of the list of suggested reforms. There isn’t a perfect correspondence between the subheadings of the post and the reforms list, so not all reforms listed will be 100% relevant to the section in question.

Finally, we have tried (imperfectly) to be reasonably precise in our wording, and we ask that before criticising an argument of ours, commenters ensure that it is an argument that we are in fact making.

Main

Summary: The use of the term “value-alignment” in the EA community hides an implicit community orthodoxy. When people say “value-aligned” they typically do not mean a neutral “alignment of values”, nor even “agreement with the goal of doing the most good possible”, but a commitment to a particular package of views. This package, termed “EA orthodoxy”, includes effective altruism, longtermism, utilitarianism, Rationalist-derived epistemics, liberal-technocratic philanthropy, Whig historiography, the ITN framework, and the Techno-Utopian Approach to existential risk.

The term “value-alignment” gets thrown around a lot in EA, but is rarely actually defined. When asked, people typically say something about similarity or complementarity of values or worldviews, and this makes sense: “value-alignment” is of course a term defined in reference to what values the subject is (un)aligned with. You could just as easily speak of alignment with the values of a political party or a homeowner’s association.[6]

However, the term’s usage in EA spaces typically has an implicit component: value-alignment with a set of views shared and promoted by the most established and powerful components of the EA community. Thus:

  • Value-alignment = the degree to which one subscribes to EA orthodoxy

  • EA orthodoxy = the package of beliefs and sensibilities generally shared and promoted by EA’s core institutions (the CEA, FHI, OpenPhil, etc.)[7]

    • These include, but are not limited to:

      • Effective Altruism

        • i.e. trying to “do the most good possible”

      • Longtermism

        • i.e. believing that positively influencing the long-term future is a (or even the) key moral priority of our time

      • Utilitarianism, usually Total Utilitarianism

      • Rationalist-derived epistemics

        • Most notably subjective Bayesian “updating” of personal beliefs

      • Liberal-technocratic philanthropy

      • A broadly Whiggish/​progressivist view of history

      • Cause-prioritisation according to the ITN framework

      • The Techno-Utopian Approach to existential risk, which includes for instance, and in addition to several of the above:

        • Defining “existential risk” in reference to humanity’s “long-term potential” to generate immense amounts of (utilitarian) value by populating the cosmos with vast numbers of extremely technologically advanced beings

        • A methodological framework based on categorising individual “risks”[8], estimating for each a probability of causing an “existential catastrophe” within a given timeframe, and attempting to reduce the overall level of existential risk largely by working on particular “risks” in isolation (usually via technical or at least technocratic means)

        • Technological determinism, or at least a “military-economic adaptationism” that is often underpinned by an implicit commitment to neorealist international relations theory

        • A willingness to seriously consider extreme or otherwise exceptional actions to protect astronomically large amounts of perceived future value

      • There will naturally be exceptions here – institutions employ many people, whose views can change over time – but there are nonetheless clear regularities

Note that few, if any, of the components of orthodoxy are necessary aspects, conditions, or implications of the overall goal of “doing the most good possible”. It is possible to be an effective altruist without subscribing to all, or even any, of them, with the obvious exception of “effective altruism” itself.

However, when EAs say “value-aligned” they rarely seem to mean that one is simply “dedicated to doing the most good possible”, but that one subscribes to the particular philosophical, political, and methodological views packaged under the umbrella of orthodoxy.

Suggested reforms

Below, we have a preliminary non-exhaustive list of relevant suggestions for structural and cultural reform that we think may be a good idea and should certainly be discussed further.

It is of course plausible that some of them would not work; if you think so for a particular reform, please explain why! We would like input from a range of people, and we certainly do not claim to have all the answers!

In fact, we believe it important to open up a conversation about plausible reforms not because we have all the answers, but precisely because we don’t.

Italics indicates reforms strongly inspired by or outright stolen from Zoe Cremer’s list of structural reform ideas. Some are edited or merely related to her ideas; they should not be taken to represent Zoe’s views.

Asterisks (*) indicate that we are less sure about a suggestion, but sure enough that we think they are worth considering seriously, e.g. through deliberation or research. Otherwise, we have been developing or advocating for most of these reforms for a long time and have a reasonable degree of confidence that they should be implemented in some form or another.

Timelines are suggested to ensure that reforms can become concrete. If stated, they are rough estimates, and if there are structural barriers to a particular reform being implemented within the timespan we suggest, let us know!

Categorisations are somewhat arbitrary, we just needed to break up the text for ease of reading.

Critique

Institutions

  • Funding bodies should enthusiastically fund deep critiques and other heterodox/​“heretical” work

Red Teams

  • The judging panels of criticism contests should include people with a wide variety of views, including heterodox/​”heretical” views

Epistemics

General

  • When EAs say “value-aligned”, we should be clear about what we mean

    • Aligned with what values in particular?

    • We should avoid conflating the possession of the general goal of “doing the most good” with subscription to the full package of orthodox views

  • EAs should consciously separate

    • An individual’s suitability for a particular project, job, or role

    • Their expertise and skill in the relevant area(s)

    • The degree to which they are perceived to be “highly intelligent”

    • Their perceived level of value-alignment with EA orthodoxy

    • Their seniority within the EA community

    • Their personal wealth and/​or power

  • EAs should make a point of engaging with and listening to EAs from underrepresented disciplines and backgrounds, as well as those with heterodox/​“heretical” views

Quantification

  • EAs should be wary of the potential for highly quantitative forms of reasoning to (comparatively easily) justify anything

    • We should be extremely cautious about e.g. high expected value estimates, very low probabilities being assigned to heterodox/​“heretical” views, and ruin risks

Diversity

  • People with heterodox/​“heretical” views should be actively selected for when hiring to ensure that teams include people able to play “devil’s advocate” authentically, reducing the need to rely on highly orthodox people accurately steel-manning alternative points of view

Expertise & Rigour

Rigour

  • Work should be judged on its quality, rather than the perceived intelligence, seniority or value-alignment of its author

    • EAs should avoid assuming that research by EAs will be better than research by non-EAs by default

Reading

  • Insofar as a “canon” is created, it should be of the best-quality works on a given topic, not the best works by (orthodox) EAs about (orthodox) EA approaches to the topic

    • Reading lists, fellowship curricula, and bibliographies should be radically diversified

    • We should search everywhere for pertinent content, not just the EA Forum, LessWrong, and the websites of EA orgs

    • We should not be afraid of consulting outside experts, both to improve content/​framing and to discover blind-spots

  • EAs should see fellowships as educational activities first and foremost, not just recruitment tools

Experts & Expertise

  • When hiring for research roles at medium to high levels, EA institutions should select in favour of domain-experts, even when that means passing over a highly “value-aligned” or prominent EA

Funding & Employment

Grantmaking

  • Grantmakers should be radically diversified to incorporate EAs with a much wider variety of views, including those with heterodox/​”heretical” views

  • A certain proportion EA of funds should be allocated by lottery after a longlisting process to filter out the worst/​bad-faith proposals*

    • The outcomes of this process should be evaluated in comparison to EA’s standard grantmaking methods as well as other alternatives

Employment

  • More people working within EA should be employees, with the associated legal rights and stability of work, rather than e.g. grant-dependent “independent researchers”

  • EA funders should explore the possibility of funding more stable, safe, and permanent positions, such as professorships

Governance & Hierarchy

Decentralisation

  • EA institutions should see EA ideas as things to be co-created with the membership and the wider world, rather than transmitted and controlled from the top down

  • The community health team and grantmakers should offer community groups more autonomy, independence, and financial stability

    • Community-builders should not worry about their funding being cut if they disagree with the community health team or appear somewhat “non-value-aligned”

Contact Us

If you have any questions or suggestions about this article, EA, or anything else, feel free to email us at concernedEAs@proton.me