Error
Unrecognized LW server error:
Field "fmCrosspost" of type "CrosspostOutput" must have a selection of subfields. Did you mean "fmCrosspost { ... }"?
Unrecognized LW server error:
Field "fmCrosspost" of type "CrosspostOutput" must have a selection of subfields. Did you mean "fmCrosspost { ... }"?
Wow! It’s really great to see such an in-depth response to the definitional and foundational work that’s been taking place around IIDM over the past year, plus I love your hand-drawn illustrations! As the author or co-author of several of the pieces you cited, I thought I’d share a few thoughts and reactions to different issues you brought up. First, on the distinctions and delineations between the value-neutral and value-oriented paradigms (I like those labels, by the way):
I don’t quite agree that Jess Whittlestone’s problem profile for 80K falls into what you’re calling the “value neutral” category, as she stresses at several points the potential of working with institutions that are working on “important problems” or similar. For example, she writes: “Work on ‘improving decision-making’ very broadly isn’t all that neglected. There are a lot of people, in both industry and academia, trying out different techniques to improve decision-making....However, there seems to be very little work focused on...putting the best-proven techniques into practice in the most influential institutions.” The definition of “important” or “influential” is left unstated in that piece, but from the context and examples provided, I read the intention as one of framing opportunities from the standpoint of broad societal wellbeing rather than organizations’ parochial goals.
This segues nicely into my second response,
which is that I don’t think the value-neutral version of IIDM is really much of a thing in the EA community. CES is sort of an awkward example to use because a core tenet of democracy is the idea that one citizen’s values and policy preferences shouldn’t count more than another’s; I’d argue that the impartial welfarist perspective that’s core to EA philosophy is rooted in similar ideas. By contrast, I think people in our community are much more willing to say that someorganizations’values are better than others, both because organizations don’t have the same rights as human beings and also because organizations can agglomerate disproportionate power more easily and scalably than people. I’ve definitely seen disagreement about how appropriate or effective it is to try to change organizations’ values, but not so much about the idea that they’re important to take into account in some way.[Edit: no longer endorsed, see comment below.]There is a third type of value-oriented approach that you don’t really explore but I think is fairly common in the EA community as well as outside of it: looking for opportunities to make a positive impact from an impartial welfarist perspective on a smaller scale within a non-aligned organization (e.g., by working with a single subdivision or team, or on one specific policy decision) without trying to change the organization’s values in a broader sense.
I appreciated your thought-provoking exploration of the two indirect pathways to impact you proposed. Regarding the second pathway (selecting which institutions will survive and flourish), I would propose that an additional complicating factor is that non-value-aligned institutions may be less constrained by ethical considerations in their option set, which could give them an advantage over value-aligned institution from the standpoint of maximizing power and influence.
I did have a few critiques about the section on directly improving the outcomes of institutions’ decisions:
I think the 2x2 grid you use throughout is a bit misleading. It looks like you’re essentially using decision quality as a proxy for institutional power, and then concluding that intentions x capability = outcomes. But decision quality is only one input into institutional capabilities, and in the short term is dominated by institutional resources—e.g., the government of Denmark might have better average decision quality than the government of the United States, but it’s hard to argue that Denmark’s decisions matter more. For that reason, I think that selecting opportunities on the basis of institutional power/positioning is at least as important as value alignment. The visualization approach you took in the “A few overwhelmingly harmful institutions” graph seems to be on the right track in this respect.
One issue you don’t really touch on except in a footnote is the distinction between stated values and de facto values for institutions, or internal alignment among institutional stakeholders. For example, consider a typical private health insurer in the US. In theory, its goal is to increase the health and wellbeing of millions of patients—a highly value-aligned goal! Yet in practice, the organization engages in many predatory practices to serve its own growth, enrichment of core stakeholders, etc. So is this an altruistic institution or not? And does bringing its (non-altruistic) actions into greater alignment with its (altruistic) goals count as improving decision quality or increasing value alignment under your paradigm?
While overall I tend to agree with you that a value-oriented approach is better, I don’t think you give a fair shake to the argument that “value-aligned institutions will disproportionately benefit from the development of broad decision-making tools.” It’s important to remember that improving institutional decision-making in the social sector and especially from an EA perspective is a very recent concept. The professional world is incredibly siloed, and it’s not hard at all for me to imagine that ostensibly publicly available resources and tools that anyone could use would, in practice, be distributed through networks that ensure disproportionate adoption by well-intentioned individuals and groups. I believe that something like this is happening with Metaculus, for example.
One final technical note: you used “generic-strategy” in a different way that we did in the “Which Institutions?” post—our definition imagines a specific organization that is targeted through a non-specific strategy, whereas yours imagines a specific strategy not targeted to any specific organization. I agree that the latter deserves its own label, but suggest a different one than “generic-strategy” to avoid confusion with the previous post.
I’ve focused mostly on criticisms here for the sake of efficiency, but I really was very impressed with this article and hope to see more writing from you in the future, on this topic and others!
Thank you for this response! I think I largely agree with you, and plan to add some (marked) edits as a result. More specifically,
On the 80K problem profile:
I think you are right; they are value-oriented in that they implicitly argue for the targeted approach. I do think they could have make it a little clearer, as much (most?) of the actual work they recommend or list as an example seems to be research-style. The key (and important) exception that I ignored in the post is the “3. Fostering adoption of the best proven techniques in high-impact areas” work they recommend, which I should not have overlooked. (I will edit that part of my post, and likely add a new example of research-level value-neutral IIDM work, like a behavioral science research project.)
“I don’t think the value-neutral version of IIDM is really much of a thing in the EA community”
Once again, I think I agree, although I think there are some rationality/decision-making projects that are popular but not very targeted or value-oriented. Does that seem reasonable? The CES example is quite complicated, but I’m not sure that I think it should be disqualified here. (To be clear, however, I do think CES seems to do very valuable work—I’m just not exactly sure how to evaluate it.)
Side note, on “a core tenet of democracy is the idea that one citizen’s values and policy preferences shouldn’t count more than another’s”
I agree that this is key to democracy. However, I do think it is valid to discuss to what extent voter’s values align with actual global good (and I don’t think this opinion is very controversial). For instance, voters might be more nationalistic than one might hope, they might undervalue certain groups’ rights, or they might not value animal or future lives. So I think that, to understand the actual (welfare) impact of an intervention that improves a government’s ability to execute its voters’ aims, we would need to consider more than democratic values. (Does that make sense? I feel like I might have misinterpreted what you were trying to say, a bit, and am not sure that I am explaining myself properly.) On the other hand, it’s possible that good government decision-making is bottlenecked more by its ability to execute its voters’ aims than it is by the voters’ values’ ethical alignment—but I still wish this were more explicitly considered.
“It looks like you’re essentially using decision quality as a proxy for institutional power, and then concluding that intentions x capability = outcomes.”
I think I explained myself poorly in the post, but this is not how I was thinking about it. I agree that the power of an institution is (at least) as important as its decision-making skill (although it does seem likely that these things are quite related), but I viewed IIDM as mostly focused on decision-making and set power aside. If I were to draw this out, I would add power/scope of institutions as a third axis or dimension (although I would worry about presenting a false picture of orthogonality between power and decision quality). The impact of an institution would then be related to the relevant volume of a rectangular prism, not the relevant area of a rectangle. (Note that the visualizing approach in the “A few overwhelmingly harmful institutions” image is another way of drawing volume or a third dimension, I think.) I might add a note along these lines to the post to clarify things a bit.
About “the distinction between stated values and de facto values for institutions”
You’re right, I am very unclear about this (and it’s probably muddled in my head, too). I am basically always trying to talk about the de facto values. For instance, if a finance company whose only aim is to profit also incidentally brings a bunch of value to the world, then I would view it as value-aligned for the purpose of this post. To answer your questions about the typical private health insurance company, “does bringing its (non-altruistic) actions into greater alignment with its (altruistic) goals count as improving decision quality or increasing value alignment under your paradigm”—it would count as increasing value alignment, not improving decision quality.
Honestly, though, I think this means I should be much more careful about this term, and probably just clearly differentiate between “stated-value-alignment” and “practical-value-alignment.” (These are terrible and clunky terms, but I cannot come up with better ones on the spot.) I think also that my own note about “well-meaning [organizations that] have such bad decision quality that they are actively counterproductive to their aims” clashes with the “value-alignment” framework. I think that there is a good chance that it does not work very well for organizations whose main stated aim is to do good (of some form). I’ll definitely think more about this and try to come back to it.
“The professional world is incredibly siloed, and it’s not hard at all for me to imagine that ostensibly publicly available resources and tools that anyone could use would, in practice, be distributed through networks that ensure disproportionate adoption by well-intentioned individuals and groups. I believe that something like this is happening with Metaculus, for example.”
This is a really good point (and something I did not realize, probably in part due to a lack of background). Would you mind if I added an excerpt from this or a summary to the post?
On your note about”generic-strategy”: Apologies for that, and thank you for pointing it out! I’ll make some edits.
Note: I now realize that I have basically inverted normal comment-response formatting in this response, but I’m too tired to fix it right now. I hope that’s alright!
Once again, thank you for this really detailed comment and all the feedback—I really appreciate it!
It does, and I admittedly wrote that part of the comment before fully understanding your argument about classifying the development of general-use decision-making tools as being value-neutral. I agree that there has been a nontrivial focus on developing the science of forecasting and other approaches to probability management within EA circles, for example, and that those would qualify as value-neutral using your definition, so my earlier statement that value-neutral is “not really a thing” in EA was unfair.
Yeah, I also thought of suggesting this, but think it’s problematic as well. As you say, power/scope is correlated with decision quality, although more on a long-term time horizon than in the short term and more for some kinds of organizations (corporations, media, certain kinds of nonprofits) than others (foundations, local/regional governments). I think it would be more parsimonious to just replace decision quality with institutional capabilities on the graphs and to frame DQ in the text as a mechanism for increasing the latter, IMHO. (Edited to add: another complication is that the line between institutional capabilities that come from DQ and capabilities that come from value shift is often blurry. For example, a nonprofit could decide to change its mission in such a way that the scope of its impact potential becomes much larger, e.g., by shifting to a wider geographic focus. This would represent a value improvement by EA standards, but it also means that it might open itself up to greater possibilities for scale from being able to access new funders, etc.)
No problem, go ahead!
I’ve been skeptical of much of the IIDM work I’ve seen to date. By contrast, from a quick skim, this piece seemed pretty good to me because it has more detailed models of how IIDM may or may not be useful, and is opinionated in a few non-obvious but correct-seeming ways. I liked this a lot – thanks for publishing!
Like, if anyone feels like handing out prizes for good content, I’d recommend that this piece of work should receive a $10k prize (though perhaps I’d want to read it in full before fully recommending).
Hi Lizka, WOW – Thank you for writing this. Great to see Rethink Priorities working on this. Absolutely loving the diagrams here.
I have worked in this space for a number of years mostly here, have been advocating for this cause within EA since 2016 and advised both Jess/80K and the effective institutions project on their writeups. Thought I would give some quick feedback. Let me know if it is useful.
I thought your disentanglement did a decent job. Here are a few thought I had on it.
I really like how you split IIDM into “A technical approach to IIDM” and “A value-aligning approach to IIDM.”
However I found the details of how you split it to be very confusing. It left me quite unsure what goes into what bucket. For example intuitively I would see increasing the “accuracy of governments” (i.e. aligning governments with the interests of the voters) as “value-aligning” yet you classify it as “technical”.
That said despite this, I very much agreed with the conclusion that “value-oriented IIDM makes more sense than value-neutral IIDM” and the points you made to that effect.
I didn’t quite understand what “(1) IIDM can improve our intellectual and political environment” was really getting at. My best guess is that by (1) you mean work that only indirectly leads to “(3) improved outcomes”. So value-oriented (1) would look like general value spreading. Is that correct?
I agree with “for the sake of clarity … we should generally distinguish between ‘meta EA’ work and IIDM work”. That said I think it is worth bearing in mind that on occasion the approaches might not be that different. For example I have been advising the UK government on how to asses high-impact risks which is relevant for EAs too.*
One institution can have many parts. Might be a thing to highlight if you do more disentanglement. E.g. Is a new office for future generations within a government, a new institution or improving an existing institution?
One other thought I had whilst reading.
I think it is important not to assign value to IIDM based on what is “predictable”.
For example you say “it would be extremely hard to produce candidate IIDM interventions that would have sufficiently predictable outcomes via this pathway, as the outcomes would depend on many very uncertain factors.” Predictions do matter but one of the key cases for IIDM is that it offers a solution to the unpredictable, the unknown unknows, to the uncertainty of the EA (and especially longtermist) endeavour. All the advice on dealing with high-uncertainty and things that are hard to predict suggest that interventions like IIDM are the kinds of interventions that should work – as set out by Ian David Moss here (from this piece).
Finally, at points you seemed uncertain about tractability of this work. I wanted to add that so far I have found it much much easier than I expected. Eg you say “it is possible that shifting the aims of institutions is generally very difficult or that the potential benefits from institutions is overwhelmingly bottlenecked by decision-making ability, rather than by the value-alignment of institutions’ existing aims”. (I am perhaps still confused about what you would count as shifting aims Vs decision-making ability see my point 1. above, but) my rough take on this is that I have found shifting the aims of government to be fairly easy and that there are not too many decision-making bottlenecks.
So super excited to see more EA work in this space.
* Oddly enough, despite being in EA for years, I think I have found it easier to influence the UK government to get better at risk identification work than the EA community. Not sure what to do about that. Just wanted to say that I would love to input if RP is working in this space.
I’m curious why you think aligning governments with the interests of voters is value-aligning rather than technical. I can see that being the case for autocratic regimes, but isn’t that the whole point of representative democracy?
(not the author)
4. When I hear “(1) IIDM can improve our intellectual and political environment”, I’m imagining something like if the concept of steelmanning becomes common in public discourse, we might expect that to indirectly lead to better decisions by key institutions.
Nice post : )
I mostly agree with your points, though am a bit more optimistic than it seems like you are about untargeted, value-neutral IIDM having a positive impact.
Your skepticism about this seems to be expressed here:
I think this is true, but it still seems like the aims of institutions are pro-social as a general matter—x-risk and animal suffering in your examples are side effects that aren’t means to the ends of the institutions, which are ‘increase biosecuirty’ and ‘make money’, and if improving decisionmaking helps orgs get at their ends more efficiently then we should think they will have fewer bad side effects if they have better decisonmaking. Also generally orgs’ aims (e.g. “make money”) will presuppose the firm’s, and therefore humanity’s survival, so it still seems good to me as a general matter for orgs to be able to pursue their aims more effectively.
Really nice and useful exploration, and I really liked your drawings.
FWIW, I intuitively would’ve drawn the institution blob in your sketch higher, i.e. I’d have put fewer than (eyeballing) 30% of institutions in the negatively aligned space (maybe 10%?). In moments like this, including a quick poll into the forum to get a picture what others think would be really useful.
Other spontaneous ideas, besides choosing more representative candidates:
increased coherence of the institution could lead to an overall stronger link between its mandate and its actions
increased transparency and coherence could reduce corruption and rent-seeking
Given what I said beforehand, I’d be interested in learning more about examples of harmful institutions that have generally high capacity.
Thank you for this comment!
I won’t redraw/re-upload this sketch, but I think you are probably right.
That’s a really good idea, thank you! I’ll play around with that.
Thank you for the suggestions! I think you raise good points, and I’ll try to come back to this.
An update: after a bit of digging, I discovered this post, “Should marginal longtermist donations support fundamental or intervention research?”, which contains a discussion on a topic that is quite close to “should EA value foundational (science/decision theory) research,” (in the pathway (1) section of my post). The conclusions of the post I found do not fit into my vague impressions of “the consensus.” In particular, the conclusion of that post is that longtermist research hours should often be spent on fundamental research (which is defined by its goals).
(Disclaimer: the author, Michael, is employed at Rethink Priorities, where I am interning. I don’t know if he still endorses this post or its conclusions, but the post seems relevant here and very valuable as a reference.)
fwiw, I think I still broadly endorse the post and its conclusions.
(I only skimmed your post, so can’t comment on how my views and my post aligns/conflicts with the views in your post.)
Thanks! Some brief ‘deep’ counterpoints. I don’t see how “decision-making quality” and
“values” or/which is implied by “value alignment” can be orthogonal to each other, thus facilitating the above graphs. To my mind bad values would promote bad decisions. (And to my mind bad values would have some relation and thus alignment with ‘our’ values, hopefully a negative one, but most likely not a priori without a relation). Relatedly, I also don’t really believe in the existence of “value-neutrality”, and moreover I think it is a dangerous, or more mildly counter-productive, concept to deploy (with effects that are misleadingly regarded as ‘neutral’ and perhaps attended less to), e.g. the economy might be—or often is—regarded as neutral, yet very significantly disregards the interests of future generations, or non-human animals.
I suppose the post’s content fits with moral relativists’ and anti-realists’ worldviews, but with more difficulty with moral realists’, or then just to relate to a way that people often ‘happen to talk’.
Follow-up: Perhaps to put altruism and effectiveness—properly understood—on the y-axis and x-axis respectively would be better, i.e. communicate what we would want to communicate and not suffer from the above-mentioned shortcomings?
Thanks for the comment! I feel funny saying this without being the author, but feel like the rest of my comment is a bit cold in tone, so thought it’s appropriate to add this :)
I lean more moral anti-realist but I struggle to see how the concept of “value alignment” and “decision-making quality” are not similarly orthogonal from a moral realist view than an anti-realist view.
Moral realist frame: “The more the institution is intending to do things according to the ‘true moral view’, the more it’s value-aligned.”
“The better the institutions’s decision making process is at predictably leading to what they value, the better their ‘decision-making quality’ is.”
I don’t see why these couldn’t be orthogonal in at least some cases. For example, a terrorist organization could be outstandingly good at producing outstandingly bad outcomes.
Still, it’s true that the “value-aligned” term might not be the best, since some people seem to interpret it as a dog-whistle for “not following EA dogma enough” [link] (I don’t, although might be mistaken). “Altruism” and “Effectiveness”as the x and y axes would suffer from the problem mentioned in the post that it could alienate people coming to work on IIDM from outside the EA community. For the y-axis, ideally I’d like some terms that make it easy to differentiate between beliefs common in EA that are uncontroversial (“let’s value people’s lives the same regardless of where they live”), and beliefs that are more controversial (“x-risk is the key moral priority of our times”).
About the problematicness of ” value-neutral”: I thought the post gave enough space for the belief that institutions might be worse than neutral on average, marking statements implying the opposite as uncertain. For example crux (a) exists in this image to point out that if you disagree with it, you would come to a different conclusion about the effectiveness of (A).
(I’m testing out writing more comments on the EA forum, feel free to say if it was helpful or not! I want to learn to spend less time on these. This took about 30 minutes.)
Thanks for the post and for taking the time! My initial thoughts on trying to parse this are below, I think it will bring mutual understanding further.
You seem to make a distinction between intentions on the y-axis and outcomes on the x-axis. Interesting!
The terrorist example seems to imply that if you want bad outcomes you are not value-aligned (aligned to what? to good outcomes?). They are value-aligned from their own perspective. And “terrorist” is also not a value-neutral term, for example Nelson Mandela was once considered one, which would I think surprise most people now.
If we allow “from their own perspective” then “effectiveness” would do (and “efficiency” to replace the x-axis), but it seems we don’t, and then “altruism” (or perhaps “good”, with less of an explicit tie to EA?) would without the ambiguity “value-aligned” brings on whether or not we do [allow “from their own perspective”].
(As not a moral realist, the option of “better value” is not available, so it seems one would be stuck with “from their own perspective” and calling the effective terrorist value-aligned, or moving to an explicit comparison to EA values, which I was supposing was not the purpose, and seems to be even more off-putting via the mentioned alienating shortcoming in communication.)
Next to value-aligned being suboptimal, which I also just supported further, you seem to agree with altruism and effectiveness (I would now suggest “efficiency” instead) as appropriate labels, but agree with the author about the shortcoming for communicating to certain audiences (alienation), with which I also agree. For other audiences, including myself, the current form perhaps has shortcomings. I would value clarity more, and call the same the same. An intentional opaque-making change of words might additionally come across as deceptive, and as aligned with one’s own ideas of good, but not with such ideas in a broader context. And that I think could definitely also count as / become a consequential shortcoming in communication strategy.
And regarding the non-orthogonality, I was—as a moral realist -more thinking along the lines of: being organized (etc., etc.), is presumably a good value, and it would also improve your decision-making (sort of considered neutrally)...