Thanks! Some brief âdeepâ counterpoints. I donât see how âdecision-making qualityâ and âvaluesâ or/âwhich is implied by âvalue alignmentâ can be orthogonal to each other, thus facilitating the above graphs. To my mind bad values would promote bad decisions. (And to my mind bad values would have some relation and thus alignment with âourâ values, hopefully a negative one, but most likely not a priori without a relation). Relatedly, I also donât really believe in the existence of âvalue-neutralityâ, and moreover I think it is a dangerous, or more mildly counter-productive, concept to deploy (with effects that are misleadingly regarded as âneutralâ and perhaps attended less to), e.g. the economy might beâor often isâregarded as neutral, yet very significantly disregards the interests of future generations, or non-human animals.
I suppose the postâs content fits with moral relativistsâ and anti-realistsâ worldviews, but with more difficulty with moral realistsâ, or then just to relate to a way that people often âhappen to talkâ.
Follow-up: Perhaps to put altruism and effectivenessâproperly understoodâon the y-axis and x-axis respectively would be better, i.e. communicate what we would want to communicate and not suffer from the above-mentioned shortcomings?
Thanks for the comment! I feel funny saying this without being the author, but feel like the rest of my comment is a bit cold in tone, so thought itâs appropriate to add this :)
I lean more moral anti-realist but I struggle to see how the concept of âvalue alignmentâ and âdecision-making qualityâ are not similarly orthogonal from a moral realist view than an anti-realist view.
Moral realist frame: âThe more the institution is intending to do things according to the âtrue moral viewâ, the more itâs value-aligned.â
âThe better the institutionsâs decision making process is at predictably leading to what they value, the better their âdecision-making qualityâ is.â
I donât see why these couldnât be orthogonal in at least some cases. For example, a terrorist organization could be outstandingly good at producing outstandingly bad outcomes.
Still, itâs true that the âvalue-alignedâ term might not be the best, since some people seem to interpret it as a dog-whistle for ânot following EA dogma enoughâ [link] (I donât, although might be mistaken). âAltruismâ and âEffectivenessâas the x and y axes would suffer from the problem mentioned in the post that it could alienate people coming to work on IIDM from outside the EA community. For the y-axis, ideally Iâd like some terms that make it easy to differentiate between beliefs common in EA that are uncontroversial (âletâs value peopleâs lives the same regardless of where they liveâ), and beliefs that are more controversial (âx-risk is the key moral priority of our timesâ).
About the problematicness of â value-neutralâ: I thought the post gave enough space for the belief that institutions might be worse than neutral on average, marking statements implying the opposite as uncertain. For example crux (a) exists in this image to point out that if you disagree with it, you would come to a different conclusion about the effectiveness of (A).
(Iâm testing out writing more comments on the EA forum, feel free to say if it was helpful or not! I want to learn to spend less time on these. This took about 30 minutes.)
Thanks for the post and for taking the time! My initial thoughts on trying to parse this are below, I think it will bring mutual understanding further.
You seem to make a distinction between intentions on the y-axis and outcomes on the x-axis. Interesting!
The terrorist example seems to imply that if you want bad outcomes you are not value-aligned (aligned to what? to good outcomes?). They are value-aligned from their own perspective. And âterroristâ is also not a value-neutral term, for example Nelson Mandela was once considered one, which would I think surprise most people now.
If we allow âfrom their own perspectiveâ then âeffectivenessâ would do (and âefficiencyâ to replace the x-axis), but it seems we donât, and then âaltruismâ (or perhaps âgoodâ, with less of an explicit tie to EA?) would without the ambiguity âvalue-alignedâ brings on whether or not we do [allow âfrom their own perspectiveâ].
(As not a moral realist, the option of âbetter valueâ is not available, so it seems one would be stuck with âfrom their own perspectiveâ and calling the effective terrorist value-aligned, or moving to an explicit comparison to EA values, which I was supposing was not the purpose, and seems to be even more off-putting via the mentioned alienating shortcoming in communication.)
Next to value-aligned being suboptimal, which I also just supported further, you seem to agree with altruism and effectiveness (I would now suggest âefficiencyâ instead) as appropriate labels, but agree with the author about the shortcoming for communicating to certain audiences (alienation), with which I also agree. For other audiences, including myself, the current form perhaps has shortcomings. I would value clarity more, and call the same the same. An intentional opaque-making change of words might additionally come across as deceptive, and as aligned with oneâs own ideas of good, but not with such ideas in a broader context. And that I think could definitely also count as /â become a consequential shortcoming in communication strategy.
And regarding the non-orthogonality, I wasâas a moral realist -more thinking along the lines of: being organized (etc., etc.), is presumably a good value, and it would also improve your decision-making (sort of considered neutrally)...
Thanks! Some brief âdeepâ counterpoints. I donât see how âdecision-making qualityâ and
âvaluesâ or/âwhich is implied by âvalue alignmentâ can be orthogonal to each other, thus facilitating the above graphs. To my mind bad values would promote bad decisions. (And to my mind bad values would have some relation and thus alignment with âourâ values, hopefully a negative one, but most likely not a priori without a relation). Relatedly, I also donât really believe in the existence of âvalue-neutralityâ, and moreover I think it is a dangerous, or more mildly counter-productive, concept to deploy (with effects that are misleadingly regarded as âneutralâ and perhaps attended less to), e.g. the economy might beâor often isâregarded as neutral, yet very significantly disregards the interests of future generations, or non-human animals.
I suppose the postâs content fits with moral relativistsâ and anti-realistsâ worldviews, but with more difficulty with moral realistsâ, or then just to relate to a way that people often âhappen to talkâ.
Follow-up: Perhaps to put altruism and effectivenessâproperly understoodâon the y-axis and x-axis respectively would be better, i.e. communicate what we would want to communicate and not suffer from the above-mentioned shortcomings?
Thanks for the comment! I feel funny saying this without being the author, but feel like the rest of my comment is a bit cold in tone, so thought itâs appropriate to add this :)
I lean more moral anti-realist but I struggle to see how the concept of âvalue alignmentâ and âdecision-making qualityâ are not similarly orthogonal from a moral realist view than an anti-realist view.
Moral realist frame: âThe more the institution is intending to do things according to the âtrue moral viewâ, the more itâs value-aligned.â
âThe better the institutionsâs decision making process is at predictably leading to what they value, the better their âdecision-making qualityâ is.â
I donât see why these couldnât be orthogonal in at least some cases. For example, a terrorist organization could be outstandingly good at producing outstandingly bad outcomes.
Still, itâs true that the âvalue-alignedâ term might not be the best, since some people seem to interpret it as a dog-whistle for ânot following EA dogma enoughâ [link] (I donât, although might be mistaken). âAltruismâ and âEffectivenessâas the x and y axes would suffer from the problem mentioned in the post that it could alienate people coming to work on IIDM from outside the EA community. For the y-axis, ideally Iâd like some terms that make it easy to differentiate between beliefs common in EA that are uncontroversial (âletâs value peopleâs lives the same regardless of where they liveâ), and beliefs that are more controversial (âx-risk is the key moral priority of our timesâ).
About the problematicness of â value-neutralâ: I thought the post gave enough space for the belief that institutions might be worse than neutral on average, marking statements implying the opposite as uncertain. For example crux (a) exists in this image to point out that if you disagree with it, you would come to a different conclusion about the effectiveness of (A).
(Iâm testing out writing more comments on the EA forum, feel free to say if it was helpful or not! I want to learn to spend less time on these. This took about 30 minutes.)
Thanks for the post and for taking the time! My initial thoughts on trying to parse this are below, I think it will bring mutual understanding further.
You seem to make a distinction between intentions on the y-axis and outcomes on the x-axis. Interesting!
The terrorist example seems to imply that if you want bad outcomes you are not value-aligned (aligned to what? to good outcomes?). They are value-aligned from their own perspective. And âterroristâ is also not a value-neutral term, for example Nelson Mandela was once considered one, which would I think surprise most people now.
If we allow âfrom their own perspectiveâ then âeffectivenessâ would do (and âefficiencyâ to replace the x-axis), but it seems we donât, and then âaltruismâ (or perhaps âgoodâ, with less of an explicit tie to EA?) would without the ambiguity âvalue-alignedâ brings on whether or not we do [allow âfrom their own perspectiveâ].
(As not a moral realist, the option of âbetter valueâ is not available, so it seems one would be stuck with âfrom their own perspectiveâ and calling the effective terrorist value-aligned, or moving to an explicit comparison to EA values, which I was supposing was not the purpose, and seems to be even more off-putting via the mentioned alienating shortcoming in communication.)
Next to value-aligned being suboptimal, which I also just supported further, you seem to agree with altruism and effectiveness (I would now suggest âefficiencyâ instead) as appropriate labels, but agree with the author about the shortcoming for communicating to certain audiences (alienation), with which I also agree. For other audiences, including myself, the current form perhaps has shortcomings. I would value clarity more, and call the same the same. An intentional opaque-making change of words might additionally come across as deceptive, and as aligned with oneâs own ideas of good, but not with such ideas in a broader context. And that I think could definitely also count as /â become a consequential shortcoming in communication strategy.
And regarding the non-orthogonality, I wasâas a moral realist -more thinking along the lines of: being organized (etc., etc.), is presumably a good value, and it would also improve your decision-making (sort of considered neutrally)...