I agree with what you say in the last paragraph, including the highlighting of autonomy/placing value on it (whether in a realist or anti-realist way).
I’m not convinced by what you said about the effects of belief in realism vs anti-realism.
If you hold fixed people’s first-order views, not just about axiology but also about practical norms, then their metaethics makes no further difference.
Sure, but that feels like it’s begging the question.
Let’s grant that the people we’re comparing already have liberal intuitions. After all, this discussion started in a context that I’d summarize as “What are ideological risks in EA-related settings, like the FTX/SBF setting?,” so, not a setting where authoritarian intuitions are common. Also, the context wasn’t “How would we reform people who start out with illiberal intuitions” – that would be a different topic.
With that out of the way, then, the relevant question strikes me as something like this:
Under which metaethical view (if any) – axiological realism vs axiological anti-realism – is there more of a temptation for axiologically certain individuals with liberal intuitions to re-think/discount these liberal intuitions so as to make the world better according to their axiology?
Here’s how I picture the axiological anti-realist’s internal monologue:
“The point of liberal intuitions is to prevent one person from imposing their beliefs on others. I care about my axiological views, but, since I have these liberal intuitions, I do not feel compelled to impose my views on others. There’s no tension here.”
By contrast, here’s how I picture the axiological realist:
“I have these liberal intuitions that make me uncomfortable with the thought of imposing my views on others. At the same time, I know what the objectively correct axiology is, so, if I, consequentialist-style, do things that benefit others according to the objectively correct axiology, then there’s a sense in which that will be better for them than if I didn’t do it. Perhaps this justifies going against the common-sense principles of liberalism, if I’m truly certain enough and am not self-deceiving here? So, I’m kind of torn...”
I’m not just speaking about hypotheticals. I think this is a dynamic that totally happens with some moral realists in the EA context. For instance, back when I was a moral realist negative utilitarian, I didn’t like that my moral beliefs put my goals in tension with most of the rest of the world, but I noticed that there was this tension. It feels like the tension disappeared when I realized that I have to agree to disagree with others about matters of axiology (as opposed to thinking, “I have to figure out whether I’m indeed correct about my high confidence, or whether I’m the one who’s wrong”).
Sure, maybe the axiological realist will come up with a for-them compelling argument why they shouldn’t impose the correct axiology on others. Or maybe their notion of “correct axiology” was always inherently about preference fulfillment, which you could say entails respecting autonomy by definition. (That said, if someone were also counting “making future flourishing people,” as “creating more preference fulfillment,” then this sort of axiology is at least in some possible tension with respecting the autonomy of present/existing people.) ((Also, this is just a terminological note, but I usually think of preference utilitarianism as a stance that isn’t typically “axiologically realist,” so I’d say any “axiological realism” faces the same issue with there being at least a bit of tension with belief in and and valuing autonomy in practice.))
When I talked about whether there’s a “clear link” between two beliefs, I didn’t mean that the link would be binding or inevitable. All I meant is that there’s some tension that one has to address somehow.
That was the gist of my point, and I feel like the things you said in reply were perhaps often correct but they went past the point I tried to convey. (Maybe part of what goes into this disagreement is that you might be strawmanning what I think of as “anti-realism” with “relativism”.)
Here’s how I picture the axiological anti-realist’s internal monologue:
“The point of liberal intuitions is to prevent one person from imposing their beliefs on others. I care about my axiological views, but, since I have these liberal intuitions, I do not feel compelled to impose my views on others. There’s no tension here.”
By contrast, here’s how I picture the axiological realist:
“I have these liberal intuitions that make me uncomfortable with the thought of imposing my views on others. At the same time, I know what the objectively correct axiology is, so, if I, consequentialist-style, do things that benefit others according to the objectively correct axiology, then there’s a sense in which that will be better for them than if I didn’t do it. Perhaps this justifies going against the common-sense principles of liberalism, if I’m truly certain enough and am not self-deceiving here? So, I’m kind of torn...”
Right, this tendentious contrast is just what I was objecting to. I could just as easily spin the opposite picture:
(1) A possible anti-realist monologue: “I find myself with some liberal intuitions; I also have various axiological views. Upon reflection, I find that I care more about preventing suffering (etc.) than I do about abstract tolerance or respect for autonomy, and since I’m an anti-realist I don’t feel compelled to abide by norms constraining my pursuit of what I most care about.”
(2) A possible realist monologue: “The point of liberal norms is to prevent one person from imposing their beliefs on others. I’m confident about what the best outcomes would be, considered in abstraction from human choice and agency, but since it would be objectively wrong and objectionable to pursue these ends via oppressive or otherwise illicit means, I’ll restrict myself to permissible means of promoting the good. There’s no tension here.”
The crucial question is just what practical norms one accepts (liberal or otherwise). Proposing correlations between other views and bad practical norms strikes me as an unhelpful—and rather bias-prone—distraction.
I agree with what you say in the last paragraph, including the highlighting of autonomy/placing value on it (whether in a realist or anti-realist way).
I’m not convinced by what you said about the effects of belief in realism vs anti-realism.
Sure, but that feels like it’s begging the question.
Let’s grant that the people we’re comparing already have liberal intuitions. After all, this discussion started in a context that I’d summarize as “What are ideological risks in EA-related settings, like the FTX/SBF setting?,” so, not a setting where authoritarian intuitions are common. Also, the context wasn’t “How would we reform people who start out with illiberal intuitions” – that would be a different topic.
With that out of the way, then, the relevant question strikes me as something like this:
Under which metaethical view (if any) – axiological realism vs axiological anti-realism – is there more of a temptation for axiologically certain individuals with liberal intuitions to re-think/discount these liberal intuitions so as to make the world better according to their axiology?
Here’s how I picture the axiological anti-realist’s internal monologue:
“The point of liberal intuitions is to prevent one person from imposing their beliefs on others. I care about my axiological views, but, since I have these liberal intuitions, I do not feel compelled to impose my views on others. There’s no tension here.”
By contrast, here’s how I picture the axiological realist:
“I have these liberal intuitions that make me uncomfortable with the thought of imposing my views on others. At the same time, I know what the objectively correct axiology is, so, if I, consequentialist-style, do things that benefit others according to the objectively correct axiology, then there’s a sense in which that will be better for them than if I didn’t do it. Perhaps this justifies going against the common-sense principles of liberalism, if I’m truly certain enough and am not self-deceiving here? So, I’m kind of torn...”
I’m not just speaking about hypotheticals. I think this is a dynamic that totally happens with some moral realists in the EA context. For instance, back when I was a moral realist negative utilitarian, I didn’t like that my moral beliefs put my goals in tension with most of the rest of the world, but I noticed that there was this tension. It feels like the tension disappeared when I realized that I have to agree to disagree with others about matters of axiology (as opposed to thinking, “I have to figure out whether I’m indeed correct about my high confidence, or whether I’m the one who’s wrong”).
Sure, maybe the axiological realist will come up with a for-them compelling argument why they shouldn’t impose the correct axiology on others. Or maybe their notion of “correct axiology” was always inherently about preference fulfillment, which you could say entails respecting autonomy by definition. (That said, if someone were also counting “making future flourishing people,” as “creating more preference fulfillment,” then this sort of axiology is at least in some possible tension with respecting the autonomy of present/existing people.) ((Also, this is just a terminological note, but I usually think of preference utilitarianism as a stance that isn’t typically “axiologically realist,” so I’d say any “axiological realism” faces the same issue with there being at least a bit of tension with belief in and and valuing autonomy in practice.))
When I talked about whether there’s a “clear link” between two beliefs, I didn’t mean that the link would be binding or inevitable. All I meant is that there’s some tension that one has to address somehow.
That was the gist of my point, and I feel like the things you said in reply were perhaps often correct but they went past the point I tried to convey. (Maybe part of what goes into this disagreement is that you might be strawmanning what I think of as “anti-realism” with “relativism”.)
Right, this tendentious contrast is just what I was objecting to. I could just as easily spin the opposite picture:
(1) A possible anti-realist monologue: “I find myself with some liberal intuitions; I also have various axiological views. Upon reflection, I find that I care more about preventing suffering (etc.) than I do about abstract tolerance or respect for autonomy, and since I’m an anti-realist I don’t feel compelled to abide by norms constraining my pursuit of what I most care about.”
(2) A possible realist monologue: “The point of liberal norms is to prevent one person from imposing their beliefs on others. I’m confident about what the best outcomes would be, considered in abstraction from human choice and agency, but since it would be objectively wrong and objectionable to pursue these ends via oppressive or otherwise illicit means, I’ll restrict myself to permissible means of promoting the good. There’s no tension here.”
The crucial question is just what practical norms one accepts (liberal or otherwise). Proposing correlations between other views and bad practical norms strikes me as an unhelpful—and rather bias-prone—distraction.