I’m much more confident about the (positive wellbeing + suffering) vs neither trade than intra-suffering trades. It sounds right that something like the tradeoff you describe follows from the most intuitive version of my model, but I’m not actually certain of this; like maybe there is a system that fits within the bounds of the thing I’m arguing for that chooses A instead of B (with no money pumps/very implausible conclusions following)
Ok interesting! I’d be interested in seeing this mapped out a bit more, because it does sound weird to have BOS be offsettable with positive wellbeing, positive wellbeing to be not offsettable with NOS, but BOS and NOS are offsetable with each other? Or maybe this isn’t your claim and I’m misunderstanding
2) Well the question again is “what would the IHE under experiential totalization do?” Insofar as the answer is “A”, I endorse that. I want to lean on this type of thinking much more strongly than hyper-systematic quasi-formal inferences about what indirectly follows from my thesis.
Right, but if IHE does prefer A over B in my case while also preferring the “neither” side of the [positive wellbeing + NOS] vs neither trade then there’s something pretty inconsistent right? Or a missing explanation for the perceived inconsistency that isn’t explained by a lexical threshold.
I think it’s possible that the answer is just B because BOS is just radically qualitatively different from NOS.
I think this is plausible but where does the radical qualitative difference come from? (see comments RE: formalising the threshold).
Maybe most importantly I (tentatively?) object to the term “barely” here because under the asymptotic model I suggest, the value of subtracting arbitrarily small amount of suffering instrument ϵ from the NOS state results in no change in moral value at all because (to quote myself again) “Working in the extended reals, this is left-continuous: limis→is∗ϕ(is)=∞+=ϕ(is∗)”
Sorry this is too much maths for my smooth brain but I think I’d be interested in understanding why I should accept the asymptotic model before trying to engage with the maths! (More on this below, under “On the asymptotic compensation schedule”)
So in order to get BOS, we need to remove something larger than ϵ, and now it’s a quasi-empirical question of how different that actually feels from the inside. Plausibly the answer is that “BOS” (scare quotes) doesn’t actually feel “barely” different—it feels extremely and categorically different
Can you think of one generalisable real world scenario here? Like “I think this is clearly non-offsetable and now I’ve removed X, I think it is clearly offsetable”
And I’ll add that insofar as the answer is (2) and NOT 3, I’m pretty inclined to update towards “I just haven’t developed an explicit formalization that handles both the happiness trade case and the intra-suffering trade case yet” more strongly than towards “the whole thing is wrong, suffering is offsetable by positive wellbeing”—after all, I don’t think it directly follows from “IHE chooses A” that “IHE would choose the 70 years of torture.” But I could be wrong about this! I 100% genuinely think I’m literally not smart enough to intuit super confidently whether or a formalization that chooses both A and no torture exists. I will think about this more!
Cool! Yeah I’d be excited to see the formalisation; I’m not making a claim that the whole thing is wrong, more making a claim that I’m not currently sufficiently convinced to hold the view that some suffering cannot be offsetable. I think while the intuitions and the hypotheticals are valuable, like you say later, there are a bunch of things about this that we aren’t well placed to simulate or think about well, and I suspect if you find yourself in a bunch of hypotheticals where you feel like your intuitions differ and you can’t find a way to resolve the inconsistencies then it is worth considering the possibility that you’re not adequately modelling what it is like to be the IHE in at least one of the hypotheticals
I more strongly want to push back on (2) and (3) in the sense that I think parallel experience, while probably conceptually fine in principle, really greatly degrades the epistemic virtue of the thought experiment because this literally isn’t something human brains were/are designed to do or simulate.
Yeah reasonable, but presumably this applies to answers for your main question[1] too?
Suppose the true value of exchange is at 10 years of happiness afterwards; this seems easier for our brains to simulate than if the true exchange rate is at 100,000 years of happiness, especially if you insist on parallel experiences. Perhaps it is just very difficult to be scope sensitive about exactly how much bliss 1E12 years of bliss is!
And likewise with (3), the self interest bit seems pretty epistemically important.
can you clarify what you mean here? Isn’t the IHE someone who is “maximally rational/makes no logical errors, have unlimited information processing capacity, complete information about experiences with perfect introspective access, and full understanding of what any hedonic state would actually feel like”?
On formalising where the lexical threshold is you say:
I agree it is imporant! Someone should figure out the right answer! Also in terms of practical implementation, probably better to model as a probability distribution than a single certain line.
This is reasonable, and I agree with probability distribution given uncertainty, but I guess it feels hard to engage with the metaphysical claim “some suffering in fact cannot be morally justified (“offset”) by any amount of happiness” and their implications if you are so deeply uncertain about what counts as NOS. I guess my view is that conditional on physicalism then whatever combination of nociceptor / neuron firing and neurotransmitter release / you can think of, this is a measurable amount. some of these combinations will cross the threshold of NOS under your view, but you can decrease all of those in continuous ways that shouldn’t lead to a discontinuity in tradeoffs you’re willing to make. It does NOT mean that the relationship is linear, but it seems like there’s some reason to believe it’s continuous rather than discontinuous / has an asymptote here. And contra your later point:
“I literally don’t know what the threshold is. I agree it would be nice to formalize it! My uncertainty isn’t much evidence against the view as a whole”
I think if we don’t know where a reasonable threshold is it’s fine to remain uncertain about it, but I think that’s much weaker than accepting the metaphysical claim! It’s currently based just on the 70 years of worst-possible suffering VS ~infinite bliss hypothetical. Because your uncertainty about the threshold means I can conjure arbitrarily high numbers of hypotheticals that would count as evidence against your view in the same way your hypothetical is considered evidence for your view.
On the asymptotic compensation schedule
I disagree that it isn’t well-justified in principlle, but maybe I should have argued this more thoroughly. It just makes a ton of intuitive sense to me but possibly I am typical-minding.
As far as I can tell, you just claim that it creates an asymptote and label it the correct view right? But why should it grow without bound? Sorry if I’ve missed something!
And I’m pretty sure you’re wrong about the second thing—see point 3 a few bullets up. It seems radically less plausible to me that the true nature of ethics involves discontinuous i_s vs i_h compensation schedules.
I was unclear about the “doesn’t seem to meaningfully change the unintuitive nature of the tradeoffs your view is willing to endorse” part you’re referring to here, and I agree RE: discontinuity. What I’m trying to communicate is that if someone isn’t convinced by the perceived discontinuity of NOS being non-offsettable and BOS being offsettable, a large subset of them also won’t be very convinced by the response “the radical part is in the approach to infinity, (in your words: the compensation schedule growing without bound (i.e., asymptotically) means that some sub-threshold suffering would require 10^(10^10) happy lives to offset, or 1000^(1000^1000). (emphasis added)”.
Because they could just reject the idea that an extremely bad headache (but not a cluster headache), or a short cluster headache episode, or a cluster headache managed by some amount of painkiller, etc, requires 1000^(1000^1000) happy lives to offset.
I guess this is just another way of saying “it seems like you’re assuming people are buying into the asymptotic model but you haven’t justified this”.
Ok interesting! I’d be interested in seeing this mapped out a bit more, because it does sound weird to have BOS be offsettable with positive wellbeing, positive wellbeing to be not offsettable with NOS, but BOS and NOS are offsetable with each other? Or maybe this isn’t your claim and I’m misunderstanding
This is what kills the proposal IMO, and EJT also pointed this out. The key difference between this proposal and standard utilitarianism where anything is offsettable isn’t the claim that that NOS is worse than TREE(3) or even 10^100 happy lives, since this isn’t a physically plausible tradeoff we will face anyway. It’s that once you believe in NOS, transitivity compels you to believe it is worse than any amounts of BOS, even a variety of BOS that, according to your best instruments, only differs from NOS in the tenth decimal place. Then once you believe this, the fact that you use a utility function compels you to create arbitrary amounts of BOS to avoid a tiny probability of a tiny amount of NOS.
Ok interesting! I’d be interested in seeing this mapped out a bit more, because it does sound weird to have BOS be offsettable with positive wellbeing, positive wellbeing to be not offsettable with NOS, but BOS and NOS are offsetable with each other? Or maybe this isn’t your claim and I’m misunderstanding
Right, but if IHE does prefer A over B in my case while also preferring the “neither” side of the [positive wellbeing + NOS] vs neither trade then there’s something pretty inconsistent right? Or a missing explanation for the perceived inconsistency that isn’t explained by a lexical threshold.
I think this is plausible but where does the radical qualitative difference come from? (see comments RE: formalising the threshold).
Sorry this is too much maths for my smooth brain but I think I’d be interested in understanding why I should accept the asymptotic model before trying to engage with the maths! (More on this below, under “On the asymptotic compensation schedule”)
Can you think of one generalisable real world scenario here? Like “I think this is clearly non-offsetable and now I’ve removed X, I think it is clearly offsetable”
Cool! Yeah I’d be excited to see the formalisation; I’m not making a claim that the whole thing is wrong, more making a claim that I’m not currently sufficiently convinced to hold the view that some suffering cannot be offsetable. I think while the intuitions and the hypotheticals are valuable, like you say later, there are a bunch of things about this that we aren’t well placed to simulate or think about well, and I suspect if you find yourself in a bunch of hypotheticals where you feel like your intuitions differ and you can’t find a way to resolve the inconsistencies then it is worth considering the possibility that you’re not adequately modelling what it is like to be the IHE in at least one of the hypotheticals
Yeah reasonable, but presumably this applies to answers for your main question[1] too?
Suppose the true value of exchange is at 10 years of happiness afterwards; this seems easier for our brains to simulate than if the true exchange rate is at 100,000 years of happiness, especially if you insist on parallel experiences. Perhaps it is just very difficult to be scope sensitive about exactly how much bliss 1E12 years of bliss is!
can you clarify what you mean here? Isn’t the IHE someone who is “maximally rational/makes no logical errors, have unlimited information processing capacity, complete information about experiences with perfect introspective access, and full understanding of what any hedonic state would actually feel like”?
On formalising where the lexical threshold is you say:
This is reasonable, and I agree with probability distribution given uncertainty, but I guess it feels hard to engage with the metaphysical claim “some suffering in fact cannot be morally justified (“offset”) by any amount of happiness” and their implications if you are so deeply uncertain about what counts as NOS. I guess my view is that conditional on physicalism then whatever combination of nociceptor / neuron firing and neurotransmitter release / you can think of, this is a measurable amount. some of these combinations will cross the threshold of NOS under your view, but you can decrease all of those in continuous ways that shouldn’t lead to a discontinuity in tradeoffs you’re willing to make. It does NOT mean that the relationship is linear, but it seems like there’s some reason to believe it’s continuous rather than discontinuous / has an asymptote here. And contra your later point:
I think if we don’t know where a reasonable threshold is it’s fine to remain uncertain about it, but I think that’s much weaker than accepting the metaphysical claim! It’s currently based just on the 70 years of worst-possible suffering VS ~infinite bliss hypothetical. Because your uncertainty about the threshold means I can conjure arbitrarily high numbers of hypotheticals that would count as evidence against your view in the same way your hypothetical is considered evidence for your view.
On the asymptotic compensation schedule
As far as I can tell, you just claim that it creates an asymptote and label it the correct view right? But why should it grow without bound? Sorry if I’ve missed something!
I was unclear about the “doesn’t seem to meaningfully change the unintuitive nature of the tradeoffs your view is willing to endorse” part you’re referring to here, and I agree RE: discontinuity. What I’m trying to communicate is that if someone isn’t convinced by the perceived discontinuity of NOS being non-offsettable and BOS being offsettable, a large subset of them also won’t be very convinced by the response “the radical part is in the approach to infinity, (in your words: the compensation schedule growing without bound (i.e., asymptotically) means that some sub-threshold suffering would require 10^(10^10) happy lives to offset, or 1000^(1000^1000). (emphasis added)”.
Because they could just reject the idea that an extremely bad headache (but not a cluster headache), or a short cluster headache episode, or a cluster headache managed by some amount of painkiller, etc, requires 1000^(1000^1000) happy lives to offset.
I guess this is just another way of saying “it seems like you’re assuming people are buying into the asymptotic model but you haven’t justified this”.
“Would you accept 70 years of the worst conceivable torture in exchange for any amount of happiness afterward?”
This is what kills the proposal IMO, and EJT also pointed this out. The key difference between this proposal and standard utilitarianism where anything is offsettable isn’t the claim that that NOS is worse than TREE(3) or even 10^100 happy lives, since this isn’t a physically plausible tradeoff we will face anyway. It’s that once you believe in NOS, transitivity compels you to believe it is worse than any amounts of BOS, even a variety of BOS that, according to your best instruments, only differs from NOS in the tenth decimal place. Then once you believe this, the fact that you use a utility function compels you to create arbitrary amounts of BOS to avoid a tiny probability of a tiny amount of NOS.