Error
Unrecognized LW server error:
Field "fmCrosspost" of type "CrosspostOutput" must have a selection of subfields. Did you mean "fmCrosspost { ... }"?
Unrecognized LW server error:
Field "fmCrosspost" of type "CrosspostOutput" must have a selection of subfields. Did you mean "fmCrosspost { ... }"?
Distinguish substantive vs procedural rationality. Procedural rationality = following neutrally describable processes like āconsidering competing evidenceā, āavoiding incoherenceā, etc. -- roughly corresponding to intelligence. Substantive rationality = responding correctly to normative reasonsāroughly corresponding to wisdom.
The Parfitian view is that the two come apart (i.e., supporting the orthogonality thesis). Future Tuesday Indifference may be procedurally rational, or compatible with perfect intelligence, and yet still objectively crazy (unwise). More generally, thereās nothing in non-naturalist moral realism that implies that intelligent agents per se are likely to converge on the normative truth. (I discuss this more in Knowing What Matters. I think we can reasonably take ourselves to be on the right track, but thatās because of our substantive starting points, not the mere fact of our general intelligence.)
Re substantive vs procedural rationality, procedural rationality just seems roughly like instrumental rationality. For the reasons I explain, Iād expect AI to be rational in general, not just instrumentally so. Do you think ignorance of the modal facts would be possible for an arbitrarily smart agent? Iād think the moral facts would be like the modal facts in that theyād figure them out. I think that when we are smart we can figure things out and they are more likely to be true. The reason I believe modal rationalism, for example, is that there is some sense in which I feel Iāve grasped it, which wouldnāt be possible if I were much less smart.
Depends whether procedural rationality suffices for modal knowledge (e.g. if false modal views are ultimately incoherent; false moral views certainly donāt seem incoherent).
Smartness might be necessary for substantive insights, but doesnāt seem sufficient. There are plenty of smart philosophers with substantively misguided views, after all.
A metaphor: think of belief space as a giant spider web, with no single center, but instead a large number of such ācentralā clusters, each representing a maximally internally coherent and defensible set of beliefs. We start off somewhere in this web. Reasoning leads us along a strand, typically in the direction of greater coherenceāi.e., towards a cluster. But if the clusters are not differentiated in any neutrally-recognizable wayāthe truths do not glow in a way that sets them apart from ideally coherent falsehoodsāthen thereās no guarantee that philosophical reasoning (or āintelligenceā) will lead you to the truth. All it can do is lead you towards greater coherence.
Thatās still worth pursuing, because the truth sure isnāt going to be somewhere incoherent. But it seems likely that from most possible starting points (e.g. if chosen arbitrarily), the truth would be forever inaccessible.
I think I just disagree about what reasoning is. I think that reasoning does not just make our existing beliefs more coherent, but allows us to grasp new deep truths. For example, I think that an anti-realist who didnāt originally have the FTI irrational intuition could grasp it by reflection, and that one can, over time, discover that some things are just not worth pursuing and others are.
This is not what Parfit is arguing. The Future Tuesday Indifference thought experiment is part of Parfitās defense of irreducible normativity:
if the subjectivist position that acting rationally is reducible to acting in accordance with some consistent set of beliefs and preferences is true, then Future Tuesday Indifference is rational
Future Tuesday Indifference is irrational
So the subjectivist position is false
Our certain hedonistās problem is not an epistemic one: itās not that they donāt know what pain is like on Tuesday, or that theyāre not smart enough to realize that Tuesday is an arbitrary label and in truth days are simply days, or that theyāve failed to reach reflective equilibrium. Theyāre acting in perfect accord with their coherent extrapolated volitionāthe problem is just that itās a ridiculous thing to want.
Assuming that a sufficiently intelligent agent would necessarily be rational in this sense to argue against the orthogonality thesis is circular.
I agree with everything youāve said after the sentence āThis is not what Parfit is arguing.ā But how does that conflict with the things I said?
If yefreitor is saying what I planned to say, the simpler version is just āthereās nothing āirrationalā about having a utility function that says āno experience matters every Tuesday.āā It certainly wouldnāt seem to be a good instrumental value, but if thatās your terminal value function thatās what it is.
No, they literally have no negative (or positive) experience on Tuesdays, unless the experience on Tuesdays affects their experience on different days.
??? āObjectively worth pursuing?ā Where did that come from? Certainly not a Tuesday-impartial utility function, which is the only āobjectiveā thing Iām seeing here? I didnāt see where you clearly explain this through a short ctrl+f for āobjective.ā
I agree one could have that value in theory. My claim is that if one were very rational, they would not. Note that, contrary to your indication, they do have experience on Tuesday, and their suffering feels just as bad on a Tuesday as on another day. They just have a higher order indifference to future suffering. I claim that what is objectively worth pursuing is indifferent to the day of the week.
Parfitās position (and mine) is that Future Tuesday Indifference is manifestly irrational. But this has little to do with what sort of preferences sufficiently intelligent agents can have.
No, thatās explicitly ruled out in the setup. They have experiences on Tuesday, those experiences have the usual valenceāthey just fail to act accordingly. Hereās the full context from Reasons and Persons:
I think our disagreement is that I think that superintelligences would be rational and avoid FTI for the same reason theyād be epistemically rational and good at reasoning in general.
Three things:
This seems to me kind of a weird statement of the thesis, ācould in principleā being too weak.
If I understand, youāre not actually denying that just about any combination of intelligence and values could in principle occur. As you said, we can take a fact like the truth of evolution and imagine an extremely smart being thatās wrong about that specific thing. Thereās no obvious impossibility there. It seems like the same would go for basically any fact or set of facts, normative or not.
I take it the real issue is one of probability, not possibility. Is an extremely smart being likely to accept what seem like glaringly obvious moral truths (like āyou shouldnāt turn everyone into paperclipsā) in virtue of being so smart?
(2) I was surprised to see you say your case depended completely on moral realism. Of course, if youāre a realist, it makes some sense to approach things that way. Use your background knowledge, right?
But I think even an anti-realist may still be able to answer yes to the question above, depending on how the being in question is constructed. For example, I think something in this anti-orthogonality vein is true of humans. They tend to be constructed so that understanding of certain non-normative facts puts pressure on certain values or normative views: If you improve a humanās ability to imaginatively simulate the experience of living in slavery (a non-moral intellectual achievement), they will be less likely to support slavery, and so on.
This is one direction I kind of expected you to go at some point after I saw the Aaronson quote mention āthe practical versionā of the thesis. That phrase has a flavor of, āEven if the thesis is mostly true because there are no moral facts to discover, it might still be false enough to save humanity.ā
(3) But perhaps the more obvious the truths, the less intelligence matters. The claim about slavery is clearer to me than the claim that learning more about turning everyone into paperclips would make a person less likely to do so. It seems hard to imagine a person so ignorant as to not already appreciate all the morally relevant facts about turning people into paperclips. Itās as if, when the moral questions get so basic, intelligence isnāt going to make a difference. Youāve either got the values or you donāt. (But Iām a committed anti-realist, and Iām not sure how much thatās coloring these last comments.)
I think 1 is right.
2 I agree that it would depend on how the being is constructed. My claim is that itās plausible that theyād be moral by default just by virtue of being smart.
3 I think there is a sense in which I haveāand most modern people haveāunlike most people historically, grasped the badness of slavery.