Error
Unrecognized LW server error:
Field "fmCrosspost" of type "CrosspostOutput" must have a selection of subfields. Did you mean "fmCrosspost { ... }"?
Unrecognized LW server error:
Field "fmCrosspost" of type "CrosspostOutput" must have a selection of subfields. Did you mean "fmCrosspost { ... }"?
One possibility is that there aren’t many risks that are truly unknown, in the sense that they fall outside of the categories Toby enumerates, for the simple reason that some of those categories are relatively broad, so cover much of the space of possible risks.
Even if that were true, there might still be (fine-grained) risks we haven’t thought about within those categories, however—e.g. new ways in which AI could cause an existential catastrophe.
Great post!
I don’t think this is true. The Doomsday Argument literature (Carter, Leslie, Gott etc.) mostly considers the probability of extinction independently of any specific risks, so these authors’ estimates implicitly involve an assessment of unknown risks. Lots of this writing was before there were well-developed cases for specific risks. Indeed, the Doomsday literature seems to have inspired Leslie, and then Bostrom, to start seriously considering specific risks.
Leslie explicitly considers unknown risks (p.146, End of the World):
As does Bostrom (2002):
I agree that the literature on the Doomsday Argument involves an implicit assessment of unknown risks, in the sense that any residual probability mass assigned to existential risk after deducting the known x-risks must fall under the unknown risks category. (Note that our object-level assessment of specific risks may cause us to update our prior general risk estimates derived from the Doomsday Argument.)
Still, Michael’s argument is not based on anthropic considerations, but on extrapolation from the rate of x-risk discovery. These are two very different reasons for revising our estimates of unknown x-risks, so it’s important to keep them separate. (I don’t think we disagree; I just thought this was worth highlighting.)
Related to this, I find anthropic reasoning pretty suspect, and I don’t think we have a good enough grasp on how to reason about anthropics to draw any strong conclusions about it. The same could be said about choices of priors, e.g., MacAskill vs. Ord where the answer to “are we living at the most influential time in history?” completely hinges on the choice of prior, but we don’t really know the best way to pick a prior. This seems related to anthropic reasoning in that the Doomsday Argument depends on using a certain type of prior distribution over the number of humans who will ever live. My general impression is that we as a society don’t know enough about this kind of thing (and I personally know hardly anything about it). However, it’s possible that some people have correctly figured out the “philosophy of priors” and that knowledge just hasn’t fully propagated yet.
Thanks — I agree with this, and should have made clearer that I didn’t see my comment as undermining the thrust of Michael’s argument, which I find quite convincing.
Thanks for this perspective! I’ve heard of the Doomsday Argument but I haven’t read the literature. My understanding was that the majority belief is that the Doomsday Argument is wrong, we just haven’t figured out why it’s wrong. I didn’t realize there was substantial literature on the problem, so I will need to do some reading!
I think it is still accurate to claim that very few sources have considered the probability of unknown risks relative to known risks. I’m mainly basing this off the Rowe & Beard literature review, which is pretty comprehensive AFAIK. Leslie and Bostrom discuss unknown risks, but without addressing their relative probabilities (at least Bostrom doesn’t, I don’t have access to Leslie’s book right now). If you know of any sources that address this that Rowe & Beard didn’t cover, I’d be happy to hear about them.
Note that it’s not just the Doomsday Argument that may give one reason for revising one’s x-risk estimates beyond what is suggested by object-level analyses of specific risks. See this post by Robin Hanson for an enumeration of the relevant types of arguments.
I am puzzled that these arguments do not seem to be influencing many of the most cited x-risk estimates (e.g. Toby’s). Is this because these arguments are thought to be clearly flawed? Or is it because people feel they just don’t know how to think about them, and so they simply ignore them? I would like to see more “reasoning transparency” about these issues.
It’s also worth noting that some of these “speculative” arguments would provide reason for revising not only overall x-risks estimates, but also estimates of specific risks. For example, as Katja Grace and Greg Lewis have noted, a misaligned intelligence explosion cannot be the Great Filter. Accordingly, insofar as the Great Filter influences one’s x-risk estimates, one should shift probability mass away from AI risk and toward risks that are plausible Great Filters.
In line with Matthew’s comment, think it’s true that several sources discuss the possibility of unknown risks or discuss the total risk level (which should presumably include unknown risks). But I’m also not aware of any sources apart from Ord and Pamlin & Armstrong which give quantitative estimates of unknown risk specifically. (If someone knows of any, please add them to my database!)
I’m also not actually sure if I know of any other sources which even provide relatively specific qualitative statements about the probability of unknown risks causing existential catastrophe. I wouldn’t count the statements from Leslie and Bostrom, for example.
So I think Michael’s claim is fair, at least if we interpret it as “no other sources appear to have clearly addressed the likelihood of unknown x-risks in particular, which implies that most others do not give unknown risks serious consideration.”
I agree with all the points you make in the “Implications” section. I also provide some relevant ideas in Some thoughts on Toby Ord’s existential risk estimates, in the section on how Ord’s estimates (especially the relatively high estimates for “other” and “unforeseen” risks) should update our career and donation decisions.
I was confused by this claim. Wouldn’t you put risks from AI, risks from nanotechnology, and much of biorisk in the category of “x-risks that depend on future technology”, rather than currently possibly yet unrecognised risks? And wouldn’t you say that effort thinking about possible sources of risk has helped us identify and mitigate (or strategise about how to mitigate) those three risks? If so, I’d guess that similar efforts could help us with similar types of risks in future, as well.
I agree. I’d also add that I think a useful way to think about this is in terms of endogenous vs exogenous increases in deliberate efforts to reduce x-risk, where the former essentially means “caused by people like us, because they thought about arguments like this”, and the latter essentially means “caused by other people or other events” (e.g., mainstream governments and academics coming to prioritise x-risk more for reasons other than the influence of EA). The more we expect exogenous increases in deliberate efforts to reduce x-risk, the less we should prioritise unknown risks. The same is not true with regards to endogenous increases, because deprioritising this area makes the endogenous increases unlikely.
(This distinction came to my mind due to SjirH making a similar distinction in relation to learning about giving opportunities.)
This is similar to discussions about how “sane” society is, and how much we should expect problems to be solved “by default”.
You seem to be using this data point to support the argument “There are many risks we’ve discovered only recently, and we should be worried about these and about risks we have yet to discover.” That seems fair.
But I think that, if you’re using that data point in that way, it’s probably worth also more explicitly noting the following: People were worried about existential catastrophe from nuclear war via mechanisms that we now know don’t really warrant concern (as alluded to in the Ord quote you also provide). So there seems to be a companion data point which supports the argument “In the past, many risks that were recently discovered later turned out to not be big deals, so we should be less worried about recently discovered or to-be-discovered risks than we might otherwise think.”
(And I’d say that both data points provide relatively weak evidence for their respective arguments.)
Thanks for this post—I think this is a very important topic! I largely agree that this argument deserves substantial weight, and that we should probably think more about unknown existential risks and about how we should respond to the possibility of them.
(I’m not sure if the following is disagreeing with you or just saying something separate yet consistent)
I agree that recently discovered risks seem the largest source of existential risk, and that two plausible explanations of that are “technological growth introduces increasingly significant risks” and “we have bad priors in general and wider error bars for probability estimates about recently discovered risks than about risks we discovered earlier”. But it also seems to me that we can believe all that and yet still think there’s a fairly high chance we’ll later realise that the risks we recently discovered are less risky than we thought, and that future unknown risks in general aren’t very risky, even if we didn’t have bad priors.
Essentially, this is because, if we have a few decades or centuries without an existential catastrophe (and especially if this happens despite the development of highly advanced AI, nanotechnology, and biotech), that’ll represent new evidence not just about those particular risks, but also about how risky advanced technologies are in general and how resilient to major changes civilization is in general.
Thousands of years where asteroids could’ve wiped humanity out but didn’t updates us towards thinking extinction risk from asteroids is low. I think that, in the same way, developing more things in the reference class of “very advanced tech with hard-to-predict consequences” without an existential catastrophe occurring should update us towards thinking existential risk from developing more things like that is relatively low.
I think we can make that update in 2070 (or whatever) without saying our beliefs in 2020 were foolish, given what we knew. And I don’t think we should make that update yet, as we haven’t yet seen that evidence of “general civilizational stability” (or something like that).
Does that make sense?
Admin: The “Pamlin & Armstrong (2015)” link was broken — I updated it.