Should strong longtermists really want to minimize existential risk?

Strong longtermists believe there is a non-negligible chance that the future will be enormous. For example, earth-originating life may one day fill the galaxy with digital minds. The future therefore has enormous expected value, and concern for the long-term should almost always dominate near-term considerations, at least for those decisions where our goal is to maximize expected value.

It is often stated that strong longtermism reduces in practice to the goal: “minimize existential risk at all costs”. I argue here that this is inaccurate. I claim that a more accurate way of summarising the strong longtermist goal is: “minimize existential risk at all costs conditional on the future possibly being very big”. I believe the distinction between these two goals has important practical implications. The strong longtermist goal may actually conflict with the goal of minimizing existential risk unconditionally.

In the next section I describe a thought experiment to demonstrate my claim. In the following section I argue that this is likely to be relevant to the actual world we find ourselves in. In the final section I give some concluding remarks on what we should take away from all this.

The Anti-Apocalypse Machine

The Earth is about to be destroyed by a cosmic disaster. This disaster would end all life, and snuff out all of our enormous future potential.

Fortunately, physicists have almost settled on a grand unified theory of everything that they believe will help them build a machine to save us. They are 99% certain that the world is described by Theory A, which tells us we can be saved if we build Machine A. But there is a 1% chance that the correct theory is actually Theory B, in which case we need to build Machine B. We only have the time and resources to build one machine.

It appears that our best bet is to build Machine A, but there is a catch. If Theory B is true, then the expected value of our future is many orders of magnitude larger (although it is enormous under both theories). This is because Theory B leaves open the possibility that we may one day develop slightly-faster-than-light travel, while Theory A being true would make that impossible.

Due to the spread of strong longtermism, Earth’s inhabitants decide that they should build Machine B, acting as if the speculative Theory B is correct, since this is what maximizes expected value. Extinction would be far worse in the Theory B world than the Theory A world, so they decide to take the action which would prevent extinction in that world. They deliberately choose a 99% chance of extinction over a 1% chance, risking all of humanity, and all of humanity’s future potential.

The lesson here is that strong longtermism gives us the goal to minimize existential risk conditional on the future possibly being very big, and that may conflict with the goal to minimize existential risk unconditionally.

Relevance for the actual world

The implication of the above thought experiment is that strong longtermism tells us to look at the set of possible theories about the world, pick the one in which the future is largest, and, if it is large enough, act as if that theory were true. This is likely to have absurd consequences if carried to its logical conclusion, even in real world cases. I explore some examples in this section.

The picture becomes more confusing when you consider theories which permit the future to have infinite value. In Nick Beckstead’s original thesis, On the Overwhelming Importance of Shaping the Far Future, he explicitly singles out infinite value cases as examples of where we should abandon expected value maximization, and switch to using a more timid decision framework instead. But even if strong longtermists are only reckless in large finite cases, that should still be enough for them to be forced to adopt extremely speculative scientific theories (using ‘adopt’ as a shorthand for ‘act as if this theory were true’).

Out of all our scientific knowledge, the 2nd law of thermodynamics is arguably one of the principles that is least likely to be proven wrong. But we can’t completely rule out the possibility that counter-examples will one day be found. The 2nd law also puts strict limits on how big the future can be. I claim strong longtermists should therefore act as if the 2nd law will turn out to be false. The same goes for any other currently understood physical limit on our growth, such as the idea that information cannot travel faster than light.

This may well have practical implications for the work that strong longtermists are currently doing on existential risk. For example, perhaps it implies that they should adopt a big distrust of the scientific establishment. There are already people on the internet who claim to have built machines which violate the 2nd law of thermodynamics. The scientific establishment have largely been ignoring these amateur scientists’ claims so far. If strong longtermists should act as if these amateur scientists are correct to reject the 2nd law, then that might mean putting less weight on the opinions of the scientific establishment, and more weight on the opinions of these amateurs.

It could be fairly objected that in a world where the 2nd law of thermodynamics is false, it is more likely to be overturned by mainstream physics, than by a random youtuber. If this is true, then perhaps the preceding claim that strong longtermists should distrust the scientific establishment goes too far. Nevertheless, strong longtermists should still act as if the 2nd law will one day turn out to be broken, or that faster-than-light travel will one day turn out to be possible, since we can’t rule these possibilities out completely, and they contain enormous expected value. I find it hard to believe that a commitment to such fundamental and unlikely beliefs would not have any practical implications.

On a more practical level, once we condition on the future potentially being enormous, that should lead us to overestimate humanity’s ability to coordinate to solve global problems, relative to what our estimation would have been without this conditioning, since such coordination will surely be necessary for us to spread throughout the galaxy. This overestimation may then lead us to different prioritisations among the current existential risks we face, than if we were just trying to minimize existential risk unconditionally.

Overall, I think we should expect the attempts of strong longtermists to reduce existential risk to be hindered, at least to some extent, if they are committed to adopting descriptions of the world which permit the largest possible future value, rather than descriptions of the world which are most likely to be correct.

I believe that the goals “minimize existential risk” and “minimize existential risk conditional on a possibly big future” are likely to conflict in practice, not just in principle.

Conclusion

Hopefully it is clear that this post is intended to be taken as a critique of strong longtermism, rather than as a recommendation that we should abandon the 2nd law of thermodynamics. I believe the take away here should be that possible futures involving enormous numbers of digital minds should feature less heavily in our prioritisation decisions than they do in the standard strong longtermist framework.