Saving lives in normal times is better to improve the longterm future than doing so in catastrophes?

You can give me anonymous feedback about anything you want here.

Summary

  • Interventions in the effective altruism community are usually assessed under 2 different frameworks, existential risk mitigation, and nearterm welfare improvement. It looks like 2 distinct frameworks are needed given the difficulty of comparing nearterm and longterm effects. However, I do not think this is quite the right comparison under a longtermist perspective, where most of the expected value of one’s actions results from influencing the longterm future, and the indirect longterm effects of saving lives outside catastrophes cannot be neglected.

  • In this case, I believe it is better to use a single framework for assessing interventions saving human lives in catastrophes and normal times. One way of doing this, which I consider in this post, is supposing the benefits of saving one life are a function of the population size.

  • Assuming the benefits of saving a life are proportional to the ratio between the initial and final population, and that the cost to save a life does not depend on this ratio, it looks like saving lives in normal times is better to improve the longterm future than doing so in catastrophes.

  • Many in the effective altruism community argue for focussing on existential risk mitigation given its high neglectedness relative to its importance, i.e. high ratio between potential (astronomical) benefits and (tiny) amount of resources currently being dedicated to it. However, this does not appear to be the right measure of whether a problem is neglected in the relevant sense. One has to consider not only the benefits of solving a problem, but also multiply these by their probability, and arguably expect spending to be proportional to the product.

  • The above is in contrast to Nick Bostrom’s maxipok rule.

    • I question this as follows under my framework:

      • The benefits of saving a life tend to infinity as the post-catastrophe population goes to 0, i.e. as one is increasingly confident it would cause human extinction.

      • Yet, the probability of the severity of a catastrophe tends to 0 as severity increases.

      • The 2 points above have opposing effects in terms of the expected value of saving a human life, and one cannot consider each of them in isolation because they are not independent.

      • I think the expected value of saving a human life tends to 0 under reasonable assumptions as the severity of the catastrophe increases.

    • As a 1st approximation, I believe all interventions could be fairly assessed based on a single metric like the number of disability-adjusted life years averted per dollar, as calculated through a standard cost-effectiveness analysis.

Introduction

Interventions in the effective altruism community are usually assessed under 2 different frameworks:

  • Existential risk mitigation, where the goal is often decreasing extinction or global catastrophic risk associated with catastrophes in the next few decades. For example, the interventions supported by the Long-Term Future Fund (LTFF).

  • Nearterm welfare improvement, where the goal is improving the lives of animals or humans in the next few decades. For example, the interventions supported by the Animal Welfare Fund and GiveWell’s funds.

In agreement with the above and its commitment to worldview diversification, Open Philanthropy says:

So far, the focus areas we have selected fall into one of two broad categories: Global Health and Wellbeing (GHW) and Global Catastrophic Risks (GCR). We summarize the key differences between these portfolios as follows:

  • While GCR grants tend to be evaluated based on something like “How much this grant reduces the chance of a catastrophic event that endangers billions of people”, GHW grants tend to be evaluated based on something like “How much this grant increases health (denominated in e.g. life-years) and/​or wellbeing, worldwide.”

  • The GHW team places greater weight on evidence, precedent, and track record in its giving; the GCR team tends to focus on problems and interventions where evidence and track records are often comparatively thin. (That said, the GHW team does support a significant amount of low-probability but high-upside work like policy advocacy and scientific research.)

  • The GCR team’s work could be hugely important, but it’s very hard to answer questions like “How will we know whether this work is on track to have an impact?” We can track intermediate impacts and learn to some degree, but some key premises likely won’t become very clear for decades or more. (Our primary goal is for catastrophic events not to happen, and to the extent we succeed, it can be hard to learn from the absence of events.) By contrast, we generally expect the work of the GHW team to be more likely to result in recognizable impact on a given ~10-year time frame, and to be more amenable to learning and changing course as we go.

It looks like 2 distinct frameworks are needed given the difficulty of comparing nearterm and longterm effects. However, I do not think this is quite the right comparison under a longtermist perspective, where most of the expected value of one’s actions results from influencing the longterm future, and the indirect longterm effects of saving lives outside catastrophes cannot be neglected. In this case, I believe it is better to use a single framework for assessing interventions saving human lives in catastrophes and normal times. One way of doing this, which I consider in this post, is supposing the benefits of saving one life are a function of the population size.

Note Rethink Priorities’ cross-cause cost-effectiveness model (CCM) also estimates the cost-effectiveness of interventions in the 2 aforementioned categories under the same metric (DALY/​$). Nevertheless, I would say it underestimates the benefits of saving lives in normal times by neglecting their indirect longterm effects, which should be considered if one holds a longtermist view.

Expected value density of the benefits and cost-effectiveness of saving a life

Denoting the population at the start and end of any period of 1 year (e.g. a calendar year) by and , which I will refer to as the initial and final population, I assume:

  • follows a Pareto distribution (power law), which means its probability density function (PDF) is , where is the tail index.

    • Given the probability of being higher than a given ratio is , one can determine from , where and are the tail risks respecting the ratios and .

    • I would not agree with assuming a Pareto distribution for the reduction in population as a fraction of the initial population ().

      • This fraction is limited to 1, whereas a Pareto distribution can take an arbitrarily large value.

      • So assuming a Pareto would mean a reduction in population could exceed the initial population, which does not make sense.

  • The benefits of saving a life are , where:

    • are the benefits of saving a life in normal times.

    • is the elasticity of the benefits with respect to the ratio between the initial and final population.

    • for the benefits to increase as the final population decreases.

  • The cost of saving a life is , where:

    • is the cost to save a life in normal times.

    • is the elasticity of the cost with respect to the ratio between the initial and final population.

Consequently, the expected value density of:

  • The benefits of saving a life is .

  • The cost-effectiveness of saving a life is .

I guess:

  • , as Carl Shulman seems to suggest in a post on the flow-through effects of saving a life that the respective indirect longterm effects, in terms of the time by which humanity’s trajectory is advanced, are inversely proportional to the population when the life is saved[1].

  • , since I feel like the effect of having more opportunities to save lives over periods with lower final population is roughly offset by the greater difficulty of preparing to take advantage of those opportunities.

As a result, the expected value densities of the benefits and cost-effectiveness of saving a life are both proportional to , which tends to 0 as the final population decreases. So it looks like saving lives in normal times is better to improve the longterm future than doing so in catastrophes.

I got values for of 0.153 (= ln(0.0905/​0.01)/​ln(2.02*10^6/​1.11)) and 0.0835 (= ln(0.20/​0.06)/​ln(2.02*10^6/​1.11)) based on the predictions of the superforecasters and experts of the Existential Risk Persuasion Tournament (XPT) for the global catastrophic and extinction risk between 2023 and 2100:

  • Global catastrophic risk refers to more than 10 % of humans dying, so is 1.11 (= 1/​(1 − 0.1)).

  • Extinction risk refers to less than 5 k people surviving, and the global population is projected to be 10.1 billion in 2061 ((2023 + 2100)/​2 = 2061.5), which is in the middle of the relevant period, so is 2.02 M (= 10.1*10^9/​(5*10^3)).

  • Superforecasters predicted a global catastrophic and extinction risk of and , and experts and (see Table 1).

For the tail indices I mentioned, the expected value densities of the benefits and cost-effectiveness of saving a life are 34.8 % (= (10^3)^-0.153) and 56.2 % (= (10^3)^-0.0835) as high for a final population 0.1 % (= 10^-3) as large. Nonetheless, I would say the above tail indices are overestimates, as I suspect extinction risk was overestimated in the XPT. The annual risk of human extinction from nuclear war from 2023 to 2050 estimated by the superforecasters, domain experts, general existential risk experts, and non-domain experts of the XPT is 602 k, 7.23 M, 10.3 M and 4.22 M times mine.

The tail index would be 0.313 (= ln(0.0905/​0.001)/​ln(2.02*10^6/​1.11)) for my guess of an extinction risk from 2023 to 2100 of around 0.1 % (essentially from transformative artificial intelligence), and superforecasters’ global catastrophic risk of 9.05 %. In this case, the expected value densities of the benefits and cost-effectiveness of saving a life would be 11.5 % (= (10^3)^-0.313) as high for a final population 0.1 % as large.

Existential risk mitigation is not neglected?

Many in the effective altruism community argue for focussing on existential risk mitigation given its high neglectedness relative to its importance, i.e. high ratio between potential (astronomical) benefits and (tiny) amount of resources currently being dedicated to it. However, this does not appear to be the right measure of whether a problem is neglected in the relevant sense. One has to consider not only the benefits of solving a problem, but also multiply these by their PDF, and arguably expect spending to be proportional to the product, i.e. to the expected value density of the benefits. Focussing on this has implications for what is fairly/​unfairly neglected. To illustrate:

  • In the context of war deaths, the tail index is “1.35 to 1.74, with a mean of 1.60”. So the PDF of the benefits is proportional to “deaths”^-2.6 (= “deaths”^-(“tail index” + 1)).

  • So I think spending should a priori be proportional to “deaths”^-1.6 (= “deaths”*“deaths”^-2.6).

  • As a consequence, if the goal is minimising war deaths[2], spending to save lives in wars 1 k times as deadly should be 0.00158 % (= (10^3)^(-1.6)) as large.

Here is an example of a discussion which apparently does not account for the probability of the benefits being realised. In Founders Pledge’s report Philanthropy to the Right of Boom, Christian Ruhl says:

suppose for the sake of argument that a nuclear terrorist attack could kill 100,000 people, and an all-out nuclear war could kill 1 billion people. All else equal, in this scenario it would be 10,000 times more effective to focus on preventing all-out war than it is to focus on nuclear terrorism.[8]

Here is footnote 8:

All else, obviously, is not equal. Questions about the tractability of escalation management are crucial.

Here is the paragraph following what I quoted above:

Generalizing this pattern, philanthropists ought to prioritize the largest nuclear wars (again, all else equal) when thinking about additional resources at the margin.

However, as I showed, the Pareto distribution describing war deaths is such that the expected value density of deaths decreases with war severity. So, all else equal except for the probability of a war having a given size, it may well be better to address smaller wars.

Here is another example where the probability of the benefits being realised is seemingly not adequately considered. Cotton-Barratt 2020 says “it’s usually best to invest significantly into strengthening all three defence layers”:

  • Prevention. “First, how does it start causing damage?”.

  • Response. “Second, how does it reach the scale of a global catastrophe?”.

  • Resilience. “Third, how does it reach everyone?”.

“This is because the same relative change of each probability will have the same effect on the extinction probability”. I agree with this, but I wonder whether tail risk is the relevant metric. I think it is better to look into the expected value density of the cost-effectiveness of saving a life, accounting for indirect longterm effects as I did. I predict this expected value density to be higher for the 1st layers, which respect a lower severity, but are more likely to be requested. So, to equalise the marginal cost-effectiveness of additional investments across all layers, it may well be better to invest more in prevention than in response, and more in response than in resilience.

Maxipok is not ok?

My conclusion that saving lives in normal times is better to improve the longterm future than doing so in catastrophes is in contrast to Nick Bostrom’s maxipok rule:

Maximize the probability of an “OK outcome,” where an OK outcome is any outcome that avoids existential catastrophe.

I question the above as follows under my framework:

  • The benefits of saving a life tend to infinity as the post-catastrophe population goes to 0, i.e. as one is increasingly confident it would cause human extinction.

  • Yet, the PDF of the severity of the catastrophe (ratio between the pre- and post-catastrophe population) tends to 0 in that case.

  • The 2 points above have opposing effects in terms of the expected value of saving a human life, and one cannot consider each of them in isolation because they are not independent.

  • The 2 points are encapsulated in the expected value density of the benefits of saving a life, which I think tends to 0 under reasonable assumptions as the severity of the catastrophe increases.

I wonder which of the following heuristics is better to improve the longterm future. Maximising:

  • The probability of nearterm human welfare not dropping to zero (roughly as in maxipok).

  • Nearterm human welfare.

  • Nearterm welfare (not only of humans, but all sentient beings, including animals and beings with artificial sentience).

Maxipok matches the 1st of these to the extent nearterm risk of human extinction is a good proxy for nearterm existential risk[3], so I guess Nick would suggest the 1st of the above is the best heuristic. Nevertheless, it naively seems to me that the last one is broader:

  • It depends on the whole distribution of welfare instead of just its left tail.

  • It encompasses all sentient beings instead of just humans.

So a priori I would conclude maximising nearterm welfare is the best heuristic among the above. Accordingly, as a 1st approximation, I believe all interventions could be fairly assessed based on a single metric like the number of disability-adjusted life years (DALYs) averted per dollar, as calculated through a standard cost-effectiveness analysis[4] (CEA). Ideally, cost-effectiveness analyses would account for effects on non-human beings too.

  1. ^

    Here is the relevant excerpt:

    For example, suppose one saved a drowning child 10,000 years ago, when the human population was estimated to be only in the millions. For convenience, we’ll posit a little over 7 million, 1/​1000th of the current population. Since the child would add to population pressures on food supplies and disease risk, the effective population/​economic boost could range from a fraction of a lifetime to a couple of lifetimes (via children), depending on the frequency of famine conditions. Famines were not annual and population fluctuated on a time scale of decades, so I will use 20 years of additional life expectancy.

    So, for ~ 20 years the ancient population would be 17,000,000th greater, and economic output/​technological advance. We might cut this to 110,000,000 to reflect reduced availability of other inputs, although increasing returns could cut the other way. Using 110,000,000 cumulative world economic output would reach the same point ~ 1500,000th of a year faster. An extra 1500,000th of a year with around our current population of ~7 billion would amount to an additional ~14,000 life -years, 700 times the contemporary increase in life years lived. Moreover, those extra lives on average have a higher standard of living than their ancient counterparts.

    Readers familiar with Nick Bostrom’s paper on astronomical waste will see that this is a historical version of the same logic: when future populations will be far larger, expediting that process even slightly can affect the existence of many people. We cut off our analysis with current populations, but the greater the population this growth process will reach, the greater long-run impact of technological speedup from saving ancient lives.

  2. ^

    This is a decent goal in the model I presented previously if the post-catastrophe population is not too different from the pre-catastrophe one.

  3. ^

    Human extinction may not be an existential catastrophe if it is followed by the emergence of another intelligent species or caused by benevolent artificial intelligence.

  4. ^

    I prefer focusing on wellbeing-adjusted life years (WELLBYs) instead of DALYs, but both work for the point I am making as long as they are calculated through a standard CEA. Elliott Thornley and Carl Shulman argued longtermists should commit to acting in accordance with a catastrophe policy driven by standard cost-benefit analysis (CBA), but only in the political sphere, whereas I think the same holds in other contexts.