# Biases in our estimates of Scale, Neglectedness and Solvability?

I suspect there are a few errors that using logarithmic scales for our estimates of Scale, Neglectedness and Solvability can make us prone to, although I’m not sure how bad this is in practice in the EA community. I describe three such possible errors. I also describe another error related only to correlations between the factors and not the log scale. In fact, *as a rule*, we aren’t accounting for possible correlations between the factors, or we’re effectively assuming the factors aren’t correlated, and there’s no general constant upper bound on how much this could bias our estimates. As a general trend, the more uncertainty in the factors, the greater the bias can be.

These could affect not just cause area analyses, but also grant-making and donations informed by these factors, as done by the Open Philanthropy Project.

**Background**

Scale/Importance, Neglectedness/Crowdedness and Solvability/Tractability are estimated on a logarithmic scale (log scale) and then added together. See the 80,000 Hours article about the framework. What we really care about is their product, on the linear scale (the regular scale), but since , the logarithm of the product of the factors is the sum of the log scale factors, so using the sum of log scale factors makes sense:

**1. Using the expected values of the logarithms instead of the logarithms of the expected values, biasing us against high-risk high-reward causes.**

If there’s uncertainty in any of the linear scale factors (or their product), say , then using the expected value of the logarithm *underestimates* the quantity we care about, , because the logarithm is a strictly concave function, so we have (the reverse) Jensen’s inequality:

When there is a lot of uncertainty, we may be prone to underestimating the log scale factors. The difference can be significant. Consider a 10% probability of 1 billion and a 90% probability of 1 million; in base 10, that’s 14.5 vs 18.4. 80,000 Hours uses base , incrementing by 2 for every factor of 10, so this would be closer to 7 vs 9, which is quite significant, since some of the top causes differ by around this much or less (according to 80,000 Hours’ old estimates).

People could also make this mistake without actually doing any explicit expected value calculations. They might think “this cause looks like somewhere between a 3 and a 5 on Tractability, so I’ll use 4 as the average”, while having a symmetric distribution centred at 4 in mind (i.e. the distribution looks the same if you reflect it left and right through 4). This actually corresponds to using a skewed distribution, with more probability given to the lower values, whereas a uniform distribution over the interval would give you 4.7 on a log scale. That being said, I think it does make sense to generally have these distributions more skewed towards lower values on the linear scale and we might otherwise be biased towards more symmetric distributions over the linear scale, so these two biases could work in opposite directions. Furthermore, we might already be biased in favour of high-risk high-reward interventions, since we aren’t sufficiently skeptical and are subject to the optimizer’s curse.

The solution is to always make sure to deal with the uncertainty before taking logarithms, or be aware that a distribution over the log scale corresponds to a distribution that’s relatively more skewed towards lower values in the linear scale.

**2. Upwards bias of log scale factors, if the possibility of negative values isn’t considered.**

Logarithms can be negative, but I’ve never once seen a negative value for any of the log scale factors (except in comments on 80,000 Hours’ problem framework page). If people mistakenly assume that negative values are impossible, this might push up their view of what kind of values are “reasonable”, i.e. “range bias”, or cause us to prematurely filter out causes that would have negative log scale factors.

80,000 Hours does provide concrete examples for the values on the log scale factors, which can prevent this. It’s worth noting that according to their Crowdedness score, $100 billion dollars in annual spending corresponds to a 0, but Health in poor countries gets a 2 on Neglectedness with “The least developed countries plus India spend about $300 billion on health each year (PPP).” So, should this actually be negative? Maybe it shouldn’t, if these resources aren’t generally being spent effectively, or there aren’t a million people working on the problem.

We could also be biased away from further considering causes (or actions) that have negative log scale factors that make up for them with other factors. In particular, some small acts of kindness/decency or helping an individual could be very low in Scale, but when they’re basically free in terms of time or resources spent, like thanking people when they do things for you, they’re still worth doing. However, I do expect that writing an article defending small acts of kindness probably has much less impact than writing one defending one of the typical EA causes, and may even undermine EA if it causes people to focus on small acts of kindness which aren’t basically free. $1 is not free, and is still generally better given to an EA charity, or to make doing EA more sustainable for you or someone else.

Furthermore, when you narrow the scope of a cause area further and further, you expect the Scale to decrease and the Neglectedness to increase. At the level of individual decisions, the Scale could often be negative. At the extreme end,

New cause area: review a s

*pecific*EA article and provide feedback to the author.New cause area: donate

*this*$1 to the Against Malaria Foundation.

**3. Neglecting the possibility that work on a cause could do more harm than good, if the possibility of undefined logarithms isn’t considered.**

In this case, where the argument is negative (or 0), the logarithm is not defined. If people are incorrectly taking expectations after taking logarithms instead of before as in 1, then they should expect undefined values. The fact that we aren’t seeing them could be a sign that they aren’t seriously considering the possibility of net harm. If log scale values are incorrectly assumed to always to be defined, this is similar to the range bias in 2, and could bias our estimates upwards.

On the other hand, if they are correctly taking expectations before logarithms, then since the logarithm is undefined if and only if its argument is negative (or 0), if it would have been undefined, it’s because work on the cause does more harm than good in expectation, and we wouldn’t want to pursue it anyway. So, as in the above, if log scale values are incorrectly assumed to always be defined, then this may also prevent us from considering the possibility that work on a cause does more harm than good in expectation, and could also also bias our estimates upwards.

Note: I’ve rewritten this section since first publishing this post on the EA Forum to consider more possibilities of biases.

**Bonus: Ignoring correlations between factors.**

Summary: When there’s uncertainty in the factors and they correlate positively, we may be underestimating the marginal cost-effectiveness. When there’s uncertainty in the factors and they correlate negatively, we may be overestimating the marginal cost-effectiveness.

What we care about is , but the expected value of the product is not in general equal to the product of the expected values, i.e.

They are equal if the terms are uncorrelated, e.g. independent. If it’s the case that the higher in Scale a problem is, the lower in Solvability it is, i.e. the better it is to solve, the harder it is to make *relative* progress, we might *overestimate* the overall score, since, as an example, letting be a uniform random variable on , and , also uniform over the same interval, but perfectly anticorrelated, then we have

Of course, this is less than a factor of 2 in the linear scale, so less than a difference of 1 on the log scale.

(EDIT: everything below was added later.)

However, the quotient can be *arbitrarily* large (or arbitrarily small), so there’s no constant upper bound on how wrong we could be. For example, let and with probability , and and with probability . Then,

and the quotient goes to as goes to , since the left-hand side goes to and the right-hand side goes to .

On the other hand, if the distributions are more concentrated over the same interval, the gap is lower. Comparing to the first example, let have probability density function over and again . Then, we have:

With anticorrelation, I think this would bias us *towards* higher-risk higher-reward causes.

On the other hand, positive correlations lead to the opposite bias and underestimation. If , identically, and uniform over , we have:

The sign of the difference between and is the same as the sign of the correlation, since we have for the covariance between and ,

and the correlation is

where .

So, as a general trend, the more uncertainty in the factors, the greater the possibility for larger biases.

**Possible examples.**

One example could be that we don’t know the exact Scale of wild animal suffering, in part because we aren’t sure which animals are actually sentient, and if it does turn out that many more animals are sentient than expected, that might mean that relative progress on the problem is harder. It could actually turn out to be the opposite, though; if we think we could get more cost-effective methods to address wild invertebrate suffering than for for wild vertebrate suffering (invertebrates are generally believed to be less (likely to be) sentient than vertebrates, with a few exceptions), then the Scale and Solvability might be *positively* correlated.

Similarly, there could be a relationship between the Scale of a global catastrophic risk or x-risk and its Solvability. If advanced AI can cause value lock-in, how long the effects last might be related to how difficult it is to make relative progress on aligning AI, and more generally, how powerful AI will be is probably related to both the Scale and Solvability of the problem. How bad climate change or a nuclear war could be might be related to its Solvability, too, if worse risks are relatively more or less difficult to make progress on.

**Some independence?**

It might sometimes be possible to define a cause, its scope and the three factors in such a way that and are independent (and uncorrelated), or at least independent (or uncorrelated) given . For example, in the case of wild animal suffering, we should include all funding towards invertebrate welfare towards even if it turns out to be the case that invertebrates aren’t sentient. However, is defined in terms of the other two factors, so should not in general be expected to be independent from or uncorrelated with either. The independence of and (given ) and the law of iterated expectations allow us to write

However, should, in theory, consider future work, too, and the greater in or a problem is, the more future resources we might expect to go to it in the future, so might actually be *anticorrelated* with the other two factors.

This is something to keep in mind for marginal cost-effectiveness analysis, too.

**Bounding the error.**

Without actually calculating the covariance and thinking about how the factors may depend on one another, we have the following bound on the difference in terms of of the variances (by the Cauchy-Schwarz inequality, which is also why correlations are bounded between −1 and 1):

and this bounds the bias in our estimate. We can further divide each side by to get

and so

which bounds our bias as a ratio without considering any dependence between and . Taking logarithms bounds the difference .

Other bounds can be derived using the maximum values (or essential suprema) and minimum values (or essential infima) of and , or Hölder’s inequality, which generalizes the Cauchy-Schwarz inequality. For example, assuming is always positive (or nonnegative), because , we have

and so, dividing by

The inequality is reversed if .

Similarly, if and one of the other two factors are always positive (or nonnegative, and is always nonnegative), then

and so, dividing by ,

And again, the inequality is reversed if .

Here, we can take to be any of , or , and and to be the other two. So, if at most one factor ranges over multiple orders of magnitude, then taking the product of the expected values should be within a few order of magnitudes of the expected value of the product, since we can use the above bounds with as the factor with the widest log scale range.

With some care, I think similar bounds can be derived when is not always positive (or nonnegative), i.e. can be negative.

Finally, if one of the three factors, say , is independent of the other two (or the product of the other two), then we have , and it suffices to bound , since

The end of this post will be beyond my math til next year, so I’m glad you wrote it :) Have you given thought to the pre-existing critiques of the ITN framework? I’ll link to my review of them later.

In general, ITN should be used as a rough, non-mathematical heuristic. I’m not sure the theory of cause prioritization is developed enough to permit so much mathematical refinement.

In fact, I fear that it gives a sheen of precision to what is truly a rough-hewn communication device. Can you give an example of how an EA organization presently using ITN could improve their analysis by implementing some of the changes and considerations you’re pointing out?

Thanks!

I think I’ve looked at a few briefly.

I think the framework is mostly fine theoretically, based on the formal definitions of the (linear scale) terms and their product as a cost-effectiveness estimate. I’d imagine the concerns are more with the actual interpretation and application. For example, Neglectedness is calculated based on current resouces, not also projected resources, so an issue might not be really neglected, because you expect resources to increase in the future without your intervention. This accounts partially for “urgency”.

I think EA orgs are mostly not making the mistakes I describe in each of sections 1, 2 and 3, but the solution is pretty straightforward: consider the possibility of negative outcomes, and take expectations before taking logarithms.

For the Bonus section, my suggestion would be to give (independent) distributions for each of the factors, and check the bounds I describe in “Bounding the error.” to check how sensitive the analysis could be to dependencies, and if it’s not sensitive enough to change your priorities, then proceed as usual. If you find that it could be sensitive enough and you think there may be dependencies, model dependencies in the distributions and actually calculate/estimate the expected value of the product (using Guesstimate, for example, but keeping in mind that extremely unlikely outcomes might not get sampled, so if most of the value is in those, your estimate will usually be way off).

Or you can just rely on cost-effectiveness analyses for specific interventions when they’re available, but they aren’t always available.

Here is that review I mentioned. I’ll try and add this post to that summary when I get a chance, though I can’t do justice to all the mathematical details.

If you do give it a glance, I’d be curious to hear your thoughts on the critiques regarding the shape and size of the marginal returns graph. It’s these concerns that I found most compelling as fundamental critiques of using ITN as more than a rough first-pass heuristic.

I’ve added a summary at the start of the Bonus section you could use:

(And this is because we’re taking the product of the expected values rather than the expected value of the product and not explicitly modelling correlations between terms.)

Is this about Neglectedness assuming diminishing marginal returns? I think if you try to model the factors as they’re defined formally and mathematically by 80,000 Hours, Tractability can capture effects in the opposite direction and, e.g. increasing marginal returns. At any rate, if the terms, as defined mathematically, are modelled properly (and assuming no division by zero issues), then when we take the product, we get Good done / extra resources, and there’s nothing there that implies an assumption of diminishing or increasing marginal returns, so if Neglectedness assumes diminishing marginal returns, then the other factors assume increasing marginal returns to compensate.

How many extra resources we consider in Neglectedness could be important, though, and it could be the case that Good done / extra resources is higher or lower depending on the size of “extra resources”. I think this is where we would see diminishing or increasing marginal returns, but no assumption either way.

This post was awarded an EA Forum Prize; see the prize announcement for more details.

My notes on what I liked about the post, from the announcement:

Thank you for giving this some thought. The impression I have from this is that the concerns you raise relate to either spurious levels of accuracy or edge cases.

But maybe my impression is wrong. If you could provide some specific examples where either 80k or another org has made a wrong decision because of these considerations, I would find this piece much more interesting.

I’ve also edited the Bonus section. As a rule, we aren’t accounting for correlations between the factors, and there’s no general upper bound on how much this could bias our estimates.

I’ve also thought of a few possible examples for the Bonus section here, and added them to the post.

I wrote above that 80,000 Hours should perhaps have given “Health in poor countries” a negative rating for Neglectedness. No other examples come to mind.

I’m more worried about cause area analyses written by individuals (mostly) on their own than the work of EA orgs, which typically goes through more review and is done more formally.

Re bonus section: Note that we are (hopefully) taking expectations over our estimates for importance, neglectedness and tractability, such that general correlations between the factors between causes do not necessarily cause a problem. However, it seems quite plausible that our estimation errors are often correlated because of things like the halo effect.

Edit: I do not fully endorse this comment any more, but still belief that the way we model the estimation procedure matters here. Will edit again, once I am less confused.

Maybe one example could be that we don’t know the exact Scale of wild animal suffering, in part because we aren’t sure which animals are actually sentient, and if it does turn out that many more animals are sentient than expected, that might mean that relative progress on the problem is harder. It could actually be the opposite, though; if we think we could get more cost-effective methods to address wild invertebrate suffering than for for wild vertebrate suffering (invertebrates are generally believed to be less (likely to be) sentient than vertebrates, with a few exceptions), then the Scale and Solvability might be

positivelycorrelated.Similarly, there could be a relationship between the Scale of a global catastrophic risk or x-risk and its Solvability. If advanced AI can cause value lock-in, how long the effects last might be related to how difficult it is to make relative progress on aligning AI, and more generally, how powerful AI will be is probably related to both the Scale and Solvability of the problem. How bad climate change or a nuclear war could be might be related to its Solvability, too, if worse risks are relatively more or less difficult to make progress on.

Note: I’ve rewritten section 3 since first publishing this post on the EA Forum to consider more possibilities of biases.