Formalizing the cause prioritization framework

When prioritizing causes, what we ultimately care about is how much good we can do per unit of resources. In formal terms, we want to find the causes with the highest marginal utility per dollar, MU/​$ (or, marginal cost-effectiveness). The Importance-Tractability-Neglectedness (ITN) framework has been used as a way of calculating MU/​$ by estimating its component parts. In this post I discuss some issues with the current framework, propose a modified version, and consider a few implications.

80,000 Hours defines ITN as follows:

  • Importance = utility gained /​ % of problem solved

  • Tractability = % of problem solved /​ % increase in resources

  • Neglectedness = % increase in resources /​ extra $

With these definitions, multiplying all three factors gives us utility gained /​ extra $, or MU/​$ (as the middle terms cancel out). However, I will make two small amendments to this setup. First, it seems artificial to have a term for “% increase in resources”, since what we care about is the per-dollar effect of our actions.[1] Hence, we can instead define tractability as “% of problem solved /​ extra $”, and eliminate the third factor from the main definition. So to calculate MU/​$, we simply multiply importance and tractability:

This defines MU/​$ as a function of the amount of resources allocated to a problem, which brings me to my second amendment. Apart from the above definition, 80k defines ‘neglectedness’ informally as the amount of resources allocated to solving a problem. This definition is confusing, because the everyday meaning of ‘neglected’ is “improperly ignored”. To say that a cause is neglected intuitively means that it is ignored relative to its cost-effectiveness. But if neglectedness is supposed to be a proxy for cost-effectiveness, this everyday meaning is circular. And really, how useful is the advice to focus on causes that have been improperly ignored? This should go without saying.

I suggest we instead use “crowdedness” to mean the amount of resources allocated to a problem. This captures intuitions about diminishing returns (other things equal, a more crowded cause is less cost-effective), uses an absolute rather than a relative standard, and avoids the problem of having the technical definition conflict with the everyday meaning.

Thus, our revised framework is now ITC:

  • Importance = utility gained /​ % of problem solved

  • Tractability = % of problem solved /​ extra $

  • Crowdedness = $ allocated to the problem

So how does crowdedness fit into this setup, if it’s not part of the main definition? Intuitively, tractability will be a function of crowdedness: the % of the problem solved per dollar will vary depending on how many resources are already allocated. This is the phenomenon of diminishing marginal returns, where the first dollar spent on a problem is more effective in solving it than is the millionth dollar. Hence, crowdedness tells us where we are on the tractability function.

A graphical approach

Let’s see how this works graphically. First, we start with tractability as a function of dollars (crowdedness), as in Figure 1. With diminishing marginal returns, “% solved/​$” is decreasing in resources.

Next, we multiply tractability by importance to obtain MU/​$ as a function of resources, in Figure 2. Assuming that Importance = “utility gained/​% solved” is a constant[2], all this does is change the units on the y-axis, since we’re multiplying a function by a constant.

Now we can clearly see the amount of good done for an additional dollar, for every level of resources invested. To decide whether we should invest more in a cause, we calculate the current level of resources invested, then evaluate the MU/​$ function at that level of resources. We do this for all causes, and allocate resources to the highest MU/​$ causes, ultimately equalizing MU/​$ across all causes as diminishing returns take effect. (Note the similarity to the utility maximization problem from intermediate microeconomics, where you choose consumption of goods to maximize utility, given their prices and subject to a budget constraint.)

While MU/​$ is sufficient for prioritizing across causes, we can also look at total utility, by integrating the MU/​$ function over resources spent. Figure 3 plots the total utility gained from spending on a problem, as a function of resources spent. Note that the slope is equal to MU/​$, which is decreasing in $.

Implications

(1) All three factors in the ITC framework are necessary to draw a conclusion about which cause is best. Consider this passage from the 80k article:

[M]ass immunisation of children is an extremely effective intervention to improve global health, but it is already being vigorously pursued by governments and several major foundations, including the Gates Foundation. This makes it less likely to be a top opportunity for future donors.

This last sentence is not strictly true. To be precise, all we can say is that other things equal, a cause with more resources has lower MU/​$. That is, for two causes with the same MU/​$ function, the cause with higher resources will be farther along the function, and hence have a lower MU/​$. If other things are not equal, the cause with more resources may have a higher or lower MU/​$. (And generally, if a cause scores low on one of the three factors, it can still have the highest MU/​$, through high scores on one or both of the other two factors.)

(2) With this setup, we can clearly see how MU/​$ depends on context (in particular, resources spent). To make up a hypothetical example, AI risk might have had the highest MU/​$ in 2013, but the funding boost from OpenAI pushed it down the tractability curve to a lower value of MU/​$. Hence, claims about “cause C is the highest priority” should be framed as “cause C is the highest priority, given current funding levels”. We should expect the “best” cause (defined as highest MU/​$) to change over time as spending changes, which we could indicate by using a time subscript, .

(3) This model also incorporates Joey Savoie’s argument about using the limiting factor instead of importance. Here, a limiting factor would show up as strongly diminishing returns in the tractability function at some level of spending. That is, the percent of the problem solved per dollar would drop off sharply after spending some level of resources on the problem.

(4) The systemic change critique argues that the standard cause prioritization framework cannot handle increasing marginal returns. For example, large-scale political reform yields no results until a critical mass is reached and massive change occurs. But in fact this is easily modeled as a tractability function (Fig. 1) that is increasing for some part of its domain. That is, when nearing the critical mass, each additional dollar solves a larger percent of the problem than the previous dollar. While this case requires a different decision rule than “allocate resources to the cause with the highest MU/​$”, it is a straightforward extension of the standard model.

Conclusion

I propose a model of cost-effectiveness using Importance, Tractability, and Crowdedness. Tractability is a function of crowdedness, and multiplying importance and tractability gives us marginal utility per dollar. So is the 80k model wrong? No. I simply find it more intuitive to think about tractability as “% of problem solved /​ extra $” instead of “% of problem solved /​ % increase in resources”, and this is the resulting model.


Notes

[1] Also, the Neglectedness term “% increase in resources /​ extra $” is always equal to (1/​resources)%, which seems a bit redundant. That is, given resources, an extra dollar always increases your resources by . Eg, given $100, an extra dollar increases your resources by 1%.

[2] This seems to be a definitional issue: we can define importance as a constant, so that “utility gained /​ % of problem solved” is a constant function of “% of problem solved”. That is, solving 1% of the problem just means gaining 1% of the total utility from solving the entire problem.