When modeling with uncertainty we often care about the expected value of our result. In CEAs, in particular, we often try to estimate E[effectcost]. This is different from both E[costeffect]−1 and E[effect]E[cost] (which are also different from each other). [1] The goal of this post is to make this clear.
One way to simplify this is to assume that the cost is constant. So we only have uncertainty about the effect. We will also assume at first that the effect can only be one of two values, say either 1 QALY or 10 QALYs with equal probability.
Expected Value is defined as the weighted average of all possible values, where the weights are the probabilities associated with these values. In math notation, for a random variable X,
E[X]=∑xP(X=x)⋅x,
where x are all of the possible values of X.[2] For non-discrete distributions, like a normal distribution, we’ll change the sum with an integral.
Coming back to the example above, we seek the expected value of effect over cost. As the cost is constant, say C dollars, we only have two possible values:
which is not1E[effectcost]=C211$QALY, a smaller amount.
The point is that generally 1E[X]≠E[1X]. In fact, we always have 1E[X]≤E[1X] with equality if and only if X is constant.[3]
Another common and useful example is when X is lognormally distributed with parameters μ,σ2. That means, by definition, that lnX is normally distributed with expected value and variance μ,σ2 respectively. The expected value of X itself is a slightly more complicated expression:
E[X]=eμ+σ22.
Now the fun part: 1X is also lognormally distributed! That’s because ln1X=−lnX. Its parameters are −μ,σ2 (why?) and so we get
E[1X]=e−μ+σ22≠e−μ−σ22=1E[X]
In fact, we see that the ratio between these values is
One way to imagine this is that if we sample X many times we will observe each possible value x roughly P(X=x) of the times. So the expected value would indeed generally be approximately the average value of many independent samples.
1/E(X) is not E(1/X)
When modeling with uncertainty we often care about the expected value of our result. In CEAs, in particular, we often try to estimate E[effectcost]. This is different from both E[costeffect]−1 and E[effect]E[cost] (which are also different from each other). [1] The goal of this post is to make this clear.
One way to simplify this is to assume that the cost is constant. So we only have uncertainty about the effect. We will also assume at first that the effect can only be one of two values, say either 1 QALY or 10 QALYs with equal probability.
Expected Value is defined as the weighted average of all possible values, where the weights are the probabilities associated with these values. In math notation, for a random variable X,
E[X]=∑xP(X=x)⋅x,where x are all of the possible values of X.[2] For non-discrete distributions, like a normal distribution, we’ll change the sum with an integral.
Coming back to the example above, we seek the expected value of effect over cost. As the cost is constant, say C dollars, we only have two possible values:
E[effectcost]=∑e,cP(effect=e,cost=c)ec==121 QALYC$+1210QALYC$==121 QALY+1210 QALYC$==E[effect]C$=5.5CQALY$.In this case we do have E[effectcost]=E[effect]E[cost], but as we’ll soon see that’s only because the cost is constant. What about E[costeffect]?
E[costeffect]=∑e,cP(effect=e,cost=c)ce==12C$1 QALY+12C$10QALY==C$(1211 QALY+12110QALY)==C$E[1effect]=C1120$QALY,which is not 1E[effectcost]=C211$QALY, a smaller amount.
The point is that generally 1E[X]≠E[1X]. In fact, we always have 1E[X]≤E[1X] with equality if and only if X is constant.[3]
Another common and useful example is when X is lognormally distributed with parameters μ,σ2. That means, by definition, that lnX is normally distributed with expected value and variance μ,σ2 respectively. The expected value of X itself is a slightly more complicated expression:
E[X]=eμ+σ22.Now the fun part: 1X is also lognormally distributed! That’s because ln1X=−lnX. Its parameters are −μ,σ2 (why?) and so we get
E[1X]=e−μ+σ22≠e−μ−σ22=1E[X]In fact, we see that the ratio between these values is
E[1X]1E[X]=eσ2.See Probability distributions of Cost-Effectiveness can be misleading for relevant discussion. There are arguably reasons to care about the two alternatives E[costeffect]−1 or E[effect]E[cost] rather than E[effectcost], which are left for a future post.
One way to imagine this is that if we sample X many times we will observe each possible value x roughly P(X=x) of the times. So the expected value would indeed generally be approximately the average value of many independent samples.
Due to Jensen’s Inequality.