In one of my comments above, I say this:
I will caveat this by saying that in my opinion it makes sense for estimation purposes to discount or shrink estimates of highly uncertainty quantities, which I think many advocates of AI as a cause fail to do and can be fairly criticized for. But the issue is a quantitative one, and so can come out either way. I think there is a difference between saying that we should heavily shrink estimates related to AI due to their uncertainty and lower quality evidence, vs saying that they lack any evidence whatsoever.
I feel like my position is consistent with what you have said, I just view this as part of the estimation process. When I say “E[benefits(A)] > E[benefits(B)]” I am assuming these are your best all-inclusive estimates including regularization/discounting/shrinking of highly variable quantities. In fact I think its also fine to use things other than expected value or in general use approaches that are more robust to outliers/high-variance causes. As I say in the above quote, I also think it is a completely reasonable criticism of AI risk advocates that they fail to do this reasonably often.
If you properly account for uncertainty, you should pick the certain cause over the uncertain one even if a naive EV calculation says otherwise
This is sometimes correct, but the math could come out that the highly uncertain cause area is preferable after adjustment. Do you agree with this? That’s really the only point I’m trying to make!
I don’t think the difference here comes down to one side which is scientific and rigorous and loves truth against another that is bias and shoddy and just wants to sneak there policies through in an underhanded manner with no consideration for evidence or science. Analyzing these things is messy, and different people interpret evidence in different ways or weigh different factors differently. To me this is normal and expected.
I’d be very interested to read your explainer, it sounds like it addresses a valid concern with arguments for AI risk that I also share.
Good post. I have two general themes I’d like to comment on:
Analogies for cause prioritization
Your analysis covers several perspectives on this phenomenon, if we focus on the “actual performance” perspective, this is pretty similar to multi-armed bandits. One pattern that I think is present in strategies for these types of problems is the idea of spreading out actions across the different possibilities (explore vs exploit and all that). It wouldn’t necessarily make sense to commit to one “arm” (or cause) early on when information is low. This “spreading out” across options is one way of dealing with uncertainty.
A similar idea comes up in another potential anology for cause prioritization, financial investing. We can think about optimizing a portfolio and its allocation to achieve good returns relative to risk, rather than trying to pick the single highest return asset. Thus we get concepts like disversification.
I find this stock-picking analogy helpful for thinking about how “neglectedness” is often treated in practice. I’ve often found myself skeptical of arguments for and from neglectedness, and I feel the way it is applied in practice doesn’t really align with the classic “diminishing returns” conception. I think the way neglectedness is treated in practice ends up being more like how an investor with a high risk tolerance might view a risky asset. Riskier assets are expected to have higher returns, investors with lower risk tolerance would staturate low-risk/high-return options quickly, leaving risker investments “neglected”. Thus an investor with high risk tolerance can find good opportunities that would be unappealing to other less risk tolerant investors by going to higher risk assets. I think this captures the spirit of what “neglected” cause areas have often looked like in EA, more speculative but where some EAs have a strong feeling that they caould have outsized impact.
If I can read between the lines a bit, under this anology EA pivoting more into AI is kind of like an investor who wants higher returns putting more of their portfolio in small cap growth stocks that are risker but which the investor thinks will result in higher return. One downside of this is decreased diversification. Another possible option would be to hold a more diversified portfolio but use leverage.
In-model vs Out-of-model robustness
I think this gets at a distinction that is worth calling out, in-model vs out-of-model robustness.
In my experience with cost-benefit analysis, both reading EA related ones and in industry, it is fairly common to propose a “median” scenario and also a “pessimistic” scenario, and provide estimates for these cases. The point is usually that since even the pessimistic scenario looks good, the analysis shows that the proposed intervention is robustly beneficial. This has a two-fold problem:
First, usually the reason to think that the “pessimistic” scenario is ’pessimistic is just that it uses parameter values that reduce the estimated benefit below the “median” scenario. It’s unclear sometimes why that means the estimate is robustly lower than the actual benefit. This is the in-model robustness.
Despite the fact that I think this is an issue, sometimes it may be perceived as (or actually be) a somewhat unfair critique. All models are wrong, we have to use what we have to make estimates. This can result in polarized views of what an estimate shows. For a person who likes the intervention and has a gut feeling it is good, the “median” estimate makes a ton of sense and this seems like a very reasonable approach. For a skeptic, it seems prone to over-estimation for the reasons you highlight in the post. Moving the parameters so that your estimate is 25% lower doesn’t turn garbage into non-garbage.
However, there is another source of error lurking in the background. What about costs that you haven’t included? The potential for the intervention to backfire that isn’t considered in any scenario? The hidden assumption that hasn’t been tested in the “pessimistic” scenario? This is out-of-model robustness.
I think the polarization when it comes to in-model robustness causes proponents or fans of an idea or intervention to over-estimate robustness even when in-model robustness is high, because they implicitly credit the (perceived) in-model robustness to the out-of-model robustness.
In my view, the whole “rule high stakes in, not out” idea in practice will result in systematically doing this a lot, which I think makes it a bad heuristic for approaching these types of situations. One way to think about this is it encourages us to focus on specific high-volatility “assets” and thus lacks diversification.