CE: Rigorously prioritizing the top health security (biosecurity) ideas
Every year at Charity Entrepreneurship (CE), we try to find the best interventions for impact-focused charities to launch through our Incubation Program. Our research in 2022 focused on health security (which includes biosecurity) and on large-scale global health.
This post outlines:
An introduction to health security at CE. What CE means by health security exactly, some of CE’s views on preventing catastrophic and existential risks, and why we picked “health security” as a topic.
Our prioritization of the top health security interventions. Why rigorous prioritization of ideas is needed to be effective in the world (especially for longtermist cause areas), and CE’s approach to this within health security.
Our results. The long list of nearly 200 ideas, our top six shortlisted ideas, and our final recommendations.
If you’d like to learn how to conduct similar prioritization research, and how to evaluate the results that come out of starting high-impact interventions, we have just launched our Research Training Program that you can apply to by July 17, 2023.
Introduction to health security at CE
What we mean by health security
We took a fairly broad interpretation of health security. Our initial definition was: anything that minimizes the danger of acute public health events.
This included several broad categories:
Pandemic prevention
Pandemic preparedness
General risk preparedness for acute public health events
Antimicrobial resistance
Health system strengthening
Other / meta ideas (e.g., better biosecurity forecasting)
This included looking at both natural pandemics and engineered pandemics (which could be due to accidental or deliberate misuse of biotech).
How much does CE value preventing catastrophic/existential risks?
This may seem like a bit of a cop-out answer, but there are a wide range of views on this topic among CE staff. There is no clear overarching organizational view or policy. This section gives a rough sense of how “CE staff” might tend to think about these topics, but probably doesn’t accurately reflect any individual staff member’s stance.
In general, CE’s staff tend to apply a very high level of empirical rigor to decision making. Staff tend to trust empirical evidence (e.g., RCTs, M&E of existing charities, base-rate forecasting, etc.) above other kinds of evidence, particularly valuing such evidence above theoretical reasoned arguments. That said, staff tend to accept that making good decisions requires robust/cluster thinking and look for cases where many kinds of evidence align, including theoretical reasoning. Along the same lines, staff are likely to think that doing good is really difficult and having some ongoing measurable evidence of impact is probably required.
In general, CE staff believe that preventing extinction is a worthwhile endeavor. However, given the above, staff are likely to be skeptical about:
The ability to know what the biggest future risks are, especially where risk estimates rely on reasoned speculation about future technologies.
The success of any organization that doesn’t have a clear path to measure the impact of its activities.
We took all these views into consideration and chose to focus on health security, including biorisks. This focus allows us (and future CE charities) to explore risk areas that have at least some chance of being globally catastrophic, but also have clear historical precedents and evidence of less extreme catastrophes. Additionally, we expected to find a number of options within biorisk that could be impactful in addressing catastrophic biorisks while also demonstrating health impacts in the short run, meaning a new charity should be able to track and demonstrate impact.
In short, we were hoping to find interventions in this sweet spot:
Aiming for this sweet spot led to a bit of a mixed approach. For example:
In our quantitative modeling of the cost-effectiveness of different interventions, we considered lives saved over the next 50 years. This is not to say that we did not care about lives post-50 years, just that this 50-year metric appeared to be the best proxy for our decision making. Given that pandemic risks are likely to be heavy-tailed, we should expect a reasonable correlation between interventions that save many lives and interventions that prevent extreme risks.
In our theoretical work mapping out the various theories of change, we mostly focused on understanding the plausible theories of change relevant to avoiding globally catastrophic biological risks. So if an intervention looked to be robustly good on multiple of these theories of change, we were more excited by it.
In practice, we probably aimed a bit to the left of this ‘sweet spot’, tending to favor more near-termist interventions. In the end (as discussed below), we somewhat missed this sweet spot in our final recommendations, one of which fell off to the right and one off to the left.
Going forward, I expect it is unlikely that CE will focus on another cause area that crosses into existential risk prevention in the near future. I think we will want to see how these charities fare first.
Prioritizing the top health security interventions
Prioritization matters
There are many ways to do good in the world. Some of those paths to impact are likely to be much better than others. Ben Todd explores this in this excellent post. He concludes that “the best of all interventions in an area are about 10 times more effective than the mean, and perhaps as much as 100 times” and “using data to identify the best interventions can perhaps boost your impact in the area by 3–10 times compared to picking randomly”.
CE expects successful charities to run for many years. We believe that putting in the time upfront to choose good ideas can significantly magnify their impact, perhaps by an order of magnitude. When CE considers a new cause area, we typically list 200-400 ideas and carry out a six-month research process to pick the very best idea we can from that list.
To date, a number of lists have been made of high-impact biosecurity interventions (for example, here and here), but it is very unclear how these have been prioritized. Across the nonprofit sector, thousands of people work on health security and thousands more on preventing catastrophic risks more generally. Yet as far as I know (please correct me if I am wrong), the research outlined in this post represents the first time anyone has systematically carried out and published an exercise prioritizing the most important interventions and technologies for preventing health security events, biorisks, or any existential risk.
I, therefore, want to showcase it a bit as an example of what could be done, and discuss some of the lessons learned.
How we prioritized in a data-poor environment
The typical CE research round begins by compiling a list of hundreds of ideas. We then gradually narrow the list down through a series of research rounds, each of which examine the best options from the previous round in greater depth.
Health security was a hard area to research. On the whole, it is a high-uncertainty and data-poor cause area – much more so than global health, and somewhat more so than animal welfare (the two areas that CE researches the most).
Here are three things we did to mitigate this:
As with all our research, we looked to address uncertainty by using a robust/cluster thinking approach. This means we looked for places where different sources of evidence all point in the same direction. Although concrete data on the impact of different interventions was poor, there were many sources of evidence that we could, and did, pull upon. These included: some global health data, data about past pandemics, CE’s institutional knowledge about what makes a good charity, collated expert opinions, case studies of similar projects, theoretical reasoning, cost-effectiveness analyses, forecasts, and so on.
We worked a bit more closely with more experts. We sought the opinion of experts, both from inside and outside of the effective altruism community, and from different sectors (government, multilateral organizations, NGO, and research). We asked for inputs at more stages of our research than in past research rounds, and spoke to more experts overall. We are extremely grateful for the views and input of everyone who helped with our research.
We did more theoretical work. CE’s standard research process is very bottom-up, listing lots of ideas and narrowing on the best ones. We did a bit more top-down work to supplement this, where we broadly mapped out the health security space (or various sub-parts of this space) and different theories of change in the space, and looked for the gaps and ideas that cut across multiple theories of change. Some examples of this work can be found on the slides at: 2022 Health Security _ Addressing pandemics, theories of change.
I think prioritizing in a data-poor environment is achievable
In the aforementioned post on prioritization, Ben Todd suggests that the benefits of prioritization likely require being “in an at least somewhat data-rich area”. So, did we find this to be the case?
I came away from this research with a sense that we managed to prioritize fairly successfully, despite the data scarcity. Overall, I am a little more confident in our prioritization of different intervention types (where CE has institutional expertise) than our prioritization of key pandemic-addressing technologies. Of course, my sense of success is not strong evidence of the value of prioritization, and certainly not enough for me to be able to put a precise estimate on the expected increase in value due to having carried out a prioritization exercise. However, it does make me optimistic about the value of other prioritization exercises.
The results of our research
Our long list
You can see our long lists at: 2022 Bio _ Health Security _ Summary of ideas and research progress tracker
There is a list of about 150 intervention ideas on the ‘IDEA LIST’ tab. There is also a list of about 30 ideas for new technologies (broken up further into direct research interventions and advocacy interventions for various technologies) on the ‘TECH IDEAS’ tab.
Each idea was evaluated with a score, indicating the likelihood of recommendation. It is important to exercise caution when interpreting these scores, as there may be some artifacts that affect their accuracy. For instance, some ideas may have lower scores than warranted because we already know someone is working on starting that idea.
A number of columns of notes and details have been redacted from this spreadsheet, or were carried out on other documents. If you need access to more of our research, do get in touch.
Some ideas on this list were contributed by people and organizations outside of Charity Entrepreneurship. We have not credited every idea, but are highly grateful to those who added ideas and supported our research at all stages.
Our shortlist and the final ideas
You can see our tabulated summary of our top six ideas here:
2022 Health Security _ Decision-making spreadsheet _ Version to share / make public
The final ideas we looked into and recommended for launching:
Policy advocacy for subscription models to drive the production of new antimicrobials
Advocating for academic guidelines to limit dual-use research of concern
An idea we were very excited about, but were not sure that CE is best positioned to get it off the ground, was:
3. R&D for high-priority biotech: PoC tests (and PPE)
Other ideas that came high up on our prioritisation exercise and for which we wrote deep dive reports but ultimately did not recommend for launching were:
4. Advocacy for investment for high-priority biosecurity technologies
5. Policy advocacy to reduce deforestation to prevent zoonotic pandemics
6. Modeling the drivers of antimicrobial resistance
Additionally, there were areas of research which emerged from this list that straddled the border between health system strengthening for health security purposes and general global health interventions. The one that lead to a full research report was:
7. Water chlorination in LMICs
We have already written about our recommended ideas here: CE: Announcing our 2023 Charity Ideas. Apply now! - EA Forum.
We will be publishing more posts, including information about our other ideas, in the coming weeks.
Missing the ‘sweet spot’
When choosing charity ideas, we want to reduce the number of failure modes. This means that although we may pick a radical area for a charity to work in (e.g., shrimp welfare), or a radical approach to driving change (e.g., mass media), we are unlikely to pick both.
As discussed above, we aimed for interventions that hit a ‘sweet spot’ between global catastrophic risk prevention and more indisputable health impacts.
Ultimately, we ended up picking top charity ideas that fell a little outside this sweet spot in both directions. One of our top ideas was justified more by its effect on preventing globally catastrophic biorisks (radical cause area), although hopefully in a way that ongoing impact measurement and feedback is possible (less radical approach) and the other idea is justified more by the effect it has on near-term health outcomes (less radical area), although with a fairly long theory of change requiring global policy changes (radical approach).
Thanks
We wish to express our deep gratitude to the research team for their remarkable effort and dedication in conducting this research. In particular, we would like to thank Akhil Bansal, Emma Buckland, Lin Bowker-Lonnecker, Melanie Basnak and Moritz Von Knebel – all of whom helped research this cause area.
We are also grateful to the many others who have contributed to this research in various ways, including the experts who provided their expertise and guidance in specific areas. Each of these individuals is recognized and thanked in the report specifically related to their contributions. This research would not have been possible without the generous support of all of these individuals, and we are deeply appreciative of their contributions.
- 11 Jan 2024 7:11 UTC; 2 points) 's comment on Exaggerating the risks (Part 13: Ord on Biorisk) by (
Thanks for doing this, I appreciate the transparency in the calculations and write-ups. I have a few comments.
One possible considerations in biorisk is Kevin Esvelt’s argument about declining costs of gene synthesis and the potential democratisation of bio WMD. I think these risks are extremely troubling in the next 10-20 years, and much larger than other biorisks, but I don’t see much discussion of this argument in your post, and that might affect your conclusions.
I’m a bit surprised to see deforestation in there as a top priority. Have any pandemics in the past been caused by deforestation? this is a genuine question, I don’t know the answer. If I were focused on zoonoses, I think I would start with wild animal markets and factory farms, not deforestation, as this seems to have been what drove eg SARS and MERS. I notice that you assume that it costs $17 to avert a tonne of CO2 through forestry protection. After looking into this for some time, Johannes Ackva and I concluded that these numbers are too high probably by 1-2 orders of magnitude, and that lot of forestry protection probably has no effect. see the climate change page of the giving what we can website.
Hi John.
Thank you for the feedback and comments.
On deforestation. Just to be clear the result of our prioritisation exercise was our top recommendations (ideas 1-2) on subscription models for new antibiotics and stopping dangerous dual use research. The ideas 4-7 (including the deforestation one) did well in our early prioritisation but ultimately we did not recommend them. I have made a minor edit to the post to try to make this clearer.
The stopping deforestation report idea was originally focused on limiting the human animal interface to prevent zoonotic pandemics (which did well in our prioritisation). Then in the report we prioritise between the ways one might go about stopping zoonoses. The summary is:
Unclear for animals
Positive for climate
High
Unfortunately the full report is not quite ready for publication. Hopefully the full report will be available soon.
Or $ per tCO2 were from two sources:
US$ 0.87-2.60/tCO2, from https://www.cifor.org/knowledge/publication/6188/
Average from REDD+ report at: https://forum.effectivealtruism.org/posts/LmjsyFvQavLQnYSTp/the-redd-framework-for-reducing-deforestation-and-mitigating#fnrefoyfcvbmm7tl
Sorry that we missed your estimate.
We didn’t look into gene synthesis risks so might have missed something there. Although potentially a charity working on reducing dual use research could play a role in limiting these risks.
Yes sorry that’s me not reading properly.
Thanks for this! I appreciate the transparency, and all the work that has gone into it from so many people! I was curious about one part of the AMR pull funding CEA (cells C22 and C30), which says: “Based on limited data, it seems like advocacy can increase chance success by 10-15%”
I was wondering what data that was based on? In the sheet the 10-15% is used as percentage points, such that the chance of the PASTEUR Act passing goes from 7.5% to 20% (alternate estimate 10% to 17.5%). The implied 2.6X (alternate 1.75X) multiplier as a result of advocacy seems quite high, but I’m unsure. My intuition’s just based on the size and complexity of the U.S. legislative machinery, but for professional lobbying this 2020 post suggests a central estimate of 2.5% change (95%CI 0% to 5%), and this LSE study suggested an average impact of 0.05 percentage points. I’d weakly expect a new non-profit to be less effective than the average of professional lobbyists.
On its own it doesn’t look like this concern changes the bottom line though—even with just a 0.03 percentage point impact, it is still $35/DALY!
Hi Ben, happy to justify this. I was responsible for the alternate estimate of 10%-17.5%
– –
These numbers here are consistent with our other estimates of policy change. Other estimates were easier (but not easy) to justify as they were in better evidenced areas and tended to range between 5% and 40% (see 2022 ideas here). Areas with the best data were road safety policy, where we looked at looked at 84 case studies finding a 48% chance of policy success, and food fortification policy, where we looked at 62 case studies (in Annex) with a 47% chance of success – we scaled these numbers downwards but they are not low.
Other EA estimates are also in this ballpark – e.g. Nuno’s estimates on OpenPhil criminal justice reform are in the same ball park and give 7-50% and 1-10% chance of policy change success. Also personal experience suggests reasonably high numbers too – CE has one long-running policy charity (LEEP) and it seems to have been pretty successful, driving policy change within 6 months. My own experience was also fairly successful. I think this allows us to put relatively high priors on policy change of 5 to 20 percentage points increase.
That said my 7.5 percentage points increase here was guesswork. I wish I had more time to look into it but it was driven by intuitions based of case studies and experience, that is mostly not US based. I would be open to hearing from experienced US policy advocates that it is too high (or too low).
– –
On the articles you link to:
The informational lobbying post estimates that lobbying raises the chance of policy change from “very low” to 2.5%. I think this is consistent with a raise from 10% to 17.5%.
My view on policy (based on working in the field) is that the most impact per $ comes when advocating for a policy issue that might just happen anyway – e.g. advocating for something where the political door is ajar and it just needs a nudge. I think moving the needle from 15%-40% is about as easy as moving the needle from 1%-5%.
I am unfamiliar with the LSE study and will have to have a look at it. Maybe it will lead me to be more pessimistic.
– –
Note 1: I work for CE. Note 2: I think policy change in the animals welfare space is a bit different so assume I am not talking about animal welfare work in any of the above.
That all makes sense and seems thorough, thanks!
Just a note here as the author of that lobbying post you cite: the CEA including the 2.5% change in chance of success is intended to be illustrative — well, conservative, but it’s based on nothing more than a rough sense of effect magnitude from having read all those studies for the lit review. The specific figures included in the CEA are very rough. As Stephen Clare pointed out in the comments, it’s also probably not realistic to have modeled that is normal on the [0,5] 95% CI.
Thanks for the clarification!
On the DURC CEA:
Is there any further calculation for B10 - the ‘Expected lives lost due to DURC per year over next 50 years’? I expect this is where a lot of the juice is, and would be a central component for the charity’s advocacy work, but I don’t see the working (maybe it’s in the idea report soon to be published?). The cell formula suggests maybe the assumption is one pandemic killing 80M in the 50 year period (yielding the 1.6M/yr estimate). Is this right?
It looks like ‘leakage %’ is the risk that a dangerous agent within DURC ends up causing harm. For leakage via accidents, were rates of LAIs or other models used to calibrate the estimate? Or just based on deforestation + priors, as it appears?
Was malicious use considered, or is it assumed within leakage estimates? I think this is a big component of the risk from DURC. The furore over horse-pox synthesis was less about the chance that horse-pox would escape Evans’ lab—it was that it showed the possibility and the technique that actors could use to access smallpox. Similarly, the worry over the mouse-pox IL-4 experiment was less that the engineered mouse-pox would escape into the wild, and more that they demonstrated a considerable step actors could take towards inducing vaccine resistance in poxviruses.
Given CE is already recommending this idea, it’s probably moot whether misuse was directly modelled or not (though it will come up for the eventual charity!). If it wasn’t, I now wonder about the bio ideas CE didn’t recommend, and whether modelling misuse would change those bottom-lines.
Hi Ben! With the benefit of hindsight, I realise we could’ve been more clear on what “leakage” means in this context, given that the topic matter might suggests we are talking about lab leaks. We’re not! In our model, lab leak rates would only factor into our estimate of how many deaths will be caused by DURC in the future. Leakage in the CEA refers to the risk that new guidelines in academic communities might be “leaky” in that researchers might choose to migrate to other jurisdictions, countries or privately owned labs (though few of these exist on the BSL levels we are most concerned with) or worse yet, move their research underground. Hence, our CEA discounts the estimate of how many lives could be saved by including this possibility.
Ah okay, thanks for the correction! In which case I think ~all my questions apply to the B10 figure then.
Another question on the AMR CEA (will look at the DURC one eventually!), was the likelihood of antibiotics being developed without the PASTEUR Act modelled as part of the counterfactual? I don’t think I understand the full sheet yet, but it looks like the benefits of the PASTEUR Act are based on a globally distributed supply of new antibiotics for 50 years, and are compared to a world where no new antibiotics are ever developed and AMR continues as projected. Is that right?
I would have to check this with Akhil, the lead author, but my understanding is that this CEA compares a case where the PASTEUR act passes with a business as usual case where very few (but not zero) new anitbiotics are developed.
I agree this is probably overly-optimistic as we can and probably should assume that someone is likely to do something about antibiotic resistance in the next few decades. Good spot!
And thank you for the great questions and for looking over things in such detail.
I can’t find these posts on the forum (I checked the pos history of both of this post’s authors). Could you please point me towards them?
Thanks for sharing!
Would Rethink Priorities’ 20 concrete projects for reducing existential risk count?