Thoughts on Toby Ordās policy & research recommendations
In Appendix F of The Precipice, Ord provides a list of policy and research recommendations related to existential risk (reproduced here). This post contains lightly edited versions of some quick, tentative thoughts I wrote regarding those recommendations in April 2020 (but which I didnāt post at the time).
Overall, I very much like Ordās list, and I donāt think any of his recommendations seem bad to me. So most of my commentary is on things I feel are arguably missing.
Regarding āother anthropogenic risksā
Ordās list includes no recommendations specifically related to any of what he calls āother anthropogenic risksā, meaning:
ādystopian scenariosā
nanotechnology
āback contaminationā from microbes from planets we explore
aliens
āour most radical scientific experimentsā
(Some of his āGeneralā recommendations would be useful for those risks, but there are no recommendations specifically targeted at those risks.)
This is despite the fact that Ord estimates a ~1 in 50 chance that āother anthropogenic risksā will cause existential catastrophe in the next 100 years. Thatās ~20 times as high as his estimate for each of nuclear war and climate change (~1 in 1000), and ~200 times as high as his estimate for all ānatural risksā put together (~1 in 10,000). (Note that Ordās ānatural risksā includes supervolcanic eruption, asteroid or comet impact, and stellar explosion, but does not include āānaturallyā arising pandemicsā. See here for Ordās estimates and some commentary on them.)
Meanwhile, Ord includes 10 recommendations specifically related to ānatural riskās, 7 related to nuclear war, and 8 related to climate change. Those recommendations do all look to me like good recommendations, and like things āsomeoneā should do. But it seems odd to me that there are that many recommendations for those risks, yet none specifically related to a category Ord seems to think poses many times more existential risk.
Perhaps itās just far less clear to Ord what, concretely, should be done about āother anthropogenic risksā. And perhaps he wanted his list to only include relatively concrete, currently actionable recommendations. But I expect that, if we tried, we could find or generate such recommendations related to dystopian scenarios and nanotechnology (the two risks from this category Iām most concerned about).
So one thing Iād recommend is someone indeed having a go at finding or generating such recommendations! (I might have a go at that myself for dystopias, but probably only at least 6 months from now.)
Similarly, Ord has no recommendations specifically related to what he called āānaturallyā arising pandemicsā (as opposed to āengineered pandemicsā), which he estimates as posing as much existential risk over the next 100 years as all ānatural risksā put together (~1 in 10,000). (Again, note that he doesnāt include āānaturallyā arising pandemicsā as a ānatural riskā.)
This is despite the fact that, as noted above, he has 10 recommendations related to ānatural risksā. This also seems somewhat strange to me.
That said, one of Ordās recommendations for āEmerging Pandemicsā would also help with āānaturallyā arising pandemicsā. (This is the recommendation to āStrengthen the WHOās ability to respond to emerging pandemics through rapid disease surveillance, diagnosis and control. This involves increasing its funding and powers, as well as R&D on the requisite technologies.ā) But the other five recommendations for āEmerging Pandemicsā do seem fairly specific to emerging rather than ānaturallyā arising pandemics.
Regarding engineered pandemics
Ord recommends āIncreas[ing] transparency around accidents in BSL-3 and BSL-4 laboratories.ā āBSLā refers to ābiosafety levelā, and 4 is the highest it gets.
In Chapter 5, Ord provides some jawdropping/āhilarious/āhorrifying tales of accidents even among labs following the BSL-4 standards (including two accidents in a row for one lab). So Iām very much on board with the recommendation to increase transparency around those accidents.
But I was a little surprised to see that Ord didnāt also call for things like:
introducing more stringent standards (to prevent rather than be transparent about accidents),
introducing more monitoring and enforcement of compliance with those standards, and/āor
restricting some kinds of research as too dangerous for even labs following the highest standards
Some possible reasons why he may not have called for such things:
He may have worried thereād be too much pushback, e.g. from the bioengineering community
He may have thought those things just actually would be net-negative, even if not for pushback
But Iād guess (with low confidence) that at least something along the lines of the three āmissing recommendationsā mentioned aboveāand beyond what Ord already recommendsāwould probably help reduce biorisk, if done as collaboratively with the relevant communities as is practical.
Regarding existential risk communication
One of Ordās recommendations is to:
Develop better theoretical and practical tools for assessing risks with extremely high stakes that are either unprecedented or thought to have extremely low probability.
I think this is a great recommendation. (See also Database of existential risk estimates.) That recommendation also made me think that another strong recommendation might be something like:
Develop better approaches, incentives, and norms for communicating about risks with extremely high stakes that are either unprecedented or thought to have extremely low probability.
That sounds a bit vague, and Iām not sure exactly what form such approaches, incentives, or norms should take or how one would implement them. (Though I think that the same is true of the recommendation of Ordās which inspired this one.)
That proposed recommendation of mine was in part inspired by the COVID-19 situation, and more specifically by the following part of an 80,000 Hours Podcast episode. (which also gestures in the direction of concrete implications of my proposed recommendations).
Rob Wiblin: The alarm [about COVID-19] could have been sounded a lot sooner and we could have had five extra weeks to prepare. Five extra weeks to stockpile food. Five extra weeks to manufacture more hand sanitizer. Five extra weeks to make more ventilators. Five extra weeks to train people to use the ventilators. Five extra weeks to figure out what the policy should be if things got to where they are now.
Work was done in that time, but I think a lot less than could have been done if we had had just the forecasting ability to think a month or two ahead, and to think about probabilities and expected value. And this is another area where I think we could improve a great deal.
I suppose we probably wonāt fall for this exact mistake again. Probably the next time this happens, the world will completely freak out everywhere simultaneously. But we need better ability to sound the alarm, potentially greater willingness actually on the part of experts to say, āIām very concerned about this and people should start taking action, not panic, but measured action now to prepare,ā because otherwise itāll be a different disaster next time and weāll have sat on our hands for weeks wasting time that could have saved lives. Do you have anything to add to that?
Howie Lempel: I think one thing that we need as a society, although I donāt know how to get there, is an ability to see an expert say that they are really concerned about some risk. They think it likely wonāt materialize, but it is absolutely worth putting a whole bunch of resources into preparing, and seeing that happen and then seeing the risk not materialize and not just cracking down on and shaming that expert, because thatās just going to be what happens most of the time if you want to prepare for things that donāt occur that often.
Regarding AI risk
Here are Ordās four policy and research recommendations under the heading āUnaligned Artificial Intelligenceā:
Foster international collaboration on safety and risk management.
Explore options for the governance of advanced AI.
Perform technical research on aligning advanced artificial intelligence with human values.
Perform technical research on other aspects of AGI safety, such as secure containment or tripwires.
These all seem to me like excellent suggestions, and Iām glad Ord has lent additional credibility and force to such recommendations by including them in such a compelling and not-wacky-seemingbook. (I think Human Compatible and The Alignment Problem were also useful in a similar way.)
But I was also slightly surprised to not see explicit mention of, for example:
Work to actually understand what human values actually are, how theyāre structured, which aspects of them we do/āshould care about, etc.
E.g., much of Stuart Armstrongās research, or some work thatās more towards the philosophical rather than technical end
āAgent foundationsā/āādeconfusionā/āMIRI-style research
Further formalisation and critique of the various arguments and models about AI risk
Perhaps the first two of the āmissing recommendationsā I mentioned were actually meant to be implicit in Ordās third and fourth recommendations
Perhaps Ord has good reasons to not see these recommendations as especially worth mentioning
Perhaps Ord thought heād be unable to concisely state such recommendations (or just the MIRI-style research one) in a way that would sound concrete and clearly actionable to policymakers
Any shortlist of a personās top recommendations will inevitably fail to 100% please all readers
You can see a list of all the things Iāve written that summarise, comment on, or take inspiration from parts of The Precipice here.
Thoughts on Toby Ordās policy & research recommendations
In Appendix F of The Precipice, Ord provides a list of policy and research recommendations related to existential risk (reproduced here). This post contains lightly edited versions of some quick, tentative thoughts I wrote regarding those recommendations in April 2020 (but which I didnāt post at the time).
Overall, I very much like Ordās list, and I donāt think any of his recommendations seem bad to me. So most of my commentary is on things I feel are arguably missing.
Regarding āother anthropogenic risksā
Ordās list includes no recommendations specifically related to any of what he calls āother anthropogenic risksā, meaning:
ādystopian scenariosā
nanotechnology
āback contaminationā from microbes from planets we explore
aliens
āour most radical scientific experimentsā
(Some of his āGeneralā recommendations would be useful for those risks, but there are no recommendations specifically targeted at those risks.)
This is despite the fact that Ord estimates a ~1 in 50 chance that āother anthropogenic risksā will cause existential catastrophe in the next 100 years. Thatās ~20 times as high as his estimate for each of nuclear war and climate change (~1 in 1000), and ~200 times as high as his estimate for all ānatural risksā put together (~1 in 10,000). (Note that Ordās ānatural risksā includes supervolcanic eruption, asteroid or comet impact, and stellar explosion, but does not include āānaturallyā arising pandemicsā. See here for Ordās estimates and some commentary on them.)
Meanwhile, Ord includes 10 recommendations specifically related to ānatural riskās, 7 related to nuclear war, and 8 related to climate change. Those recommendations do all look to me like good recommendations, and like things āsomeoneā should do. But it seems odd to me that there are that many recommendations for those risks, yet none specifically related to a category Ord seems to think poses many times more existential risk.
Perhaps itās just far less clear to Ord what, concretely, should be done about āother anthropogenic risksā. And perhaps he wanted his list to only include relatively concrete, currently actionable recommendations. But I expect that, if we tried, we could find or generate such recommendations related to dystopian scenarios and nanotechnology (the two risks from this category Iām most concerned about).
So one thing Iād recommend is someone indeed having a go at finding or generating such recommendations! (I might have a go at that myself for dystopias, but probably only at least 6 months from now.)
(See also posts tagged global dystopia, atomically precise manufacturing, or space.)
Regarding naturally arising pandemics
Similarly, Ord has no recommendations specifically related to what he called āānaturallyā arising pandemicsā (as opposed to āengineered pandemicsā), which he estimates as posing as much existential risk over the next 100 years as all ānatural risksā put together (~1 in 10,000). (Again, note that he doesnāt include āānaturallyā arising pandemicsā as a ānatural riskā.)
This is despite the fact that, as noted above, he has 10 recommendations related to ānatural risksā. This also seems somewhat strange to me.
That said, one of Ordās recommendations for āEmerging Pandemicsā would also help with āānaturallyā arising pandemicsā. (This is the recommendation to āStrengthen the WHOās ability to respond to emerging pandemics through rapid disease surveillance, diagnosis and control. This involves increasing its funding and powers, as well as R&D on the requisite technologies.ā) But the other five recommendations for āEmerging Pandemicsā do seem fairly specific to emerging rather than ānaturallyā arising pandemics.
Regarding engineered pandemics
Ord recommends āIncreas[ing] transparency around accidents in BSL-3 and BSL-4 laboratories.ā āBSLā refers to ābiosafety levelā, and 4 is the highest it gets.
In Chapter 5, Ord provides some jawdropping/āhilarious/āhorrifying tales of accidents even among labs following the BSL-4 standards (including two accidents in a row for one lab). So Iām very much on board with the recommendation to increase transparency around those accidents.
But I was a little surprised to see that Ord didnāt also call for things like:
introducing more stringent standards (to prevent rather than be transparent about accidents),
introducing more monitoring and enforcement of compliance with those standards, and/āor
restricting some kinds of research as too dangerous for even labs following the highest standards
Some possible reasons why he may not have called for such things:
He may have worried thereād be too much pushback, e.g. from the bioengineering community
He may have thought those things just actually would be net-negative, even if not for pushback
He may have felt that his other recommendations would effectively accomplish similar results
But Iād guess (with low confidence) that at least something along the lines of the three āmissing recommendationsā mentioned aboveāand beyond what Ord already recommendsāwould probably help reduce biorisk, if done as collaboratively with the relevant communities as is practical.
Regarding existential risk communication
One of Ordās recommendations is to:
I think this is a great recommendation. (See also Database of existential risk estimates.) That recommendation also made me think that another strong recommendation might be something like:
That sounds a bit vague, and Iām not sure exactly what form such approaches, incentives, or norms should take or how one would implement them. (Though I think that the same is true of the recommendation of Ordās which inspired this one.)
That proposed recommendation of mine was in part inspired by the COVID-19 situation, and more specifically by the following part of an 80,000 Hours Podcast episode. (which also gestures in the direction of concrete implications of my proposed recommendations).
Regarding AI risk
Here are Ordās four policy and research recommendations under the heading āUnaligned Artificial Intelligenceā:
These all seem to me like excellent suggestions, and Iām glad Ord has lent additional credibility and force to such recommendations by including them in such a compelling and not-wacky-seeming book. (I think Human Compatible and The Alignment Problem were also useful in a similar way.)
But I was also slightly surprised to not see explicit mention of, for example:
Work to actually understand what human values actually are, how theyāre structured, which aspects of them we do/āshould care about, etc.
E.g., much of Stuart Armstrongās research, or some work thatās more towards the philosophical rather than technical end
āAgent foundationsā/āādeconfusionā/āMIRI-style research
Further formalisation and critique of the various arguments and models about AI risk
Along the lines of this, this, this, or the sources listed here
But this isnāt really a criticism, because:
Perhaps the first two of the āmissing recommendationsā I mentioned were actually meant to be implicit in Ordās third and fourth recommendations
Perhaps Ord has good reasons to not see these recommendations as especially worth mentioning
Perhaps Ord thought heād be unable to concisely state such recommendations (or just the MIRI-style research one) in a way that would sound concrete and clearly actionable to policymakers
Any shortlist of a personās top recommendations will inevitably fail to 100% please all readers
You can see a list of all the things Iāve written that summarise, comment on, or take inspiration from parts of The Precipice here.