This is a blog post, not a research report, meaning it was produced relatively quickly and is not to Rethink Priorities’ typical standards of substantiveness and careful checking for accuracy.
Summary
This post is a shallow exploration of some technological developments that might occur and might increase risks from nuclear weapons—especially existential risk or other risks to the long-term future. This is one of many questions relevant to how much to prioritize nuclear risk relative to other issues, what risks and interventions to prioritize within the nuclear risk area, and how that should change in future. But note that, due to time constraints, this post isn’t comprehensive and was less thoroughly researched and reviewed than we’d like.
For each potential development, we provide some very quick, rough guesses about how much and in what ways the development would affect the odds and consequences of nuclear conflict (“Importance”), the likelihood of the development in the coming decade or decades (“Likelihood/Closeness”), and how much and in what ways thoughtful altruistic actors could influence whether and how the technology is developed and used (“Steerability”).
These tentative bottom line beliefs are summarized in the table below:
Category | Technological Development | Importance | Likelihood / Closeness | Steer-ability |
Bomb types and production methods | Radiological weapons | Medium | Medium/ High | Medium/ Low |
Pure fusion bombs | Medium | Medium/ Low | Medium | |
High-altitude electromagnetic pulse (HEMP) | Medium | Medium/ Low | Medium/ Low | |
Neutron bombs | Low | Medium/ Low | Medium | |
Methods for production and design | Atomically precise manufacturing (APM) | High | Low | Medium |
AI-assisted production/design | Medium/ High | Medium/ Low | Medium | |
Other developments in methods for production/design | ? | ? | ? | |
Delivery systems | Hypersonic missiles/glide vehicles | Medium/Low | Medium/ High | Medium/ Low |
More accurate nuclear weapons | Medium | Medium | Medium/ Low | |
Long-range conventional strike capabilities | Medium/Low | Medium/ High | Medium/ Low | |
Detection and defense | Better detection of nuclear warhead platforms, launchers, and/or delivery vehicles | Medium/High | Medium/ High | Medium/ Low |
Missile defense systems | Medium | Medium | Medium/ Low | |
AI and cyber | Advances in AI capabilities | Medium/ High | Medium/ High | Medium |
Cyberattack (or defense) capabilities | Medium/ High | Medium/ High | Medium | |
Advances in autonomous weapons | Medium | Medium/ High | Medium | |
More integration of AI with NC3 systems | Medium | Medium | Medium | |
Non-nuclear warmaking advances | Anti-satellite weapons (ASAT) | Medium/ Low | Medium | Medium/ Low |
“Space planes” and other (non-ASAT) space capabilities | Medium/ Low | Medium | Medium/ Low |
Note that:
Each “potential technological development” is really more like a somewhat wide area in which a variety of different types and levels of development could occur, which makes the ratings in the above table less meaningful and more ambiguous.
“Importance” is here assessed conditional on the development occurring, so will overstate the importance of thinking about or trying to steer unlikely developments.
In some cases (e.g, “More accurate nuclear weapons”), the “Importance” score accounts for potential risk-reducing effects as well.
“Likelihood/Closeness” is actually inelegantly collapsing together two different things, making our ratings of developments on that criterion less meaningful. E.g., one development could be moderately likely to occur quite soon and moderately likely to occur never, while another is very likely to occur in 15-25 years but not before then.
Some of the topics this post discusses involve or are adjacent to information hazards (especially attention hazards), as is the case with much other discussion of technological developments that might occur and might increase risks. We ask that readers remain mindful of that when discussing or writing about this topic.
Epistemic status & directions for further research
In 2021, Michael did some initial research for this post and wrote an outline and rough notes. But he pivoted away from nuclear risk research before having time to properly research and draft this. We (Michael and Will) finished a rough version of this post in 2022, since that seemed better than it never being published at all, but then didn’t get around to publishing till 2023. As such, this is just a very incomplete starting point, may contain errors, and may be outdated. It could be quite valuable for someone to spend more time:
learning about other possible technological developments worth paying attention to,
doing more thorough and careful research on the developments we discuss,
thinking more about their implications for how much to prioritize nuclear risk reduction and what interventions and policies to pursue, and/or
talking to and getting feedback from various experts
See also Research project idea: Technological developments that could increase risks from nuclear weapons.
How to engage with this post
The full post can be found here. It’s ~23,000 words, much of which is extensive quotes without added commentary from us. Some quotes are relevant to multiple sections and hence are repeated. But each section or subsection should make sense by itself, so readers should feel free to read only the sections that are of interest to them, to skim, and to skip repeated quotes and “Additional notes” sections. You can navigate to sections using the links in the summary table above.
Scope of this post
This post is focused on what potential technological developments could increase risks from nuclear weapons. As such, this post is not necessarily claiming that these technologies will be net harmful overall, nor that nuclear risk will increase in future overall; both of those claims are plausibly true and plausibly false, and we don’t assess them here. Here are some further notes on what this post is vs isn’t focusing on and claiming:
We’re not claiming any of these developments are guaranteed—in fact, we think several are unlikely or would only happen a long time from now.
We don’t address things that could increase risk from nuclear weapons but that aren’t technological developments.
We mostly focus on possible new technological developments, setting aside proliferation of existing technologies or changes in how those technologies are deployed.
However, sometimes these lines are blurry, and we did end up discussing some things that may be more like deployment than development (e.g., integration of AI with NC3).
We don’t focus on discussing ways these technological developments might also decrease nuclear risks, whether their net effect might be a decrease in nuclear risk (even if they could also have important risk-increasing effects), or other technological developments that could decrease nuclear risks.
In a few places we do touch on those points, but we didn’t set out to do so, and thus there’s far more that could be usefully said than what we’ve said in this post.
We had hoped to discuss what could and should be done to influence whether and how these technological developments occur and are used, and what other implications these potential developments might have for what risks and interventions to prioritize in the nuclear risk space. But ultimately we ran out of time, and hence this post only contains extremely preliminary and patchy discussion of those questions.
As expressed here, we think those questions would likely be one of the most valuable directions for further research building off this post, and potentially one of the most valuable directions for further research on nuclear risk in general.
For more general discussion of possible goals and interventions related to nuclear risk, see 8 possible high-level goals for work on nuclear risk and Shallow review of approach to reducing risks from nuclear weapons.
Acknowledgements
Michael’s work on this post was supported by Rethink Priorities (though he ended up pivoting to other topics before having time to get this up to RP’s usual standards). Will helped with the research and editing in a personal capacity. We’re grateful to Ben Snodin, Damon Binder, Fin Moorhouse, and Jeffrey Ladish for feedback on an earlier draft. Mistakes are our/Michael’s own.
If you are interested in RP’s work, please visit our research database and subscribe to our newsletter.
An in-our-view interesting tangential point: It might decently often be the case that a technological development initially increases risks but then later increases risk by a smaller margin or even overall reduces risks.
One reason this can happen is that developments may be especially risky in the period before states or other actors have had time to adjust their strategies, doctrine, procedures, etc. in light of the development.
(This seems in some ways reminiscent of the Collingridge dilemma or the “pacing problem”.)
Another possible reason is that a technology may be riskiest in the period when it is just useful enough to be deployed but not yet very reliable.
Geist and Lohn (2018) suggest this might happen, for the above two reasons, with respect to AI developments and nuclear risk:
“Workshop participants agreed that the riskiest periods will occur immediately after AI enables a new capability, such as tracking and targeting or decision support about escalation. During this break-in period, errors and misunderstandings are relatively likely. With time and increased technological progress, those risks would be expected to diminish. If the main enabling capabilities are developed during peacetime, then it may be reasonable to expect progress to continue beyond the point at which they could be initially fielded, allowing time for them to increase in reliability or for their limitations to become well understood. Eventually, the AI system would develop capabilities that, while fallible, would be less error-prone than their human alternatives and therefore be stabilizing in the long term”
Just remembered that Artificial Intelligence and Nuclear Command, Control, & Communications: The Risks of Integration was written and published after I initially drafted this, so Will and I’s post doesn’t draw on or reference this, but it’s of course relevant too.
This is a good post. Thank you for sharing. I disagree somewhat with your framework, because I think it is extremely important to differentiate factors that increase the likelihood of armed conflicts between nuclear powers and factors that increase the risk of nuclear escalation given a conventional conflict. I think that you’ve over-focused on the latter, and that drivers of the former are fairly important
For example, your analysis of UAVs and UUVs doesn’t consider a risk I find highly salient: mutual misunderstanding of escalatory strength. That is, if the US shoots down an uncrewed Chinese intelligence balloon over US airspace, the escalatory action was China sending the balloon at all. If the US had shot down a crewed Chinese stealth fighter, reactions would have been very different. This holds even if the capabilities of the fighter and the balloon were identical.
Now, if the sole impact of UAVs is that there’s a new step on the escalation ladder, that would probably be slightly beneficial. But if there’s a step on the escalation ladder and Chinese and American political leadership disagree on where that step is, the potential for a situation to turn into a shooting conflict that takes lives increases substantially.
A similar point about escalation uncertainty can be raised by cybersecurity capabilities: militaries across the globe have made some steps towards defining how they think about cyberattacks. I believe that the most explicit statement on the topic comes from the French, but there are also advantages to strategic ambiguity, and genuine uncertainty about how publics in both authoritarian and democratic states would react to a cyberattack that, say, impaired the power grid.
Cybersecurity has the additional problem of, in the view of some experts, having incentives towards more provocative action, with a bias towards attackers under some circumstances.
As always, I do not speak for my employer or the US government.