Probability of massive unemployment in the United States based on historical data
The views expressed here are my own, not those of my employers.
Summary
The 5th percentile, median and 95th percentile annual unemployment rate in the United States (US) from 1948 to 2023 were 3.54 %, 5.55 % and 9.02 %.
There has been a very weak upward trend in the annual unemployment rate in the US, with the coefficient of determination (R^2) of the linear regression of it on the year being 3.54 %.
The tail distribution of the annual unemployment rate in the US bends downwards in a log-log plot. I got a similar pattern for the annual conflict and epidemic/pandemic deaths as a fraction of the global population.
For the Pareto distributions fit to the 10.5 % (8) rightmost points of the tail distribution in a log-log plot, the annual probability of this being:
At least 10 % is 2.04 %.
100 % is 2.11*10^-10.
I suspect many in the global catastrophic risk community underweight the outside view outlined in this post, and therefore overestimate the probability of massive unemployment.
Methods
I fit Pareto distributions (power laws) to the 2, 3, … and 76 (= 2023 − 1948 + 1) highest values of the annual unemployment rate in the US from 1948 to 2023. To do this:
I get the slope and intercept of linear regressions of the above sections of the tail distribution on the annual unemployment rate in the US.
Since the tail distribution of a Pareto distribution is P(X > x) = (“minimum”/x)^“tail index”, ln(P(X > x)) = “tail index”*ln(“minimum”) - “tail index”*ln(x) = “intercept” + “slope”*ln(x), so I determine the parameters of the Pareto distributions from:
“Tail index” = -“slope”.
“Minimum” = e^(“intercept”/“tail index”).
Then I obtain the annual probability of an annual unemployment rate in the US of at least 10 %, 25 %, 50 %, 75 % and 100 %. Note unemployment requires not only not having a job, but also being actively looking for one:
Someone in the labor force is defined as unemployed if they were not employed during the survey reference week, were available for work, and made at least one active effort to find a job during the 4-week survey period.
The calculations are in this Sheet.
Results
Historical data
Time series
Basic stats
Statistic | Annual unemployment rate in the US |
Mean | 5.70 % |
Minimum | 2.93 % |
5th percentile | 3.54 % |
10th percentile | 3.71 % |
Median | 5.55 % |
90th percentile | 7.89 % |
95th percentile | 9.02 % |
Maximum | 9.71 % |
Linear regression on the year
Linear regression of the annual unemployment rate in the US on the year | ||
Slope (pp/century) | Intercept (pp) | Coefficient of determination |
1.39 | -21.9 | 3.54 % |
Tail distribution
Tail risk based on Pareto distributions
Discussion
Historical data
The 5th percentile, median and 95th percentile annual unemployment rate in the US from 1948 to 2023 were 3.54 %, 5.55 % and 9.02 %. The maximum was 9.71 % in 1982.
There has been a very weak upward trend in the annual unemployment rate in the US, with the R^2 of the linear regression of it on the year being 3.54 %.
The tail distribution of the annual unemployment rate in the US bends downwards in a log-log plot. I got a similar pattern for the annual conflict and epidemic/pandemic deaths as a fraction of the global population.
Tail risk
My estimates for the annual probability of an annual unemployment rate in the US of:
At least 10 % range from 0.105 % to 6.54 %.
100 % range from 1.32*10^-88 to 0.00749 %.
For the Pareto distributions fit to the 10.5 % (8) rightmost points of the tail distribution in a log-log plot, the annual probability of this being:
At least 10 % is 2.04 %.
100 % is 2.11*10^-10.
I believe the rightmost points of the tail should get the most weight to predict tail risk, which implies trusting lower estimates more. On the other hand, I expect inside view factors related to transformative AI (TAI) to push estimates upwards. All in all, I suspect many in the global catastrophic risk community underweight the outside view outlined in this post, and therefore overestimate the probability of massive unemployment.
- 20 Oct 2024 16:44 UTC; 4 points) 's comment on Why Stop AI is barricading OpenAI by (
Thank you for the post! How much weight do you think should one allocate to the inside and outside view respectively in order to develop a comprehensive estimate of the potential future unemployment rate?
Your calculations look fancy and all of that, but it seems inappropriate to me to be putting so much weight on historical data as you are doing. Especially because I think this ignores the apparent fact that the development of intelligent systems that are more capable than humans has never occurred in history. This fundamentally changes the game.
The more the world changes, I think the less weight one should be putting on the outside view (needs more nuance). People are scared, people don’t update in the face of new evidence, people dislike change.
I know you are not saying that the inside view doesn’t matter, but I am concerned that a post like this anchors people toward a base rate that is a lot lower than what things will actually be like. It reinforces status quo bias. And this is frustrating to me because so many people don’t seem to understand the seriousness of our situation.
I think it makes a lot of sense to reason bottom-up when thinking about topics like these, and I actually disagree with you a lot. It seems to be that there is a deeply correlated failure happening in the AI safety community. In my view, people are putting way too much weight onto the outside view. I am happy to elaborate.
Thank you for sparking this discussion.
Thanks for the comment, Johan!
It is hard for me to answer this. It depends on the methodology used to produce the inside view estimate. If this is just a guess from someone working on AI safety, I would put very little weight on it. If it is the output of a detailed quantitative empirical model like Epoch AI’s, I could as a 1st approximation ignore the estimates from my post (although I would have to check the model to know).
Task automation has been happening for a long time (with the unemployment rate still being low), and one can think about advanced AI as a continuation of that trend. In addition, the definition of unemployment I used requires both not having a job and being actively looking for one. For sometime in the next few decades to centuries, I predict negligible human unemployment and roughly total AI automation (i.e. almost no human workers). I guess humans will just be happy letting the AIs do everything, and whoever wants to have a job (which will be a little bit of a fake job, as AIs would be able to perform the tasks more efficiently) will also have the chance to do it, i.e. there will be basically no humans actively looking for a job, and having no success (i.e. essentially no unemployment). More pessimistically, it is also possible to have almost total homelessness with negligible unemployment in a dystopian scenario where humans gave up looking for jobs because AIs are so much better, but still kind enough to give humans a subsistence income.
According to Table 1 (2) of Hanson 2000, the global economy used to double once every 230 k (224 k) years in hunting and gathering period of human history. Today it doubles once every 20 years or so[1]. Despite a much higher growth rate, the unemployment rate is still relatively low. So I do not think one can predict massive unemployment solely on the basis of AI boosting economic growth. Note am discussing what could happen in the real world, not what could happen in the absence of any mitigation actions.
What is your median annual unemployment rate in the US in 2025, 2026 or 2027? If much higher than now, I am happy to set up a bet with you where:
I give you 10 k€ if the rate is higher than your median.
You give me 10 k€ if the rate is lower than your median, which I would donate to animal welfare interventions.
My medians are not far from the ones suggested by historical data below, although I would want to think more about them if I was to bet 10 k€.
Thanks for engaging too!
The doubling time for 3 % annual growth is 23.4 years (= LN(2)/LN(1.03)).
Thank you for your reply!
Summary: My main intention in my previous comment was to share my perspective on why relying too much on the outside view is problematic (and, to be fair, that wasn’t clear because I addressed multiple points). While I think your calculations and explanation are solid, the general intuition I want to share is that people should place less weight on the outside view, as this article seems to suggest.
I wrote this fairly quickly, so I apologize if my response is not entirely coherent.
Emphasizing the definition of unemployment you use is helpful, and I mostly agree with your model of total AI automation, where no one is necessarily looking for a job.
Regarding your question about my estimate of the median annual unemployment rate: I haven’t thought deeply enough about unemployment to place a bet or form a strong opinion on the exact percentage points. Thanks for the offer, though.
To illustrate the main point in my summary, I want to share a basic reasoning process I’m using.
Assumptions:
Most people are underestimating the speed of AI development.
The new paradigm of scaling inference-time compute (instead of training compute) will lead to rapid increases in AI capabilities.
We have not solved the alignment problem and don’t seem to be making progress quickly enough (among other unsolved issues).
An intelligence explosion is possible.
Worldview implications of my assumptions:
People should take this development much more seriously.
We need more effective regulations to govern AI.
Humanity needs to act now and ambitiously.
To articulate my intuition as clearly as possible: the lack of action we’re currently seeing from various stakeholders in addressing the advancement of frontier AI systems seems to be, in part, because they rely too heavily on the outside view for decision-making. While this doesn’t address the crux of your post ( but it prompted me to write my comment initially), I believe it’s dangerous to place significant weight on an approach that attempts to make sense of developments we have no clear reference classes for. AGI hasn’t happened yet, so I don’t understand why we should lean heavily on historical data to assess such a novel development.
What’s currently happening is that people are essentially throwing their arms up and saying, “Uh, the probabilities are so low for X or Y impact of AGI, so let’s just trust the process.” If people placed more weight on assumptions like those above, or reasoned more from first principles, the situation might look very different. Do you see? My issue is with putting too much weight on the outside view, not with your object-level claims.
I am open to changing my mind on this.