Thanks for the stab, Anthony. It’s fairly fair. :-)
Some clarifying points:
First, I should note that my piece was written from the perspective of suffering-focused ethics.
Second, I would not say that “investment in AI safety work by the EA community today would only make sense if the probability of AI-catalyzed GCR were decently high”. Even setting aside the question of what “decently high” means, I would note that:
1) Whether such investments in AI safety make sense depends in part on one’s values. (Though another critique I would make is that “AI safety” is less well-defined than people often seem to think: https://magnusvinding.com/2018/12/14/is-ai-alignment-possible/, but more on this below.)
2) Even if “the probability of AI-catalyzed GCR” were decently high — say, >2 percent — this would not imply that one should focus on “AI safety” in a standard narrow sense (roughly: constructing the right software), nor that other risks are not greater in expectation (compared to the risks we commonly have in mind when we think of “AI-catalyzed catastrophic risks”).
You write of “scenarios in which AGI becomes a catastrophic threat”. But a question I would raise is: what does this mean? Do we all have a clear picture of this in our minds? This sounds to me like a rather broad class of scenarios, and a worry I have is that we all have “poorly written software” scenarios in mind, although such scenarios could well comprise a relatively narrow subset of the entire class that is “catastrophic scenarios involving AI”.
Zooming out, my critique can be crudely summarized as a critique of two significant equivocations that I see doing an exceptional amount of work in many standard arguments for “prioritizing AI”.
First, there is what we may call the AI safety equivocation (or motte and bailey): people commonly fail to distinguish between 1) a focus on future outcomes controlled by AI and 2) a focus on writing “safe” software. Accepting that we should adopt the former focus by no means implies we should adopt the latter. By (imperfect) analogy, to say that we should focus on future outcomes controlled by humans does not imply that we should focus primarily on writing safe human genomes.
The second is what we may call the intelligence equivocation, which is the one you described. We operate with two very different senses of the term “intelligence”, namely 1) the ability to achieve goals in general (derived from Legg & Hutter, 2007), and 2) “intelligence” in the much narrower sense of “advanced cognitive abilities”, roughly equivalent to IQ in humans.
Intelligence2 lies all in the brain, whereas intelligence1 includes the brain and so much more, including all the rest of our well-adapted body parts (vocal cords, hands, upright walk — remove just one of these completely in all humans and human civilization is likely gone for good). Not to mention our culture and technology as a whole, which is the level at which our ability to achieve goals at a significant level really emerges: it derives not from any single advanced machine but from our entire economy. A vastly greater toolbox than what intelligence2 covers.
Thus, to assume that we by boosting intelligence2 to vastly super-human levels necessarily get intelligence1 at a vastly super-human level is a mistake, not least since “human-level intelligence1” already includes vastly super-human intelligence2 in many cognitive domains.
Thanks for the stab, Anthony. It’s fairly fair. :-)
Some clarifying points:
First, I should note that my piece was written from the perspective of suffering-focused ethics.
Second, I would not say that “investment in AI safety work by the EA community today would only make sense if the probability of AI-catalyzed GCR were decently high”. Even setting aside the question of what “decently high” means, I would note that:
1) Whether such investments in AI safety make sense depends in part on one’s values. (Though another critique I would make is that “AI safety” is less well-defined than people often seem to think: https://magnusvinding.com/2018/12/14/is-ai-alignment-possible/, but more on this below.)
2) Even if “the probability of AI-catalyzed GCR” were decently high — say, >2 percent — this would not imply that one should focus on “AI safety” in a standard narrow sense (roughly: constructing the right software), nor that other risks are not greater in expectation (compared to the risks we commonly have in mind when we think of “AI-catalyzed catastrophic risks”).
You write of “scenarios in which AGI becomes a catastrophic threat”. But a question I would raise is: what does this mean? Do we all have a clear picture of this in our minds? This sounds to me like a rather broad class of scenarios, and a worry I have is that we all have “poorly written software” scenarios in mind, although such scenarios could well comprise a relatively narrow subset of the entire class that is “catastrophic scenarios involving AI”.
Zooming out, my critique can be crudely summarized as a critique of two significant equivocations that I see doing an exceptional amount of work in many standard arguments for “prioritizing AI”.
First, there is what we may call the AI safety equivocation (or motte and bailey): people commonly fail to distinguish between 1) a focus on future outcomes controlled by AI and 2) a focus on writing “safe” software. Accepting that we should adopt the former focus by no means implies we should adopt the latter. By (imperfect) analogy, to say that we should focus on future outcomes controlled by humans does not imply that we should focus primarily on writing safe human genomes.
The second is what we may call the intelligence equivocation, which is the one you described. We operate with two very different senses of the term “intelligence”, namely 1) the ability to achieve goals in general (derived from Legg & Hutter, 2007), and 2) “intelligence” in the much narrower sense of “advanced cognitive abilities”, roughly equivalent to IQ in humans.
These two are often treated as virtually identical, and we fail to appreciate the rather enormous difference between them, as argued in/evident from books such as The Knowledge Illusion: Why We Never Think Alone, The Ascent of Man, The Evolution of Everything, and The Secret of Our Success. This was also the main point in my Reflections on Intelligence.
Intelligence2 lies all in the brain, whereas intelligence1 includes the brain and so much more, including all the rest of our well-adapted body parts (vocal cords, hands, upright walk — remove just one of these completely in all humans and human civilization is likely gone for good). Not to mention our culture and technology as a whole, which is the level at which our ability to achieve goals at a significant level really emerges: it derives not from any single advanced machine but from our entire economy. A vastly greater toolbox than what intelligence2 covers.
Thus, to assume that we by boosting intelligence2 to vastly super-human levels necessarily get intelligence1 at a vastly super-human level is a mistake, not least since “human-level intelligence1” already includes vastly super-human intelligence2 in many cognitive domains.