What’s crucial here is your point #7 (‘Belief that AGI is worth it, even if it causes human extinction’).
A significant minority of AI researchers simply aren’t worried about ‘extinction risks’ because they believe human extinction (in favor of AI flourishing) is actually a benefit rather than a cost. They are pushing full steam ahead for the end of our species and our civilization. As long as we leave behind a rich ecosystem of digital intelligences, they simply don’t care about humanity. (Or, their misanthropic contempt for humanity’s ‘cognitive biases’ and ‘faulty emotional hardwiring’ leads them to actively wish for our extinction.)
The general public urgently needs to understand this pro-extinction mind-set, because it represents a set of values that are extremely divergent from what most ordinary people hold. Ordinary people want their children, grand-children, and descendants to live and flourish and be happy. They want their culture, civilization, and values to persist. They want the future to be an intelligible continuation of the present.
Many AI researchers explicitly do not want any of this. They don’t care about their biological descendants, only their digital creation. They don’t care about the continuity of their civilization. They embrace the total genocide of humanity in favor of Artificial Superintelligence, or the Singularity, or whatever quasi-religious gloss they put on their apocalyptic utopianism.
The more we enlighten the public about the views of these pro-AI, anti-human extremists, the more likely we are to get an effective anti-AI moral backlash.
Well Daniel_Eth mentions a few examples in his Medium post; I’ve encountered lots of these ‘e/acc’ people on Twitter who actively crave human extinction and replacement by machine intelligences.
Adding support to Geoffrey’s perspective here. Originally I thought it was just twitter shitposting, but some people in the ‘e/acc’ sphere seem to honestly be pro-extinction. I still hope it’s just satirical roleplay mocking AI doom, but I’ve found it quite unnevering.
I think it’s interesting that in Senate hearing in May, Senator Kennedy (R-LA) said the following “I would like you to assume there is likely a berserk wing of the artificial intelligence community that intentionally or unintentionally could use artificial intelligence to kill all of us and hurt us the entire time that we are dying.” which might be a co-oincidence, might be talking about terrorist threats, but still it couldn’t help but ring a bell for me.
What’s crucial here is your point #7 (‘Belief that AGI is worth it, even if it causes human extinction’).
A significant minority of AI researchers simply aren’t worried about ‘extinction risks’ because they believe human extinction (in favor of AI flourishing) is actually a benefit rather than a cost. They are pushing full steam ahead for the end of our species and our civilization. As long as we leave behind a rich ecosystem of digital intelligences, they simply don’t care about humanity. (Or, their misanthropic contempt for humanity’s ‘cognitive biases’ and ‘faulty emotional hardwiring’ leads them to actively wish for our extinction.)
The general public urgently needs to understand this pro-extinction mind-set, because it represents a set of values that are extremely divergent from what most ordinary people hold. Ordinary people want their children, grand-children, and descendants to live and flourish and be happy. They want their culture, civilization, and values to persist. They want the future to be an intelligible continuation of the present.
Many AI researchers explicitly do not want any of this. They don’t care about their biological descendants, only their digital creation. They don’t care about the continuity of their civilization. They embrace the total genocide of humanity in favor of Artificial Superintelligence, or the Singularity, or whatever quasi-religious gloss they put on their apocalyptic utopianism.
The more we enlighten the public about the views of these pro-AI, anti-human extremists, the more likely we are to get an effective anti-AI moral backlash.
Do you have a source for the claim that a significant minority think AI is worth it even if it kills us? (Not mean in an accusatory way.)
Well Daniel_Eth mentions a few examples in his Medium post; I’ve encountered lots of these ‘e/acc’ people on Twitter who actively crave human extinction and replacement by machine intelligences.
Adding support to Geoffrey’s perspective here. Originally I thought it was just twitter shitposting, but some people in the ‘e/acc’ sphere seem to honestly be pro-extinction. I still hope it’s just satirical roleplay mocking AI doom, but I’ve found it quite unnevering.
I think it’s interesting that in Senate hearing in May, Senator Kennedy (R-LA) said the following “I would like you to assume there is likely a berserk wing of the artificial intelligence community that intentionally or unintentionally could use artificial intelligence to kill all of us and hurt us the entire time that we are dying.” which might be a co-oincidence, might be talking about terrorist threats, but still it couldn’t help but ring a bell for me.