For Pause AI or Stop AI to succeed, pausing /​ stopping needs to be a viable solution. I think some AI capabilities people who believe in existential risk may (perhaps?) be motivated by the thought that the risk of civilisational collapse is high without AI, so it’s worth taking the risk of misaligned AI to prevent that outcome.
If this really is cruxy for some people, it’s possible this doesn’t get noticed because people take it as a background assumption and don’t tend to discuss it directly, so they don’t realize how much they disagree and how crucial that disagreement is.
For Pause AI or Stop AI to succeed, pausing /​ stopping needs to be a viable solution. I think some AI capabilities people who believe in existential risk may (perhaps?) be motivated by the thought that the risk of civilisational collapse is high without AI, so it’s worth taking the risk of misaligned AI to prevent that outcome.
If this really is cruxy for some people, it’s possible this doesn’t get noticed because people take it as a background assumption and don’t tend to discuss it directly, so they don’t realize how much they disagree and how crucial that disagreement is.