I feel like I’m taking crazy pills.
It appears that many EAs believe we shouldn’t pause AI capabilities until it can be proven to have < ~ 0.1% chance of X-risk.
Put less confusingly, it appears many EAs believe we should allow capabilities development to continue despite the current X-risks.
This feels obviously a terrible thing to me.
What are the best reasons EA shouldn’t be pushing for an indefinite pause on AI capabilities development??
Messy practical reasons.
I agree with Larks that most of us would press a magic button to slow down AI progress on dangerous paths.
But we can’t, which raises two problems:
Tractability.
Effectiveness of moderate success. If you get a non-global slowdown, or a slowdown that ends too early, or a slowdown regime that’s evadable, or if you differentially slow cautious labs, or even if you just differentially slow leading labs, the effect is likely net-negative. (Increasing multipolarity among labs + differentially boosting less-cautious actors +
compute overhangenabling rapidly scaling up training compute. See Slowing AI: Foundations.)(I’d be excited to talk about proposals more specific than ‘push for a pause,’ or outcomes more specific than ‘pause until proven <0.1% doom.’ Who is doing the pausing; what are the rules? Or maybe you don’t have specific proposals/outcomes in mind, in which case I support you searching for new great ideas, but it’s not like others haven’t tried and failed already.)
Thanks for the comment Zach.
1. Can you elaborate on your comment “Tractability”?
2. I’m less worried about multipolarity because the leading labs are so far ahead AND I have short timelines (~ 10 years). My guess is if you had short timelines, you might agree?
3. If we had moderate short term success, my intuition is that we’ve actually found an effective strategy that could then be scaled. I worry that your thinking is basically pointing to ‘it needs to be an immediately perfect strategy or don’t bother!’
Pushing a magic button would be easy; affecting the real world is hard. Even if slowing is good, we should notice whether there exist tractable interventions (or: notice interventions’ opportunity cost).
Nope, my sense is that DeepMind, OpenAI, and Anthropic do and will have a small lead over Meta, Inflection, and others, such that I would be concerned (re increasing multipolarity among labs) about slowing DeepMind, OpenAI, and Anthropic now. (And I have 50% credence on human-level AI [noting this is underspecified] within 9 years.)
Yeah, maybe, depending. I’m relatively excited about “short term success” that seems likely to support the long-term policy regimes I’m excited about, like global monitoring of compute and oversight of training runs with model evals for dangerous capabilities and misalignment, maybe plus a compute cap. I fear that most pause-flavored examples of “short term success” won’t really support great long-term plans. (Again, I’d be excited to talk about specific proposals/outcomes/interventions.)
This sequence is still in progress but is the best collection of resources that I know of regarding slowing AI (including an indefinite pause).
Thank you Ben I will take a look :)
Worth cross-posting to the EA Forum. @Zach Stein-Perlman
I think the best reason is that it’s not within the Overton window :)
Last I checked, the whole point of the Overton window is that you can only shift it by advocating for ideas outside of it.
I’m confused—where is the evidence it is outside the overton window?
Don’t really think there is any; in fact, there’s plenty of evidence to the contrary, from the polls I’ve seen.
We found 51% support, 25% opposition in our polling here.
the overton window is a crutch for EAs who don’t believe in the power of advocacy
I’ve worked in advocacy for EA causes for a bit, so I definitely believe in the power of it, but I also think the Overton window is a pretty crucial consideration for folks who are trying to mobilize the public. I’m guessing this is a popular view among people who work in advocacy for EA causes, but I might be wrong.
To be fair, I do think there could be value in making bold asks outside the Overton window. James Ozden has a really good piece about this. I think groups like DxE and PETA have done this for the animal movement, and it seems totally plausible to me that this has had a net positive effect.
But on the other hand, I think lots of the tangible changes we’ve seen for farmed animals have come from the incremental welfare asks that groups like Mercy For Animals and The Humane League focus on (disclaimer: I worked at the latter). The fact that these groups have been very careful to keep their asks within the Overton window has had the benefit of (1) helping advocates gain broad-based public support; and (2) getting corporations and policymakers on board and willing to actually adopt the changes they are asking for.
It seems likely to me that the second point applies for AI safety, but I’m not sure about the first and would probably need to see more polling or message testing to know. Nonetheless I suspect these concerns might be part of why the AI pause ask hasn’t been as widely adopted among EAs (although a number of them did sign the FLI letter).
The public is very concerned about powerful AI and want something done about it.
If anyone is outside the overton window its EAs.
I agree that the public has been pretty receptive to AI safety messaging. Much more than I would have expected a few years ago.
It sounds like you already have some takes on this question — in that case, it could be worth writing something up to make the case for why EAs should be advocating for a pause. I’d be happy to offer feedback if you do.
Thats very generous of you, thanks Tyler!