No, that’s not what I mean. I mean we should use other examples of the form “you ask an AI to do X, and the AI accomplishes X by doing Y, but Y is bad and not what you intended” where Y is not as bad as an extinction event.
I understand—and agree with—the overall point being made about “don’t just talk about the extreme things like paperclip maximizers”, but I’m still thrown off by the statement that “the mechanisms needed to prevent [paperclip maximizers] are the same as those needed to prevent the less severe and more plausible-sounding scenarios”
No, that’s not what I mean. I mean we should use other examples of the form “you ask an AI to do X, and the AI accomplishes X by doing Y, but Y is bad and not what you intended” where Y is not as bad as an extinction event.
I understand—and agree with—the overall point being made about “don’t just talk about the extreme things like paperclip maximizers”, but I’m still thrown off by the statement that “the mechanisms needed to prevent [paperclip maximizers] are the same as those needed to prevent the less severe and more plausible-sounding scenarios”
Hm, yeah, I see where you’re coming from. Changed the phrasing.