Hmm, I would argue than an AI which, when asked, causes human extinction is not aligned, even if it did exactly what it was told.
Yea, I think I’d classify that as a different thing. I see alignment typically as a “mistake” issue, rather than as a “misuse” issue. I think others here often use the phrase similarly.
Hmm, I would argue than an AI which, when asked, causes human extinction is not aligned, even if it did exactly what it was told.
Yea, I think I’d classify that as a different thing. I see alignment typically as a “mistake” issue, rather than as a “misuse” issue. I think others here often use the phrase similarly.