Once we add caveats like “what we would want / intend after sufficient rational reflection,” my sense is that “values” just captures that more intuitively.
I in fact don’t want to add in those caveats here: I’m suggesting that we tell our AI system to do what we short-term want. (Of course, we can then “short-term want” to do more rational reflection, or to be informed of true and useful things that help us make moral progress, etc.)
I agree that “values” more intuitively captures the thing with all the caveats added in.
I in fact don’t want to add in those caveats here: I’m suggesting that we tell our AI system to do what we short-term want. (Of course, we can then “short-term want” to do more rational reflection, or to be informed of true and useful things that help us make moral progress, etc.)
I agree that “values” more intuitively captures the thing with all the caveats added in.