“EA-according-to-their-own values”, i.e. E, is just instrumental rationality, right?
ETA: or maybe you’re thinking instead of something like actually internalizing/adopting their explicit values as ends, which does seem like an important separate step?
I was meaning “instrumental rationality applied to whatever part of their values is other-affecting”.
I think this is especially important to pull out explicitly relative to regular instrumental rationality, because the feedback loops are less automatic (so a lot of the instrumental rationality people learn by default is in service of their prudential goals).
“EA-according-to-their-own values”, i.e. E, is just instrumental rationality, right?
ETA: or maybe you’re thinking instead of something like actually internalizing/adopting their explicit values as ends, which does seem like an important separate step?
I was meaning “instrumental rationality applied to whatever part of their values is other-affecting”.
I think this is especially important to pull out explicitly relative to regular instrumental rationality, because the feedback loops are less automatic (so a lot of the instrumental rationality people learn by default is in service of their prudential goals).