I think this is a plausible principle to have, but it trades off against some other pretty plausible principles
I wasn’t involved in making this benchmark but fwiw it feels pretty reasonable to me to separate the measurement of an attribute from the policy decision about how that attribute should trade off against other things. (Indeed, I expect that AI developers will be unbothered by creating models which cause animals harm if that provides economic benefits to them.)
I wasn’t involved in making this benchmark but fwiw it feels pretty reasonable to me to separate the measurement of an attribute from the policy decision about how that attribute should trade off against other things. (Indeed, I expect that AI developers will be unbothered by creating models which cause animals harm if that provides economic benefits to them.)