Sorry for the very delayed reply to this. I meant to reply at the time and then it slipped my mind!
Yes, youâve summarised my position perfectly, I like those diagrams!
I guess my deeper point was that I wasnât sure there was any meaningful way to say something like âX is twice as painful as Yâ without defining it via choices among gambles or durations. You say for humans it seems real, but does it? I can definitely introspect and discover that X is more painful than Y, but Iâm not sure I can introspect and discover that it is N times as painful. Where does that number come from?
Although as I was thinking more about how to justify this, I started thinking about other sensory experiences, like sound. Is it meaningful to say that âX feels twice as loud as Yâ, in a sense that doesnât have to line up with the intensity of the physical sound wave? And then I remembered my physics lessons from way back, and realised the answer might be yes. I was definitely taught that the reason we measure sound volume on a log scale (decibels) is because it lines up better with our sensory perception of it (you have to square the intensity of the sound wave in order to double the perceived intensity). But if this is true then it means there is some sense in which we can introspect and say âX sounds twice as loud as Yâ, even though the underlying sound wave might not be twice as intense. And if that is the case then maybe this should also be true for pain.
Iâm still very uncertain about this though. If I listened to different sounds and tried to place them on a numerical scale, Iâm not really sure what it is that Iâd actually be doing.
Itâs really cool that youâve done this and released the code!
Am I understanding right that the givewell baseline youâre trying to beat used GPT, while your approach uses Claude? How can you be sure that the improvements arenât due to the model choice, rather than the architecture?