This is a great point, and I do think it’s an important question for participants to consider; I should switch the last reading for something covering this. The bottleneck is just finding a satisfactory reading—I’m not totally happy with any of the posts covering this, but maybe AGI safety from first principles is the closest to what I want.
This is a great point, and I do think it’s an important question for participants to consider; I should switch the last reading for something covering this. The bottleneck is just finding a satisfactory reading—I’m not totally happy with any of the posts covering this, but maybe AGI safety from first principles is the closest to what I want.
Actually, Joe Carlsmith does it better in Is power-seeking AI an existential risk? So I’ve swapped that in instead.