Thank you for laying that out, that is elucidatory. And behind all this I guess is the belief that if we don’t suceed in “technical alignment”, the default is that the AI will be “aligned” to an alien goal, the pursuit of which will involve humanities disempowerment or destruction? If this was the belief, I could see why you would find technical alignment superior.
I, personally, don’t buy that this will be the default: I think the default will be some shitty approximation of the goals of the corporation that made it, localised mostly to the scenarios it was trained in. From the point of view of someone like me, technical alignment actually sounds dangerous to pursue: it would allow someone to imbue an AI with world domination plans and potentially actually succeed.
Thank you for laying that out, that is elucidatory. And behind all this I guess is the belief that if we don’t suceed in “technical alignment”, the default is that the AI will be “aligned” to an alien goal, the pursuit of which will involve humanities disempowerment or destruction? If this was the belief, I could see why you would find technical alignment superior.
I, personally, don’t buy that this will be the default: I think the default will be some shitty approximation of the goals of the corporation that made it, localised mostly to the scenarios it was trained in. From the point of view of someone like me, technical alignment actually sounds dangerous to pursue: it would allow someone to imbue an AI with world domination plans and potentially actually succeed.