Hi Robert, thanks for your perspective on this. I love your YouTube videos by the way—very informative and clear, and helpful for AI alignment newbies like me.
My main concern is that we still have massive uncertainty about what proportion of ‘alignment with all humans’ can be solved by ‘alignment with one human’. It sounds like your bet is that it’s somewhere above 50% (maybe?? I’m just guessing); whereas my bet is that it’s under 20% -- i.e. I think that aligning with one human leaves most of the hard problems, and the X risk, unsolved.
And part of my skepticism in that regard is that a great many humans—perhaps most of the 8 billion on Earth—would be happy to use AI to inflict harm, up to and including death and genocide, on certain other individuals and groups of humans. So, AI that’s aligned with frequently homicidal/genocidal individual humans would be AI that’s deeply anti-aligned with other individuals and groups.
Hi Robert, thanks for your perspective on this. I love your YouTube videos by the way—very informative and clear, and helpful for AI alignment newbies like me.
My main concern is that we still have massive uncertainty about what proportion of ‘alignment with all humans’ can be solved by ‘alignment with one human’. It sounds like your bet is that it’s somewhere above 50% (maybe?? I’m just guessing); whereas my bet is that it’s under 20% -- i.e. I think that aligning with one human leaves most of the hard problems, and the X risk, unsolved.
And part of my skepticism in that regard is that a great many humans—perhaps most of the 8 billion on Earth—would be happy to use AI to inflict harm, up to and including death and genocide, on certain other individuals and groups of humans. So, AI that’s aligned with frequently homicidal/genocidal individual humans would be AI that’s deeply anti-aligned with other individuals and groups.