This post is a great exemplar for why the term “AI alignment” has proven a drag on AI x-risk safety. The concern is and has always been that AI would dominate humanity like humans dominate animals. All of the talk about aligning AI to “human values” leads to pedantic posts like this one arguing about what “human values” are and how likely AIs are to pursue them.
Is there a particular part of my post that you disagree with? Or do you think the post is misleading. If so, how?
I think there are a lot of ways AI could go wrong, and “AIs dominating humans like how humans dominate animals” does not exhaust the scope of potential issues.
This post is a great exemplar for why the term “AI alignment” has proven a drag on AI x-risk safety. The concern is and has always been that AI would dominate humanity like humans dominate animals. All of the talk about aligning AI to “human values” leads to pedantic posts like this one arguing about what “human values” are and how likely AIs are to pursue them.
Is there a particular part of my post that you disagree with? Or do you think the post is misleading. If so, how?
I think there are a lot of ways AI could go wrong, and “AIs dominating humans like how humans dominate animals” does not exhaust the scope of potential issues.