I’m curious to dig into this a bit more, and hear why you think these seem like fairy tales to you (I’m not saying that I disagree...).
I wonder if this comes down to different ideas of what “solve alignment” means (I see you put it in quotes...)
1) Are you perhaps thinking that realistic “solutions to alignment” will carry a significant alignment tax? Else why wouldn’t ~everyone adopt alignment techniques (that align AI systems with their preferences/values)?
2) Another source of ambiguity: there are a lot of different things people mean by “alignment”, including: * AI is aligned with objectively correct values * AI is aligned with a stakeholder and consistently pursues their interests * AI does a particular task as intended/expected Is one of these in particular (or something else) that you have in mind here?
I’m curious to dig into this a bit more, and hear why you think these seem like fairy tales to you (I’m not saying that I disagree...).
I wonder if this comes down to different ideas of what “solve alignment” means (I see you put it in quotes...)
1) Are you perhaps thinking that realistic “solutions to alignment” will carry a significant alignment tax? Else why wouldn’t ~everyone adopt alignment techniques (that align AI systems with their preferences/values)?
2) Another source of ambiguity: there are a lot of different things people mean by “alignment”, including:
* AI is aligned with objectively correct values
* AI is aligned with a stakeholder and consistently pursues their interests
* AI does a particular task as intended/expected
Is one of these in particular (or something else) that you have in mind here?