If they are aligned, then surely our future selves can figure this out?
I think it’s entirely plausible we just don’t care to figure it out, especially if we have some kind of singleton scenario where the entity in control decides to optimize human/personal welfare at the expense of other sentient beings. Just consider how humans currently treat animals and now imagine that there is no opportunity for lobbying for AI welfare, we’re just locked into place.
Ultimately, I am very uncertain, but I would not say that solving AI alignment/control will “surely” lead to a good future.
I think it’s entirely plausible we just don’t care to figure it out, especially if we have some kind of singleton scenario where the entity in control decides to optimize human/personal welfare at the expense of other sentient beings. Just consider how humans currently treat animals and now imagine that there is no opportunity for lobbying for AI welfare, we’re just locked into place.
Ultimately, I am very uncertain, but I would not say that solving AI alignment/control will “surely” lead to a good future.