I’ve been wondering whether AGI independently discovering valence realism could be a “get out clause” for alignment. Maybe this could even happen in a convergent manner with natural abstraction?
Current theme: default
Less Wrong (text)
Less Wrong (link)
I’ve been wondering whether AGI independently discovering valence realism could be a “get out clause” for alignment. Maybe this could even happen in a convergent manner with natural abstraction?