Interesting re belief in hell being a key factor, I wasn’t thinking about that.
It seems like the whole AI x-risk community has latched onto “align AI with human values/intent” as the solution, with few people thinking even a few steps ahead to “what if we succeeded”? I have a post related to this if you’re interested.
possibly the future economy will be so much more complicated that it will still make sense to have some distributed information processing in the market rather than have all optimisation centrally planned
I think there will be distributed information processing, but each distributed node/agent will be a copy of the central AGI (or otherwise aligned to it or shares its values), because this is what’s economically most efficient, minimizes waste from misaligned incentives and so on. So there won’t be the kind of value pluralism that we see today.
I assume we won’t be able to know with high confidence in advance what economic model will be most efficient post-ASI.
There’s probably a lot of other surprises that we can’t foresee today. I’m mostly claiming that post-AGI economics and governance probably wont look very similar to today’s.
It seems like the whole AI x-risk community has latched onto “align AI with human values/intent” as the solution, with few people thinking even a few steps ahead to “what if we succeeded”? I have a post related to this if you’re interested.
I think there will be distributed information processing, but each distributed node/agent will be a copy of the central AGI (or otherwise aligned to it or shares its values), because this is what’s economically most efficient, minimizes waste from misaligned incentives and so on. So there won’t be the kind of value pluralism that we see today.
There’s probably a lot of other surprises that we can’t foresee today. I’m mostly claiming that post-AGI economics and governance probably wont look very similar to today’s.