Firstly, even with a misaligned AGI, itsintended objectiveswould provide some disincentive to act in ways that lead to disastrous consequences for humanity
You aren’t factoring in mesa-optimisation or reward hacking in this section. Or the AI being much more powerful than humans (cf we don’t trade with ants).
You aren’t factoring in mesa-optimisation or reward hacking in this section. Or the AI being much more powerful than humans (cf we don’t trade with ants).