Finally, even if the AGI initially acts in destructive ways, amending its programming to avoid causing unacceptable harm wouldnotbe particularly difficult, as long as we are able to appeal to itsfundamental objectives.
In this section it seems like you are assuming that corrigibility is solved, as well as outer alignment. These are major unsolved problems in AI Alignment. I don’t want to sound too mean, but I get the impression that your 300 hours of reading didn’t include the AGI Safety Fundamentals syllabus.
In this section it seems like you are assuming that corrigibility is solved, as well as outer alignment. These are major unsolved problems in AI Alignment. I don’t want to sound too mean, but I get the impression that your 300 hours of reading didn’t include the AGI Safety Fundamentals syllabus.