From how Paul Christiano frames it, it seems like it’s “create AGI, and make sure it’s aligned.”
I think that’s basically right. I believe something like was Eliezer’s plan too, way back in the day, but then he updated to believing that we don’t have the basic ethical, decision theoretic, and philosophical stuff figured out that’s prerequisite to actually making a safe AGI. More on that in his Rocket Alignment Dialogue.
I think that’s basically right. I believe something like was Eliezer’s plan too, way back in the day, but then he updated to believing that we don’t have the basic ethical, decision theoretic, and philosophical stuff figured out that’s prerequisite to actually making a safe AGI. More on that in his Rocket Alignment Dialogue.