I tried to figure out whether MIRI’s directions for AI alignment were good, by reading a lot of stuff that had been written online; I did a pretty bad job of thinking about all this.
I’m curious about why you think you did a bad job at this. Could you roughly explain what you did and what you should have done instead?
I’m curious about why you think you did a bad job at this. Could you roughly explain what you did and what you should have done instead?