I tried to figure out whether MIRI’s directions for AI alignment were good, by reading a lot of stuff that had been written online; I did a pretty bad job of thinking about all this.
I’m curious about why you think you did a bad job at this. Could you roughly explain what you did and what you should have done instead?
I don’t understand this. Have you written about this or have a link that explains it?