This is somewhat inspired by a variety of Twitter people saying that Eliezer Yudkowsky shouldn’t be trusted because he made bad predictions in the past (arbitrarily chosen examples here and here) but I am also interested in the question from the perspective of whether alignment strategies relying on AI being more tool-like are promising.
This is somewhat inspired by a variety of Twitter people saying that Eliezer Yudkowsky shouldn’t be trusted because he made bad predictions in the past (arbitrarily chosen examples here and here) but I am also interested in the question from the perspective of whether alignment strategies relying on AI being more tool-like are promising.