We have a lot of philosophers and philosophically-minded people in EA, but only a tiny number of them are working on philosophical issues related to AI safety. Yet from my perspective as an AI safety researcher, it feels like there are some crucial questions which we need good philosophy to answer (many listed here; I’m particularly thinking about philosophy of mind and agency as applied to AI, a la Dennett). How do you think this funnel could be improved?
We have a lot of philosophers and philosophically-minded people in EA, but only a tiny number of them are working on philosophical issues related to AI safety. Yet from my perspective as an AI safety researcher, it feels like there are some crucial questions which we need good philosophy to answer (many listed here; I’m particularly thinking about philosophy of mind and agency as applied to AI, a la Dennett). How do you think this funnel could be improved?