Stuart Russell is probably the most prominent example.
I think Dan Hendryks is doing good work in this area as well, as well as a bunch of people on the AI alignment team at DeepMind.
But yea, it’d be great if a lot more ML researchers/engineers engaged with the AI x-risk arguments and alignment research.
Stuart Russell is probably the most prominent example.
I think Dan Hendryks is doing good work in this area as well, as well as a bunch of people on the AI alignment team at DeepMind.
But yea, it’d be great if a lot more ML researchers/engineers engaged with the AI x-risk arguments and alignment research.