This argument proves too much. The same could be said of “go and donate your money, this (list of charities we think are most effective) is the way to do it”.
My takeaway was that messages which could be spread include: “we should worry about conflict between misaligned AI and all humans”, “AIs could behave deceptively, so evidence of safety might be misleading, “AI projects should establish and demonstrate safety (and potentially comply with safety standards) before deploying powerful systems”, “alignment research is prosocial and great” and “we’re not ready for this”. (I excluded “it might be important for companies and other institutions to act in unusual ways”, because I agree this doesn’t seem like a straightforward message to spread).
The answer is probably (a).
“Disproportionate” seems like it boils down to an object-level disagreement about relative cause prioritisation between AI safety and other causes.
These don’t seem very compelling to me.
This argument proves too much. The same could be said of “go and donate your money, this (list of charities we think are most effective) is the way to do it”.
My takeaway was that messages which could be spread include: “we should worry about conflict between misaligned AI and all humans”, “AIs could behave deceptively, so evidence of safety might be misleading, “AI projects should establish and demonstrate safety (and potentially comply with safety standards) before deploying powerful systems”, “alignment research is prosocial and great” and “we’re not ready for this”. (I excluded “it might be important for companies and other institutions to act in unusual ways”, because I agree this doesn’t seem like a straightforward message to spread).
The answer is probably (a).
“Disproportionate” seems like it boils down to an object-level disagreement about relative cause prioritisation between AI safety and other causes.