Are you talking about adversarial ML/adversarial examples? If so, that is certainly an area that’s relevant to long-term AI safety; e.g. many proposals for aligning AGI include some adversarial training. In general, I’d say many areas of ML have some relevance to safety, and it mostly depends on how you pick your research project within an area.
First, check out this post: https://www.lesswrong.com/posts/YDF7XhMThhNfHfim9/ai-safety-needs-great-engineers
Are you talking about adversarial ML/adversarial examples? If so, that is certainly an area that’s relevant to long-term AI safety; e.g. many proposals for aligning AGI include some adversarial training. In general, I’d say many areas of ML have some relevance to safety, and it mostly depends on how you pick your research project within an area.