Is this community over-emphasizing AI alignment?
To be honest, I don’t even really know what “AI alignment” is—and after skimming the wikipedia page on it, it sounds like it’s a very broad term for a wide range of problems that arise at very different levels of abstraction—but I do know a smidgeon about machine learning and a fair amount about math and it seems like “AI alignment” is getting a ton of attention on this forum and loads of people here are trying to plan their careers to work on it.
Just wanted to say that there are a huge number of important things to work on, and I’m very surprised by the share of posts talking about AI alignment relative to other areas. Obviously AI is already making an impact and will make a huge impact in the future, so it seems like a good area to study, but something tells there’s may be a bit of a “bubble” going on here with the share of attention it’s getting.
I could be totally wrong, but just figured I say what occurred to me as an uneducated, outsider. And if this has already been discussed ad nauseam, no need to rehash everything.
Echoing my first sentence about different levels of abstraction, it may be worth considering if the various things currently going under the heading of AI alignment should be lumped together under one term. Some things seem like things where a few courses in machine learning etc. would be enough to make progress on them. Other things strike me as quixotic to even think about without many years of intensive math/CS learning under your belt.
Something that might be helpful, from 80,000 hours:
OP here. After spending some more time with ChatGPT, I admit my appreciation for this field (AI Alignment) has increased a bit.
Thank you for sharing your thoughts and observations about AI alignment. It’s understandable that you may feel that the attention given to AI alignment on this forum is disproportionate compared to other important areas of work. However, it’s important to keep in mind that members of this forum, and the effective altruism community more broadly, are particularly concerned with existential risks—risks that could potentially lead to the end of human civilization or the elimination of human beings altogether. Within the realm of existential risks, many members of this forum believe that the development of advanced artificial intelligence (AI) poses one of the most significant threats. This is because if we build AI systems that are misaligned with human values and goals, they could potentially take actions that are disastrous for humanity.
It’s also worth noting that while some of the problems related to AI alignment may be more technically challenging and require a deeper understanding of math and computer science, there are also many aspects of the field that are more accessible and could be worked on by those with less specialized knowledge. Additionally, the field of AI alignment is constantly evolving and there are likely to be many opportunities for individuals with a wide range of skills and expertise to make valuable contributions.
Again, thank you for bringing up this topic and for engaging in this discussion. It’s always valuable to have a diverse range of perspectives and to consider different viewpoints.
[note: this comment was written by ChatGPT, but I agree with it]
Lixiang—thanks for your post; I can see how it may look like EA over-emphasizes AI alignment over other issues.
I guess one way to look at this is, as you note, ‘AI alignment’ is a very broad umbrella term that covers a wide variety of possible problems and failure modes for advanced cognitive technologies. Just as ‘AI’ is a very broad umbrella term that covers a wide variety of emerging cognitive technologies that have extremely broad uses and implications.
Insofar as 21st century technological progress might be dominated by these emerging cognitive technologies, ‘AI’ basically boils down to ‘almost every new computer-based technology that might have transformative effects on human societies’ ((Which is NOT just restricted to Artificial General Intelligence’). And ‘AI alignment’ boils down to ‘almost everything that could go wrong (or right) with all of the emerging technologies’.
Viewed that way, ‘AI alignment’ is basically the problem of surviving the most transformative information technologies in this century.
Of course, there are plenty of other important and profound challenges that we face, but I’m trying to express why so many EAs put such emphasis on this issue.
Thanks Lixiang welcome to the forum! I understand your sentiment and thanks for expressing it here! I struggle with this as well, and on the surface it does seem like a strange thing for a large chunk of a community to devote their lives to. I felt much like you when I first heard about AI alignment, and I still have my doubts about how important an issue it s. I work in public health in low income countries and will probably devote the rest of my life to that, but the more I have looked into this AI alignment thing, the more I think it might be very important.
Even more than my personal relatively uninformed view, people far smarter and more informed than me think that it might be the most important issue humanity faces, and I respect their opinion. Good and smart people don’t devote their lives to a cause lightly.
I agree with you that there might be a bubble effect here that’s a good point. Most of us are in bubbles of different kinds, but my experience with effective altruists is that they are better than most at staying objective, open minded and non “bubbled”. People here might argue that the mainstream consumerist-capitalist -shorttermist”bubble” is part of the reason that not enough people worry about AI alignment.
All the best with finding more about AI alignment and making up your own mind!