EAs seeking to maximize their expected impact should reconsider working on AI safety directly.
I am not arguing that working on AI safety is bad, but that we don’t know how effective working in this area might be.
Adopting a worldview diversification strategy might offer a more robust approach to address this issue, especially for early-career individuals.
In addition to diversifying on a community level, there are two promising pathways to achieve this: pursuing a meta career and earning to give.
Disclaimer: This post challenges widely held beliefs in EA and may feel unsettling. If you’re currently facing personal difficulties, consider reading it at another time. The reason I am sharing this post is because I think it might save you many hours, help you realize your mission as an Effective Altruist more effectively, and because I wish someone had shared this with me before I started working on AI safety.
Introduction
If you are reading this, it is likely that you subscribe to the core principles of EA and that you have spent a decent amount of your time figuring out how to use your career to maximise your expected impact. Figuring out how to do that isn’t easy and it takes agency, an open mind and the courage to think again. Thank you for being willing to do this.
Like me, you likely have ended up concluding that working on AI safety is one of the best ways to do the most good. Your thought process might have been quite similar to mine:
Since the future could be so vast, we ought to put a large amount of resources into ensuring that it goes well. This future should be filled with many happy beings who are free of unnecessary suffering. Also, we could create a radically better state for humanity and other sentient beings if we only have enough time.
If we go extinct in the process, we lose our entire potential and everything the universe could have become.
Hence, we ought to buy ourselves enough time to figure things out and uphold our potential.
As a next step, you likely thought about which cause to work on to contribute to this goal. Biosecurity, climate change, AI safety, great power conflict.
It turns out that I considered it most likely for humanity to be eradicated through a rogue ASI. This is because there is a large difference between 99% of humanity dying through a nuclear disaster and 100% through a rogue AI. With 1% or 0.1% surviving, our potential isn’t destroyed. And then there is still the chance of a suffering risk.
Hence, from an expected impact maximization perspective, working on AI safety, with the aim of reducing existential risk or preventing an s-risk, seems like the single best option.
Throughout the last two years of having only worked on AI safety, and having thought a lot about it, I have updated toward not being certain about this at all anymore. In fact, I think there are good reasons to believe that reducing x-risk might be actively harmful.
There are two reasons that I want to focus on to illustrate why the expected value of working on AI safety isn’t obviously good.
1. Complex Cluelessness
The first major source of doubt, even for those who do believe in the moral salience of preserving humanity, is the sheer complexity of the future. Accurately predicting downstream consequences of actions that attempt to positively steer the long-term future is notoriously difficult, if not impossible. An initiative that seems beneficial today can have wild, unintended ramifications. If you help ensure humanity’s survival and contribute to reducing existential risk, you may also inadvertently help proliferate factory farming deep into the universe, environmental devastation across multiple planets, or even countless digital simulations that generate enormous suffering. One concrete example is Open Philanthropy’s initial decision to support OpenAI with millions of dollars, which may have inadvertently increased the risk of extinction by laying the foundation for a now-soon-to-be-for-profit company, with unclear moral alignment. This is just one example that shows that it is often very difficult to predict the influence of your actions, even in the relative short-term.
For a more detailed critique of x-risk reduction efforts being positive in expectation, I recommend reading this post. If you want to learn more about complex cluelessness, you can take a look at this page.
2. Moral Uncertainty
The second significant reason for doubting the value of AI safety and interventions aimed at reducing existential risk is moral uncertainty. What if the moral framework justifying the preservation of humanity’s future at any cost is not the correct one? There are many moral frameworks, and it seems unreasonable to categorically dismiss certain perspectives simply because they “don’t resonate.”
Even within utilitarianism, the divergence between standard and negative utilitarian perspectives is particularly relevant. Traditional utilitarianism balances the creation of happiness with the reduction of suffering, while negative utilitarianism exclusively prioritizes reducing suffering. This distinction can lead to vastly different conclusions.
Consider this: today, an estimated 100 billion animals are slaughtered every year, the vast majority in factory farms. If these beings experience pain and suffering, it’s plausible that, should humanity colonize space and pursue protein for a “balanced diet,” we could perpetuate animal suffering on an astronomical scale.
You might argue that humanity will expand its moral circle or develop cost-effective lab-grown meat. However, even if the chance of this not happening is as low as 10%, and you give some weight to utilitarianism, the expected value of AI safety’s contribution to reducing existential risk still isn’t clear. And it might even be extremely negative.
Here is a calculation conducted by OpenAI o1 to provide an example of how you might arrive at the conclusion that, both under classical and negative utilitarianism, space colonisation could be seen negatively. This does not even account for other worldviews or the argument that creating happy beings might not outweigh the negative experiences present in the universe (elaboration).
How a worldview diversification strategy can address this issue
To be clear, I am not arguing that AI safety is definitively bad. It might turn out to be the single most valuable pursuit ever. However, I am arguing that we simply don’t know, given moral uncertainty and complex cluelessness. A broad outcome distribution indicates that working on AI safety could bring extremely positive outcomes (e.g., preventing extinction and enabling flourishing) or extremely negative outcomes (e.g., perpetuating astronomical suffering), but we currently lack reliable estimates of how probable each scenario is.
We’re unsure how large each tail is. We’re also unsure how to weigh outcomes across different moral frameworks—for instance, how to compare a small probability of vast suffering against a larger probability of moderate flourishing.
The existence of many plausible yet conflicting worldviews and the complexity of future causal chains mean that any “best guess” at the EV is laden with extreme uncertainties.
Using the INTP framework (importance, neglectedness, tractability, and personal fit), this essentially means reconsidering how we evaluate causes to work on. Specifically, when selecting a cause—particularly in the context of longtermist causes—we should place significantly less emphasis on the “importance” domain due to moral uncertainty. Similarly, we should de-emphasize the “tractability” domain because of complex cluelessness.
In light of this uncertainty, a promising strategy to hedge against it is a worldview diversification strategy (WDS). A WDS means allocating resources—time, money, influence—across multiple plausible causes and moral frameworks, proportional to how likely you think each framework might be correct and how large the stakes appear.
WDS seems to maximize expected impact by balancing the risks of over-committing to a single worldview against the benefits of spreading resources across multiple plausible frameworks, thereby ensuring that we don’t overlook substantial potential value from worldviews we may be underestimating.
“When accounting for strong uncertainty and diminishing returns, worldview diversification can maximize expected value even when one worldview looks “better” than the others in expectation. One way of putting this is that if we were choosing between 10 worldviews, and one were 5x as good as the other nine, investing all our resources in that one would – at the relevant margin, due to the “diminishing returns” point – be worse than spreading across the ten.”
What are the practical implications?
Logistically, there are three main ways to pursue a diversification strategy with your career:
1. Meta career
This involves working on interventions that enhance the overall effectiveness of the EA community or other high-impact initiatives. Examples include: Global priorities research, EA community building, Improving philanthropic decision-making, Consulting to boost the productivity of organizations, and more.
2. Earning to give
This means pursuing a job with high-earning potential and donating a significant portion of your income to fund impactful work. By doing so, you can directly support critical projects and individuals working on high-priority issues.
3. Specializing in a specific cause
This option involves, ironically, committing to a particular cause, such as AI safety. While this doesn’t diversify your individual efforts, you contribute to the broader community-level diversification strategy pursued across the EA movement. This approach combines the advantages of having specialized experts in each area with the broad coverage needed to navigate uncertainty.
Which of these paths should you pursue?
Right from the start, I want to emphasize that this is not intended to be career advice. Instead, it aims to help you explore different possibilities and considerations in your cause prioritization while encouraging you to think through various perspectives.
I highly recommend forming your own view and treating this as supplementary material, since this list is certainly not exhaustive.
Meta Career
While every pathway is unique, the general idea behind a meta career is to work on interventions that increase the resources available for more effective diversification or cause prioritization, either globally or within the EA community.
Pros
You don’t need to commit to a specific set of worldviews, allowing you to diversify your impact across multiple cause areas. This helps address challenges like moral uncertainty and complex cluelessness.
Even beyond EA, you might end up supporting projects or organizations outside the typical EA canon.
You are likely to work with value-aligned people, which can be personally rewarding.
Cons
You have less control over how your resources are being diversified and may struggle to account for varying credences in different worldviews.
Large coordination or meta initiatives run the risk of creating single points of failure—for instance, if the entire movement’s strategic direction is steered by just a few meta organizations or research agendas.
If you don’t consider extinction through AGI as a likely scenario, some meta pathways may be significantly less impactful due to automation risks (e.g., Global Priorities Research may be more at risk of automation than EA Community Building).
Further Reading
Explore more about specific meta pathways and their pros and cons through these resources:
Specialising in a specific cause (using AI safety as an example[1])
Pros
You can build in-depth expertise and become a key knowledge resource in the field, potentially enabling you to have a very large impact.
Working on globally neglected areas, like AI safety, involves addressing issues of potentially enormous importance. There are many plausible scenarios where this work makes sense. Even with low tractability, the EV often remains high due to the large stakes and the worldview bet involved.
Collaborating with value-aligned individuals in a community with shared worldviews provides a sense of belonging and the opportunity to contribute to something larger than yourself.
You can more easily impress people in your social circles by tackling a global problem with big consequences.
Cons
Committing to a specific worldview means that once you’ve built substantial specialized capital (credentials, networks, reputation), it can become psychologically and professionally challenging to pivot away, even if your views about its importance or tractability change.
By committing to a specific pathway, you are more susceptible to confirmation bias and the sunk-cost fallacy when encountering new evidence that should prompt an update. This can make it harder—similar to what happened with Yann LeCun—to reconsider or move away from the cause.
It is generally very difficult to understand whether you are even contributing to the objectives of x-risk reduction, or whether you are making things worse (Open Philanthropy example) [2].
Your mental health could suffer from losing hope in the prospects of AI safety or feeling paralyzed by uncertainty about what to work on. This, for obvious reasons, may reduce your effectiveness compared to pursuing meta work or earning to give.
When you should consider choosing / continuing this path:
You believe it makes sense to place a bet on worldviews where AI safety and x-risk reduction are good in expectation.
Short timelines and an established position (or strong personal fit for this cause) make the cost of pivoting too high.
You think that in scenarios where AGI is successfully aligned and scaled to safe superintelligence, your ability to influence outcomes will be minimal, or that global problems will be resolved very quickly (hence, reducing the need for earning to give).
Earning to give
Pros
You can easily adjust your giving proportions over time as new evidence emerges and the world changes. This allows you to fund work you are most confident in or to support organizations that specialize in cause prioritization, which can allocate your donations effectively [3].
A wide variety of high-income career options increases your chances of finding a path that you can excel at.
In worlds where there isn’t a fast takeoff from AGI to ASI (and you have the necessary capital), you might contribute more effectively to direct causes like AI safety by outsourcing work to highly intelligent AI agents.
Cons
Personal wealth accumulation may lead to value drift, reducing the likelihood of donating significant amounts or maintaining commitment to giving.
If you have short AGI timelines and a high p(doom), starting with earning to give is unlikely to impact cause areas in the short term unless you already hold a high-paying job.
You are less likely to work alongsidevalue-aligned people with shared aspirations of “doing good better,” which can be demotivating.
In worlds where AGI and ASI are successfully aligned, large donations may have little impact because global problems might be solved rapidly.
When you should consider choosing this path:
You are highly uncertain about the value of specific cause areas and don’t find it appealing to contribute to diversification at a community level (i.e., working on something when you’re uncertain about its impact).
You believe you can diversify more effectively by donating to organizations like Open Philanthropy, which can allocate funds on your behalf, especially given the challenges of moral uncertainty.
You have medium to long-term AGI timelines or a low p(doom), or you’re uncertain about the timelines and risks altogether.
You have a strong personal fit for a high-earning career pathway and feel more motivated to pursue this compared to working on direct causes or meta-level interventions.
An important consideration
No path is guaranteed to have a positive impact, and each has their downsides. There does not seem to be a clear winner among these three options. Much depends on your personal fit and how well you think you could excel in a given pathway.
Having said that, I do want to share one possibility that seems quite plausible to me.
Progress in AI shows no signs of slowing down or hitting a wall. In fact, it’s highly plausible that progress will accelerate with the new paradigm of scaling inference-time compute. In just a few years, we will (in my estimation) see AI systems capable of performing nearly any form of intellectual labor more effectively than humans. Assuming humanity survives and remains in control, real-world results will likely be achieved more effectively through capital and financial resources rather than trying to compete with hyperintelligent AI systems. Please refer to this post for an elaboration on this point.[4] This would be a strong argument in favour of pursuing earning to give.
Final Reflections
AI safety is often seen as a top priority in the EA community, but its expected value is far from certain due to moral uncertainty and complex cluelessness. While working on AI safety contributes to the community’s broader diversification strategy, other pathways—such as specialising in more neglected causes, meta careers or earning to give—may offer equally or even more promising ways to increase your impact.
Your personal fit plays a crucial role in determining the best path for you and given the arguments laid out in this post, I encourage you to weigh these more heavily as you might have done before.
Cause prioritisation isn’t easy, and I bet that I haven’t made it easier with this post. However, being an EA isn’t easy, and my goal wasn’t to provide easy answers but to provoke deeper reflection and encourage a more robust exploration of these ideas. By questioning common assumptions I hope we can collectively navigate the uncertainties and complexities to make better-informed decisions about where to focus our efforts.
I invite you to critically engage with these reflections and share your thoughts. I would love to explore together with you how we can increase our impact in the face of profound moral uncertainty and cluelessness.
Thank you for reading! I am very grateful to be part of a community that takes the implications of ideas seriously.
Invitation to a Discussion Session
Because this is a nuanced topic—one with truly high stakes—I am thinking of organising at least one online discussion where we can explore these ideas together. Anyone is welcome, whether you’re convinced, on the fence, or firmly opposed.
I am grateful to Jan Wehner, Leon Lang, Josh Thorsteinson and Jelle Donders for providing thoughtful and critical feedback on this post. Their insights were very valuable in challenging my ideas and helping me refine my writing. While their feedback played a significant role in improving this work, it’s important to note that they may not endorse the views expressed here. Any remaining errors or shortcomings are entirely my own.
A general note for people still choosing their career path:
If you accept the arguments of moral uncertainty and complex cluelessness, and you recognize that the overall importance of AI safety remains unclear after considering these points, relying solely on the community diversification strategy to justify continuing work in AI safety becomes less compelling. The “importance” domain should carry less weight in your decision-making, while neglectedness and personal fit should be given greater emphasis. Assuming an equal level of personal fit, this could lead you to favor pivoting toward areas that are more neglected or where progress is easier to achieve. Within the EA community, AI safety already receives substantial attention and resources.
However, the argument for diversification at a broader community level also applies within AI safety, offering greater assurance that your work contributes to the overall diversification of the AI safety community, even if you can’t be sure about how good the specific thing is you are working on.
The development of AGI will likely introduce many new crucial new considerations that render current diversification efforts within the EA community less effective. In such scenarios, you could have a greater impact by funding neglected and important work or alternative approaches to a cause.
A counterargument to this reasoning is that it is highly uncertain whether you would be able to significantly influence society post-AGI, even with substantial capital. Speculating about the state of affairs once AGI is developed is inherently difficult. Furthermore, if we succeed in building aligned ASI, it is unclear to what extent humans would be needed to solve global problems, as ASI might handle these tasks far more effectively.
Rethinking the Value of Working on AI Safety
Summary
EAs seeking to maximize their expected impact should reconsider working on AI safety directly.
I am not arguing that working on AI safety is bad, but that we don’t know how effective working in this area might be.
Adopting a worldview diversification strategy might offer a more robust approach to address this issue, especially for early-career individuals.
In addition to diversifying on a community level, there are two promising pathways to achieve this: pursuing a meta career and earning to give.
Disclaimer: This post challenges widely held beliefs in EA and may feel unsettling. If you’re currently facing personal difficulties, consider reading it at another time. The reason I am sharing this post is because I think it might save you many hours, help you realize your mission as an Effective Altruist more effectively, and because I wish someone had shared this with me before I started working on AI safety.
Introduction
If you are reading this, it is likely that you subscribe to the core principles of EA and that you have spent a decent amount of your time figuring out how to use your career to maximise your expected impact. Figuring out how to do that isn’t easy and it takes agency, an open mind and the courage to think again. Thank you for being willing to do this.
Like me, you likely have ended up concluding that working on AI safety is one of the best ways to do the most good. Your thought process might have been quite similar to mine:
Since the future could be so vast, we ought to put a large amount of resources into ensuring that it goes well. This future should be filled with many happy beings who are free of unnecessary suffering. Also, we could create a radically better state for humanity and other sentient beings if we only have enough time.
If we go extinct in the process, we lose our entire potential and everything the universe could have become.
Hence, we ought to buy ourselves enough time to figure things out and uphold our potential.
As a next step, you likely thought about which cause to work on to contribute to this goal. Biosecurity, climate change, AI safety, great power conflict.
It turns out that I considered it most likely for humanity to be eradicated through a rogue ASI. This is because there is a large difference between 99% of humanity dying through a nuclear disaster and 100% through a rogue AI. With 1% or 0.1% surviving, our potential isn’t destroyed. And then there is still the chance of a suffering risk.
Hence, from an expected impact maximization perspective, working on AI safety, with the aim of reducing existential risk or preventing an s-risk, seems like the single best option.
Throughout the last two years of having only worked on AI safety, and having thought a lot about it, I have updated toward not being certain about this at all anymore. In fact, I think there are good reasons to believe that reducing x-risk might be actively harmful.
There are two reasons that I want to focus on to illustrate why the expected value of working on AI safety isn’t obviously good.
1. Complex Cluelessness
The first major source of doubt, even for those who do believe in the moral salience of preserving humanity, is the sheer complexity of the future. Accurately predicting downstream consequences of actions that attempt to positively steer the long-term future is notoriously difficult, if not impossible. An initiative that seems beneficial today can have wild, unintended ramifications. If you help ensure humanity’s survival and contribute to reducing existential risk, you may also inadvertently help proliferate factory farming deep into the universe, environmental devastation across multiple planets, or even countless digital simulations that generate enormous suffering. One concrete example is Open Philanthropy’s initial decision to support OpenAI with millions of dollars, which may have inadvertently increased the risk of extinction by laying the foundation for a now-soon-to-be-for-profit company, with unclear moral alignment. This is just one example that shows that it is often very difficult to predict the influence of your actions, even in the relative short-term.
For a more detailed critique of x-risk reduction efforts being positive in expectation, I recommend reading this post. If you want to learn more about complex cluelessness, you can take a look at this page.
2. Moral Uncertainty
The second significant reason for doubting the value of AI safety and interventions aimed at reducing existential risk is moral uncertainty. What if the moral framework justifying the preservation of humanity’s future at any cost is not the correct one? There are many moral frameworks, and it seems unreasonable to categorically dismiss certain perspectives simply because they “don’t resonate.”
Even within utilitarianism, the divergence between standard and negative utilitarian perspectives is particularly relevant. Traditional utilitarianism balances the creation of happiness with the reduction of suffering, while negative utilitarianism exclusively prioritizes reducing suffering. This distinction can lead to vastly different conclusions.
Consider this: today, an estimated 100 billion animals are slaughtered every year, the vast majority in factory farms. If these beings experience pain and suffering, it’s plausible that, should humanity colonize space and pursue protein for a “balanced diet,” we could perpetuate animal suffering on an astronomical scale.
You might argue that humanity will expand its moral circle or develop cost-effective lab-grown meat. However, even if the chance of this not happening is as low as 10%, and you give some weight to utilitarianism, the expected value of AI safety’s contribution to reducing existential risk still isn’t clear. And it might even be extremely negative.
Here is a calculation conducted by OpenAI o1 to provide an example of how you might arrive at the conclusion that, both under classical and negative utilitarianism, space colonisation could be seen negatively. This does not even account for other worldviews or the argument that creating happy beings might not outweigh the negative experiences present in the universe (elaboration).
How a worldview diversification strategy can address this issue
To be clear, I am not arguing that AI safety is definitively bad. It might turn out to be the single most valuable pursuit ever. However, I am arguing that we simply don’t know, given moral uncertainty and complex cluelessness. A broad outcome distribution indicates that working on AI safety could bring extremely positive outcomes (e.g., preventing extinction and enabling flourishing) or extremely negative outcomes (e.g., perpetuating astronomical suffering), but we currently lack reliable estimates of how probable each scenario is.
We’re unsure how large each tail is. We’re also unsure how to weigh outcomes across different moral frameworks—for instance, how to compare a small probability of vast suffering against a larger probability of moderate flourishing.
The existence of many plausible yet conflicting worldviews and the complexity of future causal chains mean that any “best guess” at the EV is laden with extreme uncertainties.
Using the INTP framework (importance, neglectedness, tractability, and personal fit), this essentially means reconsidering how we evaluate causes to work on. Specifically, when selecting a cause—particularly in the context of longtermist causes—we should place significantly less emphasis on the “importance” domain due to moral uncertainty. Similarly, we should de-emphasize the “tractability” domain because of complex cluelessness.
In light of this uncertainty, a promising strategy to hedge against it is a worldview diversification strategy (WDS). A WDS means allocating resources—time, money, influence—across multiple plausible causes and moral frameworks, proportional to how likely you think each framework might be correct and how large the stakes appear.
WDS seems to maximize expected impact by balancing the risks of over-committing to a single worldview against the benefits of spreading resources across multiple plausible frameworks, thereby ensuring that we don’t overlook substantial potential value from worldviews we may be underestimating.
A section in Holden Karnofsky’s post alludes to this quite well:
“When accounting for strong uncertainty and diminishing returns, worldview diversification can maximize expected value even when one worldview looks “better” than the others in expectation. One way of putting this is that if we were choosing between 10 worldviews, and one were 5x as good as the other nine, investing all our resources in that one would – at the relevant margin, due to the “diminishing returns” point – be worse than spreading across the ten.”
What are the practical implications?
Logistically, there are three main ways to pursue a diversification strategy with your career:
1. Meta career
This involves working on interventions that enhance the overall effectiveness of the EA community or other high-impact initiatives. Examples include: Global priorities research, EA community building, Improving philanthropic decision-making, Consulting to boost the productivity of organizations, and more.
2. Earning to give
This means pursuing a job with high-earning potential and donating a significant portion of your income to fund impactful work. By doing so, you can directly support critical projects and individuals working on high-priority issues.
3. Specializing in a specific cause
This option involves, ironically, committing to a particular cause, such as AI safety. While this doesn’t diversify your individual efforts, you contribute to the broader community-level diversification strategy pursued across the EA movement. This approach combines the advantages of having specialized experts in each area with the broad coverage needed to navigate uncertainty.
Which of these paths should you pursue?
Right from the start, I want to emphasize that this is not intended to be career advice. Instead, it aims to help you explore different possibilities and considerations in your cause prioritization while encouraging you to think through various perspectives.
I highly recommend forming your own view and treating this as supplementary material, since this list is certainly not exhaustive.
Meta Career
While every pathway is unique, the general idea behind a meta career is to work on interventions that increase the resources available for more effective diversification or cause prioritization, either globally or within the EA community.
Pros
You don’t need to commit to a specific set of worldviews, allowing you to diversify your impact across multiple cause areas. This helps address challenges like moral uncertainty and complex cluelessness.
Even beyond EA, you might end up supporting projects or organizations outside the typical EA canon.
You are likely to work with value-aligned people, which can be personally rewarding.
Cons
You have less control over how your resources are being diversified and may struggle to account for varying credences in different worldviews.
Large coordination or meta initiatives run the risk of creating single points of failure—for instance, if the entire movement’s strategic direction is steered by just a few meta organizations or research agendas.
If you don’t consider extinction through AGI as a likely scenario, some meta pathways may be significantly less impactful due to automation risks (e.g., Global Priorities Research may be more at risk of automation than EA Community Building).
Further Reading
Explore more about specific meta pathways and their pros and cons through these resources:
Global Priorities Research
Building Effective Altruism
Founding Impactful Projects
Specialising in a specific cause (using AI safety as an example[1])
Pros
You can build in-depth expertise and become a key knowledge resource in the field, potentially enabling you to have a very large impact.
Working on globally neglected areas, like AI safety, involves addressing issues of potentially enormous importance. There are many plausible scenarios where this work makes sense. Even with low tractability, the EV often remains high due to the large stakes and the worldview bet involved.
Collaborating with value-aligned individuals in a community with shared worldviews provides a sense of belonging and the opportunity to contribute to something larger than yourself.
You can more easily impress people in your social circles by tackling a global problem with big consequences.
Cons
Committing to a specific worldview means that once you’ve built substantial specialized capital (credentials, networks, reputation), it can become psychologically and professionally challenging to pivot away, even if your views about its importance or tractability change.
By committing to a specific pathway, you are more susceptible to confirmation bias and the sunk-cost fallacy when encountering new evidence that should prompt an update. This can make it harder—similar to what happened with Yann LeCun—to reconsider or move away from the cause.
It is generally very difficult to understand whether you are even contributing to the objectives of x-risk reduction, or whether you are making things worse (Open Philanthropy example) [2].
Your mental health could suffer from losing hope in the prospects of AI safety or feeling paralyzed by uncertainty about what to work on. This, for obvious reasons, may reduce your effectiveness compared to pursuing meta work or earning to give.
When you should consider choosing / continuing this path:
You believe it makes sense to place a bet on worldviews where AI safety and x-risk reduction are good in expectation.
Short timelines and an established position (or strong personal fit for this cause) make the cost of pivoting too high.
You think that in scenarios where AGI is successfully aligned and scaled to safe superintelligence, your ability to influence outcomes will be minimal, or that global problems will be resolved very quickly (hence, reducing the need for earning to give).
Earning to give
Pros
You can easily adjust your giving proportions over time as new evidence emerges and the world changes. This allows you to fund work you are most confident in or to support organizations that specialize in cause prioritization, which can allocate your donations effectively [3].
A wide variety of high-income career options increases your chances of finding a path that you can excel at.
In worlds where there isn’t a fast takeoff from AGI to ASI (and you have the necessary capital), you might contribute more effectively to direct causes like AI safety by outsourcing work to highly intelligent AI agents.
Cons
Personal wealth accumulation may lead to value drift, reducing the likelihood of donating significant amounts or maintaining commitment to giving.
If you have short AGI timelines and a high p(doom), starting with earning to give is unlikely to impact cause areas in the short term unless you already hold a high-paying job.
You are less likely to work alongside value-aligned people with shared aspirations of “doing good better,” which can be demotivating.
In worlds where AGI and ASI are successfully aligned, large donations may have little impact because global problems might be solved rapidly.
When you should consider choosing this path:
You are highly uncertain about the value of specific cause areas and don’t find it appealing to contribute to diversification at a community level (i.e., working on something when you’re uncertain about its impact).
You believe you can diversify more effectively by donating to organizations like Open Philanthropy, which can allocate funds on your behalf, especially given the challenges of moral uncertainty.
You have medium to long-term AGI timelines or a low p(doom), or you’re uncertain about the timelines and risks altogether.
You have a strong personal fit for a high-earning career pathway and feel more motivated to pursue this compared to working on direct causes or meta-level interventions.
An important consideration
No path is guaranteed to have a positive impact, and each has their downsides. There does not seem to be a clear winner among these three options. Much depends on your personal fit and how well you think you could excel in a given pathway.
Having said that, I do want to share one possibility that seems quite plausible to me.
Progress in AI shows no signs of slowing down or hitting a wall. In fact, it’s highly plausible that progress will accelerate with the new paradigm of scaling inference-time compute. In just a few years, we will (in my estimation) see AI systems capable of performing nearly any form of intellectual labor more effectively than humans. Assuming humanity survives and remains in control, real-world results will likely be achieved more effectively through capital and financial resources rather than trying to compete with hyperintelligent AI systems. Please refer to this post for an elaboration on this point.[4] This would be a strong argument in favour of pursuing earning to give.
Final Reflections
AI safety is often seen as a top priority in the EA community, but its expected value is far from certain due to moral uncertainty and complex cluelessness. While working on AI safety contributes to the community’s broader diversification strategy, other pathways—such as specialising in more neglected causes, meta careers or earning to give—may offer equally or even more promising ways to increase your impact.
Your personal fit plays a crucial role in determining the best path for you and given the arguments laid out in this post, I encourage you to weigh these more heavily as you might have done before.
Cause prioritisation isn’t easy, and I bet that I haven’t made it easier with this post. However, being an EA isn’t easy, and my goal wasn’t to provide easy answers but to provoke deeper reflection and encourage a more robust exploration of these ideas. By questioning common assumptions I hope we can collectively navigate the uncertainties and complexities to make better-informed decisions about where to focus our efforts.
I invite you to critically engage with these reflections and share your thoughts. I would love to explore together with you how we can increase our impact in the face of profound moral uncertainty and cluelessness.
Thank you for reading! I am very grateful to be part of a community that takes the implications of ideas seriously.
Invitation to a Discussion Session
Because this is a nuanced topic—one with truly high stakes—I am thinking of organising at least one online discussion where we can explore these ideas together. Anyone is welcome, whether you’re convinced, on the fence, or firmly opposed.
If you would like to be notified about this, please share with me your email address here https://forms.gle/hKv6A9jDS8dSiRrk8
Acknowledgements
I am grateful to Jan Wehner, Leon Lang, Josh Thorsteinson and Jelle Donders for providing thoughtful and critical feedback on this post. Their insights were very valuable in challenging my ideas and helping me refine my writing. While their feedback played a significant role in improving this work, it’s important to note that they may not endorse the views expressed here. Any remaining errors or shortcomings are entirely my own.
A general note for people still choosing their career path:
If you accept the arguments of moral uncertainty and complex cluelessness, and you recognize that the overall importance of AI safety remains unclear after considering these points, relying solely on the community diversification strategy to justify continuing work in AI safety becomes less compelling. The “importance” domain should carry less weight in your decision-making, while neglectedness and personal fit should be given greater emphasis. Assuming an equal level of personal fit, this could lead you to favor pivoting toward areas that are more neglected or where progress is easier to achieve. Within the EA community, AI safety already receives substantial attention and resources.
However, the argument for diversification at a broader community level also applies within AI safety, offering greater assurance that your work contributes to the overall diversification of the AI safety community, even if you can’t be sure about how good the specific thing is you are working on.
The development of AGI will likely introduce many new crucial new considerations that render current diversification efforts within the EA community less effective. In such scenarios, you could have a greater impact by funding neglected and important work or alternative approaches to a cause.
A counterargument to this reasoning is that it is highly uncertain whether you would be able to significantly influence society post-AGI, even with substantial capital. Speculating about the state of affairs once AGI is developed is inherently difficult. Furthermore, if we succeed in building aligned ASI, it is unclear to what extent humans would be needed to solve global problems, as ASI might handle these tasks far more effectively.