Cosmic AI safety

A significant portion of our impact may occur through future interactions with, or influences on, civilizations or superpowers beyond Earth. Therefore, our potential impact on other cosmic actors plausibly deserves to be a key consideration in our decisions, and it plausibly deserves greater attention and priority than it receives currently.

In this essay, I will briefly seek to support and explore these claims. The core point I am making about devoting greater consideration to our potential effects on cosmic actors is not limited to AI safety. In principle, it applies more broadly, such as when it comes to how we devise and implement our institutions, norms, decision procedures, and even our values. However, the consideration becomes particularly weighty in contexts where our decisions are likely to influence a long-lasting value lock-in, and hence the consideration seems especially important in the context of AI safety.

(The term “cosmic AI safety” might sound overly grand or lofty, yet one could alternatively call it “non-parochial AI safety”: AI safety that does not exclude or neglect potential impacts on other cosmic actors.)

The ideas I explore in this essay are similar to ideas explored in Nick Bostrom’s recent draft “AI Creation and the Cosmic Host” as well as in Anders Sandberg’s “Game Theory with Aliens on the Largest Scales”, though my focus is somewhat different. I am not claiming to say anything particularly original here; I am simply trying to say something that seems to me important and neglected.

1. Why influence on cosmic actors seems plausible

There are various reasons and world models that suggest that future influence on cosmic actors is quite plausible.

1.1 Our apparent earliness

We have emerged after around 13.8 billion years in a universe that may support the emergence of life for hundreds of billions of years or more. This apparent earliness lends some plausibility — perhaps even a high probability — to the prospect of encountering powerful cosmic actors in the future, even if no such actors exist today.

To be sure, it is not clear whether we are in fact early, as this depends partly on which kinds of star systems can give rise to life (cf. Burnetti, 2016; 2017). Yet it still seems quite plausible that we have appeared early in the window in which the universe can support the emergence of life, which in itself seems a sufficient reason to take potential impacts on cosmic actors into account.

1.2 Grabby aliens

A more specific model that implies future influence on cosmic actors is the grabby aliens model (Olson, 2015; Hanson et al., 2021; Cook, 2022). The grabby aliens model is to some extent built on our apparent earliness, so it is not wholly independent of the point made above, yet it still helps to provide some added structure and salience as to what encounters with future cosmic actors might roughly look like, as can be seen in the video below.


In particular, the grabby aliens model implies a future in which any hypothetical Earth-originating cosmic actor would most likely encounter several other cosmic actors. This could potentially result in cooperation or conflict (or anything in between) depending in part on the actors’ respective aims and values.

Indeed, it is conceivable that such eventual “inter-cosmic-actor” encounters are the most consequential events in the cosmos from an ethical standpoint. By extension, it is conceivable that our nudges on such eventual cosmic encounters, even if only marginal, could ultimately be the most consequential part of our impact. (Though note that this is not my core claim here; my claim is about our potential impact on other cosmic actors in general — not the encounters of grabby aliens in particular — and my claim is merely that such potential impact is plausibly a key consideration.)

Additionally, as Robin Hanson has noted, even if humanity goes extinct and is not succeeded by expansionist AI, it is still conceivable that we could have a non-negligible influence on other grabby aliens. For instance, we might provide relevant information, and perhaps even serve as a coordination point, for other grabby civilizations that have each observed us but not yet encountered each other (Hanson, 2021).

1.3 Silent cosmic rulers

An alternative to the grabby aliens model is one that allows expansionist aliens to be quiet (cf. silent cosmic rulers). This model is different in some key respects. For example, on this model, we are likely to encounter just one cosmic superpower, and that superpower could already be here. In fact, if a notable fraction of expansionist aliens are quiet, such a quiet superpower would likely be here already (cf. the anthropic argument found here).

In such a scenario, there would all but surely be a vast power difference between them and humanity, yet there could still be a potential for meaningful human influence, including, and perhaps especially, through AI safety in particular. After all, given the potentially vast reach of such a hypothetical enveloping superpower, even just a marginal influence on that superpower could still be large in absolute terms.

1.4 Evidential cooperation in large worlds

On some decision theories, direct contact and causal influence are not required for the possibility of influencing and cooperating with other cosmic actors in some real sense. This is a key idea behind what has been termed “evidential cooperation in large worlds” — cooperation with causally disconnected actors elsewhere in the universe (Oesterheld, 2017; Treutlein, 2023). If such influence is possible, it is plausibly of great importance given its potential scope, including in the context of AI safety.

1.5 Other kinds of cosmic actors

In his draft on “AI Creation and the Cosmic Host”, Bostrom lists various other kinds of cosmic actors that could conceivably exist, including ones that fall outside a “conventional” naturalistic worldview, such as simulators and theistic gods. Even if we place a very low probability on the existence of these kinds of actors, it might still be a weak additional reason to give consideration to our potential impact on other cosmic actors, broadly construed.

2. Objections

2.1 This seems overly speculative

I sympathize strongly with this sentiment, and part of me indeed has a negative reaction to this entire topic for this very reason. That said, while many of the possibilities raised above are speculative, the probability that we can have a meaningful influence on cosmic actors in at least one of the ways outlined above does not seem extremely low. For example, it seems fairly likely that advanced life will be more prevalent in the future than it is today (e.g. since it will have had more time to emerge), and our existence at this point in time makes it seem plausible that powerful cosmic actors will be sufficiently prevalent to eventually run into each other (as suggested, for example, by the grabby aliens model).

2.2 This seems intractable

The tractability of influencing potential cosmic actors is admittedly unclear, yet this is hardly a good reason to dismiss the issue. On the contrary, given the apparent plausibility and the potential scale of such influence, we appear to have good reason to explore the issue further, so as to get a better sense of its tractability and how we may act wisely in light of this prospect.

3. Potential biases

It might be easy to neglect considerations about our potential impact on other cosmic agents, even for reasons that we ourselves do not endorse on reflection. For example, it may be easy to neglect such potential impact because it feels far away in time and space; such distance might make it feel less important, even if we at another level hold that what happens far away in time and space is not inherently less important than what happens nearby. In short, spatial and temporal biases may distort our assessments of the importance of this consideration.

Similarly, there might be hidden biases of partiality that distort our priorities (this can be a bias relative to a stated commitment to impartiality, or even just a commitment to some measure of impartiality). For instance, in the context of AI safety, we might be inclined to downplay considerations about far-future impacts beyond Earth, and to instead focus exclusively on, say, maximizing the chances of our own survival, or reducing human suffering. And while aims like “maximize the chances of our own survival” and “create the best outcomes over all time and space” may be convergent to some extent, it would be extremely surprising if they were anything close to perfectly convergent.

Impartial AI safety would plausibly give strong consideration to our potential impact on other cosmic agents, whereas AI safety that exclusively prioritizes, say, human survival or human suffering reduction would probably not give it strong consideration, if indeed any consideration at all. So the further we diverge from ideals of impartiality in our practical focus, the more likely we may be to neglect our potential impact on other cosmic agents.

4. What can we do?

In broad terms, perhaps the main thing we can do at this point to improve our expected impact on other cosmic actors is to have more of a focus on such potential impact. That is, it may be helpful to try to make such potential impact more salient, to orient more toward it, to do more research on it, and so on. In particular, it seems worth giving greater prominence to this consideration when we seek to design or otherwise shape future AI systems, such as by exploring how future AI systems could be better optimized for potential interactions with other cosmic actors.

Relatedly, it might be helpful to look at familiar problems and focus areas through the lens of “potential impact on cosmic actors”. For instance, how can we best reduce risks of astronomical future suffering (s-risks) if most of our impact will occur through, or in interaction with, other cosmic actors? This question could be explored for a variety of potential cosmic actors and types of influence.

In addition, as Bostrom also stresses, it may be worth adopting an attitude of humility, such as by updating toward an attitude of cosmic considerateness, as well as by being “modest, willing to listen and learn” (Bostrom, 2022, sec. 39d; 2024).

Those loose suggestions notwithstanding, it is very much an open question how to improve our expected impact on other cosmic actors. More research is needed.

5. Potential impact on cosmic actors as an opportunity

Finally, it is worth noting that while the prospect of influencing other cosmic actors might feel dizzyingly abstract and perhaps even overwhelming, it can also be viewed as an opportunity and as something that can be approached in concrete and realistic steps. After all, if our counterfactual influence on other cosmic actors represents a large fraction of our expected impact, it would indeed, almost by definition, be a massive opportunity.