Existential risks are not just about humanity
By MichaelAđ¸ @ 2020-04-28T00:09 (+35)
This post was written for Convergence Analysis. This post highlights and analyses existing ideas more than proposing new ones.
In The Precipice, Toby Ord writes:
An existential catastrophe is the destruction of humanityâs longterm potential.
An existential risk is a risk that threatens the destruction of humanityâs longterm potential.
Iâve previously discussed some distinctions and nuances relevant to these concepts. This post will focus on:
- The idea that these concepts are really about the destruction of the potential of humanity or its âdescendantsâ; they're not necessarily solely about human wellbeing, nor just Homo sapiensâ potential.
- The implications of that, including for how âbadâ an existential catastrophe might be
The potential of humanity and its âdescendantsâ
When explaining his definitions, Ord writes:
my focus on humanity in the definitions is not supposed to exclude considerations of the value of the environment, other animals, successors to Homo sapiens, or creatures elsewhere in the cosmos. It is not that I think only humans count. Instead, it is that humans are the only beings we know of that are responsive to moral reasons and moral argument - the beings who can examine the world and decide to do what is best. If we fail, that upwards force, that capacity to push towards what is best or what is just, will vanish from the world.
Our potential is a matter of what humanity can achieve through the combined actions of each and every human. The value of our actions will stem in part from what we do to and for humans, but it will depend on the effects of our actions on non-humans too. If we somehow give rise to new kinds of moral agents in the future, the term âhumanityâ in my definition should be taken to include them.
This makes two points clear:
- An existential catastrophe is not solely about the destruction of the potential for human welfare, flourishing, achievement, etc. Instead, itâs about humanityâs potential to bring about or protect whatever turns out to be of value. This may include, among many other things, the welfare of other beings.
- More specifically, itâs about the potential of humanity or its âdescendantsâ,[1] not just âHomo sapiensâ, to bring about or protect whatever turns out to be of value.
In line with that second point, Bostromâs (2012) definition is:
An existential risk is one that threatens the premature extinction of Earth-originating intelligent life or the permanent and drastic destruction of its potential for desirable future development (emphasis added)
And Bostrom also writes:
Above, we defined "humanity" as Earth-originating intelligent life rather than as the particular biologically defined species Homo sapiens. The reason for focusing the notion of existential risk on this broader concept is that there is no reason to suppose that the biological species concept tracks what we have reason to value. If our species were to evolve, or use technology to self-modify, to such an extent that it no longer satisfied the biological criteria for species identity (such as interbreedability) with contemporary Homo sapiens, this need not be in any sense a catastrophe.
If we wished to more explicitly capture the above two points in our definitions, we could expand Ordâs definitions to:
An existential catastrophe is the destruction of the long-term potential humanity (or its âdescendantsâ) has to cause morally valuable outcomes.
An existential risk is a risk that threatens the destruction of the long-term potential humanity (or its âdescendantsâ) has to cause morally valuable outcomes.[2]
However, these tweaks would also make the definitions longer, perhaps âweirderâ sounding, and perhaps less emotionally resonant. So Iâm not suggesting that theyâre all-things-considered improvements to Ordâs definitions.
Hereâs another option that might avoid those issues, while capturing the two points noted above:
An existential catastrophe is the destruction of the long-term potential for value in the universe.
An existential risk is a risk that threatens the destruction of the long-term potential for value in the universe.
But I donât think those definitions would quite work, for reasons explained in the following section.
What about moral agents other than humanity or its âdescendantsâ?
As noted, Ord focuses on humanity in his definitions because âhumans are the only beings we know of that are responsive to moral reasons and moral argumentâ (emphasis added). And he adds âIf we somehow give rise to new kinds of moral agents in the future, the term âhumanityâ in my definition should be taken to include them.â
But it seems plausible that there are now, or will be in future, other âmoral agentsâ which developed or will develop independently of us. I see three ways this could occur.
Firstly, as Bostrom (2009) notes:
It is possible that if humanity goes extinct, another intelligent species might evolve on Earth to fill the vacancy. The fate of such a possible future substitute species, however, would not strictly be part of the future of humanity.
Secondly, it seems possible that roughly the same thing could occur if humanity doesnât go extinct but for some reason fully departs the Earth, despite it remaining habitable. (Incidentally, these possibilities for âEarth-originating intelligent lifeâ which arises independently of humanity is why I use the phrase âhumanity or its âdescendantsââ when discussing existential risks, instead of Bostromâs âEarth-originating intelligent lifeâ.)
Thirdly, it seems possible that there currently are, or in future will be, extraterrestrial intelligent life that would classify as âmoral agentsâ (see Dissolving the Fermi Paradox for discussion).
Thus, it seems possible - though of course highly speculative - that value (or disvalue) in the universe could be created without humanity or its descendants. So if we defined an existential catastrophe as âthe destruction of the vast majority of the long-term potential for value in the universeâ, that would have the strange implications that:
- The destruction of the potential of humanity or its descendants wouldnât necessarily count as an existential catastrophe.
- Whether that counts as an existential catastrophe would depend in part on the likelihood that there is or will later be âmoral agentsâ unrelated to humans, and what such life would do with âour partâ of the universe.
So I favour sticking with Ordâs definitions, and just being aware that they mean something like âAn existential catastrophe is the destruction of the long-term potential humanity (or its âdescendantsâ) has to cause morally valuable outcomes.â
But this also has strange implications mirroring the above:
-
At least in theory, an âexistential catastropheâ could not be overwhelmingly bad, or perhaps not bad at all.
-
At least in theory, how bad an âexistential catastropheâ would be depends in part on the likelihood that there is or will later be âmoral agentsâ unrelated to humans, and what such life would do with âour partâ of the universe.
- E.g., it might be less bad than weâd naively think, if such agents are likely to exist and to create value using some of the resources we would otherwise have used.
- Or it might be even worse than weâd naively think, if such agents are likely to exist and do net-negative things with those resources if weâre not around.
But those strange implications seem worth accepting, partly to stay consistent with established definitions and to avoid the aforementioned implications the alternative definitions would also have. And in any case, in practice, it doesnât seem like the possibility of âmoral agentsâ unrelated to humanity should substantially affect how strongly we value reducing existential risks. This is for reasons related to:
- the probability of such other moral agents
- uncertainty about whether their actions would be of positive or negative moral value in any case
- the seemingly far greater decision-relevance of other considerations.
For interesting discussion of related points, see The expected value of extinction risk reduction is positive.
Conclusion
To summarise:
-
An existential catastrophe can be thought of as the destruction of the long-term potential humanity (or its âdescendantsâ) has to cause morally valuable outcomes.
- Itâs not necessarily just about human wellbeing, or just about Homo sapiensâ potential.
-
An existential catastrophe is also not necessarily the destruction of the long-term potential for value in the universe.
- This is due to the speculative possibility of current or future âmoral agentsâ unrelated to humanity, which could create value (or disvalue) in our absence.
- But it doesnât seem like this possibility should play a key role in our decision-making in practice.
Plausibly, reducing existential risks should be a top priority of our time. One way to improve our existential risk reduction efforts is to clarify and sharpen our thinking, discussion, and research, and one way to do that is to clarify and sharpen our key concepts. I hope this post has helped on that front.
In an upcoming post, Iâll discuss another question related to the concept of existential risks: What if humanity maintains its potential, but doesnât use it well?
My thanks to David Kristoffersson and Justin Shovelain for useful comments and suggestions.
This is one of a series of posts I plan to write that summarise, comment on, or take inspiration from parts of The Precipice. You can find a list of all such posts here.
The phrase âhumanity (or our descendants)â is used by Ord on page 382, and a similar phrase is used on page 395. âŠď¸
Unimportantly, Iâve also added a hyphen in âlong-termâ in these definitions. See footnote 2 here.
Personally, Iâd also be inclined to say âthe vast majority of the long-term potentialâ; see here. âŠď¸