Some considerations for different ways to reduce x-risk

By Jacy @ 2016-02-04T03:21 (+28)

I believe the far future is a very important consideration in doing the most good, but I don’t focus on reducing extinction risks like those from unfriendly artificial intelligence. This post introduces and outlines some of the key considerations that went into that decision, and leaves discussion of the best answers to those considerations for future work.

The different types of x-risk

In “Astronomical Waste”, written by Nick Bostrom in 2003, an existential risk (commonly referred to as a x-risk) is described as “one where an adverse outcome would either annihilate Earth-originating intelligent life or permanently and drastically curtail its potential.” In effective altruism and rationalist circles, we have most commonly addressed x-risks of the former type, those that could “annihilate Earth-originating intelligent life.” Let’s call those population risks.


The other risks could “permanently and drastically curtail [Earth-originating intelligent life’s] potential.” Let’s assume our “potential” is to create a very large and very good civilization, with “very large” referencing the number of sentient beings and “very good” being up to the reader. (I’ll assume that you believe there are at least some scenarios with a very large civilization that wouldn’t be very good, such as those filled with more suffering than happiness.)


I think humanity could fail in either or both of these respects. In this post, I’ll focus on the risk of creating a very large but not very good civilization, but I think the other types are worth exploring some other time. We’ll use the term quality risk to mean a risk of having a very large civilization that’s not very good.


Note that considering quality risks explicitly isn’t a new idea at all. Nick Beckstead wrote about it in August 2014, referring to previous mentions of it in 2009, 2013, and another in 2014. Beckstead explicitly states, “Astronomical waste may involve changes in quality of life, rather than size of population.”

“I think we should focus our efforts almost entirely on improving the expected value of the far future. Should I work on quality risks or extinction risks?”

I think more people in the effective altruism and rationalist communities should be asking themselves this question. I think people often make the jump from “the far future is very important” to “I should work on the most important extinction risks” too quickly.

“But what if I’m not exclusively focused on the far future?”

There are reasonable justifications for not focusing entirely on the far future, such as concerns about the tractability of making a substantive difference, and wanting to give limited weight to linear arguments like that most commonly used to justify focusing on the far future. I personally give significant weight to both far future and near-term outcomes. But for this post, I’ll focus exclusively on far future impact.

“But why can’t I just reduce both sorts of x-risk?”

Some activities might reduce both quality and extinction risks. For example, getting more people involved in EA might increase the number of people working in each area. Also, research that increases the likelihood of friendly AI might not only reduce the risk of extinction, but also might affect the relative likelihoods of different non-extinction AI scenarios, some of which might be better than others. I think this is super interesting, but for this post, I’ll only consider the tradeoff itself.

Considerations

To better understand whether to focus on quality risks or extinction risks, I think there are a number of informative questions we can ask. Unfortunately, the questions that seem most useful if we had answers also seem quite intractable and little research has been done on them.


Because quality risk is such a diverse category, I’ll use widening our moral circles as an example of how we could increase the expected value of the far future given we continue to exist, mainly because it’s what I see as most promising and have considered most fully. By this, I mean, “Increasing the extent to which we accommodate the interests of all sentient individuals, rather than just those who are most similar to us or have the most power in society.” It seems that narrow moral circles and a lack of concern for all sentient beings could lead civilization to be much worse, given it continues to exist, posing a substantial quality risk.


Here’s a list of some particularly interesting considerations, in my opinion. Explanations are included for considerations with less obvious relevance, and I think there’s a good chance I’m leaving out a few important ones or not breaking them down optimally (e.g. “tractability” could be segmented into several different important considerations).









Future research and discussion

Further analyses of these considerations and introduction of new ones could be quite valuable [Edit: As noted in the comments, some analyses do exist for some of these questions. I could have been clearer about that.]. My best guesses on these questions — which are largely just intuitions at this point — lead me to favor working on widening our moral circles instead of reducing extinction risk, but I could change my mind on that and I hope others would also be willing to do so. It’s also worth considering the prioritization of finding better answers to these questions, even if they seem mostly intractable.

I worry that many of us who focus on long-term impact haven’t given much thought to these considerations and mostly just went with the norms of our social circles. Upon considering them now, I think it’s tempting to just settle them in the direction that favors our current activities, and I hope people try to avoid doing so.



undefined @ 2016-02-04T22:39 (+19)

This is a critically important and neglected topic, and I'm glad you wrote about it. I've written about this distinction before but I think you did a much better job of explaining why it matters.

Here are some more writings on the subject, along with a summary of my favorite points from each article:

Michael Bitton: Why I Don't Prioritize GCRs

Brian Tomasik: Values Spreading Is Often More Important than Extinction Risk

Paul Christiano: Against Moral Advocacy

Paul Christiano: Why Might the Future Be Good?

undefined @ 2016-02-05T10:57 (+10)

I'd double-upvote this if I could. Providing (high-quality) summaries along with links is a great pro-social norm!

undefined @ 2016-02-05T02:09 (+1)

A couple of remarks:

GCR prevention only matters if they will happen soon enough

The very same, from a future perspective, applies to values-spreading.

Most people are incentivized to prevent extinction but not many people care about my/our values

This is a suspiciously antisocial approach that only works if you share Brian's view that not only are their no moral truths for future people to (inevitably) discover, but nonetheless it is very important to promote one's current point of view on moral questions over whatever moral views are taken in the future.

Jacy @ 2016-02-05T07:22 (+6)

The very same, from a future perspective, applies to values-spreading.

Why do you think that? There are different values we can change that seem somewhat independent.

This is a suspiciously antisocial approach

That seems mean and unfair. Having different values than the average person doesn't make you antisocial or suspicious; it just makes you different. In fact, I'd say most EAs have different values than average :)

undefined @ 2016-02-04T11:17 (+9)

This certainly gets quite a bit of attention in internal conversations at the Future of Humanity Institute. Bostrom discussed it when first(?) writing about existential risk in 2001, under the name shrieks. Note I wouldn't recommend reading that paper except for historical interest -- his more modern exposition in Existential Risk Prevention as Global Priority is cleaner and excellent. I think your quality risk coincides with Bostrom's notion of flawed realisation, although you might also mean to include subsequent ruination. Could you clarify?

Anyhow I'll give my view briefly:

undefined @ 2016-02-05T10:50 (+4)

When I get multiple downvotes I like to use this to learn not to do things which people find unhelpful. Often I can go back and re-read my comment and work out what people didn't like. Here I'm not so sure -- something about tone? The fact that I gave my own framing of the issue more than building on the framing in the OP? Mixing two unrelated points (history of discussion and my views now) in one comment?

I'd appreciate pointers from anyone who downvoted, or who didn't but felt a temptation to. I don't want to discuss whether my post 'deserved' downvotes, I just want to understand what about it would drive them.

undefined @ 2016-02-05T17:48 (+2)

The downvoting throughout this thread looks funny. Absent comments, I'd view it as a weak signal.

Jacy @ 2016-02-04T17:02 (+1)

"Quality risk" is meant to include both of those ideas, just any situation where we get "very large" (~"technologically mature") but not "very good."

undefined @ 2016-02-04T05:31 (+3)

Michael Dickens wrote about quality risks vs existential risks here and here.

Jacy @ 2016-02-04T06:08 (+2)

Thanks for noting. I should have included links to those and other existing materials in the post. Was just trying to go quickly and show an independent perspective.

undefined @ 2016-02-04T22:47 (+1)

I talk about values spreading rather than quality risks, but they're similar since the most commonly discussed way to mitigate quality risks is via values spreading.

(I would actually be interested in a discussion of what we could do to mitigate quality risks other than values spreading.)

undefined @ 2016-08-10T19:41 (+2)

Thanks a lot for this article! I just wanted to link to Lukas Gloor's new paper on Fail-Safe AI, which discusses the reduction of "quality future-risks" in the context of AI safety. It turns out that there might be interventions that are less directed at achieving a perfect outcome, but instead try to avoid the worst outcomes. And those interventions might be more tractable (because they don't aim at such a tiny spot in value-space) and more neglected than other work on the control problem. https://foundational-research.org/wp-content/uploads/2016/08/Suffering-focused-AI-safety.pdf

undefined @ 2016-02-04T22:54 (+2)

Do you see any specific examples where reducing other types of existential risks could increase quality risks?

undefined @ 2016-02-05T10:54 (+8)

Moving towards political singleton, and increasing surveillance technology, both look like they should help to reduce risks of human extinction. But they may well increase the risk of locking in a value system which is suboptimal, whereas a more varied society could do better in expectation (particularly after different parts of the society trade with each other).

undefined @ 2016-02-05T00:06 (+5)

If you expect the far future to be net negative in expectation, then reducing existential risk necessarily increases quality risk. In this essay I list some reasons why the far future might be net negative:

  • We sustain or worsen wild animal suffering on earth.
  • We colonize other planets and fill them with wild animals whose lives are not worth living.
  • We create lots of computer simulations of extremely unhappy beings.
  • We create an AI with evil values that creates lots of suffering on purpose. (But this seems highly unlikely.)

In the essay I discuss how likely I think these scenarios are.

undefined @ 2016-02-23T10:53 (+1)

I have created a roadmap of x-risks prevention, and I think that it is complete and logically ordered. I will make longer post about it if i will be able to get enough carma. )) The pdf is here: http://immortality-roadmap.com/globriskeng.pdf

undefined @ 2016-02-05T02:54 (+1)

Thanks, this made me think more deeply about quality risks - appreciate it!

My tendency is to think that a good path forward may be to focus on goal factoring activities, namely ones that hit many types of risks at once - x-risks and quality risks. I appreciate what Owen brought up here about building up the EA movement as one way to do so. I think spreading rationality is another way to do so.

undefined @ 2016-02-04T14:44 (+1)

Both trying to change attitudes to animals, and reducing extreme poverty, have been advocated on the basis that it could create trajectory changes that improves the quality of the future.

Christiano has some thoughts here:

http://rationalaltruist.com/2013/06/13/against-moral-advocacy/

Moral value of the far future Political power in the far future

These two are discussed in Superintelligence.

Also see these sections of the original x-risk paper:

"Crunches – The potential of humankind to develop into posthumanity[7] is permanently thwarted although human life continues in some form.

Shrieks – Some form of posthumanity is attained but it is an extremely narrow band of what is possible and desirable.

Whimpers – A posthuman civilization arises but evolves in a direction that leads gradually but irrevocably to either the complete disappearance of the things we value or to a state where those things are realized to only a minuscule degree of what could have been achieved."

http://www.nickbostrom.com/existential/risks.html

"Quality of life getting ‘stuck’"

See Bostrom's comments on a negative 'singleton'.