What do staff at CEA believe? (Evidence from a rough cause prio survey from April)

By Lizka @ 2023-10-02T10:42 (+85)

In April, I ran a small and fully anonymous cause prioritization survey of CEA staff members at a CEA strategy retreat. I got 31 responses (out of around 40 people),[1] and I’m summarizing the results here, as it seems that people sometimes have incorrect beliefs about “what CEA believes.” (I don’t think the results are very surprising, though.)

Important notes and caveats:

Rough summary of the results:

  1. Approach to cause prioritization: Most people at CEA care about doing some of their own cause prioritization, although most don’t try to build up the bulk of their cause prioritization on their own.
  2. Approach to morality: About a third of respondents said that they’re “very consequentialist,” many said that they “lean consequentialist for decisions like what their projects should work on, but have a more mundane approach to daily life.” Many also said that they’re “big fans of moral uncertainty.”
  3. Which causes should be “key priorities for EA”: people generally selected many causes (median was 5), and most people selected a fairly broad range of causes. Two (of 30) respondents didn’t choose any causes not commonly classified as “longtermist/x-risk-focused” (everyone else did choose at least one, though). The top selections were Mitigating existential risk, broadly (27), AI existential security (26), Biosecurity (global catastrophic risk focus) (25), Farmed animal welfare (22), Global health (21), Other existential or global catastrophic risk (15), Wild animal welfare (11), and Generically preparing for pandemics (8). (Other options on the list were Mental health, Climate change, Raising the sanity waterline / un-targeted improving institutional decision-making, Economic growth, and Electoral reform.)
  4. Some highlights from more granular questions: 
    1. Most people selected “I think reducing extinction risks should be a key priority (of EA/CEA)” (27). Many selected “I think improving how the long-run future goes should be a key priority (of EA/CEA)” (17), and “I think future generations matter morally, but it’s hard to affect them.” (13)
    2. Most people selected “I think AI existential risk reduction should be a top priority for EA/CEA” (23) and many selected “I want to learn more in order to form my views and/or stop deferring as much” (17) and “I think AI is the single biggest issue humanity is facing right now” (13). (Some people also selected answers like “I'm worried about misuse of AI (bad people/governments, etc.), but misalignment etc. seems mostly unrealistic” and “I feel like it's important, but transformative developments / x-risk are decades away.”)
    3. Most people (22) selected at least one of “I think animal welfare work is important, similar to global health work” (16) and “I'd be working directly on animal welfare work if I didn't think existential risk reduction is a top priority.” (10)
    4. Slightly fewer than half of respondents (14) say global health work is less important than animal welfare work, while 17 selected at least one of “It's important, similar to animal welfare,” “It's is a top priority, and should be prioritized more in EA,” and “I'd be working on it directly if I didn't think existential risk reduction is a top priority.”
  5. Views relative to what the EA community is doing: respondents generally didn’t think that the EA community was over-prioritizing AI safety, animal welfare, and global health/development. For each of the three areas, about ~6 people (out of ~30 respondents) thought they were at least a bit over-prioritized. 
  6. Appendix: views relative to CEA and experience with cause prio at CEA: I had asked a variety of questions aimed at seeing if staff felt free to disagree about cause prioritization, and how staff perceived their cause prioritization relative to others at CEA. Broadly, most people didn’t think they’d be stigmatized for their cause prio but a few thought they would be (I couldn’t find real patterns in the cause prio of people who were on the higher end of this), a number of people seemed to feel some pressure to present themselves as better informed about EA content than they feel, and people had a pretty reasonable distribution of how informed on EA content they thought they were, relative to other CEA staff (e.g. about half of people thought they were above average, but this still could mean that people were still totally off), and the median respondent thought they were slightly less focused on animal welfare than other CEA staff. 

1. How people approach cause prioritization and morality

Approaches to cause prioritization

I asked “Which of the following describes your approach to cause prioritization? (select all that apply)”, and the options were: 

  1. I don't really care
  2. I defer to CEA or stakeholders
  3. I did some cause prio work for myself in the past but now mostly go off intuitions or defer
  4. I do some combination of deferring and trying to build my own cause prio
  5. I try to build a lot up on my own

Results: 

Here are the raw counts for how many people selected each response: 

Morality

I asked “Which of the following describes your approach to morality? (Please select up to two)” — 30 responses, and the options were: 

  1. I don't really have one
  2. Deontological
  3. I'm very consequentialist
  4. I lean consequentialist for decisions like what my project should work on, but I have a more mundane approach to daily life
  5. Mix of consequentialist and other things
  6. Big fan of moral uncertainty

Results: 

Response counts: 

2. What should we prioritize?

I asked a number of questions about specific causes and worldviews. 

Causes that should be priorities for EA

A. I asked “Which of the following should be key priorities for EA? (mark all that apply.)” The options were (not in this order):

  1. Mitigating existential risk, broadly
  2. AI existential security
  3. Biosecurity (global catastrophic risk focus)
  4. Farmed animal welfare
  5. Global health
  6. Other existential or global catastrophic risk
  7. Wild animal welfare
  8. Generically preparing for pandemics
  9. Mental health
  10. Climate change
  11. Raising the sanity waterline / un-targeted improving institutional decision-making
  12. Economic growth
  13. Electoral reform
  14. There was also a write-in: “Growing the number of people working on big problems (= building effective altruism?)”

Results: 

Raw counts for causes that staff selected: 

Here’s the distribution of how many people selected a given number of causes: 

Most people selected 4-8 causes.

More granular views on longtermism/existential risk reduction, AI safety, animal welfare, and global health/wellbeing/development

“Which of the following describes your approach to longtermism and existential risk reduction? (Select all that apply.)”

Most people (27) selected “I think reducing extinction risks should be a key priority (of EA/CEA)” and most of those (16 of those, as well as one other person) also put “I think improving how the long-run future goes should be a key priority (of EA/CEA).”

“Which of the following describes your beliefs on AI (safety)? (Select all that apply.)”

23 - "I think AI existential risk reduction should be a top priority for EA/CEA," 17 - "I want to learn more in order to form my views and/or stop deferring as much," 13 - "I think AI is the single biggest issue humanity is facing right now," 9 - "I am mostly deferring, but other people seem to think it's important," etc.

“Which of the following describes your views on animal-welfare-oriented work?”

16 - "I think animal welfare is important, similar to global health work," 11 - "farmed animal welfare makes a lot of sense to me, I'm less sure about wild animal welfare work," 10 - I'd be working directly on animal welfare work if I didn't think exiseexistentialntial risk reduction is a top priority," etc.

“Which of the following describes your views on global development-oriented work?”

14 - "Less important than animal welfare," 10 - "It's important, similar to animal welfare," 7 - "It's a top priority, and should be prioritized more in EA," 5 - "I'd be working on it directly if I didn't think existential risk reduction is a top priority," etc.

Views relative to what people think the focus of EA is currently on AI safety, animal welfare, and global health/wellbeing/development

TLDR: CEA staff seem to broadly think things aren’t being overprioritized, with a couple of people disagreeing.[2]

“The EA community is focusing too much on AI safety”

“The EA community is focusing too much on animal welfare”

“The EA community is focusing too much on global development-type causes”

3. Appendix: Where staff think they are relative to others at CEA, pressure to appear a certain way, etc. 

I asked “Do staff agree with “CEA’s ~official cause prio?”” I don’t really know what I was asking; we don’t really have an official cause prio (lots of people expressed something like “Do we have an official cause prio?” in the comments of this question). I expect that people interpreted “official cause prio” as “in practice, what CEA does,” although it’s possible that people were trying to guess “leadership’s” cause prio, or something like that. (I think it would have been useful to find out if people felt like they disagreed a lot.) I think I personally put “Maybe, not sure.” Anyway, here are the responses: 

Forms response chart. Question title: Do you think you mostly agree with CEA's ~official cause prio?. Number of responses: 31 responses.
55% - "Probably agree on most things," 36% - "Maybe, not sure," and some more "Yes" and "Probably disagree on some significant answers. No one picked "No, I disagree on a lot of core things." 

“Do you think a random group of CEA staff would stigmatize you for sharing any of your opinions on cause prioritization?” — 7 people are above neutral, which made me a bit worried. Most people think they wouldn’t get stigmatized:

“Do you feel pressure to present yourself as better informed about EA content than you feel?” — most people (21) feel this a decent amount:

“How informed on EA content do you think you are, relative to other CEA staff?”

What staff thought their cause prio was, relative to other CEA staff

Generally quite reasonable. Slight bimodal distribution for neartermist to longtermist, which I guess means that people think they're less average than they actually are (or that others are more neartermist/longtermist than they actually are). Pretty reasonable distribution of GHD-to-non-GHD. And people think they're a bit more not-animal-focused relative to others at CEA than they actually are, probably, but not by a lot. 
  1. ^

    This is an OK response rate, but you might imagine that the people who didn’t respond have non-representative beliefs. E.g. they’re the folks who care less about cause prioritization, or who felt that it would take a long time to fill out a survey like this because their beliefs are so complex, etc. 

  2. ^

    Here’s a combined view: 


NickLaing @ 2023-10-03T13:22 (+7)

Thanks, I'm really encouraged by the diversity of thought here, there seems to be a range of beliefs (that aren't necessarily that tightly held) across the staff spectrum.

Definitely a bit less X-risk focused than I would have expected, although perhaps this is at least partially explained by the low response rate.

Lizka @ 2023-10-04T10:31 (+4)

Thanks! 

Definitely a bit less X-risk focused than I would have expected, although perhaps this is at least partially explained by the low response rate.

I would be pretty surprised if the people who didn't respond are significantly/systematically more x-risk-focused, for what it's worth.