In Continued Defense Of Effective Altruism — Scott Alexander

By Pablo @ 2023-11-29T10:37 (+163)

This is a linkpost to https://www.astralcodexten.com/p/in-continued-defense-of-effective


elifland @ 2023-12-01T13:04 (+59)

EDIT: Scott has admitted a mistake, which addresses some of my criticism:



(this comment has overlapping points with titotal's)

I've seen a lot of people strongly praising this article on Twitter and in the comments here but I find some of the arguments weak. Insofar as the goal of the post is to say that EA has done some really good things, I think the post is right. But I don't think it convincingly argues that EA has been net positive for the world.[1]

First: based on surveys, it seems likely that most (not all!) highly-engaged/leader EAs believe GCRs/longtermist causes are the most important, with a plurality thinking AI x-risk / x-risk more general are the more important.[2] I will analyze the post from a ~GCR/longtermist-oriented worldview that thinks AI is the most important cause area during the rest of this comment; again I don't mean to suggest that everyone has it, but if something like it is held by the plurality of highly-engaged/leader EAs it seems highly relevant for the post to be convincing from that perspective.

My overall gripe is exemplified by this paragraph (emphasis mine):

And I think the screwups are comparatively minor. Allying with a crypto billionaire who turned out to be a scammer. Being part of a board who fired a CEO, then backpedaled after he threatened to destroy the company. These are bad, but I’m not sure they cancel out the effect of saving one life, let alone 200,000.

(Somebody’s going to accuse me of downplaying the FTX disaster here. I agree FTX was genuinely bad, and I feel awful for the people who lost money. But I think this proves my point: in a year of nonstop commentary about how effective altruism sucked and never accomplished anything and should be judged entirely on the FTX scandal, nobody ever accused those people of downplaying the 200,000 lives saved. The discourse sure does have its priorities.)

I'm concerned about the bolded part; I'm including the caveat for context. I don't want to imply that saving 200,000 lives isn't a really big deal, but I will discuss from the perspective of "cold hard math" .

  1. 200,000 lives equals roughly a ~.0025% reduction in extinction risk, or a ~.25% reduction in risk of a GCR killing 80M people, if we care literally zero about future people. To the extent we weight future people, the numbers obviously get much lower.
  2. The magnitude of the effect size of the board firing Sam, of which the sign is currently unclear IMO, seems arguably higher than .0025% extinction risk and likely higher than 200,000 lives if you weight the expected value of all future people >~100x of that of current people.
  3. The FTX disaster is a bit more ambiguous because some of the effects are more indirect; quickly searching for economic costs didn't find good numbers, but I think a potentially more important thing is that it is likely to some extent an indicator of systemic issues in EA that might be quite hard to fix.
  4. The claim that "I’m not sure they cancel out the effect of saving one life" seems silly to me, even if we just look at generally large "value of a life" estimates compared to the economic costs of the FTX scandal.

Now I'll discuss the AI section in particular. There is little attempt to compare the effect sizes of "accomplishments" (with each other and also with potential negatives, with just a brief allusion to EAs accelerating AGI) or argue that they are net positive. The effect sizes seem quite hard to rank to me, but I'll focus on some ones that seem important but potentially net negative (not claiming that they definitely are!), in order of their listing:

  1. "Developed RLHF, a technique for controlling AI output widely considered the key breakthrough behind ChatGPT."
    1. This is needless to say controversial in the AI safety community
  2. Got two seats on the board of OpenAI, held majority control of OpenAI for one wild weekend, and still apparently might have some seats on the board of OpenAI, somehow?
    1. As I said above the sign of this still seems unclear, and I'm confused why it's included when later Scott seems to consider it a negative
  3. Helped found, and continue to have majority control of, competing AI startup Anthropic, a $30 billion company widely considered the only group with technology comparable to OpenAI’s.
    1. Again, controversial in the AI safety community
  1. ^

    My take is that EA has more likely than not been positive, but I don't think it's that clear and either way, I don't think this post makes a solid argument for it.

  2. ^

    As of 2019, EA Leaders thought that over 2x (54% vs. 24%) more resources should go to long-term causes than short-term with AI getting the most (31% of resources), and the most highly-engaged EAs felt somewhat similarly. I'd guess that the AI figure has increased substantially given rapid progress since 2019/2020 (2020 was the year GPT-3 was released!). We have a 2023 survey of only CEA staff in which 23/30 people believe AI x-risk should be a top priority (though only 13/30 say "biggest issue facing humanity right now", vs. 6 for animal welfare and 7 for GHW). CEA staff could be selected for thinking AI is less important than those directly working on it, but would think it's more important than those at explicitly non-longtermist orgs.

Ben_West @ 2023-12-04T23:17 (+17)

seems arguably higher than .0025% extinction risk and likely higher than 200,000 lives if you weight the expected value of all future people >~100x of that of current people.

If Scott had used language like this, my guess is that the people he was trying to convince would have completely bounced off of his post.

I interpreted him to be saying something like "look Ezra Klein et al., even if we start with your assumptions and reasoning style, we still end up with the conclusion that EA is good." 

And it seems fine to me to argue from the basis of someone else's premises, even if you don't think those premises are accurate yourself.

I do think it would have been clearer if he had included a caveat like "if you think that small changes in the chance of existential risk outweigh ~everything else then this post isn't for you, read something else instead" but oh well.

elifland @ 2023-12-05T14:36 (+10)

If Scott had used language like this, my guess is that the people he was trying to convince would have completely bounced off of his post.

I mostly agree with this, I wasn't suggesting he included that specific type of language, just that the arguments in the post don't go through from the perspective of most leader/highly-engaged EAs. Scott has discussed similar topics on ACT here but I agree the target audience was likely different.

I do think part of his target audience was probably EAs who he thinks are too critical of themselves, as I think he's written before, but it's likely a small-ish fraction of his readers.

I do think it would have been clearer if he had included a caveat like "if you think that small changes in the chance of existential risk outweigh ~everything else then this post isn't for you, read something else instead" but oh well.

Agree with that. I also think if this is the intention the title should maybe be different, instead of being called "In continued defense of effective altruism" it could be called something else like "In defense of effective altruism from X perspective". The title seems to me to imply that effective altruism has been positive on its own terms.

Furthermore, people who identify as ~longtermists seemed to be sharing it widely on Twitter without any type of caveat you mentioned.

And it seems fine to me to argue from the basis of someone else's premises, even if you don't think those premises are accurate yourself.

I feel like there's a spectrum of cases here. Let's say I as a member of movement X in which most people aren't libertarians write a post "libertarian case for X", where I argue that X is good from a libertarian perspective.

  1. Even if those in X usually don't agree with the libertarian premises, the arguments in the post still check out from X's perspective. Perhaps the arguments are reframed to make to show libertarians that X will lead to positive effects on their belief system as well as X's belief system. None of the claims in the post contradict what the most influential people advocating for X think.
  2. The case for X is distorted and statements in the piece are highly optimized for convincing libertarians. Arguments aren't just reframed, new arguments are created that the most influential people advocating for X would disendorse.

I think pieces or informal arguments close to both (1) and (2) are common in the discourse, but I generally feel uncomfortable with ones closer to (2). Scott's piece is somewhere in the middle and perhaps even closer to (1) than (2) but I think it's too far toward (2) for my taste given that one of the most important claims in the piece that makes his whole argument go through may be disendorsed by the majority of the most influential people in EA.

Geoffrey Miller @ 2023-11-30T18:23 (+35)

Pablo - thanks for sharing the link to this excellent essay.

Apart from being a great intro to EA for people made semi-skeptical by the FTX and OpenAI situations, Scott Alexander's writing style is also worthy of broader emulation among EAs. It's light, crisp, clear, funny, fact-based -- it's everything that EA writing for the public should be.

Pablo @ 2023-11-30T21:12 (+9)

Scott Alexander's writing style is also worthy of broader emulation among EAs

I very much agree. Here’s a post by Scott with nonfiction writing advice.

Lizka @ 2023-11-30T11:54 (+16)

Here's a long excerpt (happy to take it down if asked, but I think people might be more likely to go read the whole thing if they see part of it): 

The only thing everyone agrees on is that the only two things EAs ever did were “endorse SBF” and “bungle the recent OpenAI corporate coup.”

In other words, there’s never been a better time to become an effective altruist! Get in now, while it’s still unpopular! The times when everyone fawns over us are boring and undignified. It’s only when you’re fighting off the entire world that you feel truly alive.

And I do think the movement is worth fighting for. Here’s a short, very incomplete list of things effective altruism has accomplished in its ~10 years of existence. I’m counting it as an EA accomplishment if EA either provided the funding or did the work, further explanations in the footnotes. I’m also slightly conflating EA, rationalism, and AI doomerism rather than doing the hard work of teasing them apart:

Global Health And Development

  • Saved about 200,000 lives total, mostly from malaria1
  • Treated 25 million cases of chronic parasite infection.2
  • Given 5 million people access to clean drinking water.3
  • Supported clinical trials for both the RTS.S malaria vaccine (currently approved!) and the R21/Matrix malaria vaccine (on track for approval)4
  • Supported additional research into vaccines for syphilis, malaria, helminths, and hepatitis C and E.5
  • Supported teams giving development economics advice in Ethiopia, India, Rwanda, and around the world.6

Animal Welfare:

  • Convinced farms to switch 400 million chickens from caged to cage-free.7
  • Things are now slightly better than this in some places! Source: https://www.vox.com/future-perfect/23724740/tyson-chicken-free-range-humanewashing-investigation-animal-cruelty
  • Freed 500,000 pigs from tiny crates where they weren’t able to move around8
  • Gotten 3,000 companies including Pepsi, Kelloggs, CVS, and Whole Foods to commit to selling low-cruelty meat.

AI:

  • Developed RLHF, a technique for controlling AI output widely considered the key breakthrough behind ChatGPT.9
  • …and other major AI safety advances, including RLAIF and the foundations of AI interpretability10.
  • Founded the field of AI safety, and incubated it from nothing up to the point where Geoffrey Hinton, Yoshua Bengio, Demis Hassabis, Sam Altman, Bill Gates, and hundreds of others have endorsed it and urged policymakers to take it seriously.11
  • Helped convince OpenAI to dedicate 20% of company resources to a team working on aligning future superintelligences.
  • Gotten major AI companies including OpenAI to work with ARC Evals and evaluate their models for dangerous behavior before releasing them.
  • Got two seats on the board of OpenAI, held majority control of OpenAI for one wild weekend, and still apparently might have some seats on the board of OpenAI, somehow?12
  • [Skipped screenshot]
  • Helped found, and continue to have majority control of, competing AI startup Anthropic, a $30 billion company widely considered the only group with technology comparable to OpenAI’s.13
  • [Skipped screenshot]
  • Become so influential in AI-related legislation that Politico accuses effective altruists of having “[taken] over Washington” and “largely dominating the UK’s efforts to regulate advanced AI”.
  • Helped (probably, I have no secret knowledge) the Biden administration pass what they called "the strongest set of actions any government in the world has ever taken on AI safety, security, and trust.”
  • Helped the British government create its Frontier AI Taskforce.
  • Won the PR war: a recent poll shows that 70% of US voters believe that mitigating extinction risk from AI should be a “global priority”.

Other:

I think other people are probably thinking of this as par for the course - all of these seem like the sort of thing a big movement should be able to do. But I remember when EA was three philosophers and few weird Bay Area nerds with a blog. It clawed its way up into the kind of movement that could do these sorts of things by having all the virtues it claims to have: dedication, rationality, and (I think) genuine desire to make the world a better place.

II.

Still not impressed? Recently, in the US alone, effective altruists have:

  • ended all gun violence, including mass shootings and police shootings
  • cured AIDS and melanoma
  • prevented a 9-11 scale terrorist attack

Okay. Fine. EA hasn’t, technically, done any of these things.

But it has saved the same number of lives that doing all those things would have.

About 20,000 Americans die yearly of gun violence, 8,000 of melanoma, 13,000 from AIDS, and 3,000 people in 9/11. So doing all of these things would save 44,000 lives per year. That matches the ~50,000 lives that effective altruist charities save yearly18.

People aren’t acting like EA has ended gun violence and cured AIDS and so on. all those things. Probably this is because those are exciting popular causes in the news, and saving people in developing countries isn’t. Most people care so little about saving lives in developing countries that effective altruists can save 200,000 of them and people will just not notice. “Oh, all your movement ever does is cause corporate boardroom drama, and maybe other things I’m forgetting right now.”

In a world where people thought saving 200,000 lives mattered as much as whether you caused boardroom drama, we wouldn’t need effective altruism. These skewed priorities are the exact problem that effective altruism exists to solve - or the exact inefficiency that effective altruism exists to exploit, if you prefer that framing.

Dawn Drescher @ 2023-11-29T12:54 (+7)

Caude.ai summary for those in a hurry:

The article argues in defense of the effective altruism movement, citing its accomplishments in areas like global health, animal welfare, and AI safety, while contending criticisms of it are overblown. It makes the case that effective altruism's commitment to evidence-based altruism that focuses on the most tractable interventions to help others is a positive development worth supporting, despite some mistakes. The article concludes the movement has had significant positive impact that outweighs the negatives.

I'll read the article itself later, so be warned that I don't know how good this summary is.

Update: The summary is correct but significantly less viscerally motivating than the original. I love it!