Quantitative Mindset (Effective Altruism Definitions Sequence)
By ozymandias @ 2025-08-01T17:45 (+32)
I.
I’ve heard it said that the two principles of rationality are:
- Do the math.
- Pay attention to the math.
This post is about the first one.
II.
When I was in high school, my best friend announced one day that she wasn’t eating microwave popcorn any more. She’d read an article that said that someone had developed lung problems after eating two bags of popcorn a day for ten years.
I was a young rationalist with charmingly naive ideas about winning people over through argument. So I pointed out to her that popcorn lung is very rare, that popcorn lung from eating microwave popcorn instead of working in a popcorn plant is even rarer, that the popcorn lung case was newsworthy specifically because it was so rare, and that if she wanted to improve her health through her diet she’d be far better served by eating more vegetables and fewer chocolate-chip cookies.
This is why I was not popular in high school.
People often behave irrationally because they don’t think mathematically. Like my high school best friend, they struggle with reasoning about how probable an event is. Often, people fail to multiply two numbers together, or divide one number by another number. And exponential growth gets almost everyone.
For example, consider lotteries. Most lotteries have about a 1 in 300 million chance that your ticket is the winning ticket. Americans spent $97.8 billion dollars on lottery tickets in 2022. For comparison, the same year we spent $7.4 billion—less than 10% of that amount—on going to the movie theater.
Buying a lottery ticket is stupid. Yes, it allows you to daydream about being rich. If you want to daydream about being rich, you can take a walk and look for spare lottery tickets that someone dropped on the ground—it doesn’t actually reduce your chance of winning the lottery by any appreciable amount.
People don’t buy lottery tickets as a form of structured daydreaming. People buy lottery tickets because they didn’t do the math—because to them 1 in 300 million is the same as 1 in 300 billion, or 1 in 300,000, or 1 in 3,000. It means there’s a chance, right?
Similar cases of failure to reason mathematically:
- “I’m only spending $7 on a coffee and pastry every day. If I cut back on that, it wouldn’t do anything for my budget, not when rent costs $2000/month.”
- “Sure, I could easily afford to buy a new dishwasher, but my jury-rigged solution only takes three minutes every time I do the dishes.”
- “All those plane crashes! Flying is so dangerous these days. I’m going to drive to visit my relatives.”
- “Here are all the precautions I take about serial killers.”
- “I bought insurance on my concert ticket so I’ll get a refund if it turns out I can’t go.”
- “Covid-19 isn’t a big problem, because there are only a few hundred cases in the United States.”
III.
People often fail to do the math about their own lives, but they really fail to do the math about doing good. You care a lot about paying your rent, so at some point it might occur to you to multiply $7 by 30 and observe that you’re spending 10% of your rent on your morning coffee. You have much less personal investment in the outcomes of your do-gooding. They benefit someone else, often someone you’ve never even met.
Indeed, people are often specifically opposed to applying quantitative reasoning to doing good. When you help others, it’s supposed to come from the heart. It’s supposed to be an authentic upswelling of emotion. You’re not supposed to be sitting there with your spreadsheets, calculating out exactly how much you can get for your dollar. Quantitative reasoning seems ungenerous, unempathetic, even cruel.
As Paul Bloom wrote in Against Empathy:
To get a sense of the innumerate nature of our feelings, imagine reading that two hundred people just died in an earthquake in a remote country. How do you feel? Now imagine that you just discovered that the actual number of deaths was two thousand. Do you now feel ten times worse? Do you feel any worse?
Now I’m going to walk through three examples of quantitative reasoning in effective altruism: one for each of the three most popular effective altruist cause areas.
Global poverty. According to Toby Ord’s The Moral Imperative For Cost-Effectiveness In Global Health, in global health, “moving money from the least effective intervention to the most effective would produce about 15,000 times the benefit, and even moving it from the median intervention to the most effective would produce about 60 times the benefit” (my bolding).[1] Again, this is within the single cause of providing healthcare to the global poor. Certain causes, such as renovating the Lincoln Center to have better acoustics ($100 million),[2] no doubt are many many more orders of magnitude less effective.
This is batshit. No one would ever make a decision that is fifteen thousand times worse than another decision if their own interests were involved, no matter how averse they are to arithmetic. But if your decisions affect someone else, it’s easy to reason with your heart rather than your head. You’re not the one who faces the consequences.
Animal advocacy. Animal Charity Evaluators has one of the finest pieces of data visualization in the effective animal advocacy movement:[3]
Farmed animals vastly outnumber pets in need of shelters, and yet pets are far more likely to receive donations than farmed animals. This chart actually makes the farmed-animal situation look better than it is. More than a quarter of farmed-animal welfare funding comes from a single effective-altruism-affiliated donor, Open Philanthropy.[4] Before effective altruism, the difference was far starker.
Sometimes animal advocates ask why effective animal advocates focus on farmed animals to the exclusion of lab animals and shelter animals. All three kinds of animals matter, after all. Other animal advocacy organizations like PETA work on all three. The answer is that we compared the size of two numbers.
Existential risk. I once had a conversation with one of my friends that went something like this:
Friend: I feel a lot more uncertain about AI than most people we know do. I think maybe 50% chance of AGI by 2050?
Me: [nodding along]
Friend: —which is why AI is one of my top cause areas, because being more uncertain about something doesn’t make it safer.
As I wrote in The Four Guys Who Wouldn’t Shut Up And Were Wrong About Everything:
Their favorite mantra was “the science is not settled.” We aren’t sure whether tobacco smoking causes cancer—or whether air pollution is the reason that some samples of rain water are as acidic as lemon juice—or whether climate change is caused by something else, like sunspots. Therefore, we shouldn’t do anything about the situation until we know for sure that something is going wrong…
However, we should often take action about problems we’re uncertain about. Those Four Guys reliably assumed that the cost/benefit analysis was always in favor of the status quo—even when it wasn’t. It’s not that costly to society to regulate the pollutants that cause acid rain, and once you do you’ll quickly find out if your theory of acid rain is accurate.
We don’t want to find out how bad nuclear winter is by causing mass famines, or whether acid rain kills fish by driving irreplaceable fish species extinct. Many decisions we make, like putting carbon into the atmosphere, are irreversible. If you are unsure of the effects of your decision, don’t do irreversible things because that’s what you’re already doing. Sometimes, we need to make decisions based on weak or preliminary evidence, because that is the evidence we have available, and the other alternative is catastrophic.
If you think there is a 10% chance of artificial general intelligence in the next ten years,[5] and a 20% chance that artificial general intelligence will drive us extinct, that’s a 2% chance of humans going extinct. Some people look at those numbers and go “wow, only a 2% chance, that’s not going to happen, nothing to worry about.” Other people look at those numbers and go “holy shit, that’s one of the most important problems in the world.”
The latter sort of person becomes an effective altruist.
IV.
The well-kept secret about cost-effective analyses is that they’re all fake.
Deworming charities give children pills that cure parasitic worm infections. A long-term randomized controlled trial showed that people dewormed as children have ~10% higher income as adults. However, this is only a single study, and epidemiologists have raised numerous critiques of its methodology (the so-called “Worm Wars”).[6] Some people characterize this situation as “deworming might not work,” which I hate. Deworming definitely works. If you take deworming pills, you will no longer be infected with parasitic worms. The dispute is about how bad it is to be infected with parasitic worms.
GiveWell’s cost-benefit analyses adjust the effect of deworming charities down by 87% to account for this dispute. Why 87%? Well, you can go read their explanation of how they came to that number, but I hope they won’t take it as an insult to their excellent organization if I say that they kind of pulled it out of their butts. A very rigorous and statistical butt-pull, to be sure! But ultimately you can’t put a concrete number on “how likely is it that this study is fake?” in the same way you can put one on “how many pills do we give out per dollar?”
GiveWell’s 2023 cost-effectiveness spreadsheet says that Deworm the World is 34.4 times as good as an equivalently sized cash transfer in Kenya, but only 8.9 times in Lagos. This is, in an important sense, made up. We don’t know that Deworm the World is precisely 34.4 times better than cash transfers in Kenya. It could be 30 times, or 36 times, or even 34.5 times.
Why, then, do a cost-benefit analysis at all?
Because if you sit down, and make your spreadsheets, and do your fancy rigorous statistical butt-pulling, what you can say is: “deworming is cheap. Deworming is so cheap that, even with the most conservative assumptions we can reasonably justify, even if you assume that in seven out of eight universes it has no long-term effects at all, it is worth doing. Deworming is so fucking cheap.”
If you don’t think with numbers, all you have is two facts. Deworming probably doesn’t have significant long-run effects. Deworming programs are very inexpensive. Does one fact outweigh the other? Who knows? Not me.
Deworming is relatively uncertain for a global public health intervention, but we still know a lot about it. But in more speculative areas like animal advocacy or AI safety, quantitative thinking is even more important. Even if your numbers are kind of made up, thinking with numbers allows you to answer questions like:
- Is this intervention so overwhelmingly good that I should do it, even given how uncertain I am?
- What downsides of this intervention do I predict? How likely do I think they are compared to the advantages?
- What assumptions am I making? Do I really believe them?
- Of the things I’m uncertain about, which have the biggest effect on the answer? How can I get more information about them?
For more on practical cost-benefit analyses, I highly recommend this post.
V.
When I was first brainstorming this series with friends, I think probably the most common suggestion I got was about this point.
“People keep thinking that I think with numbers because I don’t care,” they said. “But I think with numbers because I do care. Numbers are how I care.”
Effective altruism can seem cold and calculating. But to an effective altruist, quantitative reasoning comes from profound compassion. People are dying of treatable illnesses. Children grow up unable to read or do math. Citizens have no say in their governance. Animals are being tortured. The world is warming, and it teeters on the brink of destruction half a dozen different ways. We care, and because we care we want to stretch every dollar as far as we can. And so we want to apply to saving a world a hundredth of the amount of cleverness and skill that our society does to the correct pricing of call options on Eurodollar futures.
As Bertrand Russell probably didn’t say, “The mark of a civilized man is the capacity to read a column of numbers and weep.” I am too uncivilized for number columns myself, but I have been driven to tears by any number of charts.
So here is the second of my three core claims of effective altruism: think with numbers.
- ^
Note that GiveWell has uncovered serious errors in the DCP2, Ord's primary source here. I am not aware of any more in-depth discussion of how reliable the DCP2 is. To the best of my understanding, it is certainly true that some global health interventions are many thousands of times better than others.
- ^
David Geffen needs to experience more public shaming.
- ^
Note that "number of animals killed" is a bad metric; see here. However, I do not believe this to substantially change the results.
- ^
The farmed animal welfare movement spends about $200 million a year; I eyeballed about $50 million of grants a year using Open Philanthropy’s grants database.
- ^
Of course, many people who work on AI safety think AI is far more likely than that.
- ^
You can find out Much More Than You Wanted To Know about deworming by following the links on GiveWell’s pages about it.
Clara Torres Latorre 🔶️ @ 2025-08-15T08:54 (+3)
People keep thinking that I think with numbers because I don’t care, they said. But I think with numbers because I do care. Numbers are how I care.
This got me emotional. I bet it makes for a good focal point in a story.
SummaryBot @ 2025-08-01T19:31 (+2)
Executive summary: This reflective and persuasive post argues that quantitative reasoning—"doing the math"—is essential to effective altruism because it enables us to make vastly better decisions about doing good, even when the numbers are uncertain or incomplete, and challenges the common perception that numerical thinking is cold or unfeeling.
Key points:
- People routinely fail to apply basic quantitative reasoning in daily life (e.g. overestimating risks, underestimating costs), and this failure becomes even more acute in charitable or altruistic contexts where the stakes affect others, not themselves.
- Emotional resistance to numerical reasoning in altruism is common, as people often feel moral acts should come from the heart, not from spreadsheets—yet this resistance can lead to dramatically less effective choices.
- Quantitative differences in effectiveness can be enormous: within global health alone, some interventions are up to 15,000 times more effective than others, making numerical analysis critical to doing the most good.
- Even imperfect cost-effectiveness analyses are valuable, because they help clarify assumptions, assess robustness, and highlight when an intervention is "so cheap it's worth doing" despite uncertainty (e.g. deworming).
- Numbers are a way of caring deeply, not detaching emotionally: effective altruists use quantitative reasoning because they want to maximize impact for others—not in spite of their compassion, but because of it.
- Core claim of the post: A defining principle of effective altruism is to "think with numbers"—embracing quantitative tools not as a replacement for empathy, but as an expression of it.
This comment was auto-generated by the EA Forum Team. Feel free to point out issues with this summary by replying to the comment, and contact us if you have feedback.