State of the EA Newsletter
By Toby Tremlettđš @ 2025-05-21T11:28 (+49)
This is a brushed up memo from a CEA retreat, aimed at orienting staff to the current state of the EA Newsletter, in preparation for a marketing push.
If you want to subscribe to the EA Newsletter, you can do so here.
Purpose and goal
The EA Newsletter is a frankenstein project - built by many people, not all of whom I've met. Itâs five months off a decade old. I've been its writer and custodian for a year and change (before me it was Lizka, and before her, Aaron Gertler).
When it began in 2015, it was focused on updating its audience on developments in the effective altruism community. Over the years, the focus of the newsletter shifted in a few directions. Now it:
- Tries not to assume interest in EA as such, but rather in things EAs find interesting and important. More âX says AGI might come soon and this is important because Yâ and less âEA should encourage earning to give moreâ.
- Shares jobs and opportunities which apply to people of all career levels and various interests.
- Follows a set template with three ~250 word featured article summaries, an âin other newsâ link list, a resources section identical in every issue, job postings, announcements (events, fellowships, funding), a link to the organisational updates post, and a timeless classic.
The EA Newsletter serves:
- The person who has just heard about EA, and might be excited to donate, take an introductory level fellowship, or apply to a relevant role if they see the right link.
- The person who was once very involved in EA, and wants to stay involved in the community.
- The person who reads it for fun
This year, I want to hone in on our theory of change, and either segment the newsletter (if we get more capacity), or change its focus to suit our ideal audience. I want to have the mindset that the stakeholders are the ideal audience, not the current audience.
Two goals which guide our growth in the meantime will be:
- Increasing subscribers.
- While retaining our impact-weighted-clicks/$[1] score of $6/click.
Engagement metrics and performance
Open rate
% of the audience who open the email.
I take the month-month numbers relatively seriously, but the absolute numbers not so much. The apple MPP exclusion (necessary because of a change in how apple handles emails which began in June last year and means that emails going to iphones might register as opened automatically) makes the data look a lot weirder, but the ranking of emails from most-least opened is roughly the same with or without the exclusion.
To get a more accurate average, we can add on 50% to the open rates (to account for roughly half of global emails going to apple mail). This gets us an average open rate of around 21% for the last year.
Whatâs the average open rate for emails from non-profits?
NeonOne claims it is 28.59%.
Mailchimp claims it is 40.04%
But neither of them provide detailed data on how this drops with the size of the email list, and the majority of Mailchimp's customers have far smaller lists than ours (so their average open rate would be much higher).
Click rate
% of the audience who click on anything.
Whatâs the average click rate for emails from non-profits?
NeonOne claims it is 3.29%.
Mailchimp claims it is 3.27%
I trust click rates more, I think they are much easier to measure.
As such I can more confidently say that we are doing great!
New subscriptions and conversion
As of 21 May, we have 58,715 subscribers.
Up until the beginning of April this year, we got most of our subscribers from virtual programs, via a tickbox in their form, and a manual upload every couple of months (the upticks in this graph are all VP):
Subscribers also came from the embedded form on EA.org, but the number was negligible, far below the amount needed to replace the monthly loss of around 200-300 subscribers per newsletter.
On April 1, we changed the embed forms on EA.org and CEA.org to allow for single-opt-in (i.e., no need to verify subscription by clicking on an email). This has been transformative:
But we are still short of where we could be if we fully made use of CEAâs existing infrastructure. For example:
- The conversion rates (page visit -> subscribe) on EA.org (0.7%) and CEA.org (0.38%) could be optimised and A/B tested to be higher. I donât have evidence for this â we just havenât tried.
- We could replicate the success of the VP opt-in on other CEA projects, for example:
- The EA Forum sign-up flow (we started doing this on April 28)
- EAG and EAGx application forms
- Groups â groups have touchpoints with a bunch of people who donât become members, but might take impactful actions if they are subject to a slow drip of info and opportunities from the newsletter. Signing people up to the EA newsletter when they sign up to an event at a university or city group could be killer.
Also, it's worth noting that it's too soon to tell whether this is a real slam dunk. The new audience seems to open our intro campaign at about half the previous rate (30% open rate rather than 60%), and we'd need to change the sign-up process back if it leads to higher abuse rates (people signing up so easily they think the newsletter is spam), or bot attacks (we already had one, which was resolved, and now we have added captchas which has so far đ¤ worked).
Understanding our audience
We have two ways to understand our audience â mailchimp âdataâ and surveys. I write data in scare quotes because mailchimp demographics data is a guess, and Iâve anecdotally seen it get things wrong.
Mailchimp demographics
Caveat caveat according to mailchimp, but:
- The largest age demographic is 25-34.
- More men read the newsletter than women.
Mailchimp location data is not worth mentioning â it seems very off.
Impact survey
The Impact survey is the newsletterâs longest-running feedback survey, active since 2018. As such, it gives a better impression of some demographics than anything else we have.
Note that the CTA in the newsletter for much of this time has been âIf youâve taken action because of the Newsletter and havenât taken our impact survey, please do â it helps us improve future editions.â, so it is selected for people who have found the newsletter useful.
In some other important ways, the data from this survey wonât be representative. Obviously, we canât scale up the distribution of actions taken to cover the whole audience, but we should also be somewhat sceptical of the industry breakdowns â we might reasonably expect more students because students have easier actions to take than professionals.
What actions have you taken directly because of clicking a link in the newsletter?
The below table shows the total number of responses containing each action type (note that many responses contained more than one action).
Action Type | Total Count |
Applied to a new job, internship, or research position | 80 |
Applied to EA Global, EAGx, or another event organized by an EA nonprofit | 60 |
Attended an event or meeting organized by an EA community group | 50 |
Donated to an EA-aligned charity (unrelated to Giving What We Can) | 109 |
Taken the Giving What We Can pledge or signed up for Try Giving | 51 |
Other (explain below) | 128 |
Overall Total | 333 |
This doesn't accurately show the total actions taken by the newsletter audience (it massively low-balls it) because most of the responses to this form occurred in just three months over the past 7 years, when the survey was highlighted in the newsletter. To get a better impression of actions taken, I plan to 1) estimate actions taken based on an extrapolation of this table 2) estimate actions based on clicks in the newsletter, and 3) compare the two to triangulate[2] a better estimate.
How important is the Newsletter vs other ways you get information?
Iâm surprised by the proportion for whom the newsletter is the main form of information about effective altruism, but perhaps I should expect this survey to show a higher proportion in that camp because it is selected for people who count the newsletter as the cause of their having taken action. If you see a job in the newsletter, then later on twitter or the forum, you wonât be as likely to say the newsletter was responsible.
The industry respondents work in
Industry breakdown from impact survey (Summarised by Claude):
Industry | Count | Percentage |
Academia/Research | 91 | 19.2% |
Student | 79 | 16.6% |
Tech/Software | 65 | 13.7% |
Finance/Business | 41 | 8.6% |
Nonprofit/NGO | 24 | 5.1% |
Retired | 20 | 4.2% |
Healthcare | 18 | 3.8% |
Government/Policy | 17 | 3.6% |
Education | 10 | 2.1% |
Arts/Media | 6 | 1.3% |
Other | 104 | 21.9% |
Again, this is likely biased by the fact that the user has taken action. However, it is still interesting that the audience is predominantly academic in some way â the student part of the demographic is undoubtedly large (and some of those listed as academics may be grad students).
Where respondents live
From the impact survey, according to Claude.
Europe | 49.5% |
North America | 39.2% |
Oceania | 4.7% |
Asia | 3.8% |
South America | 2.8% |
This seems more accurate than Mailchimpâs estimations (which claim Amsterdam is our top city for readers).
Where next?
Based on analysis of where we are at, my goals over the next 6 weeks are to:
- Work with other teams at CEA to integrate newsletter sign-ups into their projects (and, if it isnât already, make the newsletter more integrated with their goals).
- Be more decisive about who our target audience should be.
- Work with someone (keeping this vague because I'm speaking to people currently) to run ads for the newsletter, reaching that target audience.
Feel free to ask more questions or suggest directions for further research into the Newsletter. This memo isn't perfectly comprehensive.
Specifically, I'm working this week on figuring out a rational Willingness To Pay per Subscriber metric (so that I can assess the price of marketing campaigns), so any suggestions there would be much appreciated.
- ^
Impact weighted clicks are a metric I'm still experimenting with, which assigns points based on a simple model where a click on a job ad = 1 point, fellowship = 0.5, Courses = 0.25 etc... It'd be ideal to have an absolute metric (i.e. a metric which could tell me how valuable the newsletter is overall), but in the mean-time, a metric which lets me compare month-to-month is valuable.
- ^
Apologies but I don't know the two-vertices word which I should be using here.
Vasco Grilođ¸ @ 2025-05-22T10:20 (+14)
Thanks for sharing, Toby!
- While retaining our impact-weighted-clicks/$[1] score of $6/click.
How does your 0.167 impact-weighted-click/$ (= 1/6) compare with the clicks per $ regarding 80 k's, Animal Advocacy Career's, and Probably Good's job boards? Since you are assuming 1 click on a job ad means 1 impact-weighted-click (footnote 1), and the above is your main metric to track cost-effectiveness, how would you update on the value of the EA newsletter if other impact-focussed job boards had way more clicks per $?
Specifically, I'm working this week on figuring out a rational Willingness To Pay per Subscriber metric (so that I can assess the price of marketing campaigns), so any suggestions there would be much appreciated.
I would ask current subscribers how much they would be willing to pay in order to maintain their subscription.
Whatâs the average click rate for emails from non-profits?
NeonOne claims it is 3.29%.
Mailchimp claims it is 3.27%
I trust click rates more, I think they are much easier to measure.
As such I can more confidently say that we are doing great!
Nitpick. I think it is more accurate to say the click rate has been quite good in April, but similar to that of nonprofits from October to March 2025.
Apologies but I don't know the two-vertices word which I should be using here [instead of triangulate].
Interpolate?
Toby Tremlettđš @ 2025-05-22T10:27 (+10)
"I would ask current subscribers how much they would be willing to pay in order to maintain their subscription."
That's interesting! I could definitely give that a go. I think that might be a bit off though because the more relevant thing is how much money I have access to in our budget/ could access by making the right case to OP (and what it would be spent on otherwise). I've also never really understood how individual willingness to pay is a good signal when people have wildly different resources available to them. Could be a useful benchmark.
Another point on that - I don't trust myself when it comes to valuing things monetarily. I'd never pay for social media or a newsletter, because it just isn't something I'm used to (I'm used to them being free), but if I actually calculate the time I'm spending on social media/ newsletters I count as a valuable use of time, and use something like this to calculate the value of my time, I get surprisingly high figure.
Fair point on click-rates!
"How does your 0.167 impact-weighted-click/$ (= 1/6) compare with the clicks per $ regarding 80 k's, Animal Advocacy Career's, and Probably Good's job boards?"
This is a good question - I should prioritise comparing these. I do expect that some of those programs (and possibly the EA opportunities board) are cheaper per click, and have a higher intent audience. The EA Newsletter hasn't been optimised for job clicks (and probably shouldn't be). I've been a bit bottlenecked by what I can most easily measure.
Vasco Grilođ¸ @ 2025-05-22T13:47 (+4)
Thanks, Toby!
That's interesting! I could definitely give that a go. I think that might be a bit off though because the more relevant thing is how much money I have access to in our budget/ could access by making the right case to OP (and what it would be spent on otherwise). I've also never really understood how individual willingness to pay is a good signal when people have wildly different resources available to them. Could be a useful benchmark.
You can ask the subscribers which fraction of their net annual income they would be willing to spend to get an annual subscription of the EA newsletter, multiply the mean fraction across the surveyed subscribers by the number of subscribers to get the total value of the newsletter in terms of doublings of net income, and divide this by 2 doublings of net income per DALY averted to get the impact of the newsletter in terms of averted DALYs. Then you could divide this impact by the cost to get the cost-effectiveness in DALY averted per $, and compare with the cost-effectiveness of GiveWell's top charities of around 0.01 DALY/$.
Another point on that - I don't trust myself when it comes to valuing things monetarily. I'd never pay for social media or a newsletter, because it just isn't something I'm used to (I'm used to them being free), but if I actually calculate the time I'm spending on social media/ newsletters I count as a valuable use of time, and use something like this to calculate the value of my time, I get surprisingly high figure.
I wonder whether people are sometimes thinking about the total instead of marginal value. I would pay for additional/marginal air because I can get air for free, but I still get lots of value from air, and would pay a lot to keep what I have relative to having nothing. Similarly, you may not mind losing access to a single newsletter due to the other abundant online content you can continue to check for free. The marginal value is what matters from the perspective of assessing counterfactual impact. The product between one's net hourly rate, and the time checking a newsletter would only be a good way of estimating its additional impact if the alternative to checking the newsletter had no value, which is not the realistic counterfactual.
This is a good question - I should prioritise comparing these. I do expect that some of those programs (and possibly the EA opportunities board) are cheaper per click, and have a higher intent audience. The EA Newsletter hasn't been optimised for job clicks (and probably shouldn't be). I've been a bit bottlenecked by what I can most easily measure.
I also expect job boards to have more clicks per $.
Toby Tremlettđš @ 2025-05-22T10:33 (+4)
Also also - this isn't included in the post but 70% of clicks on the newsletter go towards content links, i.e. articles, podcasts, videos etc... which are great resources on something EAs care about, but don't have clear CTAs. I'm not sure how to measure the value of this yet - but in surveys, many of the responses show people changing their minds/ plans because of content they read in the newsletter. Seems likely this is a very important part of the newsletter's impact, but currently it counts for 0 in my impact-weighted-clicks metric.
Toby Tremlettđš @ 2025-05-22T11:10 (+12)
Here is the very sketchy first salvo at measuring "Impact-weighted clicks" (not sure how much I endorse this anymore, but feel free to treat me as if I do endorse it in responses to the comment - that'll help me figure out what I think).
Was going to just share a doc with @Vasco Grilođ¸ but I figured there is no reason not to make it public. Very happy to get input on this as well, especially if you have experience trying to measure the value of content (outside of just "clicks on jobs")
Impact-weighted clicks
Not all EA Newsletter clicks are created equal. The EA Newsletter contains links to impactful jobs, but also interesting podcasts, videos, and articles â where the only further call to action is subscribing. This content isnât of no value, but itâs very unlikely to be the source of most of the newsletterâs value.
The metric Iâm making will give categories of clicks a value from 0-1, and then produce a total number of âimpact-weightedâ clicks, where the total amount of unique clicks each category is multiplied by the value assigned to that category.
What is the value of content clicks?
Firstly â retention. Without content (hypothesis 1), people wouldnât be interested in the EA Newsletter.
Secondly â content pipelines. If someone starts reading an EA friendly outlet, subscribes to the 80k podcast, buys a book, subscribes to rational animations, etc⌠They increase their chance of coming across EA content in the future.
Third â Changing minds. People who are aware of big problems might decide to work on them.
How should we categorise EANL clicks?
Category | % of clicks for the march newsletter |
Content | 69% |
Job ad (and job boards) | 23% |
Fellowship/ funding | 7% |
Courses | 1% |
Donation opportunities | 0% (none included) |
What value should we give to each type of click?
An obvious ranking is:
- Job ads (this person is already engaged and ready to act)
- Fellowship/ funding (likewise ready to act, but funding is short term and fellowship is the step before the impactful job)
- Courses (intent to get more involved, learn more, but step before deeper engagement)
- Donation (signals strong interest, and likely to sign them up to marketing from an effective charity)
- Content (value described above, pretty loosely valuable)
I would stand by the obvious ranking (though maybe more quibling should separate âfellowship/ fundingâ). However, we also need to provide values for each step. In the ideal world, we could base this on the drop off rate between each step in our funnel (perhaps X% of people who do a fellowship donât end up getting a relevant job, therefore fellowship clicks should be ranked as the value of job ads - X). We should come back to this, but for now, Iâll do some crude estimates.
- Job ad clicks = 1
- Letâs say 50% of fellowship graduates , or funding recipients get a relevant job. Fellowship = 0.5
- 25% of courses graduates do something relevant afterwards = 0.25
- Donation = 0.05
- Content = 0
Category | Value per click |
Content | 0 |
Job ad (and job boards) | 1 |
Fellowship/ funding | 0.5 |
Courses | 0.25 |
Donation opportunities | 0.05 |
How could the metric be better?
- If we knew more about our audience, we could multiply the impact-weighted total clicks by the âvalueâ of our current email-opening audience (i.e. each click is more valuable in expectation if the audience is highly motivated, educated, etcâŚ). Hopefully if we get a marketer involved we could learn more about this.
Hypotheses to test:
H1: âWithout content, people wouldnât be interested in reading the EA Newsletterâ. I.e. without content, overall clicks go down.
- This might be wrong - the content doesnât seem to have led to more subscribers.
- And yet - people come back to click every month.
Vasco Grilođ¸ @ 2025-05-22T13:57 (+2)
Thanks for sharing, Toby! I very much like this sort of transparency.
- This might be wrong - the content doesnât seem to have led to more subscribers.
I am not sure what you mean here. Only subscribers can see the newsletter's content, so how can content lead to more subscribers? Are you saying that content has not prevented "the monthly loss of around 200-300 subscribers per newsletter"?
James Herbert @ 2025-05-27T00:26 (+2)
Very cool! Thanks for sharing.
Iâm curious, what proportion of your readers does Mailchimp report as being based in Amsterdam? Iâm sure itâs quite misleading for various technical reasons but Iâm still intrigued.
I heard from the EA Survey team that Amsterdam was the sixth largest hub, in terms of respondents, in the 2024 survey (2.4%). The biggest hub was London, which generated something like 6% of respondents.
Toby Tremlettđš @ 2025-05-27T11:10 (+4)
Interesting! It doesn't give a %, just shows this:
Based on your comment, this seems less unlikely. But the "these locations look accurate to me" thing at the bottom suggests mailchimp isn't very confident here.
James Herbert @ 2025-05-27T14:20 (+2)
Thanks!