I am the Principal Research Director at Rethink Priorities. I lead our Surveys and Data Analysis department and our Worldview Investigation Team.
The Worldview Investigation Team previously completed the Moral Weight Project and CURVE Sequence / Cross-Cause Model. We're currently working on tools to help EAs decide how they should allocate resources within portfolios of different causes, and to how to use a moral parliament approach to allocate resources given metanormative uncertainty.
The Surveys and Data Analysis Team primarily works on private commissions for core EA movement and longtermist orgs, where we provide:
Formerly, I also managed our Wild Animal Welfare department and I've previously worked for Charity Science, and been a trustee at Charity Entrepreneurship and EA London.
My academic interests are in moral psychology and methodology at the intersection of psychology and philosophy.
Survey methodology and data analysis.
RP actually did some empirical testing on this and we concluded that people really like the name "Effective Altruism", but not the ideas, values or mission.
That's unfortunate. But I think it suggests there's scope for a new 'Centre for Effective Altruism' to push forward exciting new ideas that have more mainstream appeal, like raising awareness of the cause du jour, while the rebranded Center for ████████ continues to focus on all the unpopular stuff.
I think it was a mistake to post about "Hidden Capabilities Evals Leaked In Advance to Bioterrorism Researchers and Leaders (minor)" in a public forum... it seems too minor! Maybe if you'd included some specific examples it would be more useful.
Are introductory fellowships and other courses run by local EA groups considered under "educational courses" or does that mean things like philosophy classes, where EA might be mentioned as a concept?
There are two different things to consider here:
In most cases where someone joins EA through an activity run by their EA group, they'll select EA Group. I went through references to "group" in the comments for those who did not select EA Group and most (23/36) were not referring to an EA Group, 3 referred to a friend referring them to a group, 2 referred to seeing an ad for a group, 5 mentioned something else first introducing them to EA but a group being useful for them beginning to seriously engage and the rest were miscellaneous/uncategorisable. I went through comments of those who did select EA Group and 9/111 (8%) mentioned a fellowship. You can see a more detailed analysis of comments by those who selected EA Group in 2020 here, where 7% mentioned a fellowship, but a larger number mentioned an event or fresher's fair.
Alternatively, would it be possible to ask for country-specific data analysis?
For Estonia, there are not very clear differences (p=0.023). This is from n=37 for Estonia, so we should not be very confident about the differences.
Thanks for raising this interesting point!
If we look direction at the association between active outreach and levels of engagement, we see no strong relationship (aside from within the very small "No engagement with EA" category).
I agree that it's intuitively plausible, as you say, that people who actively sought out EA might be expected to be more inclined to high engagement.
But I think that whether the person encounters EA through active outreach (influencing selection) may often be in tension with whether the person encounters EA through a high-touch form of engagement (leading to different probability of continued engagement). For example, EA Groups are quite active outreach (ex hypothesi, leading to lower selection for EA inclination), but higher touch (potentially leading to people being more likely to get engaged), whereas a book is more passive outreach (ex hypothesi, leading to higher selection for EA inclination), but lower touch (and so may be less likely to facilitate engagement)[1].
There may be other differences between these particular two examples of course. For example, EA Groups and EA books might select in different ways.
@titotal I'm curious whether or to what extent we substantively disagree, so I'd be interested in what specific numbers you'd anticipate, if you'd be interested in sharing.
I don't say strictly 0% only because I think there's always the possibility for a few unusual cases, e.g. someone is googling how to do good and happens across an old post about EAG or their inactive local group.
I'm imagining someone googling "ethical career" 2 years from now and finding 80k, noticing that almost every recent article, podcast, and promoted job is based around AI, and concluding that EA is just an AI thing now.
I definitely agree that would eventually become the case (eventually all the older non-AI articles will become out of date). I'm less sure it will be a big factor 2 years from now (though it depends on exactly how articles are arranged on the website and so how salient it is that the non-AI articles are old).
It could also be bad even for AI safety: There are plenty of people here who were initially skeptical of AI x-risk, but joined the movement because they liked the malaria nets stuff. Then over time and exposure they decided that the AI risk arguments made more sense than they initially thought, and started switching over.
I also think this is true in general (I don't have a strong view about the net balance in the case of 80K's outreach specifically).
Previous analyses we conducted suggested that over half of Longtermists (~60%) previously prioritised a different cause and that this is consistent across time.
You can see the overall self-reported flows (in 2019) here.
Thanks Arden!
I also agree that prima facie this strategic shift might seem worrying given that 80K has been the powerhouse of EA movement growth for many years.
That said, I share your view that growth via 80K might reduce less than one would naively expect. In addition to the reasons you give above, another consideration is our finding is that a large percentage of people get into EA via 'passive' outreach (e.g. someone googles "ethical career" and finds the 80K website', and for 80K specifically about 50% of recruitment was 'passive'), rather than active outreach, and it seems plausible that much of that could continue even after 80K's strategic shift.
Our framings will probably change. It's possible that the framings we use more going forward will emphasise EA style thinking a little less than our current ones, though this is something we're actively unsure of.
As noted elsewhere, we plan to research this empirically. Fwiw, my guess is that broader EA messaging would be better (on average and when comparing the best messaging from each) at recruiting people to high levels of engagement in EA (this might differ when looking to recruit people directly into AI related roles), though with a lot of variance within both classes of message.
Maybe a slogan could be "All altruism is effective"?