Note — I’m writing this in a personal capacity, and am not representing the views of my employer.
I’m interested in the EA red-teaming contest as an idea, and there are lots of interesting critiques I’d want to read. But I haven’t seen any of those written yet. I put together a big list of critiques of EA I’d be really interested in seeing come out of the contest. I personally would be interested in writing some of these, but don’t really have time to right now, so I am hoping that by sharing these, someone else will write a good version of them. I’d also welcome people to share other critiques they’d be excited to see written in the comments here!
I think that if someone wrote all of these, there are many where I wouldn’t necessarily agree with the conclusions, but I’d be really interested in the community having a discussion about each of them and I haven’t seen that discussion happen before.
If you want to write any of these, I’m happy to give brief feedback on it, or give you a bunch of bullet-points of my thoughts on them.
Critiques of EA
- Defending person-affecting views
- Person-affecting views try to capture an intuition that is something like, “something can only be bad if it is bad for some particular person or group of people.”
- This specifically interacts with an argument for reducing existential risk that presents reducing x-risk as good because then future people will be born and have good lives. If you take a person-affecting view seriously, maybe you think it is no big deal that people won’t be born, since they aren’t any particular people, and thus not being born is not bad for them.
- Likewise, some forms of person-affecting views are generally neutral toward adding additional happy people, while a non-person-affecting total utilitarian view is in favor of adding additional happy people, all other things being neutral.
- People in the EA community have critiqued person-affecting views for a variety of reasons related to the nonidentity problem
- Person-affecting views are interesting, but pretty much universally dismissed in the EA community. However, I think a lot of people find the intuitions behind person-affecting views to be really powerful. And, if there was a convincing version of a person-affecting view, it probably would change a fair amount of longtermist prioritization.
- I’d really love to see a strong defense of person-affecting views, or a formulation of a person-affecting view that tries to address critiques made of them. This seems like a fairly valuable contribution to making EA more robust. I think I’d currently bet on the community not realigning in light of this defense, but it seems worth trying to make because the intuitions powering person-affecting views are compelling.
- Person-affecting views try to capture an intuition that is something like, “something can only be bad if it is bad for some particular person or group of people.”
- It’s getting harder socially to be a non-longtermist in EA.
- People seem to push back on the idea that EA is getting more longtermist by pointing at the increase in granting in the global health and development space.
- Despite this, a lot of people seem to have a sense that EA is pushing toward longtermism.
- I think that the argument that EA is becoming more longtermist is stronger than doubters give it credit for, but it might mostly have to do with social dynamics in the space.
- The intellectual leaders / community building efforts seem to be focused on longtermism.
- The “energy” of the space seems mostly focused on longtermist projects.
- People join EA interested in neartermist causes, and gradually become more interested in longtermist causes (on average)
- I think that these factors might be making it socially harder to be a non-longtermist who engages with the EA community, and that is an important and missing part of the ongoing discussion about EA community norms changing.
- EA is neglecting trying to influence non-EA organizations, and this is becoming more detrimental to impact over time.
- I’m assuming that EA is generally not missing huge opportunities for impact.
- As time goes on, theoretically many grants / decisions in the EA space ought to be becoming more effective, and closer to what the peak level of impact possible might be.
- If this is the case, changing EAs’ minds is becoming less effective, because the possible returns on changing their views are lower.
- Despite this, it seems like relatively little effort is put into changing the minds of non-EA funders, and pushing them toward EA donation opportunities, and a lot more effort is put into shaping the prioritization work of a small number of EA thinkers.
- If this is the case, it seems like a non-ideal strategy for the EA research community — the possible returns on changing the prioritization of EA thinkers are fairly small.
- The fact that everyone in EA finds the work we do interesting and/or fun should be treated with more suspicion.
- It's exciting and interesting to work on lots of EA topics. This should make us mildly suspicious that they are as important as we think.
- I’ve worked professionally in EA and EA-adjacent organizations since around 2016, and the entire time, I’ve found my work really really interesting.
- I know a lot of other people who find their work really really interesting.
- I’m pretty confident that what I find interesting is not that likely to overlap with what’s most important.
- It seems pretty plausible from this that I’m introducing fairly large biases into what I do because of what I find interesting, and missing a lot of opportunities for impact.
- It seems plausible that this is systematically happening across the EA space.
- Sometimes change takes a long time — EA is poorly equipped to handle that
- I’ve been involved in the EA space and adjacent communities for around 8 or so years, and throughout that time, it feels like the space has changed dramatically.
- But some really important projects probably take a long time to demonstrate progress or results.
- If the community is going to continue changing and evolving rapidly, it seems like we are not equipped to do these projects.
- There are some ways to address this (e.g. giving endowments to charities so they can operate independently for longer), but these seem underexplored in the EA space.
- Alternative models for distributing funding are probably better and are definitely under-explored in EA
- Lots of people in EA buy into the idea that groups of people make better decisions than individuals, but all our funding mechanisms are built around a small number of individuals making decisions.
- The FTX regranting program is a counterexample to this (and a good one), but still is fundamentally not that transformative, and only slightly improves the number of people making decisions about funding.
- There are lots of alternative funding models that could be explored more, and should be!
- Distribute funds across EA Funds by the number of people donating to each cause, or by people’s aggregate weighting of causes, instead of total donations (thus getting a fund distribution that represents priorities of the community instead of wealth).
- Play with projects like Donation Democracy on a larger scale.
- Trial consensus mechanisms for distributing fundings with large groups of donors (likely moderating against very unusual but good grants, but improving average grant quality).
- Pursue more active grantmaking of ideas that seem promising (not really using group decision making but still a different funding approach).
- Lots of people in EA buy into the idea that groups of people make better decisions than individuals, but all our funding mechanisms are built around a small number of individuals making decisions.
- EA funder interactions with charities often make it harder to operate an EA charity than it has to be
- I’ve worked at several EA and non-EA charities, and overall, the approach to funding in the EA space is vastly better than the non-EA world. But it still isn’t ideal, and lots of problems happen.
- Sometimes funders try to play 5d chess with each other to avoid funging each other’s donations, and this results in the charity not getting enough funding.
- Sometimes funders don’t provide much clarity on the amount of time they intend to fund organizations for, which makes it harder to operate the organization long-term or plan for the future.
- Lots of EA funding mechanisms seem basically based on building relationships with funders, which makes it much harder to start a new organization in the space if you’re an outsider.
- Relatedly, it’s harder to build these relationships without knowing a large EA vocabulary, which seems bad for bringing in new people.
- These interactions seem addressable through funders basically thinking less about how other funders are acting, and also working on longer time-horizons with grants to organizations.
- I’ve worked at several EA and non-EA charities, and overall, the approach to funding in the EA space is vastly better than the non-EA world. But it still isn’t ideal, and lots of problems happen.
- RFMF doesn’t make much sense in many of the contexts it’s used in
- Room For More Funding makes a lot of sense for GiveWell-style charity evaluation. It’s answering a straightforward question like, “how much more money could this charity absorb and continue operating at X level of cost-effectiveness with this very concrete intervention.”
- But this is not how charities that don’t do concrete interventions operate, and due to some historical reasons (like GiveWell using this term), people often ask these charities about RFMF.
- Charities estimating their own “RFMF” probably mean a variety of different things, including:
- How much money they need to keep operating for another year at their current level
- How much money they could imagine spending over the next year
- How much money they could imagine spending over XX years
- How much money a reasonable strategic plan would cost over some time period
- We need a more precise language for talking about the funding needs of research or community-building organizations, so that people can interpret these figures accurately.
- Suffering-focused longtermism stuff seems weirdly sidelined
- S-risks seem like at least a medium-big deal, but seem to be treated as basically not important.
- Lots of people in the EA space seem to believe that a large portion of future minds will be digital (e.g. here is a leader in the EA space saying they think there is an ~80% chance of this).
- If this happens, it seems totally reasonable to give some credence to worlds where lots of digital minds suffer a lot.
- This possibility seems to be taken seriously by only a few organizations (e.g. Center on Long-term Risk) but basically seems like a fringe position, and doesn’t seem represented at major grantmakers and community-building organizations.
- I think this is probably because these views are somewhat associated with really strong negative utilitarianism, but they seem also very concerning for total utilitarians.
- It seems bad to have sidelined these perspectives, and s-risks probably should be explored more, or at least discussed more openly (especially from a non-negative utilitarian perspective).
- Logical consistency seems underexplored in the EA space
- A big core and unstated premise for an EA approach is that ethics should be consistent.
- I don’t think I have particularly good reasons for thinking ethics should be consistent, especially if I adopt a moral realism that seems somewhat popular in the EA space.
- It seems like there are plausible reasons for explaining why I think ethics should be consistent that don’t have much to do with morality (e.g. maybe logic is an artifact of the way human languages are structured, etc.).
- I’d be interested in someone writing a critique of the idea that ethics have to be consistent, as it seems like it underpins a lot of EA thinking.
- There is a lot of philosophical critique of moral particularism, but I think that EA cases are interesting both because of the high degree of interest in moral realism, and because EA explicitly acts on what might be only thought experiments in other contexts (like very low-likelihood, high-EV interventions).
- If the result of a critique is that consistency is actually fairly important, that seems like it would have ramifications for the community (consistency seems to be assumed to be important, but a lot of decisions don’t seem particularly consistent).
- People are pretty justified in their fears of critiquing EA leadership/community norms
- It's actually probably bad for your career to publicly critique EA leadership, or to write some critiques in response to this contest, but the community seems to want to pretend that it isn't in the framing of these invitations for critique.
- I think I have a handful of critiques I want to make about EA that I am fairly certain would negatively impact my career to voice, even though I believe they are good faith criticisms, and I think engaging with them would strengthen EA.
- I removed a handful of items from this list because I basically thought they’d be received too negatively, and that could harm me or my employer.
- I think everything I removed was fairly good faith and important as a critique.
- I'm making a list of critiques I want to see, not even critiques I necessarily believe, but still felt like I had to remove things.
- I think it is bad that I felt the need to remove anything, but I also think it was the right decision.
- I removed a handful of items from this list because I basically thought they’d be received too negatively, and that could harm me or my employer.
- I’ve heard anecdotes about people posting what seem like reasonable critiques and being asked to take them down by leadership in the EA space (for reasons like “they are making a big deal out of something not that important”).
- I think that this is a bad dynamic, and has a lot to do with the degree of power imbalances in the EA space, and how socially and intellectually tight-knit EA leadership is.
- This seems like a dynamic that should be discussed openly and critiqued.
- Grantmakers brain-drain organizations — is this good?
- My impression of the EA job market is that people consider jobs at grantmakers to be the highest status.
- The best paying jobs in EA (maybe besides some technical roles), are at grantmakers.
- This probably causes some degree of “brain-drain” where grantmakers are able to get the most talented researchers.
- This seems like it could have some negative effects for organizations that are bad for the community.
- Grantmakers are narrowly focused on short-term decisions ("issue this grant or not?"), rather than doing longer-term or exploratory research.
- This means the most skilled researchers are taken away from exploratory questions to short-term questions, even if the skilled researcher might prioritize the exploratory research over the short-term questions.
- Grantmakers tend to be relatively secretive / quiet about their decision-making and thinking, so the research of the best researchers in the community often isn’t shared more widely (and thus can’t be more widely adopted).
- Grantmakers are narrowly focused on short-term decisions ("issue this grant or not?"), rather than doing longer-term or exploratory research.
- If this dynamic is net-negative (it has good effects too, like grantmakers making better grants!), then addressing it seems pretty important.
- Effective animal advocacy mostly neglects the most important animals
- The effective animal advocacy movement has focused mostly on welfare reforms for laying hens and broiler chickens for the last several years.
- I think this is probably partially for historical reasons — the animal advocacy movement was already a bit focused on welfare reforms, and of the interventions being pursued at the time it seemed like the most promising.
- I think that it is possible this approach has entirely missed a ton of impact for other farmed animals (especially fish and insects) and wild animals, and that prioritizing these other animals from the beginning could have been a much more effective use of that funding, even if new organizations needed to be formed, etc.
- I think in particular not working on insect farming over the last decade may come to be one of the largest regrets of the EAA community in the near future.
- This dynamic probably will continue to play out to some extent, and it seems like it could be important to address it sooner rather than later.
- Large organizations in the space seem focused only on specific strategies and specific animals, and priorities over the next few years are still on laying hens and broilers.
Something I personally would like to see from this contest is rigorous and thoughtful versions of leftist critiques of EA, ideally translated as much as possible into EA-speak. For example, I find "bednets are colonialism" infuriating and hard to engage with, but things like "the reference class for rich people in western countries trying to help poor people in Africa is quite bad, so we should start with a skeptical prior here" or "isolationism may not be the good-maximizing approach, but it could be the harm-minimizing approach that we should retreat to when facing cluelessness" make more sense to me and are easier to engage with.
That's an imaginary example -- I myself am not a rigorous and thoughtful leftist critic and I've exaggerated the EA-speak for fun. But I hope it points at what I'd like to see!
Strong upvote. I'm a former leftist and I've got a soft spot for a few unique ideas in their memeplex. I read our leftist critics whenever I can because I want them to hit the quality target I know the ideas are worth in my mind, but they never do.
If anyone reading this knows leftist critics that you think have hit a reasonable quality bar or you want to coauthor a piece for the contest where we roleplay as leftists, DM me on the forum or otherwise hit me up.
I consider myself a current leftist, and I honestly don't have a big "leftist critique of ea". Effective altruism seems uncomplicatedly good according to all the ideas I have that I consider "leftist", and leftism similarly seems good according to all the ideas that I consider EA.
Effective altruists as individuals aren't always radical leftist of course, though they are pretty much all left of center. If you press me to come up with criticisms of EA, I can think of harmful statements or actions made by high profile individuals to critique, I guess, though idk if that would be useful to anyone involved. I can also say that the community as a whole doesn't particularly escape the structural problems and interpersonal prejudices found in larger society - but it's certainly not any worse than larger society. Also EA organizations, are not totally immune to power and corruption and internal politics and things like that, these things could be pointed out too. What I am saying is, effective altruists and institutions aren't immune from things like racism and sexism and stuff like that. But that's true of most people and organizations, including leftist ones. But there's... (read more)
I definitely agree with this. Here are a bunch of ideas that are vaguely in line with this that I imagine a good critique could be generated from (not endorsing any of the ideas, but I think they could be interesting to explore):
I think that one issue is that lots of the left just isn't that utilitarian, so unless utilitarianism itself is up for debate, it seems hard to know how seriously people in the EA community will take lefty critiques (though I think that utilitarianism is worth debating!). E.g. "nobody's free until everyone is free" is fundamentally not a utilitarian claim.
For onlookers I want to point out that this doesn't read as leftist criticism.
This is very close (almost identical) to what classical conservatives say:
From:
- https://www.econtalk.org/william-macaskill-on-effective-altruism-and-
... (read more)I think the vocabulary is not fully separable from the ideology. As the latter evolves, I'd expect changes to be required in the former.
And for what it's worth, all the versions you gave are equally intellectually challenging for me to understand. The jargon is easier for some people but harder for others, most importantly to outsiders. This also means it's unfair to expect outsiders to voice their views in insider-speak.
I agree that S-risks are more neglected by EA than extinction risks, and I think the explanation that many people associate S-risks with negative utilitarianism is plausible. I'm a regular utilitarian and I've reached the conclusion that S-risks are quite important and neglected, and I hope this bucks the perception of those focused on S-risks.
Two recent, related articles by Magnus Vinding that I enjoyed reading:
Strong upvote. My personal intuitions are suffering focused, but I’m currently convinced that I ought to do whatever evidential cooperation in large worlds (ECL) implies. I don’t know exactly what that is, but I find it eminently plausible that it’ll imply that extinction and suffering are both really, really bad, and s-risks, especially according to some of the newer, more extreme definitions, even more so.
Before ECL, my thinking was basically: “I know of dozens of plausible models of ethics. They contradict each other in many ways. But none of them is in favor of suffering. In fact, a disapproval of many forms of suffering seems to be an unusually consistent theme in all of them, more consistent than any other theme that I can identify.[1] Methods to quantify tradeoffs between the models are imprecise (e.g., moral parliaments). Hence I should, for now, focus on alleviating the forms of suffering of which this is true.”
Reducing suffering – in all the many cases where doing so is unambiguously good across a wide range of ethical systems – still strikes me as at least as robust as reducing extinction risk.
- ^
... (read more)Some variation on universalizability, broadly construed, may be a conten
This seems suboptimal, particularly if more people feel like that. But it does seem fixable: I'm up for receiving things like this anonymously at this link, waiting for a random period, rewording them using GPT-3, and publishing them. Not sure what proportion of that problem that would fix, though.
The criticism contest has an anonymous submission form too.
It's not anonymous, it records the name associated with your google account. (Of course you can just create a google account with a fake name, but then you can also just make an EA forum account with a fake name and post here.)
I believe this is just the confusing way that Google handles anonymous forms. It states the account you are currently using, but then has a parenthetical indicating that the information won't be shared.
Yeah I asked em to fix this
Yeah, I think that some percentage of this problem is fixable, but I think one issue is that there are lots of important critiques that might be made from a place of privileged information, and filling in a form will be deanonymizing to some extent. I think this is especially true when an actor's actions diverge from stated values/goals — I think many of the most important critiques of EA that need to be made come from actions diverging from stated values/goals, so this seems hard to navigate. E.g. I think your recent criminal justice reform post is a pretty good example of the kind of critique I'm thinking of, but there are ones like it based on actions that aren't public or at least aren't written up anywhere that seem really important to have shared.
Related to this, I feel like a lot of people in EA lately have expressed a sentiment that they have general concerns like the one I outlined here, but can't point to specific situations. One explanation for this is that their concerns aren't justified, but another is that people are unwilling to talk about the specifics.
That being said, I think the anonymous submission form is really helpful, and glad it exists.
For what its worth, I've privately been contacted more about about this particular critique resonating with people than any other in this post by a large degree, which suggests to me that many people share this view.
Someone suggested I should mention a few of the EA critiques I'm personally working on. I've only skimmed the comment so sorry if I've missed something relevant.
Three are of longtermism (and prospectively with funding support from the Forethought Foundation).
I've also got a 'red-team' of Open Philanthropy's cause prioritisation framework. That's written and should appear within a mont... (read more)
I know that "everyone" was an intentional exaggeration, but I'd be interested to see the actual baseline statistics on a question like "do you find EA content interesting, independent of its importance?"
Personally, I find "the work EA does" to be, on average... mildly interesting?
In college, even after I found EA, I was much more intellectually drawn to random topics in psychology and philosophy, as well as startup culture. When I read nonfiction books for fun, they are usually about psychology, business, gaming, or anthropology. Same goes for the Twitter feeds and blogs I follow.
From what I've seen, a lot of people in EA have outside interests they enjoy somewhat more than the things they work on (even if the latter takes up much more of their time).
*****
Also, as often happens, I think that "EA culture" here may be describing "the culture of people who spend lots of time on EA Twitter or the Forum", rather than "the culture of people who spend a lot of their time on EA work". Members of the former group seem more likely to find their work interesting and/or fun; the people who feel more like I do probably spend their free time on other interests.
I think I agree with everything here, though I don't think the line is exactly people who spend lots of time on EA Twitter (I can think of several people who are pretty deep into EA research and don't use Twitter/aren't avid readers of the Forum). Maybe something like, people whose primary interest is research into EA topics? But it definitely isn't everyone, or the majority of people into EA.
there's an EA Twitter?
+1 to this — it's something I've been thinking about quite a bit lately, and I'm happy you mentioned it.
I'm not convinced the EA community will be able to effectively solve the problems we're keen on tackling if we mainly rely on a (relatively) small group of people who are unusually receptive to counterintuitive ideas, especially highly technical problems like AI safety. Rather, we'll need a large coalition of people who can make progress on these sorts of challenges. All else equal, I think we've neglected the value of influencing others, even if these folks might not become highly active EAs who attend conferences or whatever.
Thanks for sharing! I'd also love to read some of these critiques more fleshed out! Really appreciate that you posted bullet point summaries instead of either holding off for a more developed critique or just posting a vague list without summaries 😀
This seems very much too strong to me:
I consider myself part of the EA community and I do not dismiss PAV... I am very sympathetic to them. When I have presented these others have not been dismissive. They are usually at least mentioned as a potential important part of a balanced breakfast of moral uncertainty.
Some articles in the forum seem to be sticking up for PAV, by Michael St Jules and others:
https://forum.effectivealtruism.org/posts/cXEvzaQhQGfvFSy5Z/the-problem-of-possible-populations-animal-farming
https://forum.effectivealtruism.org/posts/2BWQ4NrCEP7a4vzaW/defending-the-procreation-asymmetry-with-conditional
Here, the author states:
(Love your post by the way)
Here are four more things that I’m somewhat skeptical of and would like someone with more time on their hands and the right brain for the topic to see whether they hold water:
- Evidential cooperation in large worlds is ridiculously underexplored considering that it might “solve ethics” as I like to epitomize it. AI safety is arguably more urgent, but maybe it can even inform that discipline in some ways. I have spent about a quarter of a year thinking about ECL, and have come away with the impression that I can almost ignore my own moral intuitions in favor of what little I think I can infer about the compromise utility function. More research is needed.
- There is a tension between (1) the rather centralized approach that the EA community has traditionally taken and that is still popular, especially outside key organizations like CEA, and the pervasive failures of planned economies historically, and between (2) the much greater success of Hayakian approaches and the coordination that is necessary to avert catastrophic coordination failures that can end our civilization. My cofounders and I have started an EA org to experiment with market mechanisms for the provision of public and commo
... (read more)+1 for S-risk. I was surprised by the lack of its discussion on 80k podcast.
I'm working on a piece on this. (It's only a "critique of longtermism" in the weak sense that I think some longtermist claims are overstated.) If someone is working on something similar or interested in giving feedback, please DM me!
This is an interesting question in itself that I would love someone to explore in more detail. I don't think it's an obviously true statement. Two give a few counterpoints:
I'd point out this attempt which was well-explained in a forum post. There is also this which I haven't really engaged with much but seems relevant. My sense is that the philosophical community has been trying to formulate a convincing person-affecting view an... (read more)
This is something that I find myself thinking about a lot. If you could wave a magic wand, what changes would you implement? I'm aware of Rethink's work to incubate the Insect Welfare Project - with that in mind, do you have any recommendations for other EAAs to help out with insect work in the meantime, even if this requires a large commitment (like starting a new org)? (I am aware of your past research... (read more)
Right now the thing we are most interested in is finding a strong candidate to work on the Insect Welfare Project full-time: https://careers.rethinkpriorities.org/en/jobs/50511
Donations would also be helpful. This kind of stuff can be harder to find financial support for than other things in EA. https://rethinkpriorities.org/donate
Perhaps some of these criticisms might be even more useful if they were framed as opportunities. For example:
I'm not sure if this matters much but I think it puts the focus on what can be done (e.g., what projects EA entrepreneurs could start), rather than on people feeling bad about what they already did and then defending it.
Thanks for writing this post, I think it raises some interesting points and I'd be interested in reading several of these critiques.
(Adding a few thoughts on some of the funding related things, but I encourage critiques of these points if someone wants to write them)
I'm not aware of this happening very much, at least between EA Funds, Open Phil and FTX (but it's plausible to me that this does happen ... (read more)
Thanks for the response!
RE 5d chess - I think I've experienced this a few times at organizations I've worked with (e.g. multiple funders saying, "we think its likely someone else will fund this, so are not/only partially funding it, though we want the entire thing funded," and then the project ends up not fully funded, and the org has to go back with a new ask/figure things out. This is the sort of interaction I'm thinking of here. It seems costly for organizations and funders. But I've got like an n=2 here, so it might just be chance (though one person at a different organization has messaged me since I posted this and said this point resonated with their experiences). I don't think this is intentional on funders part!
RE timelines - I agree with everything here. I think this is a tricky problem to navigate in general, because funders can have good reasons to not want to fund projects for extended periods.
RE vocabulary - cultural differences make sense as a good explanation too. I can think of one instance where I felt like this was especially noticeable - I encouraged a non-EA project I thought was promising to apply for funding, and they didn't get it. I pitched the funder on the... (read more)
This has felt very true for me!
I came across EA way back around 2011 when I was at university, pre-longtermism... EA at that point formalised a lot of my existing thinking/values and I made graduate career decisions in line with 80k advice at the time. I started getting more involved again about a year ago and was surprised ... (read more)
I'm confused about whether I should note my disagreements here or just wait for someone to write the proper versions.
So I'll just note one that I really want to see: I was unpersuaded by this
until I saw
Alternate funding models as a solution to the grantmaking bottleneck could be great!
Thanks, I found this list really interesting!
Although note that Will MacAskill supports lead elimination from a broad longtermist perspective:
... (read more)I would like to agree with Aaron's comment and make a stronger claim - my impression is that many EAs around me in Israel, especially those coming from a strong technical background, don't find most direct EA-work very intellectually interesting or fun (ignoring its impact).
Speaking for myself, my background is mostly in pure math and in cyber-security research / software engineering. Putting aside managerial and entrepreneurial roles, it seems to... (read more)
Could you say more about this? (My anecdata suggest that EAs typically embrace anti-realism)
I want to echo all the interest in leftist critique (usually it reduces into something about colonialism, racism, or capitalism), but from the perspective that @JulianHazell brought up, i.e. of being able to reach a wider audience. I.e. at some point, EA needs to get better at representing itself in a nontechnical manner.
Btw, I'm writing from the perspective of someone who doesn't have a job in EA, but who sees a lot of leftist leanings in organizations that I'm a part of.
My personal experience is that I doubt a point-by-point rebuttal would ch... (read more)
I’m particularly optimistic about “impact markets” here, where you get:
That model promises to gre... (read more)
What about critiques that you or another EA might not love but needs to read?
The intra-EA jargon is strong with this one, young padawan(s).