I worked as a software/product engineer at the Centre for Effective Altruism for three years, and recently became the EA Forum Project Lead. If you'd like to support our work, sign up for a 30 min user interview with someone on our team. Hearing about your experience with the Forum helps us improve the site for everyone.
In general, we'd be happy to hear any feedback you have! :) Feel free to contact us or post in this suggestion thread. You can also give us anonymous feedback via this form.
Thanks for the feedback! I think moderation is tricky and I'm relatively new at it myself. I'm sad at how long users can get stuck in the queue, and I'd love to improve how fast we resolve moderation questions, but where exactly we draw these lines will probably be a learning process for me, and we'll continue to iterate on that.
It looks like you submitted the comment on Dec 17, and our facilitator messaged you on Jan 6 (the delay partly being due to people being out for the holidays), and then they approved your comment a little over a week after messaging you. Yeah I agree that this was an edge case, and I don't think you were being malicious, but I think you could have made your point more productively by, for example, just using "torture".
I feel that using the rejected content feature would give our team more leeway to be opinionated about shaping the home page of our site (compared to now), and we'd feel somewhat free to reject things that don't fit the type of discussions we want to see. For example, it looks like LW rejects posts from new users that don't have a clear introduction. So I think if something is an edge case in the current system, then it would likely get rejected under the other system.
Hi! I just want to start by clarifying that a user’s first post/comment doesn’t go up immediately while our facilitators/moderators check for spam or a clear norm violation (such as posting flame bait/clear trolling). Ideally this process takes no more than a day, though we currently don’t have anyone checking new users outside of approximately US Eastern Time business hours.
However, some content (like your first comment) requires additional back and forth internally (such as checking with moderators) and/or with the new user. This process involves various non-obvious judgement calls, which is what caused a long delay between your submitting the comment and us reaching out to you (plus the fact that many people were out over the winter holidays). In the case of your comment, we asked you to edit it and you didn’t respond to us or edit the comment for over a week, and then our facilitator felt bad for keeping you in the queue for so long so they approved your comment.
We currently do not use the rejected content feature that LW uses. Instead, almost all[1] of the content that may have been rejected under their system ends up appearing on the rest of our site, and we currently mostly rely on users voting to make content more or less visible (for example, karma affects where a post is displayed on the Frontpage). I plan to seriously consider whether we should start using the rejected content feature here soon; if so, then I expect that we’ll have the same page set up.
I think that, if we had been using the rejected content feature, the right move would have been for us to reject your comment instead of approving it.
My guess is that there are edge cases, but in practice we keep our queue clear, so my understanding is that users are typically not in limbo for more than a few days. Things like spam are not rejected — accounts that post spam are banned.
To quickly add on to what Toby wrote: the CEA Online Team has also been redesigning effectivealtruism.org and we expect to soft launch it soon. I post quick takes when we update our half-quarterly plans, so you can follow along there. :)
For example many editions of the EA handbook spend a huge fraction of their introductions to other cause areas effectively arguing why you should work on AI instead. CEA staffers very heavily favor AI.
Just wanted to quickly add that I don't think that this is quite accurate.
My experience facilitating the Intro Fellowship using the previous version of the EA Handbook was that AI basically didn't come up until the week about longtermism, and glancing through the current version that doesn't seem to have changed. Though I welcome people to read the current version of the EA Handbook and come to their own conclusions.
The most recent relevant data on CEA staff cause prio is this post about where people are donating, and I think animal welfare is more common in that list than AI safety (though this only includes a subset of staff who were interested in participating in the post).
I've updated the public doc that summarizes the CEA Online Team's OKRs to add Q2.1 (the next six weeks).
Thanks for the suggestion! I reached out to them last week about their USAID content, and I expect to see something here from them soon. :)
If you see content you like from GiveWell in the future, I encourage you to to reach out to them and suggest that they crosspost it! You can also flag it to myself or Toby and we can reach out, though that may take longer.
Thanks for writing this Ozzie! :) I think lots of things about the EA community are confusing for people, especially relationships between organizations. As we are currently redesigning EA.org it might be helpful for us to add some explanation on that site. (I would be interested to hear if anyone has specific suggestions!)
From my own limited perspective (I work at CEA but don’t personally interact much with OP directly), your impression sounds about right. I guess my own view of OP is that it’s better to think of them as a funder rather than a collaborator (though as I said I don’t personally interact with them much so haven’t given this much thought, and I wouldn’t be surprised if others at CEA disagree). They have their own goals as an organization, and it’s not necessarily bad if those goals are not exactly aligned with the overall EA community. My understanding is that it’s very standard for projects to adapt their pitches for funders that do not have the same goals/values as them. For example, I’m not running the Forum in a way that would maximize career changes[1] (TBH I don’t think OP would want me to do this anyway), but it’s helpful to include data we have about how the Forum affects career changes when writing a funding proposal[2]. In fact, no one at OP has ever asked me to maximize career changes as a requirement before or after receiving funding, nor do I recall anyone at OP ever asking me to make any changes to the Forum (OP staff do provide feedback but I personally weigh those mostly relative to how much I think they understand the Forum — for example, I’d probably weigh Lizka’s feedback higher than anyone at OP).
I acknowledge that this is complicated by the fact that CEA likely has a unique relationship with OP (due to our large size relative to other community building orgs, long history working in this space, and the fact that our current CEO used to work at OP), so I expect that my own experience with OP does not necessarily generalize to other fundees. Also OP is the overwhelmingly largest funder for EA community building, and so the extent to which they are not aligned with the overall EA community does matter, as money straightforwardly gives them power and influence, though I don’t personally have a good picture of the practical effects.
I think that having these discussions in a public community space is valuable, so I appreciate you sharing this here!
For the sake of this comment, I'm assuming that Ozzie's description accurately describes OP's view, though I have never talked with anyone at OP about this so I don't actually know if it's accurate.
Note that I care about improving the world, and I think that getting people to do high-impact jobs is in fact a good way to make the world better.
Yeah I agree that funding diversification is a big challenge for EA, and I agree that OP/GV also want more funders in this space. In the last MCF, which is run by CEA, the two main themes were brand and funding, which are two of CEA’s internal priorities. (Though note that in the past year we were more focused on hiring to set strong foundations for ops/systems within CEA.) Not to say that CEA has this covered though — I'd be happy to see more work in this space overall!
Personally, I worry that funding diversification is a bit downstream of improving the EA brand — it may be hard for people to be excited to support EA community building projects if they feel like others dislike it, and it may be hard to convince new people/orgs to fund EA community things if they read stuff about how EA is bad. So I’m personally more optimistic about prioritizing brand-related work (one example being highlighting EA Forum content on other platforms like Instagram, Twitter, and Substack).
Thanks for flagging these! And sorry for the delayed response. :)