I currently lead EA funds.
Before that, I worked on improving epistemics in the EA community at CEA (as a contractor), as a research assistant at the Global Priorities Institute, on community building, and Global Health Policy.
Unless explicitly stated otherwise, opinions are my own, not my employer's.
You can give me positive and negative feedback here.
This fund was spun out of the Long-Term Future Fund (LTFF), which makes grants aiming to reduce existential risk. Over the last five years, the LTFF has made hundreds of grants, specifically in AI risk mitigation, totalling over $20 million. Our team includes AI safety researchers, expert forecasters, policy researchers, and experienced grantmakers. We are advised by staff from frontier labs, AI safety nonprofits, leading think tanks, and others.
More recently ARM Fund has been doing active grantmaking in AIS areas, we'll likely write more about this soon. I expect the funds to become much more differentiated in staff in the next few months (though that's not a commitment). Longer term, I'd like them to be pretty separate entities but for now they share roughly the same staff.
If you perceive any sort of downside from it, you can always remove it again.
Aren't most of the downsides and upsides to norms hard to reverse (almost by definition)? Maybe you don't think the upside is in getting other people to also participate in using the signal - but my read of the OP thinks that this is mostly about creating norms.
I've only skimmed this post, but I think this I agree with all of the main points. I would prefer EA meta orgs that provide benefits to people with money charged for some of their services. I do think the situation is significantly more complicated with orgs that receive substantial institutional funding so I think the original post applies a bit less to orgs like CEA, and more to specific EA groups or small-scale projects (including projects that the EAIF funds).
I suggested to various regional EA groups that they should try and cover some fraction of their costs from members, but there was quite a lot of negative push back (e.g. fundraising distracting them from their main jobs).[1]
EAIF is most interested in funding projects that shouldn't be funded via regular markets, or might not be noticed as being especially valuable (e.g. many public goods in the nonexcludable and nonrivalrous sense).
@Harri Besceli feel free to push back on any of this if it conflicts with your impression of how the EAIF does/should work.
I still think some version of this is workable but it's not a priority for the EAIF to figure out right now.
A few things that come to mind that I appreciate in people’s applications:
To be clear you don’t need to do any of these things to get funding, but I often find that applications are improved after people consider some of these bullet points.
Hi Markus,
For context I run EA Funds, which includes the EAIF (though the EAIF is chaired by Max Daniel not me). We are still paying out grants to our grantees — though we have been slower than usual (particularly for large grants). We are also still evaluating applications and giving decisions to applicants (though this is also slower than usual).
We have communicated this to the majority of our grantees, but if you or anyone else reading this urgently needs a funding decision (in the next two weeks), please email caleb [at] effectivealtruismfunds [dot] org with URGENT in the subject line, and I will see what I can do. Please also include:
You can also apply to one of Open Phil’s programs; in particular, Open Philanthropy’s program for grantees affected by the collapse of the FTX Future Fund may be particularly of note to people applying to EA Funds due to the FTX crash.