As Katja’s response alludes to, the non-Open-Phil chunk of GiveWell has more than doubled since 2015 (plus EA funds has gone from zero to $9M, etc.)
I see a few comments at the Reddit/LessWrong versions of this post intimating that EA does not want [much] more money, or has stopped trying to fundraise. This was not my impression at all. Is it not true that even just considering GiveWell’s top charities and near-misses, they could absorb many millions more without being saturated?
TylerMaule
Total Funding by Cause Area
Also, while Open Phil’s donations to GiveWell have remained at a similar level, the amount they direct to the EA movement as a whole has grown substantially:
Thanks for the reply, definitely gives me a lot to consider.
“Cause area” is also a pretty weird/arbitrary unit of analysis
Personally, I quite like the cause area distinction. One alternate definition I might propose is that a cause area is a subset of interventions which are plausibly cross-comparable. Direct comparisons across these cause areas are flimsy at best, and even if I felt strongly that one of them was the most effective, I would still value each of the others receiving non-trivial funding for the purposes of (a) hedging (b) worldview diversification.
Also I think the choice of what you are funding within each cause also matters a lot.
It certainly does, but so long as I donate via EA Funds or GiveWell, that decision is passed along to the very most qualified people I know of.
I’m not sure if this makes sense from a donor collaboration/coordination/cooperation standpoint
I might disagree here. Using base rate funding to inform decisions is no different than ‘neglectedness’ as a pillar of EA—If I had to be truly agnostic I suppose I’d give money to climate change, or purchasing COVID vaccines.
That 80k article is very cool, though they also seem to agree: “If the community is unresponsive to what you do, you can (mostly) take a single-player approach to working out the best action.”
a lot of these areas have very large individual donors that aren’t captured
It would be great to know more about these donors, and specifically which orgs they donate to. It’s starting to feel like a satisfactory measure of ‘fundedness’ would require a lot more future work.
I imagine your personal views about the difference in the value of cause areas will dominate this, given that causes might be 10x different whereas these gaps are only 5x at most.
The size of the gaps are dependent on my personal views, so I think we’re in agreement here.
Not taking a side here, but couldn’t you get around this by framing your values as ‘maximizing sum of global utility’? This way there is no need to make a comparison between Joe and [absence of Joe]; I can simply say that Joe’s existence has caused my objective function to increase.
I like the idea of political contributions going to charity, though I can’t help thinking about the game theory implications here:
If I (a left-leaning person who prefers charity to political donations) felt strongly that much more money would come in on the Democrat side, I imagine I’d route my usual donation through this platform under the Republican candidate.
I guess it’s difficult to imagine an actual Republican contributing to this platform unless they preferred giving to charity anyway. Arguably this platform would then only deplete the funds of one candidate (the Democrat), with much of the funds intended for charity in the first place. But still, to be clear, this would be a net positive contribution IMO.
I see now that this and a couple other points were mentioned in Repledge++. One more I would add to the list:
‘Relative advantage’ in cash vs percentage terms could be a sticking point. In the case of a $10M/$8M split, giving $2M/$0 to the respective candidates seems unfair to candidate B, because $2M is infinitely more than $0 in percentage terms. Say this money was going to ad buys, instead of running 100 vs 80 ad spots, candidate A now runs 20 spots vs zero for candidate B, and is the only candidate on the airwaves.
I would argue that a fair split would be $1.111M vs $0.889M, but I’m not sure that supporters of candidate A would agree.
Of course, if you assume that the platform is only a tiny fraction of total campaign contributions this is much less significant, but still worth a thought.
Cash Transfers as a Simple First Argument
Hi Benjamin,
I totally forgot about that article, thank you for pointing it out! That is an excellent resource.
Your concern totally makes sense. Something I’ve been thinking about lately is whether EA should make a more concerted effort to promote ‘streams’ of varying fidelity intended for audiences which are coming from very different places.
Put another way: say I have a co-worker who every year gives to traditional, community-based charitable orgs, and has never considered giving that money elsewhere. Is this person more likely to spend the time on excellent and in-depth philosophical articles + podcasts I push on them, or engage with a more direct and irrefutable appeal to logic? I tend to think that the latter can serve as a gateway to the former.
Thanks for sharing! I like the way you phrased it in the interview, I think that’s a nice way to start.
Thanks for writing, and I agree it would be great to see more like this in future.
It does seem like ‘ideal portfolio of resources’ vs ‘ideal split of funds donated this year’ can be quite a bit different—perhaps a question for next time?
(see here for some similar funding estimates)
For each person in a leadership role, there’s typically a need for at least several people in the more junior versions of these roles or supporting positions — e.g. research assistants, operations specialists, marketers, ML engineers,...I’d typically prefer someone in these roles to an additional person donating $400,000–$4 million per year
If this is true, why not spend way more on recruiting and wages? It’s surprising to me that the upper bound could be so much larger than equivalent salary in the for-profit sector.
I might be missing something, but it seems to me the basic implication of the funding overhang is that EA should convert more of its money into ‘talent’ (via Meta spending or just paying more).
- 29 Apr 2022 23:19 UTC; 8 points) 's comment on Increasing Demandingness in EA by (
Doubling costs to get +10% labour doesn’t seem like a great deal
I agree in principal, but in this case the alternative is eliminating$400k-4M of funding, which is much more expensive than doubling the salary of e.g. a research assistant.
To be clear, I am more so skeptical of this valuation than I am actually suggesting doubling salaries. But conditional on the fact that one engaged donor entering the non-profit labor force is worth >$400k, seems like the right call.
Agreed, just a function of how many salaries you assume will have to be doubled alongside to fill that one position
(a) Hopefully, doubling ten salaries to fill one is not a realistic model. Each incremental wage increase should expand the pool of available labor. If the EA movement is labor-constrained, I expect a more modest raise would cause supply to meet demand.
(b) Otherwise, we should consider that the organization was paying only half of market salary, which perhaps inflated their ‘effectiveness’ in the first place. Taking half of your market pay is itself an altruistic act, which is not counted towards the org’s costs. Presumably if these folks chose that pay cut, they would also choose to donate much of their excess salary (whether pay raise from this org, or taking a for-profit gig).
In reality I don’t think we’d want to go that close to the breakeven point—because there may be better uses of money, due to the reputation costs of unusually high salaries, and because salaries are harder to lower than to raise (and so if uncertain, it’s better to undershoot).
Good points, I agree it would be better to undershoot.
Still, even with the pessimistic assumptions, the high end of that $0.4-4M range seems quite unlikely.
Does 80k actually advise people making >$1M to quit their jobs in favor of entry-level EA work? If so, that would be a major update to my thinking.
That all seems reasonable.
Shouldn’t the displacement value be a factor though? This might be wrong, but my thinking is (a) the replacement person in the $1M job will on average give little or nothing to effective charity (b) the switcher has no prior experience or expertise in non-profit, so presumably the next-best hire there is only marginally worse?
This is very cool! I share your view that comprehensive data is an important part of my personal e2g decision-making (and can be difficult to find).
If you haven’t seen it already, this recent post by Ben Todd is probably the best source I know of as far as resource allocation.
Make a line plot of cumulative grants from Open Philanthropy (for each focus area individually and in total).
Do all the same plots I have for Open Philanthropy for EA Funds as well.
I made a rough attempt to this effect earlier this year (there you can also find a link to the source code).
I think I agree that only the last two should qualify, but presently I would assume a weaker definition is most common.
I suppose this can create a bad incentive where someone offering a counterfactual donation then has to make sure to do something not charitable with that money later on? I guess in my view a ‘counterfactual donation’ really only ever makes sense when you have a strong prior the money would not otherwise be put to similar use.
Would it not make sense to start some sort of ‘EA Venture Capital’ firm?
Surely more EAs would take this leap if provided with some runway/salary (in exchange for equity, which by this logic would be a phenomenal investment for patient philanthropy money)
I was thinking along very similar lines with ‘Limitations’ #1. It would be much better to model this as a contribution function in four dimensions, rather than only counting ‘EA dollars’.
Not only would this require more data, but one would need to assign a multiplier to each separate intervention à la GiveWell moral weights. What fraction of a ‘Global Health’ dollar is counted when Bill Gates funds vaccine research? Could be interesting for future work.