How Should a Large Donor Prioritize Cause Areas?
Introduction
The Open Philanthropy Project has made some grants that look substantially less impactful than some of its others, and some people have questioned the choice. I want to discuss some reasons why these sorts of grants might plausibly be a good idea, and why I ultimately disagree.
I believe Open Phil’s grants on criminal justice and land use reform are much less effective in expectation1 than its grants on animal advocacy and global catastrophic risks. This would naively suggest that Open Phil should spend all its resources on these more effective causes, and none on the less effective ones. (Alternatively, if you believe that the grants on US policy do much more good than the grants on global catastrophic risk, then perhaps Open Phil should focus exclusively on the former.) There are some reasons to question this, but I believe that the naive approach is correct in the end.
Why give grants in cause areas that look much less effective than others? Why give grants to lots of cause areas rather than just a few? Let’s look at some possible explanations for these questions.
Disclaimer: I don’t know anything about Open Phil’s reasoning except what’s been published online and what I’ve learned from a few conversations with employees, so I don’t claim that anything in this post correctly reflects Open Phil’s motivations. I don’t have as much information as Open Phil has, and haven’t spent nearly as much time investigating these cause areas. Open Phil may be making prioritization decisions based on information that I don’t have, or there may be important subtleties to the grantmaking process that I don’t understand. This post isn’t necessarily about what Open Phil is doing or should be doing, but more about what a very large value-aligned foundation should do. I mostly talk about Open Phil because it’s currently the largest such foundation2, and there’s a decent chance that it actually cares what I say.
Thanks to Peter Hurford and Kelsey Piper for reading drafts of this.
Contents
- Introduction
- Contents
- Reasons why these grants might be the right choice
- Combining multiple reasons
- Notes
Reasons why these grants might be the right choice
Diminishing utility over large amounts of money
Money has diminishing marginal utility–as you spend more, it usually becomes harder to find good giving opportunities3. Perhaps someone like Open Phil has so much money that its utility-maximizing strategy is to make grants in lots of cause areas. Is that reasonable? Let’s look at how much room for more funding there might be in AI safety and in farm animal advocacy, and when we should expect marginal grants in US policy to do more good than these4.
The US government spends about $6 billion annually on biosecurity5. According to a Future of Humanity Institute survey, the median respondent believed that superintelligent AI was more than twice as likely to cause complete extinction as pandemics, which suggests that, assuming AI safety isn’t a much simpler problem than biosecurity, it would be appropriate for both fields to receive a similar amount of funding. (Sam Altman, head of Y Combinator, said in a Business Insider interview, “If I were Barack Obama, I would commit maybe $100 billion to R&D of AI safety initiatives.”) Currently, less than $10 million a year goes into AI safety research.
Open Phil can afford to spend something like $200 million/year. Biosecurity and AI safety, Open Phil’s top two cause areas within global catastrophic risk, could likely absorb this much funding without experiencing much diminishing marginal utility of money. (AI safety might see diminishing marginal utility since it’s such a small field right now, but if it were receiving something like $1 billion/year, that would presumably make marginal dollars in AI safety “only” as useful as marginal dollars in biosecurity.)
To take another approach, let’s look at animal advocacy. Extrapolating from Open Phil’s estimates, its grants on cage-free campaigns are probably about ten thousand times more cost-effective than GiveDirectly (if you don’t heavily discount non-human animals, which you shouldn’t) (more on this later), and perhaps a hundred times better after adjusting for robustness. Since grants on criminal justice reform are not significantly more robust than grants on cage-free campaigns, the robustness adjustments look similar for each, so it’s fair to compare their cost-effectiveness estimates rather than their posteriors.
Open Phil’s estimate for PSPP suggests that cage-free campaigns are a thousand times more effective. If we poured way more money into animal advocacy, we’d see diminishing returns as the top interventions became more crowded, and then less strong interventions became more crowded. But for animal advocacy grants to look worse than grants in criminal justice, marginal utility would have to diminish by a factor of 1000. I don’t know what the marginal utility curve looks like, but it’s implausible that we would hit that level of diminished returns before increasing funding in the entire field of farm animal advocacy by a factor of 10 at least. If I’m right about that, that means we should be putting $100 million a year into animal advocacy before we start making grants on criminal justice reform.
(Of course, in practice we wouldn’t just throw $100 million at the problem; we’d learn as much as we could about what the marginal utility curve looks like, and try and determine at what point marginal funding for animal advocacy starts to look worse than criminal justice reform. But right now we’re nowhere close to that point.)
If you’ll allow me to make a somewhat contentious6 point, I’ll add that there are other important cause areas (like wild animal suffering) Open Phil is currently not focusing on that have tremendous scope (and probably tremendous room for funding); if it were up to me, I’d prioritize this way above US policy. Wild-animal suffering could probably absorb about $1 quadrillion per year in funding before marginal dollars there would be less effective than marginal dollars in US policy7.
Fundamental risk aversion
If you’re making lots of grants, you may want to diversify them for some reason other than diminishing marginal utility. You might do this if you are inherently risk-averse (although Open Phil has explicitly said it’s not).
People are generally risk averse because most things have diminishing utility. If you make $10,000 a year, getting another $10,000 will improve your life a lot. But if you already make $100,000, an extra $10,000 won’t matter as much. Hence, we say that money has diminishing marginal utility.
Fundamental risk aversion means that you value each unit of utility less than the previous unit–that is, you have diminishing marginal utility of utility. This makes no sense–utility can’t diminish relative to itself. For more detail, Brian Tomasik has an essay about why we should maximize expected value.
Fundamentally confused risk aversion
Suppose you value lives in the far future, and you believe that reducing extinction risk is important. But you’re unsure if it’s possible to have an effect on reducing extinction risk because it’s so hard to measure. Then might it make sense to allocate some money to x-risk and some to more established causes like criminal justice?
Unless we’re fundamentally risk-averse, diversifying this way makes no sense. We are not trying to maximize the odds we at least accomplish something good, we’re trying to maximize expected value. If criminal justice has higher expected value than x-risk, we should make all our grants there (up to the point of diminishing marginal returns)–not make any lower-expected-value bets.8
Unequal consideration of interests
In Practical Ethics, Peter Singer describes the concept of equal consideration of interests:
[W]hen we make ethical judgments, we must go beyond a personal or sectional point of view and take into account the interests of all those affected. […] This means that we weigh interests, considered simply as interests and not as my interests, or the interests of people of European descent, or of people with IQs higher than 100. This provides us with a basic principle of equality: the principle of equal consideration of interests.
The essence of the principle of equal consideration of interests is that we give equal weight in our moral deliberations to the like interests of all those affected by our actions. This means that if only X and Y would be affected by a possible act, and if X stands to lose more than Y stands to gain, it is better not to do the act. We cannot, if we accept the principle of equal consideration of interests, say that doing the act is better, despite the facts described, because we are more concerned about Y than we are about X. What the principle really amounts to is: an interest is an interest, whoever’s interest it may be.9
According to the principle of equal consideration of interests, we cannot give special weight to humans purely by virtue of the fact that they are humans; nor can we give special weight to the current generation over future generations10. But if this principle is wrong then animal advocacy and existential risks may be a lot less important than they look.
Is it reasonable to reject the principle of equal consideration of interests? I do not believe so. Some people may want to give special consideration to certain groups, but there’s no place for that when we’re adopting, in the words of Henry Sidgwick, the point of view (if I may say so) of the Universe. If we want to do the most good, this necessarily requires us to give equal consideration to all beings.
It straightforwardly follows from the principle of equal consideration of interests that future generations should be valued equally to the present generation10. It’s less obvious how we should value non-human animals. Valuing different types of animals suffers from the two envelopes problem. But however you resolve this, it’s implausible that criminal justice reform is more effective than cage-free campaigns. To see why, let’s compare the cost-effectiveness estimates for cage-free campaigns and one of Open Phil’s grants in criminal justice reform.
Open Phil made a grant to the Pew Public Safety Performance Project (PSPP), a criminal justice intervention that looked particularly strong. Open Phil believes that “PSPP’s past work is plausibly competitive with donations to our top charities” and their back-of-the-envelope calculation suggests that PSPP’s activities avert one prison year per $29. Compare this to the estimate that cage-free campaigns avert 38 to 250 years of cage confinement per dollar. Leaving aside the fact that the latter estimate looks somewhat more robust than the former11, you would have to say that averting one prison year is 1100 to 7200 times better than averting one cage year. This is only plausible if you discount chicken suffering relative to human suffering by a factor of at least 500. If we give equal consideration to all interests, I don’t believe a discount rate this high is justifiable.
But even if we violate equal consideration of interests, it’s extremely unlikely that the PSPP grant and cage-free grants would look similarly good in expectation, so this alone is insufficient to justify making lots of grants in different cause areas. It answers why you might prefer criminal justice reform to animal welfare, but if you do, why give to animal welfare at all?
Value pluralism
Even if we can’t justify pure unequal consideration of interests, perhaps we could adopt its watered-down alternative–value pluralism.
Suppose we’re 90% confident that all generations have equal moral value, and believe there’s a 10% chance that only the current generation matters. (There are other possibilities but let’s narrow it down to these two for simplicity’s sake.) There’s no agreed-upon method for handling this kind of moral uncertainty, but let’s say we handle it by assigning resources to each moral system in proportion to our confidence in it and then let each system “buy” what it wants. So if we have $1 million, we give $900,000 to our first moral system which says that all generations have equal moral value, and $100,000 to our second system which only values the current generation. Then each system can use the money on whatever intervention it believes is best, or they can trade with each other12.
If we agree that this system makes sense (and I do have some doubts about it), you could justify giving some money to causes that aren’t optimal according to your dominant value system. However, if we’re pretty confident that future generations matter (as I believe we should be), then this alone cannot justify allocating lots of resources to interventions that only help the present generation.
Combining multiple reasons
Maybe you believe that none of the above arguments fully work, but you put weak credence in some of them, and that’s enough to justify diversifying grants across many cause areas. Say you think there’s a non-trivial chance that we should apply time discounting to the value of beings in the future (so you believe the principle of equal consideration of interests might be wrong), and you also think the best cause areas have less than $100 million/year in room for more funding, but maybe not a lot less. Then it might make sense to devote significant capacity to cause areas like criminal justice, even if no single one of these factors would have been sufficient.
This is the most compelling reason we’ve looked at, but even this doesn’t have sufficient persuasive force.
- The best cause areas look sufficiently better than the alternatives that they would have to see extraordinarily rapid diminishing marginal utility for wide diversification to do more good than a narrower focus.
- Fundamental risk aversion just makes no sense.
- “Fundamentally confused” risk aversion is fundamentally confused (hence the name).
- There does not exist any compelling argument that non-human animals’ interests matter less than humans’, or that we should discount the far future13; and it’s implausible that farm animals are massively less sentient than humans.
- Similarly, under a value-pluralistic approach, we should only assign small probabilities to moral theories that heavily discount non-human animals or the far future.
Even if some combination of these reasons does look sufficiently compelling, that doesn’t mean we should start funding lots of cause areas right away. We have limited research capacity and actions now matter more than actions in the future, so it makes sense to prioritize the most important cause areas. I can think of some reasons why you might start with less important areas; I don’t find these reasons compelling, but this essay is already way longer than I had planned so let’s leave that untouched for now.
If you’re a big funder with lots of money, you should prioritize a small set of cause areas (or perhaps even a single cause area) that you believe is most effective.
Notes
-
Any particular grant may have a high probability of failing; a grant that looks really good right now might turn out to be useless. What we want to measure is the expected value of an intervention, not the true value.
For example, if intervention A is guaranteed to save one life and intervention B has a 10% chance of saving a thousand lives, then intervention A has a 90% probability of producing a better result, but intervention B is definitely better in expectation, so we should prefer intervention B. ↩
-
Greenbaum Foundation appears value-aligned, but it’s much smaller so the same considerations do not apply. ↩
-
There are exceptions to this. On a traditional economic cost curve, average cost diminishes as output increases, and then at some point it reverses and average cost begins increasing. If a firm has low output, it can get increasing marginal utility by producing more. Most charitable activities probably follow a similar pattern: initially, each dollar does more good than the previous dollar, and this trend eventually reverses. But just as few firms find themselves on the lower part of the cost curve, I would expect few interventions to have increasing marginal utility of money. ↩
-
If we’re going by Open Phil’s classification system then animal advocacy counts as US policy, but I think of it separately because its goal is to help non-human animals–this arguably makes it more different from any of Open Phil’s other focus areas than they are from each other. ↩
-
Sell, T. K., & Watson, M. (2013). Federal agency biodefense funding, FY2013-FY2014. Biosecurity and bioterrorism: biodefense strategy, practice, and science, 11(3), 196-216. ↩
-
This is a comedic understatement, I’m so funny, right guys? I may be a lunatic who wants humans to spend world GDP on reducing wild animal suffering but at least I’m self-aware about it.
Actually in all seriousness I think marginal spending on AI safety is probably better than marginal spending on wild animal suffering. That’s considered a reasonable position in some circles right? ↩
-
Justification: Let’s say if we are willing to spend $30 per prisoner per year, then we should be willing to spend at least $1 per wild vertebrate per year. There are about a quadrillion wild vertebrates. This is a bad estimate because that’s not how it works (you can’t just throw money at a problem based on how big it is and expect all your dollars to be equally effective), but it gives a rough idea of how much it might be reasonable to spend on alleviating wild animal suffering. Note that $1 quadrillion is much higher than world GDP. This is a feature, not a bug. If we used all of humanity’s economic production on preventing wild animal suffering we probably wouldn’t come anywhere close to solving the problem. ↩
-
Kelsey wrote this paragraph almost in its entirety. ↩
-
Singer, P. (2011). Practical Ethics (p. 20). Cambridge University Press. Kindle Edition. ↩
-
Although we could perhaps justify devaluing future generations’ happiness by adopting the person-affecting view. I believe this view is wrong, but this is too hard to justify here. See Beckstead, N. (2013). On the Overwhelming Importance of Shaping the Far Future. PhD Thesis. Department of Philosophy, Rutgers University. ↩ ↩2
-
This is sort of an irrelevant side point but I’m going to briefly justify it anyway.
The estimate for cage-free campaigns uses the inputs (1) spending on campaigns, (2) number of hens affected, (3) time until reforms would have happened anyway, (4) badness of battery cages. The first two numbers are based on hard data; the last two are harder to estimate.
The estimate for PSPP relies on (1) spending by PSPP (which Open Phil does not report, but uses a conservative figure), (2) number of affected prisoners (easy to measure), (3) decrease in prison population (non-robust estimate based on impact forecasts), (4) time that reforms last (non-robust assumption), (5) badness of prison life. The individual inputs for both estimates are comparable, but this estimate has one more non-robust input than the previous estimate, which introduces more variability in the final result. ↩
-
Hat tip to Carl Shulman for introducing me to this argument. ↩
-
Based on my understanding and comments like this, I believe most Open Phil employees do in fact discount non-human animals way too much, which partially explains Open Phil’s behavior, and makes me more confident that Open Phil is in fact behaving incorrectly. On the question of the value of animals, we have good reason to expect people to be heavily biased here, so the fact that people disagree with me is only extremely weak evidence that they might be right. If a well-informed person disagreed with me about, say, the effectiveness of deworming treatments, this is a complex empirical issue where I don’t particularly expect other people to be less knowledgeable or more biased than I am. But if a large part of my disagreement with Open Phil happens simply because they value animals less, we can be particularly confident that it’s wrong.
Adding to this, sometimes disagreements about the value of animal interventions ostensibly stem from empirical disagreements, when in fact they are caused by simple run-of-the-mill speciesism. For an example of this, Zach Groff writes about how it’s considered reasonable (even among some people who supposedly appropriately value animals) to give farm animals to poor people as an investment, but it wouldn’t be considered reasonable to give them, say, a child slave–even if doing so helped them economically. ↩