Many charities rely on donations to support their work addressing some of the world’s most pressing problems. We conducted a meta-review to determine what interventions work to increase charitable donations. We found 21 systematic reviews incorporating 1339 primary studies and over 2,139,938 participants. Our meta-meta-analysis estimated the average effect of an intervention on charitable donation size and incidence: r = 0.08 (95% CI [0.03, 0.12]). Due to limitations in the included systematic reviews, we are not certain this estimate reflects the true overall effect size. The most robust evidence found suggests charities could increase donations by (1) emphasising individual beneficiaries, (2) increasing the visibility of donations, (3) describing the impact of the donation, and (4) enacting or promoting tax-deductibility of the charity. We make recommendations for improving primary research and reviews about charitable donations, and how to apply the meta-review findings to increase charitable donations.
Avoid common mistakes on your manuscript.
Charities address some of the world’s most important and neglected problems (MacAskill, 2015; Singer, 2019). Some of the highest-impact (e.g., Against Malaria Foundation; GiveWell, 2021) and most famous (e.g., American Red Cross; Charity Navigator, 2022) charities rely on asking people to give money for no tangible reward (Bendapudi et al., 1996). As a result, effective fundraising is both critical and challenging for nonprofits. We conduct a meta-review of systematic reviews to identify ‘what works’ to promote charitable donations. Our aim is to provide practitioners and researchers with a resource for identifying which interventions have been investigated, which ones work, and which do not. By charitable donations, we mean the altruistic transfer of money from a person to an organisation that helps people in need (after Bekkers & Wiepking, 2011b). We catalogue systematic reviews because they: (a) search for and assess the evidence about which interventions work (Hulland & Houston, 2020; Stanley et al., 2018), (b) describe the effectiveness of interventions in a way that can be systematically compared, and (c) help practitioners and researchers understand which interventions have good external validity and generalisability (Higgins et al., 2019; Stanley et al., 2018). By synthesising systematic reviews, we can provide stronger recommendations for evidence-informed decision-making than by reviewing individual studies alone (HM Treasury, 2020).
We organise the presentation of results from our meta-review using an established and highly-cited model of drivers for charitable donations (Bekkers & Wiepking, 2011b). This narrative review proposed a model where donors are more likely to give when they are prompted to donate (solicitation) to a cause they know about (awareness of need), if the cost is low enough (costs and benefits) for the effect it has on society (altruism). According to this model, people also donate if they think doing so will make them look good in the eyes of others (reputation), make them feel good (psychological benefits), align with what is important to them (values), and make a meaningful difference (efficacy). Bekkers and Wiepking classified different interventions found in primary research into one or more of these drivers, for example, by discussing how tax deductibility decreases the costs of donation. However, unlike a systematic review, their narrative review approach did not account for publication bias or pre-register inclusion and exclusion criteria; it also did not estimate the relative effectiveness of each driver for influencing charitable donations. In our meta-review, we seek to comprehensively identify all interventions that increase charitable behaviour and that have been the focus of an existing systematic review. Because systematic reviews often include a meta-analysis, which summarises the quantitative effect size or ‘strength’ of an intervention on charitable donation behaviour, our meta-review will also assess the effectiveness of each driver (e.g., awareness, costs and benefits) in increasing charitable donation behaviour. In this review, we use the Bekkers and Wiepking (2011b) classification to identify which drivers have been the most studied, which have not, and which drivers appear to most influence charitable donation behaviour.
In this meta-review, we aim to:
We conducted a meta-review of systematic reviews using established recommendations (Becker & Oxman, 2011; Grant & Booth, 2009; Khangura et al., 2012; Pollock et al., 2017; World Health Organisation, 2017) to synthesise the literature on how to increase charitable donations. We conducted a meta-meta-analysis on any meta-analyses reported in the included systematic reviews. A meta-meta-analytic approach was necessary because it permitted the use of all available information from the original meta-analyses to calculate a pooled effect while accounting for variability at both the study and meta-analysis level. Our meta-review was prospectively registered on the Open Science Framework (https://osf.io/465ej/). Details of our search strategy including search strings, screening and selection of studies, data extraction and quality assessment, quantitative synthesis, and certainty assessment are presented in Supplementary File 1 and summarised below.
We searched Scopus, PsycINFO (Ovid), Web of Science, and Database of Abstracts of Reviews of Effects due to their broad but non-overlapping corpora, and their coverage of topic areas relevant to our research question. We conducted searches on July 17th, 2019 and March 4th, 2021. We developed terms for identifying systematic reviews informed by a comprehensive typology of review methods (Grant & Booth, 2009). Terms for charitable donations as outcomes included: altruis*, charit*, philanthro*, donat*, pledge*, or non-profit. Titles and abstracts were screened in duplicate; full-text articles were screened in duplicate; and included papers were extracted in duplicate. Disputes were resolved by discussion between reviewers, consulting a senior member of the team, if necessary.
Our inclusion criteria were (1) systematic reviews, scoping reviews, or similar reproducible reviews (i.e., those with a reproducible method section describing a searching and screening procedure); (2) reviews describing monetary charitable donations; (3) reviews assessing any population of participants in any context; and (4) written in English (due to logistical constraints) and (5) peer-reviewed (although no papers ended up being excluded on the basis of this criteria). Exclusion criteria were (1) primary research reporting new data (e.g., randomised experiments); (2) non-systematic reviews, theory papers, or narrative reviews; (3) reviews on cause-related marketing; and (4) reviews of other kinds of prosocial behaviour (e.g., honesty, non-financial donations). We also conducted forward and backward citation searching (Hinde & Spackman, 2015) via Scopus with no subject or publication requirements. We developed a data extraction template to capture information from each included review and assessed the quality of the included reviews using an abbreviated list of quality criteria drawn from AMSTAR 2 (Shea et al., 2017). We used the GRADE approach to assess the quality of the evidence across all reviews for each combination of intervention and outcome (Guyatt et al., 2011; Higgins et al., 2019; Hultcrantz et al., 2017). More information about and results of these quality assessments are available in Supplementary File 1.
Many, but not all, systematic reviews also conducted meta-analyses to quantify the size of effects on donations. So we could compare the relative size of effects between these different meta-analyses, we conducted a meta-meta-analysis, or second-order meta-analysis (Hennessy et al., 2019; Schmidt & Oh, 2013). These models are the best practice for synthesising effects across different meta-analyses because they can compare effect sizes on a common metric while accounting for variability both within- and between reviews (Hennessy et al., 2019; Schmidt & Oh, 2013). Our primary outcome was the overall pooled effect size of intervention on donation size. A secondary outcome was donation incidence—whether a donation of any size was provided—because many reviews reported on this dichotomous outcome. We extracted quantitative estimates from reviews that included meta-analyses and converted them to the most commonly used metric (r) using the compute.es package (Del Re, 2020) in R (R Core Team, 2020). We conducted a meta-meta-analysis using the meta sem package (Cheung, 2014) and msemtools packages (Conigrave, 2019). We used random-effects meta-analyses to calculate pooled effects for each mechanism and each outcome, then conducted moderation analyses to assess whether interventions were homogenous within mechanism and outcome. Raw data and code for reproducing the analyses are available at https://osf.io/465ej/.
We organise the results as follows. First, we describe the reviews identified and included through the systematic search (Table 1 and Fig. 1). Second, we present a meta-meta-analysis for the pooled effect of interventions on donation size and donation incidence (Fig. 2). Third, we organise the included interventions using the model of drivers for charitable donations from Bekkers and Wiepking (2011b) to present a meta-meta-analysis of interventions for each driver (Fig. 3) and describe each intervention in detail.
As outlined in Fig. 1, we screened 2294 unique titles and abstracts. The team subsequently screened 60 full texts for eligibility, 21 of which were included. Of the included systematic reviews, 15 included meta-analyses of either donation size or donation incidence. Characteristics and summaries of each included review are presented in Table 1. Most full-texts were excluded for being reviews that were not systematic (Weyant, 1996). Ten focused on prosocial behaviour but did not report charitable donations distinctly, so unique effects on that outcome could not be discerned (Nagel & Waldmann, 2016). Six were on organisational behaviour that did not include charitable donations (e.g., the effects of nonprofits becoming more commercial; Hung, 2020) and five were primary research (Kinnunen & Windmann, 2013) (e.g., randomised experiments; Kinnunen & Windmann, 2013). Three reviews did not report prosocial outcomes (e.g., effects of advertising on sales; Assmus et al., 1984). Quality appraisal and certainty assessment of the included reviews were conducted consistent with our pre-registered protocol. Due to limitations of space, we report the results of these assessments in detail in Supplementary File 1, including a table describing the quality assessment (Table S1) and certainty assessment (Table S2).
As shown in Fig. 2, the meta-meta-analytic pooled effect on donation size and donation incidence was small (r = 0.08, 95% CI [0.03, 0.12], K = 23). The pooled effect was calculated using meta-analyses reported in the included systematic reviews. These effects were heterogeneous between reviews (I2 2 = 0.85), meaning that the different interventions (e.g., pique, identifying recipient) had very different effects on outcomes (e.g., donation size). The effects were not moderated by the specific outcome (p = 0.12). As seen in Fig. 2, this means pooled effects were similar for donation incidence (r = 0.15, 95% CI [0.05, 0.25], K = 4) and donation size (r = 0.06, 95% CI [0.02, 0.11], K = 19). Raw effect sizes extracted from meta-analyses in the included reviews are available on the Open Science Framework (https://osf.io/465ej/).
We used a mixed-methods approach to synthesise quantitative effect size estimates with a qualitative analysis of findings according to Bekkers and Wiepking (2011b), with descriptions of each included review. We conducted a further meta-meta-analysis (Fig. 3) with interventions grouped by the mechanism ascribed by Bekkers and Wiepking (2011b). As shown in Fig. 3, the pooled effects of each hypothesised mechanism were significant, however there was large heterogeneity in the effects of each mechanism (all I 2 total > 0.70). Moderation analyses for interventions within each mechanism were all significant (each p < 0.018), suggesting that the specific design, channel, or context in which the intervention was delivered influenced the effective use of the hypothesised mechanism. In the following sections, we describe each identified behaviour change intervention organised by mechanism.
On average, strategies designed to increase awareness Footnote 1 had small to moderate effects on donations (r = 0.16, 95% CI [0.09, 0.23], K = 3). In general, charities can increase awareness and therefore donations by piquing donor interest, demonstrating the need, or identifying a victim.
Piquing interest increased both compliance (r = 0.27, 95% CI [0.19, 0.35], k = 16; Lee & Feeley, 2017) and donation size (r = 0.29, 95% CI [0.25, 0.33], k = 16; Lee & Feeley, 2017), leading to much larger total revenue (r = 0.49, 95% CI [0.45, 0.53], k = 16; Lee & Feeley, 2017). The pique technique involved asking donors for unusual amounts of money (17c instead of 10c), and was designed to break the ‘refusal script’: would-be donors were more likely to stop and ask for a rationale when an odd amount of money was requested of them (Lee & Feeley, 2017). The largest experiment involved a $3 request so the technique may have questionable ecological validity. It is unclear whether it would also work for requesting $1017 instead of $1000.
This was evaluated in three meta-analyses. Engel found that needy recipients received an increased donation size in dictator games (r = 0.13, 95% CI [0.10, 0.17], k = 69; Engel, 2011). Footnote 2 Neediness was also a mechanism that explained legitimizing paltry contributions (described below). When someone said “even a penny would help”, many donors saw the recipient as more needy, which had indirect effects on donation compliance (Bolkan & Rains, 2017). Finally, compared with causes with modest negative impact (e.g., no school books), when a problem was described as severe, certain, and calamitous (e.g., natural disaster), donation size increased regardless of whether the victim was identifiable or not (Butts et al., 2019).
Under most circumstances, donation size increased when donors were presented with a single, ‘identifiable victim’ (r = 0.13, 95% CI [0.08, 0.17], k = 47; Butts et al., 2019) than when presented with statistics or multiple recipients. This is also known as ‘compassion fade’, where a larger number of victims leads to lower perceived impact and lower expected positive affect from donating (Jenni & Loewenstein, 1997). Mediation analyses supported these hypothesised paths (Butts et al., 2019). Empathy had a smaller mediating role: while people showed slightly less empathy for a larger group of people, this lower empathy had only a small effect on donations.
On average, strategies targeting costs and benefits weakly increased donations (r = 0.08, 95% CI [0.02, 0.13], K = 5). The most influential effects appeared to be imbuing a charity with tax deductibility, but nudges or framing strategies had few effects.
One large meta-analysis of 69 studies (n = 1,418,212), examined the impact of tax deductibility on charitable donations (Peloza & Steel, 2005). Effects were reported as price elasticities which could not be converted to effect sizes. They found substantial elasticity: a tax deduction of $1 resulted in an additional $1.44 being donated to charity (confidence interval not reported). The authors found that tax deductions particularly increased the likelihood of bequests. High-income donors were no more concerned with tax deductions than lower-income donors.
In contrived experiments, when donors were told their funds would be matched (fully or partially), there was a small but non-significant increase in donations (r = 0.08, 95% CI [− 0.02, 0.17], k = 18; Engel, 2011).
Door in the face is designed to reduce the perceived cost of donating by initially presenting a high anchor (e.g., “will you donate $1000?”), then asking for something more achievable (e.g., “how about $10?”; also known as the ‘request then retreat strategy’). Evidence for this strategy appears weak: donors may be marginally more likely to say they will donate (r = 0.08, 95% CI [0.03, 0.12], k = 7; Feeley et al., 2012) but this does not translate into actual compliance (r = 0.06, 95% CI [− 0.01, 0.12], k = 15; Feeley et al., 2012).
In contrived experiments with repeated rounds, donors gave less when they were aware there would be multiple opportunities to donate (r = − 0.06, 95% CI [− 0.08, − 0.04], k = 64; Engel, 2011).
Prospect theory proposes that small losses loom larger than small gains, but framing appeals for charitable donations as ‘losses averted’ did not increase the likelihood of donations (r = − 0.01, 95% CI [− 0.10, 0.09], k = 25; Xu & Huang, 2020).
Some studies have asked whether those donating from larger pools of money (usually in contrived experiments) are more generous or more frugal. The larger of the two meta-analyses found no relationship between stake size and donations (r = − 0.01, 95% CI [− 0.09, 0.07], k = 603; Engel, 2011). But, a follow-up, more focused meta-analysis found that those endowed with more money tended to be less generous, in relative terms, than those endowed with less (r = − 0.07, 95% CI [− 0.03, − 0.12], k = 18; Larney et al., 2019). That is, when people had more, they may donate more in absolute terms, but usually donated a lower percentage of the money they held.
Three systematic reviews investigated the effect of ‘legitimizing paltry contributions’ on charitable donations (usually words like “even a penny will help”; Andrews et al., 2008; Bolkan & Rains, 2017; Lee et al., 2016). The largest of these reviews found a moderate increase in compliance (r = 0.22, 95% CI [0.17, 0.26], k = 34; Bolkan & Rains, 2017) which was offset by a decrease in the size of the average donation (r = − 0.23, 95% CI [− 0.34, − 0.12], k = 11; Bolkan & Rains, 2017). The net effect of these competing forces was a non-significant increase in total revenue (r = 0.03, 95% CI [− 0.01, 0.07], k = 18; Bolkan & Rains, 2017). Mediation analyses suggest that the technique increases the perceived neediness of the cause, but that it also vindicates those likely to donate a small amount to avoid judgement (i.e., donors high on ‘impression management’).
On average, strategies targeting efficacy increased donations (r = 0.11, 95% CI [0.01, 0.22], K = 3). Direct modelling of the desired behaviour—seeing others donate money—appears to increase donations, but general prosocial media does not.
When people saw others acting prosocially, they were more likely to imitate, including charitable donations (r = 0.22, 95% CI [0.21, 0.24], k = 40; Jung et al, 2020). Effects were consistent across media (e.g., direct observation vs watching on TV), age, gender, and culture.
Jung and colleagues (2020) looked at studies where the model performed the same behaviour (i.e., the model donated money and the dependent variable was donation too); they moderated how the model was viewed (real observation vs. via media). Coyne and colleagues (2018) instead looked at media only (TV, movies, video games, music or music videos) with explicitly prosocial content (but not necessarily donating money). Participants in one study were more likely to donate money while listening to “Love generation” (by Bob Sinclair) rather than “Rock this party” (also Bob Sinclair; Greitemeyer, 2009). This trend, however, was not reliable with small pooled effects on financial donations and a confidence interval including the null hypothesis (r = 0.09, 95% CI [− 0.04, 0.21], k = 9; Coyne et al., 2018). The media seldom demonstrated the exact behaviour being measured (i.e., ‘Love generation’ does not talk about donations); imitation and efficacy may increase the likelihood of the behaviour being observed but effects do not spill over to nearby prosocial behaviours.
As mentioned previously, correlational studies show that certain calamities appear to attract donations, regardless of interventions like ‘identifiable victims’ (Butts et al., 2019). Among dictator games, when uncertainty was added to the benefit (e.g., donating lottery tickets) there was no significant reduction in donations (r = − 0.04, 95% CI [− 0.17, 0.10], k = 7; Engel, 2011). This may not necessarily translate to different types of uncertainty, however, such as uncertainty that a charity will have an impact.
On average, strategies targeting reputation slightly increased donations (r = 0.06, 95% CI [0.00, 0.12], K = 6). In general, people are somewhat more likely to donate when there is some reputational benefit to doing so (e.g., being observed or having the donation amount visible).
After synthesising a large number of studies and participants (N > 500,000), Bradley and colleagues found that being observed significantly increased donations (r = 0.15, 95% CI [0.11, 0.20], k = 101; Bradley et al., 2018). Consistent with reputation hypotheses, effects were larger for repeat interactions, interactions with personal consequences, group social dilemmas (vs. 1:1 bargaining games), and where observation is more intense. In contrast, Engel moderated his findings by whether or not donations were concealed. He found donations decreased when concealing the donor (r = − 0.09, 95% CI [− 0.02, − 0.16], k = 52; Engel, 2011) or the amount donated (r = − 0.06, 95% CI [− 0.12, 0.00], k = 19; Engel, 2011), but effects were small.
Three systematic reviews have explored the effect of artificial surveillance cues on donor generosity (Sparks & Barclay, 2013; Nettle et al., 2013; Northover et al., 2017). Studies have typically analysed the effect of displaying images of ‘watching eyes’ on donation decisions made within economic games, but many include field experiments (e.g., eyes above ‘honesty boxes’). The largest of these reviews found negligible increases in compliance (r = 0.04, 95% CI [0.03, 0.06], k = 27; Northover et al., 2017) and donation size (r = 0.01, 95% CI [− 0.02, 0.05], k = 26; Northover et al., 2017). Effects seem to only work short-term, with few studies finding any long-term benefits (Sparks & Barclay, 2013). Overall, artificial surveillance may increase the chance of people donating something in the short term, but the best quality evidence suggests effects are small.
Making a decision as a group may increase the reputational stakes of signalling altruism but also may diffuse the reputational benefit of donating. Group decisions had no significant total influence on donations (r = − 0.05, 95% CI [− 0.28, 0.18], k = 4; Engel, 2011) but the confidence intervals are wide due to the small number of studies.
Few reviews explored the altruism mechanism proposed by Bekkers and Wiepking (2011b). Two reviews explored the crowding-out hypothesis—that donors motivated by a desire to have an impact would avoid causes already supported by governments because of diminishing marginal returns (Lu, 2016, k = 60; de Wit & Bekkers, 2017, k = 54). Neither review found decisive evidence for crowding out. A subset of the studies in the reviews had higher internal validity—they either controlled for confounding statistically or via experimental designs. These studies were more likely to suggest that government funding reduces private donations (de Wit & Bekkers, 2017; Lu, 2016), but given the small, heterogeneous effect sizes, the evidence for a relationship is weak.
We did not find reviews of interventions that could be easily classified as ‘solicitation’, ‘psychological benefits’, or ‘values’. We could not easily classify two review findings on the basis of Bekkers and Wiepking’s (2011a, 2011b) mechanisms. One review tested a range of interventions designed to promote intuitive thinking (e.g., high cognitive load), but these studies did not influence donations (r = − 0.01, 95% CI [− 0.02, 0.01], k = 60; Fromell et al., 2020). The authors argue that ‘fast’ and ‘slow’ thinking are often aligned on issues of charitable donations. Engel (2011) found that reducing the number of options available to the donor decreased the amount they donated.
Taking the findings together, and notwithstanding the limitations of the included reviews, we recommend practitioners consider the following interventions for promoting charitable donations. Examples of the source, recipient, context, channel, and content of each intervention (Lasswell, 1948; Slattery et al., 2020) are presented in Supplementary Table S3.
When interventions increased donor confidence, they tended to solicit higher donations. Effective strategies included seeing other people who donated money (Jung et al., 2020), not merely seeing people performing ‘prosocial behaviours’ (Coyne et al., 2018). Theory and preliminary findings would suggest that effects are stronger when viewing those who share our group identity (Chapman et al., 2018, 2020). Uncertainty about the benefit of a charitable donation may cause prospective donors to reduce their donation size; donation matching campaigns may cause prospective donors to slightly increase their contribution (Engel, 2011). Identifiable victims work because donors feel more confident that they could make a meaningful difference (Butts et al., 2019). Overall, the key mechanism is that if prospective donors think they can make a meaningful difference, they are more likely to donate (Butts et al., 2019).
Donors are persuaded by needy recipients (Engel, 2011). Campaigns that say things like ‘even a penny will help’ can increase the likelihood of an initial donation when it signals the ‘desperate need’ of the cause (Bolkan & Rains, 2017). Similarly, highlighting a single beneficiary (“identifiable victim”) does not change likelihood of donation behaviour if the charitable cause is obviously severe and widespread (Butts et al., 2019). Piquing a donor’s interest via odd requests (e.g., 17c) appears to work by prompting a conversation around why the donation is needed (Lee & Feeley, 2017).
Donations are more likely when donors are observed (Bradley et al., 2018), and when both they and their donation size are identified to recipients (Engel, 2011). Charities should be careful to avoid using this in a way that creates guilt or social pressure (Bennett, 1998) or in a way that is contrived/artificial (Northover et al., 2017; Sparks & Barclay, 2013). Instead, charities can use transparency as a way of facilitating pride and self-efficacy (Crocker et al., 2017), to minimise the taboo around discussing charitable donations publicly, and to help establish a social norm toward giving (Singer, 2019).
Given the large and significant price elasticity from tax-deductibility (i.e., tax-deductibility increased donations; Peloza & Steel, 2005), directing effort toward becoming tax deductible will likely pay dividends. While few studies explicitly assessed the impact of advertising deductibility, we assume that doing so may confer some benefits for donations.
Nudges usually assume that people will be more likely to donate if charities activate their ‘fast’, intuitive thinking system, but this is not the case (Fromell et al., 2020). It appears that intuitive and deliberate thinking around donations are usually aligned. Nudges and framing strategies such as artificial cues (Northover et al., 2017; Sparks & Barclay, 2013), legitimizing paltry contributions (Bolkan & Rains, 2017), ‘door-in-the-face’, (Feeley et al., 2012) are not consistently effective.
By focusing on review-level evidence we necessarily excluded primary studies that would have been useful for charity and non-profit researchers and practitioners. While a review of 1339 included primary studies would have been intractable, reviews of primary studies have sufficient granularity to look at mediators and moderators that might be useful across studies. Instead, we were beholden to the methods of the included reviews. Similarly, we were limited to the interventions selected by previous reviewers, so necessarily omitted interventions not included in any systematic reviews, even though they may inform research and practice (e.g., opt-in vs. opt-out donations; Everett et al., 2015). There may, for example, be a wealth of knowledge on interventions using the internet to drive donations, but since there have been few systematic reviews on that topic, those interventions would have been excluded from our meta-review (Bennett, 2016, 2019; Liang et al., 2014). In a similar vein, focusing on systematic reviews means we may have excluded some more recent, ‘cutting-edge’ interventions. It often takes a number of years for an intervention gaining traction and it being subject to a systematic review. For example, recent research has shown that donors may actually prefer cost-effectiveness indicators (i.e., cost per life saved) to overhead ratios (i.e., percent directed to administrative expenses) but that the latter is usually the focus of decision-making because of the ‘evaluability bias’: people weigh an attribute based on how easy it is to evaluate (Caviola et al., 2014). However, few studies have examined the effect of publishing cost-effectiveness indicators so it is not yet possible to meta-analyse these interventions. As a result, while the interventions presented in our review have been thoroughly assessed, and many have been shown to be robustly beneficial, there may be other interventions with larger effect sizes not listed here.
Our meta-review prospectively excluded grey literature and reviews in other languages. This may affect generalisability, but doing so seldom affects conclusions from meta-reviews (unlike reviews of primary studies; Ganann et al., 2010), and we excluded no reviews on the basis of this criteria (see Fig. 1). This is likely because unpublished reviews of charitable donations are less likely to use systematic search and synthesis methods. Nevertheless, there may be other reviews that contribute to this discussion that was missed by our searches and inclusion criteria.
Our review used well-validated assessments of certainty (i.e., GRADE) and review quality (i.e., an abbreviated AMSTAR2 checklist; see Supplementary File 1). These assessments allow interested readers to know the quality of the included reviews and certainty of the included findings. However, in a meta-review, these tools are again beholden to the methods of the included systematic reviews. For example, GRADE reduces the certainty of the findings if there are few randomised experiments, or if the included randomised experiments may have been subject to common experimental biases (e.g., if they were unblinded). These biases reduce the internal validity of the findings, but few included reviews formally assessed these biases. As a result, we could not conduct sophisticated assessments of the internal validity of the included without examining the methods of the 1339 primary studies. We hope future systematic reviews of primary studies more frequently assess these biases using a validated tool, like ROB2 (Sterne et al., 2019). Similarly, GRADE accounts for the external validity of the included studies—such as whether or not findings are likely to generalise to the populations or situations most practitioners are interested in. This can be a complex question requiring judgement. For example, in some cases, Mechanical Turk contractors may be representative samples, but external validity also depends on the design of the study (e.g., viewing a real advertisement vs. playing an economic game). Our ability to assess external validity was subject to the quality of the reporting in the included systematic reviews (unless we wanted to review all 1339 methods). We hope future reviews discuss the external validity of their included studies, and could consider integrating those judgements into their own certainty assessment (e.g., via GRADE). Another approach would be to assess the facilitators and barriers to successfully delivering a pilot-tested intervention to new populations and in new contexts (e.g., scale-up; Saeri et al., 2021).
One additional limitation concerns the intervention of tax deductibility (Peloza & Steel, 2005). This systematic review and meta-analysis investigated the impact of tax deductibility on charitable donations primarily in the United States, with a minority of included primary studies describing the effect in similar countries such as Canada and the United Kingdom. Given that formal tax structures and cultural values of taxation and charitable giving differ significantly between countries, and tax policy can vary over time within a given country, the substantial effect size observed in Peloza and Steel’s (2005) meta-analysis may not hold in other settings.
Increasing charitable donations could benefit society in a multitude of ways: from helping to address global poverty, health, animal suffering, climate change, human rights, and the long-term future of humanity. As a result, identifying robust strategies for promoting charitable causes can have widespread social benefits. Providing good review-level evidence is a key way that charity science can contribute to evidence-informed decision-making in this important area.
In this meta-review, we synthesised multidisciplinary literature on how to promote charitable donations. We identified a range of strategies that may increase donations and some mechanisms that may help explain their effects. These findings suggest that organisations can solicit more money by focusing on individual victims, increasing the publicity of donations, discussing the impact of the donation, and both ensuring and promoting the tax-deductibility of their charity.
Future reviews into other interventions—particularly those conducted outside of contrived experimental settings—would allow researchers and practitioners to assess the ecological validity of those interventions. Readers could have more faith in those reviews if they more consistently followed best-practice approaches to systematic reviews. Our meta-review reveals patterns and gaps within the current research, but it also identifies an array of well-researched mechanisms for promoting charitable donations. Using the findings of these reviews may increase the funds directed to some of the most important and neglected problems facing humanity.
We expect ‘compassion face’ could be categorized under a number of different mechanisms, but relied on Bekkers & Wiepking (2011a, 2011b) who explicitly identified this as ‘awareness’.
Dictator games are designed as contrived analogies for donation situations: one participant is given some money and is given the chance to donate to another with no consequences or tangible benefits. Since actual money changes hands to a relative stranger, we deemed it sufficiently analogous to real charitable donations for this review.
We gratefully acknowledge the assistance of Lucius Caviola and David Reinstein, and several anonymous reviewers for their efforts in maximising the rigor and usefulness of this manuscript
Open Access funding enabled and organized by CAUL and its Member Institutions. The authors did not receive support from any organization for the submitted work. No funding was received to assist with the preparation of this manuscript. No funding was received for conducting this study. No funds, grants, or other support were received.