Skip to main content
PLOS One logoLink to PLOS One
. 2022 Sep 22;17(9):e0273971. doi: 10.1371/journal.pone.0273971

Sure-thing vs. probabilistic charitable giving: Experimental evidence on the role of individual differences in risky and ambiguous charitable decision-making

Philipp Schoenegger 1,*, Miguel Costa-Gomes 2
Editor: Junhuan Zhang3
PMCID: PMC9499298  PMID: 36137160

Abstract

Charities differ, among other things, alongside the likelihood that their interventions succeed and produce the desired outcomes and alongside the extent that such likelihood can even be articulated numerically. In this paper, we investigate what best explains charitable giving behaviour regarding charities that have interventions that will succeed with a quantifiable and high probability (sure-thing charities) and charities that have interventions that only have a small and hard to quantify probability of bringing about the desired end (probabilistic charities). We study individual differences in risk/ambiguity attitudes, empathy, numeracy, optimism, and donor type (warm glow vs. pure altruistic donor type) as potential predictors of this choice. We conduct a money incentivised, pre-registered experiment on Prolific on a representative UK sample (n = 1,506) to investigate participant choices (i) between these two types of charities and (ii) about one randomly selected charity. Overall, we find little to no evidence that individual differences predict choices regarding decisions about sure-thing and probabilistic charities, with the exception that a purely altruistic donor type predicts donations to probabilistic charities when participants were presented with a randomly selected charity in (ii). Conducting exploratory equivalence tests, we find that the data provide robust evidence in favour of the absence of an effect (or a negligibly small effect) where we fail to reject the null. This is corroborated by exploratory Bayesian analyses. We take this paper to be contributing to the literature on charitable giving via this comprehensive null-result in pursuit of contributing to a cumulative science.

Introduction

Charitable giving has been the subject of much research across the social sciences. After all, substantial resources are spent every year on charitable ventures. For example, in 2019, US Americans donated just over 2% of GDP [1] to a total of up to 1.5 million registered non-profit organisations [2]. The vast majority of these decisions are made under conditions of risk (with known outcome distributions) as well as ambiguity (with unknown outcome distributions). Plausibly, some charitable interventions have a high probability of providing the outcome promised; think of soup kitchens–where there is little uncertainty as to the intervention’s ability to produce the outcome: If one donates a certain amount of money to a soup kitchen, a few additional litres of soup will probably be dispensed quite soon. However, other charitable interventions are more probabilistic in nature, i.e., they only have a (sometimes very) small chance of making an impact and some aspects of this calculation are difficult to articulate numerically; think of interventions aimed at reducing the risk of nuclear war. Here, the donor is confronted with substantial additional uncertainty over the likelihood of an event (like nuclear war inciting incidents) arising and ambiguity over whether the charity’s interventions will have an impact on them should they arise: For all the donor knows, there might not be an incident risking nuclear war this year and even if there was, the charity might not be able to meaningfully influence the outcome. The former level of uncertainty can frequently be stated numerically, while the latter often cannot.

In this paper, we study actual donor decisions about these two kinds of charities: The first we call “sure-thing charities”, where donations are technically still made under conditions of some risk, though their risk is quantifiable and comparatively minor. The second we call “probabilistic charities”, where donations are also made under risk (e.g., relating to the likelihood of their event of concern arising), though other aspects of this choice are made under ambiguity (e.g., relating to the chance that an intervention will help address the issue that is itself extremely rare). When making choices between highly reliable and more uncertain options, variation in individual differences may drive heterogenous behaviour. In this paper, we specifically want to investigate whether individual differences in risk and/or ambiguity attitudes, empathy, numeracy, or donor type (warm glow vs. pure altruistic type) predict donation behaviour. Understanding why donors make the choices they do about these two types of charities has implications both for the academic literature and for fundraising efforts more directly.

There has been some previous work on the role of risk and ambiguity attitudes on pro-social behaviour in more abstract contexts. For example, Vives & FeldmanHall [3] find that ambiguity attitudes robustly predict pro-social behaviour, though that risk attitudes do not. Similarly, Chen & Zhong [4] find that uncertainty increases sharing behaviour in dictator games and reduces lying in dice games. In the context of other-regarding behaviour in binary dictator games, Haisley & Weber [5] find that dictators are less likely to choose an unfair distribution when the outcome allocation is dependent upon an ambiguous lottery rather than a risky one. For further similar research see [6, 7]. Additionally, Mesa-Vazquez, Rodriguez-Lara, & Urbano [8] find that when there is uncertainty about whether a dictator’s choices are implemented, their actions are more generous.

A related but distinct literature looks at the effect of risk over donations themselves. Exley [9] find that in situations of risky donations where donors trade off personal payoffs and donations, donors give less, cf. also [10]. However, there is also some evidence for the opposite effect, namely that some types of uncertainty actually increase pro-social behaviour, cf. [11]. Further work investigates another type of uncertainty investigated in pro-social behaviour, namely uncertainty over whether any given recipient needs one’s help. Engel & Goerg [12] find that uncertainty of this type “does not deter generosity” [12, p. 51]. When excluding selfish dictators, they find that in conditions of uncertainty donations are higher than under certainty. Further, Niehaus [13] suggests that because outcomes of charitable giving are usually not observed by the donors, decisions are primarily made on the perception of the outcome rather than the actual outcome. This would also be consistent with findings that show how many factors explain donor’s reluctance to give to the most effective charities [14].

Our research builds on the literature on risk that has so far mostly employed directly controllable levels of risk in the lab. For example, in abstract game scenarios, risk can be controlled and stated precisely, for example by imposing a 50% chance of one’s donation not being implemented, or by introducing a 5% chance that one’s donation is matched. Crucially, our research is substantially different from the discussed literature primarily because we move the level of risk from precisely calculable interventions in the lab (as outlined above) to the actual charities themselves. While this introduces several design challenges, we argue that this step leads to an increased level of ecological validity of any potential finding. However, note that there is already a large literature on charitable giving generally that has a similar or higher level of external validity [1517]. However, our paper’s main contribution is the moving of our focus on risk and ambiguity to actual organisations and their interventions and away from aspects that can be controlled in the lab. Having risk and ambiguity at the level of actual charities is the level at which risk and ambiguity typically enter people’s decision-making processes; rarely are we uncertain as to whether our donation will randomly increase when we donate (as in some experimental lab studies), but we are almost always acting under uncertainty about the charity’s interventions that we consider donating to.

Additionally, there has also been a recently growing literature that adds individual differences in personality traits to standardly used predictors like risk preferences in a variety of economic contexts. For example, Knapp, Wuepper, & Finger [18] analyse the interrelations of risk preferences and personality as well as their predictive power in the context of farmer behaviour, finding that personality measures are among the best predictors of some of their economic behaviour. See also [19] on the role of the Big Five and [20] on the HEXACO model on economic behaviour, cf. also [21]. The role of personality has also been more directly studied in the specific context of charitable giving. Some find that personality traits can play a central role in charitable decision-making while others [22] see the effect as markedly smaller. Specifically, some have also found associations between empathetic character traits and donations as well as volunteering behaviour [23] while others have established a link between the Big Five personality trait of agreeableness and prosocial behaviour [24]. For further recent work on the relationship between personality traits and charitable giving, see [25, 26]. Our paper adds to this literature that aims to integrate personality traits into economics related studies and aims to advance our understanding of the role of individual difference measures to standard predictors of economic behaviour like risk and ambiguity attitudes concretely in the context of charitable giving.

In attempting the construction of an externally valid donation choice that allows us to capture the difference between sure-thing charities and probabilistic charities in actual charities that participants can donate money to, our main outcome variables of interest are (i) the choice between two real charities, one sure-thing charity and one probabilistic charity and (ii) the choice about one of those charities that has been randomly selected. Both of these charitable decision-making scenarios have strengths and weaknesses from an experimental design perspective, but we hope that they jointly allow us to better understand the role of individual differences in charitable decision-making scenarios like these. We outline the main weakness of (i) in the discussion section and argue that, overall, (ii) is a cleaner design.

In our experiment, each participant is first presented with a randomly selected pair of charities consisting of one charity of each type to control for accidental confounds relating the charity’s context as each are presented with substantial additional accurate information to increase the naturalness of the choice. Participant choices with respect to this randomly selected charity pair then allows us to isolate and capture the element of probability between the two charity types. In the second part of this experiment, we study participant behaviour when they are shown only one randomly selected charity (either sure-thing or probabilistic), which more narrowly captures the predictive value of individual differences on donation choices to charities of specific types. Overall, we find little to no evidence that individual differences in risk/ambiguity attitudes, numeracy, optimism, and donor type predict charitable giving behaviour. However, we do find that a purely altruistic donor type predicts donations to probabilistic charities when participants are either shown a sure-thing or a probabilistic charity. As such, we take this paper to be primarily reporting a null result.

Hypotheses

In this paper, we are centrally interested in what best explains donations choices regarding sure-thing and probabilistic charities as we offer an account of moving the level of risk and ambiguity from risk over donations or outcomes in abstract lab environments to risk and ambiguity over actual charities’ interventions, thus increasing the level of ecological validity (while clearly stating the design constraints and weaknesses that come with this move). To provide data on these questions, we pre-registered five concrete null hypotheses on the Open Science Framework (https://osf.io/w9gfu/).

First, we investigate whether donor choices can be explained by individual differences in risk and ambiguity attitudes. Previous work in domains such as stock market participation [27] and health-related field behaviours [28] has found that attitudes to risk and ambiguity can play significant roles. In the context of pro-social behaviour in game environments the results show that ambiguity aversion may play a role while risk aversion sometimes does not [3], though risk aversion has also been found to be “predictive for giving” [6, p. 95]. As such, we argue that given risk and ambiguity aversion have been shown to impact behaviour in many contexts including charitable giving, this makes it an a priori interesting relation to test. This hypothesis is also theoretically grounded, in that it might be the case that one’s preference not to give to charities that have a low chance of making an impact might be driven by an individual’s general risk aversion profile, or it might be that given the ambiguous nature of charitable interventions that it is only ambiguity aversion that impacts this choice. As such it is plausible that either (or both) of them may play a role in this choice. Our directional pre-registered prediction is that ambiguity and risk aversion predict donations towards sure-thing charities because those averse to risk and ambiguity may prefer charities that have clearly stated, and low-risk interventions as opposed to more ambiguous and risky ones. This is our first null hypothesis.

Null Hypothesis #1: Ambiguity and risk attitudes do not predict choices between sure-thing and probabilistic charities.

Second, we investigate whether there are a few potential further individual difference measures that may help explain donor behaviour in the context studied here. Specifically, we will focus on numeracy, optimism, donor type (warm glow vs. pure altruistic type), and empathy. First, it may be the case that basic individual differences in numeracy explain a potential preference for probabilistic charities, primarily because decision-making in situations of small probabilities (and big potential payoffs) might be particularly difficult to comprehend for those less numerically versed. For example, previous research has found that those lower in numeracy were more insensitive to proportions of donation recipients [29] and that they showed higher susceptibility to changes in numeric presentation [30]. It may as such be the case that one’s level of numeracy also meaningfully impacts behaviour in the context studied here as the probabilistic charities include interventions that have a small chance of making a large impact. Understanding these proportions plausibly requires a certain level of numeracy. Further, one may also think that a general proclivity to optimism may bias individuals towards overestimating the success of probabilistic charitable interventions, or conversely that higher pessimism may explain a preference for sure-thing charities as those promise to have a reliable impact even in the worst-case scenario. This is corroborated by previous research that draws on the German socioeconomic panel, finding that optimism predicts charitable giving in some of their models [31]. Additionally, we investigate whether estimates of donor type, i.e. warm-glow vs. pure altruistic donor type, may predict behaviour too: For those on the warm-glow part of the spectrum, giving to a charity that has an immediate and reliable impact may have a higher chance of producing such a warm glow than giving to a charity that most likely will have no direct impact (or one that will not be observable for quite a while). Conversely, pure altruists would presumably primarily care about the perceived (expected) impact on social welfare and may as such be more likely to choose probabilistic charities on expected value grounds. Lastly, empathy has previously been shown to predict charitable giving frequency in a variety of contexts [32, 33]. It may thus be that empathy also explains choices in this study, for example if participants are preferring sure-thing charities due to their easier-to-relate-to interventions.

Our directional pre-registered predictions are that high numeracy, purely altruistic motives, and tendency towards optimism predict donations towards probabilistic charities. Conversely, our directional pre-registered predictions are that empathy and warm glow motives predict donations towards sure-thing charities. These predictions are encapsulated in our second null hypothesis.

Null Hypothesis #2: Individual differences in numeracy, optimism, empathy, and donor type do not predict choices between sure-thing and probabilistic charities.

One worry with the present design that might be raised is that despite the randomisation of charity pairs, any potential effect may still be driven by some level of contextual confounding present in the charity descriptions. To address this concern, we also have a condition in which participants make a choice between two ‘context-free’ charities, in which only fundamental information as to the underlying probabilistic aspects are preserved and the remainder of context is removed, though participants are informed about all aspects of the charity after their decision. This is our third null hypothesis.

Null Hypothesis #3: The factors predicting behaviour in NH1/NH2 do not predict behaviour in the blinded choice condition.

Fourth, it may be the case that if much of the potential preference for sure-thing charities is explainable by individual difference measures related to numeracy or risk attitudes, this might open up the possibility of shifting behaviour via informational interventions aimed at this. Specifically, we investigate how exposure to expected value reasoning affects donation behaviour. It might be the case that learning about or being made aware of expected value reasoning may meaningfully impact behaviour. Our directional pre-registered prediction is that those shown the expected value reasoning treatment text will be more likely to donate to probabilistic charities than those shown a made-up theory of decision-making (midontic decision-theory), which introduces participants to a nonsensical description of a theory of decision-making with the same directional recommendations, aimed at controlling for experimenter demand.

Null Hypothesis #4: The expected-value treatment does not impact behaviours in the sure-thing vs. probabilistic charity choice more than the control condition.

Lastly, one may also be interested in donation behaviour not between these two types of charities, but rather just in the context where potential donors are presented with one such charity. This may reduce the chance of additional confounds (like worrying that the design that presents two charities is artificial in its dichotomous presentation; after all, most naturalistic decisions are not decisions between two distinct choices). It also is overall a cleaner design that brings with it less drawbacks regarding interpretation of results. As such, we also investigate all our main hypotheses in the context where they are presented only with a single, randomly selected charity (equiprobable that it is a sure-thing or a probabilistic charity). This allows us to control for the type of charity and any potential confounds that this dichotomisation and variance in background knowledge might bring with it, as well as by presenting a cleaner design overall. As above, we study both the frequency and amount of giving with the same predictors (and directoinal predictions) as outlined in the previous sections. Our fifth null hypothesis below states this in detail and basically pools null hypotheses #1 and #2 into one single hypothesis.

Null Hypothesis #5: Ambiguity and risk attitudes (as well as individual differences in numeracy, optimism, empathy, and donor type) do not predict frequency and size of donations when presented with either a sure-thing or a probabilistic charity.

Methods

Participants and procedure

We pre-registered this project on the Open Science Framework (https://osf.io/w9gfu/), where we also deposited the full data set. For this study, we recruited a total of 1506 participants via Prolific (47.8% male) at a mean age of M = 45.09, SD = 15.49 from the United Kingdom. This sample is representative of the UK’s population according to census data from the UK Office of National Statistics via Prolific’s representative samples tool. This study received ethics approval from the University of St Andrews (approval code SA15429). Overall, participants receive £2 for participation and can earn up to £1.83 in addition (over three sets of tasks), depending on their choices in the experiment. We arrived at sample size of 1506 by conducting a pre-registered a priori power analysis in which we calculate that in order to obtain .95 power to detect an effect of the size f2 = .02 (a standard measure of a small effect size which we picked as our smallest effect size of interest) at 5% alpha error probability in a linear multiple regression model, we would need to recruit a total of 652 participants. In order to account for those not wanting to donate (estimated at 33% as previous work has documented this rate to be between roughly 20% [34] and 40% [35]) and those failing comprehension questions, we aimed to recruit 1050 participants for the main condition and a further 450 for the remaining secondary conditions (for full specifications of all additional power analyses for further conditions and analyses, please see our pre-registration). In all results reported, we excluded a total of 261 participants who got more than one of the comprehension questions or attention checks incorrect. Our final exclusion criteria were: We excluded all participants who got more than comprehension/attention check incorrect. For all analyses in Final Choice, we also, in addition, excluded all those who indicated a donation that exceeded their earned endowment. This choice deviates from our pre-registered data exclusion plan. First, we did not pre-register that we would exclude data in Final Choice from participants who were stating a final donation higher than their earned endowment. As such, the final sample size used for all analyses is 1,245 participants.

In the first part of the experiment, participants provide fully informed consent, and then complete several individual difference measures relating to numeracy skills [36], tendency for optimism [37], and empathy [38]. For the completion of each these surveys, they are paid a fixed endowment of £0.25 per survey (£0.75 in total) which is at their disposal for the remainder of the study. All participants then move onto the next part, where they are randomly selected into either the Main Choice condition (70% chance), the Expected-Value Treatment condition (20%), or the Context-Free Choice condition (10%). There, they decide to allocate some or all of their endowment to one of the two randomly chosen charities (one sure-thing and one probabilistic charity), or alternatively allocate none of it to either charity. After this choice, participants are asked to state their views on the impact of the charities that they had just seen on a 100-point scale. Further, they are also asked to estimate the average views of other participants on these charities. This task was incentivised by awarding participants an additional £0.10 if their estimates are within 5 percentage points of the actual average of judgements. This additional payment was paid out after the experiment ended and was not part of any donation decision in the experiment, because determination of estimation accuracy could not be done within the experiment itself. Following this task, they enter the second main section of the experiment, in which they complete two sets of choices that measure their risk and ambiguity attitudes, specifically a bomb risk elicitation task [39, 40] and an Ellsberg urn task. Across these two tasks, they can earn up to £0.98 in a second endowment, depending on their decisions and luck. All participants are then presented with one additional donation choice at Final Choice where they can decide to allocate all, some, or none of their earned endowment to one randomly selected charity (either a sure-thing or a probabilistic charity). For a complete overview of the experimental outline, see Fig 1. For an in-detail description of all steps and measures, see the section below.

Fig 1. Experimental outline.

Fig 1

Measures

Numeracy, empathy, optimism, and donor type measures

In order to acquire measures of empathy, numeracy, and optimism, we use three validated, standardly used scales. Our numeracy measure is the 11-point numeracy scale [36] which includes three items on general numeracy and 7 items on risk numeracy specifically. We collect tendency towards optimism with the Life Orientation [37] which is a standard measure for optimism and pessimism. Further, we use the Basic Empathy Scale in Adults (BES-A) as our measure of empathy [38]. Participants are incentivised to complete these surveys by being rewarded with £0.25 for the full completion of each of the three surveys. We also collect self-reported donor type classification. To determine donor type (warm glow vs. pure altruist) we draw on Carpenter’s [41] methodology that relies on a self-reported recollection of previous charitable behaviour, classifying donors as either warm glow, pure altruist, or other motivations.

Risk and ambiguity attitudes

We estimate risk and ambiguity attitudes via two distinct, fully incentivised tasks. In order to ensure participant comprehension, participants complete a trial run of the risk attitude measure, are shown the mechanism for the ambiguity mechanism in detail via an example, and answer comprehension questions on both measures. Specifically, we measure risk attitudes via the ‘Bomb risk elicitation task’ [39, 40], adopting the mechanism for Qualtrics directly from Nielsen [40]. In this task, participants are presented with a 5x5 matrix of boxes. They are informed that one box contains a bomb, and that all other boxes yield an additional income of 2 pence. Participants can select any numbers of boxes, but if the box containing a bomb is selected and they confirm that they have selected as many boxes as they wish, their income is zero. As Crosetto & Filippin [39] point out, this measure of risk attitudes has several upsides over other alternatives in that it “requires minimal numeracy skills [and] avoids truncation of the data” [39, p. 31]. Specifically, it “measures individual-level risk attitudes by a single parameter k ∈ {0, 1, …, n}, the number of boxes collected” [42, p. 599]. Risk aversion is indicated by selecting fewer than 13 boxes and risk-seeking by selecting more than 14. In effect, participants make choices between 26 lotteries ranging from choosing 0 boxes with a certain outcome of £0 to choosing all 25 boxes with a certain outcome of £0, while any number of boxes in between has a different chance to lead to additional income. The expected value of these lotteries is bow-shaped: playing lotteries with 13 and 14 boxes has the highest expected value, with choosing more or fewer boxes reducing expected value. As such, risk-neutral participants would choose 13 or 14 boxes, risk averse participants fewer than 13, and risk-seeking participants more than 14.

We measure ambiguity attitudes by recourse to standard Ellsberg urns: Participants are presented with two urns. They can bet on both the outcome of a risky urn containing 5 red and 5 black balls and on the outcome of an ambiguous urn containing an unknown quantity or both red and black balls. We then collect reservation prices for each urn by using a Becker-DeGroot-Marschak mechanism with random integers being uniformly drawn from the interval {1,2,…,25}. This design is adapted from Halevy [43]. The ambiguity premium is calculated by taking the difference between the ambiguous urn’s reservation price and the risky urn’s reservation price [43, p. 522–523] with negative outcomes indicating ambiguity aversion (and positive outcomes indicating ambiguity proneness).

Demographic factors

We also collect a number of additional demographic factors that are chosen as additional control variables. These are partially drawn from Bekkers & Wiepking [15] and Wiepking & Bekkers [44] who survey the literature and present a number of factors that predict charitable giving in a variety of contexts. Specifically, we collect data on age, gender, education, employment, religious affiliation and practice, marital status, children, and subjective financial well-being.

Main choice and final choice outcomes

We collect two main outcome variables for this study. These rely on participant choices when presented with the option to donate all, some, or none of their earned endowment to charity. Specifically, our two main outcomes are Main Choice (where participants make choices between a sure-thing and a probabilistic charity) and Final Choice (where participants make choices about a randomly selected charity). In both cases, we collect data on whether they donate at all, if they do, which charity they choose in Main Choice (participants can choose up to one charity), and how much they donate. In other words, participants can donate a non-zero amount to either charity or not donate at all (i.e., one cannot select a charity and choose to donate ‘0’).

In Main Choice, each participant is shown exactly two charities, one sure-thing charity and one probabilistic charity. Each of these charities is drawn randomly from a pool of three charities (for a total of six individual charities and nine total charity pairs). This level of randomisation was chosen to avoid confounding this choice by the specifics of any given charity or any possible comparison effects between two specific charities. By randomising the pairs, we aim to isolate the factor of probability between the choices (sure-thing vs. probabilistic interventions) without having any potential result be influenced by the specific characteristics of any of these charities all the while keeping this choice as naturalistic as possible. Their choice here relates to null hypotheses #1 and #2.

In Final Choice, participants can again make a choice about their new endowment that they earned from the risk and ambiguity attitude tasks (ranging from £0 to £0.98, depending on their choices and luck). They are presented with one randomly selected charity (either a sure-thing charity or a probabilistic charity with equal likelihood) from the same pool as before and can again decide how much of their endowment to donate on this. We also ask participants before the uncertainty over their endowment in this section is resolved to state how much they would donate to this charity for different levels of endowments. This condition controls for a number of potential confounds in Main Choice which allows for it to answer the paper’s central question more directly and cleanly, though it being the last task of the experiment, we cannot rule out potential order effects. This choice relates to null hypothesis #5.

All six charities that are used in both Main Choice and Final Choice are recommended by GivingWhatWeCan in their 2021 charity recommendation titled “What are the best charities to donate to in 2021?” [45] to ensure that all of these charities are relatively similar in quality and reputation, and participants are informed about this fact prior to their choice. The three sure-thing charities are the SCI Foundation (deworming), GiveDirectly (direct cash transfers), and the Against Malaria Foundation (insecticide-treated bed nets). The three probabilistic charities are the Center for Health Security (epidemic/global health disaster prevention), Nuclear Threat Initiative (nuclear war prevention), and the Machine Intelligence Research Institute (artificial intelligence risk).

Our theoretical justification for this grouping is that all three sure-thing charities describe interventions that have a substantially high likelihood of success. All three interventions (deworming drugs, crash transfers, and bed nets) are administered by highly effective charities and have interventions that have been shown to have a high probability of being deployed successfully (i.e., that the cash transfer arrives, or the drug is administered). On the other side, all three probabilistic interventions (epidemic prevention, nuclear war prevention, artificial intelligence risk mitigation) are made under substantial ambiguity as to the probability distributions both of the intervention itself and the corresponding event occurring. These interventions are highly probabilistic in a way that cannot be resolved ex ante, making them qualitatively different than the sure-thing charities on a theoretical level, which is also recognised by the charities themselves and is as such part of their mission (after all, some of those risks are hard to quantify and even harder to address, but if such an intervention would be successful, it would bring with it substantially positive outcomes). For full treatment texts, see S7 Appendix.

However, we also provide an empirical justification for this grouping that supports the theoretical reasoning outlined above by conducting an auxiliary study. Note, however, that this study was conducted ex-post at the request of an anonymous reviewer and was not part of our original pre-registration. We conducted a post-hoc auxiliary study to empirically confirm our categorisation into sure-thing and probabilistic charities. This is to ensure that this distinction is not only theoretically grounded but also perceived as intended by the general public. We recruited a total of 101 participants on Prolific that had not participated in the main study, none of which failed the attention check. Participants were paid £0.75 for their participation. They were presented with all six charities and were asked to rate them on a scale from 0–10 on the likelihood that the charity’s intervention succeeds (relating to uncertainty over its interventions) and on the quantifiability of the charity’s intervention (relating to ambiguity). We also asked participants to rate the charities on their moral deservingness to keep the objective of this study relatively opaque.

We find strong support for the distinction between sure-thing charities and probabilistic charities on the basis of both uncertainty and ambiguity. See Table 1 for means, standard deviations, and medians of the uncertainty and ambiguity ratings for all six charities, with 0 indicating low probability that the charity’s intervention will succeed and a low level of quantifiability of its interventions, and 10 indicating a high probability and quantifiability. In other words, the higher the scores, the less risky and the less ambiguous the charity’s respective intervention is.

Table 1. Probability-ratings for all six charities.
Uncertainty
Mean (SD) Median
Sure-Thing Charity 1 (SCI Foundation) 7.19 (2.08) 8
Sure-Thing Charity 2 (GiveDirectly) 5.92 (2.09) 6
Sure-Thing Charity 3 (Against Malaria Foundation) 7.36 (1.83) 8
Probabilistic Charity 1 (Machine Intelligence Research Institute) 4.15 (2.46) 4
Probabilistic Charity 2 (Nuclear Threat Initiative) fd 4.09 (2.64) 4
Probabilistic Charity 3 (The Center for Health Security) 4.30 (2.50) 4
Ambiguity
Mean (SD) Median
Sure-Thing Charity 1 (SCI Foundation) 7.28 (1.97) 8
Sure-Thing Charity 2 (GiveDirectly) 5.42 (2.49) 5
Sure-Thing Charity 3 (Against Malaria Foundation) 7.24 (2.10) 8
Probabilistic Charity 1 (Machine Intelligence Research Institute) 3.63 (2.72) 3
Probabilistic Charity 2 (Nuclear Threat Initiative) fd 4.03 (2.73) 4
Probabilistic Charity 3 (The Center for Health Security) 4.69 (2.40) 4

Notes: Mean, Standard Deviation, and Median of risk and ambiguity ratings for all six charities.

We find that the data behave as generally expected, with sure-thing charities receiving higher ratings about the likelihood that their interventions will succeed as well as higher ratings for the quantifiability of their interventions, and probabilistic charities receiving lower rating correspondingly. Adding a subject’s scores for the individual charities of each bucket, we find that sure-thing charities are rated as having significantly higher probability interventions (M = 20.47, SD = 4.52) than the probabilistic charities (M = 13.96, SD = 5.83). This difference, 6.51, 95% CI [5.19, 7.82] was highly statistically significant, t(100) = 9.81, p < .001. The same picture emerges with regard to the quantifiability of the interventions, with the mean of the sum of the quantifiability scores of sure-thing charities (M = 19.83, SD = 4.60) being significantly higher than that of the probabilistic charities (M = 12.85, SD = 6.70), with the difference of 6.98, 95% CI [5.62, 8.34] also being statistically significant at t(100) = 10.17, p < .001. The effect sizes of these two differences, in Cohen’s d, is d = .98 for the probability ratings and d = 1.01 for the quantifiability rankings. This provides strong support for our theoretically based distinction between sure-thing and probabilistic charities.

Belief elicitations

We elicit 1st and 2nd order beliefs regarding the charity’s impact to investigate if participants who donate to one set of charities simply think that they are more impactful overall. Specifically, we ask participants to state their views on how impactful both charities they were presented with were on a percentage scale from 0 (Not impactful at all) to 100 (Extremely impactful). We also ask participants to estimate the average judgements regarding the impact of each charity as ascertained by the responses of all other participants. This is incentivised by paying those participants whose estimation is within 5 percentage points of the actual average an additional £0.10.

Additional treatments

We also had two additional treatments, an Expected Value Informational Treatment and a Context-Free Treatment. These were chosen to provide data that can speak to null hypothesis #4 and #3 respectively.

For our Expected Value Informational Treatment, we present participants with an introductory text of expected value reasoning. This is to test whether this simple information treatment can meaningfully shift donor behaviour. In order to control for experimenter demand effects, half of the participants in this condition are presented with a control text that introduces them to a made-up mathematical decision theory (midontic decision-theory). The control paragraphs introduce participants to this theory of decision-making via mathematically formulated but nonsensical arguments that are illustrated with the same example as the treatment text. It recommends the same type of action, namely that taking risks can be worth it if the potential outcomes are good enough. This allows us to identify whether a given effect is due to experimenter demand or not by comparing a shift in frequency of donation to sure-thing/probabilistic charities between the expected-value treatment with the Main Choice and the made-up theory treatment (midontic decision-theory) and the Main Choice. See S7 Appendix for full texts. The choices here relate to null hypothesis #4.

In the Context-free condition, participants are given the option to choose between two charities that had all context (area of intervention, geographic focus, name, etc) removed. The only characteristics left are of the probabilistic nature of their interventions as well as some generic filler text. The participants are told that the full information would be made available to them right after they made their choice. See S7 Appendix for full texts. This condition is used to provide data about to null hypothesis #3.

However, because the number of people who made donations was unexpectedly small, both of these conditions did not have the power that we calculated prior to running this study to detect a meaningful effect. This means that results of these conditions are inconclusive. We still report the full pre-registered analyses in the appendix, see S4 Appendix, but do not discuss them in the main results and discussion sections due to this reason.

Results

In Table 2, we summarise the general demographic variables and their frequencies with our sample. Generally, the sample shows high numeracy skills, with only 5.5% of participants not scoring at least 8 out of 10, and we observe a mean empathy score of M = 76.54, SD = 9.75 and mean optimism scores of M = 33.08, SD = 7.20. Further, according to Carpenter’s [41] classification of donor types, we can classify 24.3% of participants as warm-glow donors, 11.9% as pure altruist donors, and the remaining 63.9% as being motivated by some other reason.

Table 2. Demographics.

n % n %
Age Religious Participation
 18–28 247 .198  Yes 112 .090
 29–38 216 .174   No 1133 .910
 39–48 228 .183
 49–58 228 .183 Marriage Status
 59 and above 326 .262  Married 586 .471
 Not married 659 .529
Gender
 Male 617 .496 Children
 Female 618 .496  Has children 692 .556
 Other 10 .008  Does not have children 553 .444
Education Financial Wellbeing
 High school 498 .400  Finding it very difficult 58 .047
 Undergraduate 446 .358  Finding it quite difficult 149 .120
 Graduate/Professional 301 .242  Just about getting by 408 .328
 Doing alright 479 .385
Religious Affiliation  Living comfortably 151 .121
 No affiliation 893 .717
 Protestantism 162 .130 Employment
 Catholicism 120 .096  Unemployed 174 .140
 Islam 34 .027  Out of the workforce 223 .179
 Hinduism 15 .012  Part-time employment 279 .224
 Judaism 10 .008  Full-time employment 569 .457
 Buddhism 8 .006
 Sikhism 3 .002

Notes: Demographics for the full sample after exclusion of 261 participants who failed more than one attention or comprehension check.

Our results for risk and ambiguity attitudes show a strong tendency towards risk aversion and a moderately strong tendency towards ambiguity aversion. Following [42] we measure individual-level risk attitudes by the number of boxes collected, k. Based on the cumulative distribution of participant choices, we find that 83.7% of participants exhibit risk aversion (k < 13), 8.9% of participants exhibit risk neutrality (13 ≪ k ≪ 14), and 7.4% exhibit risk-seeking behaviour (k > 14). Our ambiguity attitude measure is calculated by subtracting the ambiguous urn’s reservation price from the risky urn’s reservation price with negative outcomes indicating ambiguity aversion (and positive outcomes indicating ambiguity proneness). We find that 34.4% of participants exhibit ambiguity aversion, 46.6% of participants exhibit ambiguity neutrality, and the remaining 19.0% participants exhibit ambiguity-seeking attitudes. Additionally, we do not find a significant correlation between these two measures in the full sample with Pearson’s r = -.05, p = .079.

Concerning the charitable giving outcomes in Main Choice, we find that across all conditions, 35.8% of participants make a donation. Of those making a donation, 84.7% donate to a sure-thing charity, and 15.3% donate to a probabilistic charity. The average donation of those who donate is 50.60 pence (SD = 23.98). For those donating to a sure-thing charity, the average donation is 51.50 pence (SD = 23.83), while for those donating to a probabilistic charity, it is 45.60 pence (SD = 24.43). This difference is not statistically significant, t(305) = 1.557, p = .121. We further find a strong relationship between choosing a type of charity and one’s belief in its impact as well as its estimation of the general consensus of its impact by those making a donation. Specifically, we find that choosing to donate to a probabilistic charity stands in a strong positive point-biserial correlation to judging the probabilistic charity as impactful, rpb = .424, p < .001, as well as estimating its average impact judgement, rpb = .346, p < .001. Conversely, the same choice also stands in a strong negative correlation with judging the sure-thing charity as impactful, rpb = -.314, p < .001, which is again mirrored in its estimated average impact judgement, rpb = -.267, p < .001. Looking at donation behaviour in Final Choice, we find that 29.1% of participants made a donation. 36.3% of those being presented with a sure-thing charity donated, compared to only 22.8% of those being shown a probabilistic charity. This difference is statistically significant, χ2 (1, N = 1177) = 25.66, p < .001. Of those making a donation, those being presented with a sure-thing charity donated an average of 26.58 pence (SD = 17.92) while those who were shown a probabilistic charity donated an average of 20.55 pence (SD = 16.55), a difference that is again statistically significant, t(340) = 3.45, p < .001.

First, we investigate general donation behaviour relating to Main Choice. The results presented in Table 3 speak to the central null hypotheses #1, #2, and #3. Model (1) reports the results for the Main Condition. The outcome variable is the type of charity conditional on a donation being made, with 0 being coded as the sure-thing charity and 1 as the probabilistic charity. The gender variable is coded 1 for female, all other categorical variables are coded 1 for the affirmative. As outlined above, the risk attitude measure is a discrete variable of the number of boxes opened, the ambiguity aversion is the result of the subtraction of the reservation prices. All other scales are the sum of the (re-reversed) individual items. As specified in our pre-registration, we report main regression results for binary outcomes using an OLS model and also for the corresponding logit model as a robustness check in Appendix Table 2 in S2 Appendix to check for sensitivity to functional form choice, where we find no impact of model choice. We made this choice primarily to enable clear and straightforward a priori power calculations, present more easily interpretable results, and because results are generally insensitive to model choice in most situations because the main reason against the use of linear models is that predicted probabilities may be greater than 1 or smaller than 0 which mainly occurs when the probabilities of the outcomes are extreme, not when they are in a roughly 80/20 distribution as is present in this paper [4650]. We also report a robustness check in Appendix Table 3 in S2 Appendix where we report a regression with random effects at the stimulus level, finding that our null result is also robust to this model choice.

Table 3. Regression results for charitable giving behaviour in main choice predicting choice between sure-thing and probabilistic charities.

(1)
Risk Attitude .000 (.005)
Ambiguity Aversion -.003 (.005)
Numeracy .004 (.024)
Empathy < .001 (.003)
Optimism -.004 (.003)
Donor Type
 Warm-Glow -.014 (.051)
 Pure Altruism -.062 (.064)
Donation (amount) -.001 (.001)
Age .003 (.002)
Gender .013 (.050)
Education
 Undergraduate degree .044 (.051)
 Postgraduate/Professional degree -.020 (.055)
Religion
 Protestantism -.082 (.067)
 Catholicism .012 (.073)
 Islam -.101 (.122)
 Judaism -.163 (.379)
 Buddhism -.105 (.287)
 Hinduism .378 (.276)
Religious Participation .041 (.089)
Marriage Status .045 (.050)
Children -.007 (.054)
Financial Wellbeing .005 (.024)
Employment
 Out of the workforce -.154 (.097)
 Part-time employment .003 (.085)
 Full-time employment -.054 (.079)
R2 .064
Sample size 307

Notes: OLS regression reporting unstandardised coefficients and standard errors. Outcome variable is charity choice (0 = sure-thing charity, 1 = probabilistic charity). *p < .1, **p < .05, ***p < .01, ****p < .001

We find that none of the main variables nor the demographic control variables predict charity choice in the Main Condition. We can straightforwardly conclude from Model (1) that we do not have evidence to reject the null hypotheses #1 and #2 as none of the independent variables meaningfully predict donor behaviour. However, note that Model (1) includes the control variable of amount donated that was not pre-registered but suggested by a helpful anonymous reviewer. For the pre-registered regression model without this control with no difference in results, please see S1 Appendix.

Second, we investigate general donation behaviour in Final Choice where participants were either presented with a sure-thing or a probabilistic charity. In Table 4, we report two further regression models relevant to null hypothesis #5. These are not the pre-registered ones but instead include interaction terms that we did not pre-register. For specifications of the regression models as pre-registered, see Appendix Table 6 in S3 Appendix. Model (2) predicts frequency of donation and Model (3) predicts size of donation. For Model (2), not making a donation is coded as 0 and making a donation is coded as 1. The central variables are interaction terms, where we interact the main individual difference measures with the condition. Specifically, they are coded with 0 = sure-thing charity and 1 = probabilistic charity for all our main explanatory variables. In these analyses, participants from all conditions are included and are split only by which type of charity they were presented with at Final Choice; recall that each participant here was only presented with one randomly selected charity. For all analyses of Final Choice, we both apply the general exclusion criteria (excluding anyone who has answered more than one comprehension/attention task incorrectly) as well as the additional exclusion criterion where we exclude participants who indicated a donation that was higher than their earned endowment for Final Choice.

Table 4. Regression results for charitable giving behaviour in final choice predicting frequency of giving and size of donation.

(2) (3)
Risk Attitude .002 (.005) .124 (.120)
Ambiguity Aversion -.002 (.004) -.127 (.109)
Numeracy .015 (.017) -.193 (.453)
Empathy .007**** (.002) .212**** (.048)
Optimism .003 (.003) .062 (.071)
Donor Type
 Warm-Glow .061* (.047) 2.170* (1.237)
 Pure Altruism .045 (.058) -.144 (1.525)
Condition X Risk Attitude .003 (.006) .019 (.161)
Condition X Ambiguity Aversion .002 (.005) .085 (.143)
Condition X Numeracy .006 (.018) .275 (.482)
Condition X Empathy -.002* (.002) -.086 (.056)
Condition X Optimism -.004 (.003) -.033 (.088)
Donor Type
 Condition X Warm-Glow -.034 (.063) -.993 (1.649)
 Condition X Pure Altruism .177** (.083) 2.955 (2.176)
Age .001 (.001) .047 (.030)
Gender -.050* (.030) -.747 (.788)
Education
 Undergraduate degree -.050 (.031) -1.129 (.800)
 Postgraduate/Professional degree .030 (.034) -.095 (.897)
Religion
 Protestantism .071* (.043) .153 (1.129)
 Catholicism .045 (.047) 1.468 (1.239)
 Islam .173* (.142) 6.606*** (2.317)
 Judaism .134 (.142) 2.636 (3.730)
 Buddhism .159 (.160) 4.994 (4.207)
 Hinduism .149 (.126) 7.049** (3.301)
 Sikhism .046 (.260) -2.343 (6.819)
Religious Participation .015 (.054) -.265 (1.406)
Marriage Status -.025 (.031) -.469 (.819)
Children .032 (.033) .697 (.877)
Financial Wellbeing .023 (.015) .398 (.384)
Employment
 Out of the workforce .017 (.055) -.820 (1.442)
 Part-time employment -.019 (.047) -2.211* (1.243)
 Full-time employment -.029 (.042) -1.436 (1.107)
R2 .076 .085
Sample size 1177 1177

Notes: OLS regressions reporting unstandardised coefficients and standard errors. Model (2) predicts frequency of donation and Model (3) predicts size of donation. Interaction terms interact the condition (0 = sure-thing charity, 1 = probabilistic charity) with the explanatory variables

*p < .1,

**p < .05,

***p < .01,

****p < .001

Here we find that while empathy predicts donation frequency and size overall, it is only the interaction term with pure altruistic donor type that statistically significantly predicts frequency of donation to probabilistic charities. This is in line with theoretical predictions that hold that pure altruists would be more likely to give to probabilistic charities as they primarily care about the (expected) impact on social welfare. Further, in our robustness check that reproduces Model (2) as a logit model (Appendix Table 5 in S3 Appendix), we find the same result. All other pre-registered variables (interacting with condition) do not predict donor behaviour.

Because we have found mostly null results in both Main Choice and Final Choice, we conduct the following exploratory analyses to further investigate the failure to reject the null hypotheses. Regarding the results from Main Choice, we conduct a series of equivalence tests for the multiple linear regression coefficients from Model (1) in Table 3 following Alter & Counsell [51] These allow us to make direct inferences about the absence of an effect or the presence of a negligibly small effect (i.e., an estimate of a precise zero) by providing two one-sided t-tests against an upper and a lower equivalence bound. In order for there to be sufficient evidence in favour of a negligible effect (i.e., evidence in favour of the absence of an effect or the presence of a negligibly small effect), both hypotheses have to be rejected at the .05 level. Because we did not pre-register the equivalence bounds, we report three plausible levels of the bounds in unstandardised coefficients at .01, .025, and .05 to show sensitivity of results to this choice. We report this for all our primary variables of interest from Main Choice that did not show significant effects.

These results suggest that we can be confident that the majority of our variables do not predict donor behaviour between sure-thing and probabilistic charities. Risk attitudes, empathy, and optimism show evidence in favour of a negligibly small effect at very tight equivalence bounds of (-).01, while ambiguity aversion shows this effect at (-).025, both relatively small levels. For numeracy, we can only conclude negligible effects at the (-).05 level. We remain unable to state clearly whether donor type has an effect on this choice as its results are insignificant both in the pre-registered regression analyses as well as the exploratory tests of equivalence. As such, we do not reach a conclusion as to the effect of donor type in Main Choice, but we can confidently state that the other explanatory variables of individual differences do not predict choices between sure-thing and probabilistic charities.

For the Final Choice analyses, we also conduct equivalence tests as above, see Table 6, with a focus on the interaction term coefficients from Model (2). We find that at the relatively tight (-).025 level, risk attitudes, ambiguity aversion, empathy, and optimism show a negligibly small effect in the frequency of donation model. Numeracy again shows a negligible effect at the (-).05 level and results regarding donor type (specifically warm glow) again do not allow for the conclusion of a negligible effect. We also report equivalence tests for the donation size in standardised coefficients in Appendix Table 10 in S5 Appendix, finding a similar pattern of results that shows numeracy and donor type as non-negligible effects while the remainder of variables show negligibly small effects.

These results provide us with a broad evidence base in favour of there not being an effect across many of our central explanatory variables. Also note that the results in Tables 5 and 6 are relatively robust to adjusting for multiple hypothesis testing. For example, if we adjusted our significance level to .007 following the Bonferroni method, we would still find that all significant effects remain significant at least at the (-).05 level.

Table 5. TOST for main choice coefficients.

-.01 .01 -.025 .025 -.05 .05
Risk Attitudes 2.0** 2.0** 5.0**** 5.0**** 10.0**** 10.0****
Ambiguity Aversion 1.4* 2.6*** 4.4**** 5.6**** 9.4**** 10.6****
Numeracy .58 .25 1.21 .88 2.25** 1.92**
Empathy 3.32**** 3.43**** 8.32**** 8.34**** 16.66**** 16.68****
Optimism 2** 4.66**** 7.0**** 9.67**** 15.33**** 18.0****
Warm Glow -.08 .47 .22 .76 .71 1.25
Pure Altruism -.81 1.13 -.58 1.36* -.19 1.75**

Notes: All t-test results for TOST procedures on a variety of lower and upper equivalence bounds (in unstandardized coefficients) from Model (1).

*p < .1,

**p < .05,

***p < .01,

****p < .001

Table 6. TOST for final choice coefficients.

-.01 .01 -.025 .025 -.05 .05
Condition X Risk Attitudes 2.17** 1.17 4.67**** 43.67**** 8.83**** 7.83****
Condition X Ambiguity Aversion 2.4*** 1.6* 5.4**** 4.6**** 10.4**** 9.6****
Condition X Numeracy .89 .22 1.72** 1.06* 3.11**** 2.44***
Condition X Empathy 4**** 6**** 11.5**** 13.5**** 24**** 26****
Condition X Optimism 2** 4.67**** 7**** 9.67**** 15.33**** 18****
Condition X Warm Glow -.38 .70 -.14 .94 .25 1.33*

Notes: All t-test results for TOST procedures on a variety of lower and upper equivalence bounds (in unstandardized coefficients) from Model (2).

*p < .1,

**p < .05,

***p < .01,

****p < .001

Further, we conduct exploratory Bayesian [52, 53] where we use Bayesian linear regression analyses that draw on Bayesian model averaging [54, 55]. First, in Table 7, we report results where we compare the individual models with one predictor each (the respective explanatory variable) against the null model with an intercept only. We report Bayes Factor Model Odds for the null for both Main Choice and Final Choice (frequency of donation as well as size of donation). For the analyses in Table 7, we use a uniform model prior.

Table 7. Bayes factor model odds for null model.

Main Choice Final Choice (Freq.) Final Choice (Size)
Risk Attitude 7.46
Ambiguity Aversion 7.81
Numeracy 7.44
Empathy 7.89
Optimism 3.50
Warm-Glow Type 7.94
Pure Altruistic Type 5.94
Condition X Risk Attitude .04 < .001
Condition X Ambiguity Aversion 13.41 7.57
Condition X Numeracy < .001 < .001
Condition X Empathy < .001 < .001
Condition X Optimism < .001 < .001
Condition X Warm Glow 3.82 6.23
Condition X Pure Altruism 1.46 15.32

Notes: Bayes Factor Model Odds for the Null Model with Uniform Prior for both the explanatory variables for Main Choice and the interaction terms for Final Choice.

The results in Table 7 indicate that for Main Choice, we have strong additional evidence in favour of a null effect of all variables of interest with Bayes factors of between 3.5 and 7.94. The picture is more complicated with regard to the interaction effects from Final Choice. There, we find that compared to models with ambiguity aversion or altruistic types individually, the null model is much more likely given the data. However, the other explanatory variables do not share this pattern. To further investigate the effect of our main variables on our outcomes of interest, we also report the model averaged coefficients that account for uncertainty over the estimates as well as uncertainty over model choice with all interaction effects and all control variables (age, gender, education, marital status, children, financial well-being, employment, religious affiliation, and religious participation). We report those coefficients as well as their 95% credible intervals that represent a weighted average that is weighted by the posterior probability of predictor inclusion. We use a uniform model prior and a JZS parameter prior with the default r scale of .354.

The results reported in Table 8 are in line with previous frequentist regression analyses, where we find that pure altruistic donor type predicts donations to probabilistic charities while most other interaction effects do not show any effects or only provide weak evidence in favour of them once all variables are entered into the model. As such, we conclude that the exploratory Bayesian analyses roughly corroborate our previous results.

Table 8. Bayesian linear regression coefficients and 95% credible intervals.

Final Choices (Freq.) Final Choice (Size)
Condition X Risk Attitude .000 [-.001, .005] .02 [-.03, .24]
Condition X Ambiguity Aversion .000 [-.001, .005] -.003 [-.008, .10]
Condition X Numeracy -.01 [-.02, .000] -.49 [-.89, .000]
Condition X Empathy .000 [-.002, .001] .001 [-.07, .06]
Condition X Optimism -.001 [-.005, .000] -.000 [-.03, .08]
Condition X Warm Glow -.004 [-.01, .07] .19 [-.25, 2.19]
Condition X Pure Altruism .20 [.08, .32] .90 [.000, 4.66]

Notes: Model averaged coefficients and 95% credible intervals.

Discussion

Overall, we find little to no evidence in favour of rejecting our null hypotheses, and as such this paper primarily reports a negative result: First, we do not find evidence that risk and ambiguity attitudes robustly predict charitable decision-making between sure-thing and probabilistic charities (null hypothesis #1), and we also do not find evidence that individual differences in numeracy, optimism, donor type, and empathy predict behaviour in this choice as well (null hypothesis #2). Further, we only find weak evidence predicting behaviour when participants are presented with either a sure-thing or a probabilistic charity, where we find that pure altruistic donor types significantly predict donation frequency to probabilistic charities (null hypothesis #5). Further, due to a lower than planned sample size, we are unable to conclusively evaluate the data about our no-context condition (null hypothesis #3) and with regard to the effect of the informational treatment introduced participants to expected-value reasoning (null hypothesis #4).

Importantly though, in conducting our exploratory analyses of equivalence tests and Bayesian analyses both in the context of Main Choice and Final Choice, we find that there is strong evidence in favour of negligibly small effects (or in favour of null effects) across a variety of plausibly set equivalence bounds and variable. Specifically, we find that for risk attitudes, empathy, and optimism, irrespective of the equivalence bound tested, their effects are negligibly small even at a delta level of an unstandardised coefficient at .01. For ambiguity aversion, the effect is negligible at .025, and for numeracy at .05. We do not have evidence in favour of a negligible effect for donor types. These patterns hold for both Main Choice and Final Choice with the caveat that in Final Choice, these tests were conducted with the interaction terms. We also conduct Bayesian analyses and find corroborating patterns of results. This provides evidence in favour of a null effect, something that the standard null-hypothesis testing technically cannot provide. In other words, these results suggest that risk attitudes, ambiguity attitudes, and individual differences in numeracy, empathy, and optimism have a negligible or no effect on charitable giving behaviour with regard to sure-thing and probabilistic charities (in both Main Choice and Final Choice designs). This is in itself an important finding that deserves to be recognised in the academic literature in an effort to combat file-drawer concerns and to expand our understanding of donation behaviour with regard to real charities.

Our results in Final Choice also provide us with some statistically significant effects. First, we find that empathy predicts both frequency and size of donation overall. This is in line with the previous literature on empathy and charitable giving [23, 32, 56]. Second, with regard to the central research question at hand, we find that in predicting charitable behaviour when participants are only presented with either a sure-thing or a probabilistic charity, we find that purely altruistic donor type predicts frequency of donation to probabilistic charities. This effect is directionally as predicted and is in line with theory in that pure altruists are more likely to be primarily concerned with an intervention’s impact on social welfare and given the potentially high (expected) impact that probabilistic charities might have, the results provide evidence for this claim. This suggests that, at least compared to our analyses of the Main Choice, the data in Final Choice provide at least weak evidence in favour of our alternative hypotheses.

The results thus indicate that our hypothesised variables for explaining behaviour in the context of donating to sure-thing or probabilistic charities fail to meaningfully predict actual donor behaviour. However, these results do not by themselves allow for a concrete specification of why we fail to detect an effect. It may simply be that a different factor is at play when donors make these types of decisions and that individual differences in risk/ambiguity aversion do not impact decision-making at this level or that standard laboratory measures of risk/ambiguity attitudes does not directly translate into charitable decision-making in the context studied. It may also be that while this design was meant to provide a naturally occurring decision as is possible within the constraints of an online experiment, that a contraposition of two distinct charity types impacts decision-making such that participant choices are biased in some way. For example, this could be because of distinction bias, i.e., the bias describing how our preferences and choices may be substantially distinct in conditions where we evaluate options in separation or jointly [5759]. We have some evidence in favour of this worry as at least some of the pre-registered factors predict directionally as expected in Final Choice, where only one charity was presented to participants. As such, it may be justified to put more interpretative emphasis on the results from Final Choice compared to Main Choice.

The results in Final Choice may also help reinforce the above claim in favour of the validity of this outcome measure and suggest that one reason for a failure to detect an effect in Main Choice may have been due to distinction bias related concerns, the bias that preferences and choices may differ between joint evaluation and separate evaluation modes [5759]. On their own terms, these findings suggest that pure altruists are more likely to donate to probabilistic charities compared to other donor types (like warm glow donors).

Further, it may be that participants did not pick up on the distinction between the two charities and that our outcome variable was faulty in some way. If this was the case, this would threaten the validity of all results presented here. There are several reasons to believe that this is not the case. First, the qualitative comments at the end of this study often explicitly raise concerns of probability, risk, and chance with regard to the participants’ charity choice, suggesting that they did pick up on the central difference. Second, we find that participants who favour probabilistic charities also judge them as more impactful and also estimate that the average judgements of other participants as higher than those who favour sure-thing charities. Lastly, we conducted an auxiliary study in which a new set of participants rated the charities among other things on the dimension of risk. There, we find that sure-thing charities are rated as significantly less risky and less ambiguous than probabilistic charities (p < .001). This gives us some prima facie evidence in favour of the validity of our outcome variable, though they are not (and indeed cannot be) conclusive. However, it is worth pointing out that this auxiliary study cannot help us rule out whether the two groups of charities also differed along other dimensions in the perceptions of participants.

It is also worth pointing out that while most of the previous literature discussed above uses manipulations of risk and ambiguity that are more tightly controlled (i.e. by introducing risk over donations controlled by the experimenter), this study attempted to provide a more naturalistic and thus externally valid picture of actual risky and ambiguous charitable giving, and failing to provide evidence in favour of this may also be partly explained by the general difficulty that highly abstract concepts in perfectly controlled laboratory games have when they are moved towards more externally valid contexts, cf. [60]. As such, failing to provide further evidence in favour of the impact of individual differences in, say, risk attitudes, on charitable decision making may be best understood as being, at least in part, explainable by a failure of experimental measures to generalise above and beyond their original context (as the outcome variables in this study were actual charities and not, like similar literature before, allocation decisions in dictator games where risk can be varied between conditions quite precisely). This in itself is not a bad thing, and we take our finding to contribute to a cumulative science that aims to establish which measures generalise in what contexts and situations and where such generalisations fail (as is the case in this paper). Additionally, it is important to point that much of the literature that this paper builds on is based on fundamentally different experimental designs. As Frey et al. [61] have already shown, behavioural elicitation methods for risk preferences often show low correlations between each other, making it not as straight-forward as presented previously to assume a similar underlying relation of different elicitation methods. While we argue that this is not primarily a weakness of the currently, we see it as general issue with the field overall, suggesting that more specific cognitive models and theories that build stronger connections between elicitation methods, constructs, and cognitive processes influencing behaviour would be sorely needed. We thank an anonymous referee for pressing us on this point.

Moreover, there is an additional reason to significantly favour the results of Final Choice over those of Main Choice throughout the interpretation of this paper’s data. This is because the design of Main Choice has some inherent flaws that make interpretation of results difficult. We thank an anonymous referee for pressing us on this point. This is because when participants make choices between two charities, they can make choices between compositions that they are indifferent between. For example, any one donor may be indifferent between a donation of £0.25 to a sure-thing charity and a donation of £0.10 to a probabilistic charity. In this scenario, the donor would choose randomly between the two. This in itself makes it difficult to cleanly identify a preference for one charity type or the other in the design that we have pre-registered and analysed in this paper. Based on this reason, we argue that the results from Final Choice should be given greater weight compared to Main Choice throughout this paper.

This concern is even more relevant if one considers that it may be the case that donors could be indifferent between donating a positive amount to a sure-thing charity and no donation to a probabilistic charity. As before, if they are indifferent between the two and choose randomly, this inhibits straightforward interpretation of our results. This is because we exclude all participants who choose not to donate (to understand their behaviour between these two options). Further, because it is quite plausible that, given the distribution of donation choices (most people who donate donate to a sure-thing charity), we may be excluding significantly more people favouring probabilistic charities but not sure-thing charities from those who are indifferent between a non-zero donation to a sure-thing charity and a zero donation to a probabilistic (and thus choose randomly between the two).

Some of these concerns have been addressed ex-post in this paper, for example by controlling for amount donated in the analyses of Main Choice that we did not initially pre-register. While this goes some way towards addressing this concern, we argue that some fundamental design constraints of the set-up that we chose remain. As such, we have put a higher emphasis on our analyses of Final Choice (which do not fall prey to the same structural challenges) throughout this paper and argue that one ought to be generally cautious in interpreting the results of Main Choice. However, given that it was pre-registered, we continue to report it fully (and where we deviate from the pre-registered protocol, we document this in detail and provide the original analyses in the appendix). We hope that this discussion, highlighting these issues in detail, properly contextualises the results for readers.

Overall, given that that we find little or no evidence in favour of rejecting our null hypotheses as well as provide evidence in favour of a null effect across both of our main hypotheses, we take our paper to primarily report a negative result. Given our relatively high level of power that was calculated a priori, and the fact that all analysis steps were pre-registered (and where we deviated from the pre-registered protocol, we documented this and provided the pre-registered analyses in the appendix), and because our equivalence tests provided strong evidence in favour of negligibly small results, we can be relatively confident in this result (at least in the data provided by Final Choice). Because it is important for any science aiming to be cumulative and reproducible, it is imperative that null results like these are communicated openly and clearly, and we take our paper to be doing precisely that.

Limitations

One potential limitation of this study is that our final sample sizes used for analyses of Main Choice were lower than those specified in the a priori power analyses. While we recruited the exact number that we preregistered, this reduction in sample size was primarily because a lower percentage of participants decided to donate to any charity than expected, i.e., in our data set 35.8% of participants choose to make a donation, while we expected about two-thirds of participants to make a donation [34, 35] However, because our power analyses were conducted with the highly set goal of having .95 power to detect the smallest effect size of interest at f2 = .02, we were only slightly below the sample size required for the more conventional level of power at .8 (which would have required 395 participants), which should alleviate these concerns at least in part. Additionally, these sample size concerns do not apply to any analyses conducted with regard to our Final Choice data due to the difference in design. As such, while these worries ought to be brought to the attention of readers, we do think that they are ultimately manageable in scope and do not take away from the findings of this paper.

A second potential limitation is the use of a web-based sample from Prolific. While we have tried to counterbalance this concern by relying on a representative sample (at least along the dimensions of age, sex, and ethnicity), there may still be some dimension along which our sample is not representative of the population as a whole. This, in turn, means that that there may be some external validity worries inherent in using this sample, which may impact generalisability of our results.

Conclusion

In this paper, we have attempted to investigate what best explains individual charitable giving behaviour in situations where donors are presented with highly reliable charities (sure-thing charities) and significantly riskier charities (probabilistic charities). We have failed to reject most of our null hypotheses in the context of donor behaviour when choosing between the types of charities and can only offer mixed results with regard to donor choices when they are presented with either a sure-thing or a probabilistic charity. There, we find that donor type meaningfully predicts frequency of donation to probabilistic charities. Exploratory results of equivalence tests, however, provide positive evidence in favour of the absence of an effect (or the presence of a negligibly small effect) across most of our main explanatory variables. Overall, we take this study to produce a robustly negative result in that our hypothesised variables failed to predict donor behaviour, while also producing positive evidence in favour of the claim that individual differences in risk/ambiguity attitudes and individual differences in numeracy, optimism, and empathy do not predict choices with regard to these two types of charities.

Supporting information

S1 Appendix. Preregistered model specification for model (1).

(PDF)

S2 Appendix. Robustness checks for main choice.

(PDF)

S3 Appendix. Robustness checks for final choice.

(PDF)

S4 Appendix. Additional hypotheses tests with inconclusive results.

(PDF)

S5 Appendix. Additional equivalence tests.

(PDF)

S6 Appendix. Additional regression for main choice.

(PDF)

S7 Appendix. Experimental texts.

(PDF)

S1 File

(PDF)

S2 File

(DOCX)

S1 Data

(CSV)

S2 Data

(CSV)

Acknowledgments

We thank for helpful comments, suggestions, and support Kirby Nielsen, Julian Jamison, Andreas Mogensen, Maximilian Maier, Joshua Lewis, Ben Grodeck, Theron Pummer, Ruru Hoong, and Thomas Rowe, as well as participants at the 2021 Early Career Conference Programme at the Global Priorities Institute, University of Oxford.

Data Availability

The data are stored on the OSF repository and freely available here: https://osf.io/w9gfu/.

Funding Statement

We hereby declare the following source of funding. One of the authors, Philipp Schoenegger, has received a research funding from the Forethought Foundation and the Centre for Effective Altruism (they do not provide grant numbers). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. Links: https://www.forethought.org/ https://www.centreforeffectivealtruism.org/.

References

  • 1.Giving USA 2020 (2020, June). Giving USA. https://givingusa.org/giving-usa-2020-charitable-giving-showed-solid-growth-climbing-to-449-64-billion-in-2019-one-of-the-highest-years-for-giving-on-record/
  • 2.Nonprofit Sector in Brief 2019 (2020, June). National Center for Charitable Statistics. https://nccs.urban.org/publication/nonprofit-sector-brief-2019
  • 3.Vives M. L., & FeldmanHall O. (2018). Tolerance to ambiguous uncertainty predicts prosocial behavior. Nature communications, 9(1), 1–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Chen, Y., & Zhong, S. (2021). Uncertainty Motivates Morality: Evidence and Theory. Available at SSRN 3737959.
  • 5.Haisley E. C., & Weber R. A. (2010). Self-serving interpretations of ambiguity in other-regarding behavior. Games and economic behavior, 68(2), 614–625. [Google Scholar]
  • 6.Cettolin E., Riedl A., & Tran G. (2017). Giving in the face of risk. Journal of risk and uncertainty, 55(2), 95–118. [Google Scholar]
  • 7.Olschewski S., Dietsch M., & Ludvig E. A. (2019). Anti-social motives explain increased risk aversion for others in decisions from experience. Judgment and Decision making, 14(1), 58–71. [Google Scholar]
  • 8.Mesa-Vázquez E., Rodriguez-Lara I., & Urbano A. (2021). Standard vs random dictator games: On the effects of role uncertainty and framing on generosity. Economics Letters, 109981. [Google Scholar]
  • 9.Exley C. L. (2016). Excusing selfishness in charitable giving: The role of risk. The Review of Economic Studies, 83(2), 587–628. [Google Scholar]
  • 10.Garcia T., Massoni S., & Villeval M. C. (2020). Ambiguity and excuse-driven behavior in charitable giving. European Economic Review, 124, 103412. [Google Scholar]
  • 11.Kappes A., Nussberger A. M., Faber N. S., Kahane G., Savulescu J., & Crockett M. J. (2018). Uncertainty about the impact of social decisions increases prosocial behaviour. Nature human behaviour, 2(8), 573–580. doi: 10.1038/s41562-018-0372-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Engel C., & Goerg S. J. (2018). If the worst comes to the worst: Dictator giving when recipient’s endowments are risky. European Economic Review, 105, 51–70. [Google Scholar]
  • 13.Niehaus, P. (2014). A theory of good intentions. San Diego, CA: University of California and Cambridge, MA: NBER, 111.
  • 14.Caviola L., Schubert S., & Nemirow J. (2020). The many obstacles to effective giving. Judgment and Decision Making, 15(2), 159–172. [Google Scholar]
  • 15.Bekkers R., & Wiepking P. (2011). Who gives? A literature review of predictors of charitable giving part one: Religion, education, age and socialisation. Voluntary Sector Review, 2(3), 337–365. [Google Scholar]
  • 16.Karlan D., & List J. A. (2007). Does price matter in charitable giving? Evidence from a large-scale natural field experiment. American Economic Review, 97(5), 1774–1793. [Google Scholar]
  • 17.Vesterlund L. (2006). Why Do people give? In Powell W. & Steinberg R. (Eds.), The Nonprofit Sector: A Research Handbook (pp. 568–590). Yale University Press. [Google Scholar]
  • 18.Knapp L., Wuepper D., & Finger R. (2021). Preferences, Personality, Aspirations, and Farmer Behavior. Agricultural Economics, 52(6), 1–13. [Google Scholar]
  • 19.Müller J., & Schwieren C. (2012). What can the Big Five personality factors contribute to explain small-scale economic behavior? [Google Scholar]
  • 20.Hilbig B. E., & Zettler I. (2009). Pillars of cooperation: Honesty–Humility, social value orientations, and economic behavior. Journal of Research in Personality, 43(3), 516–519. [Google Scholar]
  • 21.Antoncic B., Bratkovic Kregar T., Singh G., & DeNoble A. F. (2015). The big five personality–entrepreneurship relationship: Evidence from Slovenia. Journal of small business management, 53(3), 819–841. [Google Scholar]
  • 22.Sargeant A., & Woodliffe L. (2007). Gift giving: An interdisciplinary review. International Journal of Nonprofit and Voluntary Sector Marketing, 12(4), 275–307. [Google Scholar]
  • 23.Davis M. H., Mitchell K. V., Hall J. A., Lothert J., Snapp T., & Meyer M. (1999). Empathy, expectations, and situational preferences: Personality influences on the decision to participate in volunteer helping behaviors. Journal of personality, 67(3), 469–503. doi: 10.1111/1467-6494.00062 [DOI] [PubMed] [Google Scholar]
  • 24.Ashton M. C., Paunonen S. V., Helmes E., & Jackson D. N. (1998). Kin altruism, reciprocal altruism, and the Big Five personality factors. Evolution and Human Behavior, 19(4), 243–255. [Google Scholar]
  • 25.O’Reilly N., Ayer S., Pegoraro A., Leonard B., & Rundle-Thiele S. (2012). Toward an understanding of donor loyalty: Demographics, personality, persuasion, and revenue. Journal of Nonprofit & Public Sector Marketing, 24(1), 65–81. [Google Scholar]
  • 26.Kang E., & Lakshmanan A. (2018). Narcissism and self-versus recipient-oriented imagery in charitable giving. Personality and Social Psychology Bulletin, 44(8), 1214–1227. doi: 10.1177/0146167218764658 [DOI] [PubMed] [Google Scholar]
  • 27.Bianchi M., & Tallon J. M. (2019). Ambiguity preferences and portfolio choices: Evidence from the field. Management Science, 65(4), 1486–1501. [Google Scholar]
  • 28.Sutter M., Kocher M. G., Glätzle-Rützler D., & Trautmann S. T. (2013). Impatience and uncertainty: Experimental decisions predict adolescents’ field behavior. American Economic Review, 103(1), 510–31. [Google Scholar]
  • 29.Kleber J., Dickert S., Peters E., & Florack A. (2013). Same numbers, different meanings: How numeracy influences the importance of numbers for pro-social behavior. Journal of Experimental Social Psychology, 49(4), 699–705. [Google Scholar]
  • 30.Dickert S., Kleber J., Peters E., & Slovic P. (2011). Numeracy as a precursor to pro-social behavior: The impact of numeracy and presentation format on the cognitive mechanisms underlying donation decisions. Available at: http://hdl.handle.net/1794/22050 [Google Scholar]
  • 31.Boenigk S., & Mayr M. L. (2016). The happiness of giving: Evidence from the German socioeconomic panel that happier people are more generous. Journal of Happiness Studies, 17(5), 1825–1846. [Google Scholar]
  • 32.Verhaert G. A., & Van den Poel D. (2011). Empathy as added value in predicting donation behavior. Journal of Business Research, 64(12), 1288–1295. [Google Scholar]
  • 33.Kim S. J., & Kou X. (2014). Not all empathy is equal: How dispositional empathy affects charitable giving. Journal of Nonprofit & Public Sector Marketing, 26(4), 312–334. [Google Scholar]
  • 34.Butera, L., & Horn, J. R. (2017). Good News, Bad News, and Social Image: The Market for Charitable Giving. George Mason University Interdisciplinary Center for Economic Science (ICES) Working Paper.
  • 35.Metzger L., & Günther I. (2019). Making an impact? The relevance of information on aid effectiveness for charitable giving. A laboratory experiment. Journal of Development Economics, 136, 18–33. [Google Scholar]
  • 36.Lipkus I. M., Samsa G., & Rimer B. K. (2001). General performance on a numeracy scale among highly educated samples. Medical decision making, 21(1), 37–44. doi: 10.1177/0272989X0102100105 [DOI] [PubMed] [Google Scholar]
  • 37.Scheier M. F., Carver C. S., & Bridges M. W. (1994). Distinguishing optimism from neuroticism (and trait anxiety, self-mastery, and self-esteem): a reevaluation of the Life Orientation Test. Journal of personality and social psychology, 67(6), 1063. doi: 10.1037//0022-3514.67.6.1063 [DOI] [PubMed] [Google Scholar]
  • 38.Carré A., Stefaniak N., D’ambrosio F., Bensalah L., & Besche-Richard C. (2013). The Basic Empathy Scale in Adults (BES-A): Factor structure of a revised form. Psychological assessment, 25(3), 679–691. doi: 10.1037/a0032297 [DOI] [PubMed] [Google Scholar]
  • 39.Crosetto P., & Filippin A. (2013). The “bomb” risk elicitation task. Journal of Risk and Uncertainty, 47(1), 31–65. [Google Scholar]
  • 40.Nielsen K. (2019). Dynamic risk preferences under realized and paper outcomes. Journal of Economic Behavior & Organization, 161, 68–78. [Google Scholar]
  • 41.Carpenter, J. P., 2018. The Shape of Warm Glow: Field Experimental Evidence from a Fundraiser. IZA Discussion Paper No. 11760.
  • 42.Holzmeister F., & Stefan M. (2021). The risk elicitation puzzle revisited: Across-methods (in) consistency? Experimental Economics, 24(2), 593–616. doi: 10.1007/s10683-020-09674-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Halevy Y. (2007). Ellsberg revisited: An experimental study. Econometrica, 75(2), 503–536. [Google Scholar]
  • 44.Wiepking P., & Bekkers R. (2012). Who gives? A literature review of predictors of charitable giving. Part Two: Gender, family composition and income. Voluntary Sector Review, 3(2), 217–245. [Google Scholar]
  • 45.What are the best charities to donate to in 2021? (2021). Giving What We Can. https://www.givingwhatwecan.org/best-charities-to-donate-to-2021/
  • 46.Cleary P. D., & Angel R. (1984). The analysis of relationships involving dichotomous dependent variables. Journal of Health and Social Behavior, 334–348. [PubMed] [Google Scholar]
  • 47.Dey E. L., & Astin A. W. (1993). Statistical alternatives for studying college student retention: A comparative analysis of logit, probit, and linear regression. Research in higher education, 34(5), 569–581. [Google Scholar]
  • 48.Angrist J. D., & Pischke J. S. (2008). Mostly harmless econometrics. Princeton university press. [Google Scholar]
  • 49.Hellevik O. (2009). Linear versus logistic regression when the dependent variable is a dichotomy. Quality & Quantity, 43(1), 59–74. [Google Scholar]
  • 50.Gomila R. (2020). Logistic or linear? Estimating causal effects of experimental treatments on binary outcomes using regression analysis. Journal of Experimental Psychology: General, 150(4), 700–709. [DOI] [PubMed] [Google Scholar]
  • 51.Alter, U., & Counsell, A. (2021). Equivalence Testing for Multiple Regression. https://psyarxiv.com/ugc9e/
  • 52.Rouder J. N., & Morey R. D. (2012). Default Bayes factors for model selection in regression. Multivariate Behavioral Research, 47(6), 877–903. doi: 10.1080/00273171.2012.734737 [DOI] [PubMed] [Google Scholar]
  • 53.Liang F., Paulo R., Molina G., Clyde M. A., & Berger J. O. (2008). Mixtures of g priors for Bayesian variable selection. Journal of the American Statistical Association, 103(481), 410–423. [Google Scholar]
  • 54.Hinne M., Gronau Q. F., van den Bergh D., & Wagenmakers E. J. (2020). A conceptual introduction to Bayesian model averaging. Advances in Methods and Practices in Psychological Science, 3(2), 200–215. [Google Scholar]
  • 55.Maier M., Bartoš F., & Wagenmakers E. J. (2022). Robust Bayesian meta-analysis: Addressing publication bias with model-averaging. In press at Psychological Methods. doi: 10.1037/met0000405 [DOI] [PubMed] [Google Scholar]
  • 56.Neumayr M., & Handy F. (2019). Charitable giving: What influences donors’ choice among different causes? Voluntas: International Journal of Voluntary and Nonprofit Organizations, 30(4), 783–799. [Google Scholar]
  • 57.Hsee C. K., Loewenstein G. F., Blount S., & Bazerman M. H. (1999). Preference reversals between joint and separate evaluations of options: a review and theoretical analysis. Psychological bulletin, 125(5), 576–590. [Google Scholar]
  • 58.Hsee C. K., & Zhang J. (2004). Distinction bias: misprediction and mischoice due to joint evaluation. Journal of personality and social psychology, 86(5), 680–695. doi: 10.1037/0022-3514.86.5.680 [DOI] [PubMed] [Google Scholar]
  • 59.Anvari F., Olsen J., Hung W. Y., & Feldman G. (2021). Misprediction of affective outcomes due to different evaluation modes: Replication and extension of two distinction bias experiments by Hsee and Zhang (2004). Journal of Experimental Social Psychology, 92, 104052. [Google Scholar]
  • 60.Galizzi M. M., & Navarro-Martinez D. (2019). On the external validity of social preference games: a systematic lab-field study. Management Science, 65(3), 976–1002. [Google Scholar]
  • 61.Frey R., Pedroni A., Mata R., Rieskamp J., & Hertwig R. (2017). Risk preference shares the psychometric structure of major psychological traits. Science Advances, 3(10), e1701381. doi: 10.1126/sciadv.1701381 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Junhuan Zhang

17 Feb 2022

PONE-D-21-35635Sure-Thing vs. Probabilistic Charitable Giving: Experimental Evidence On the Role of Individual Differences in Risky and Ambiguous Charitable Decision-MakingPLOS ONE

Dear Dr. Schoenegger,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Apr 03 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Junhuan Zhang, PhD

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please provide additional details regarding participant consent. In the ethics statement in the Methods and online submission information, please ensure that you have specified (1) whether consent was informed and (2) what type you obtained (for instance, written or verbal, and if verbal, how it was documented and witnessed). If your study included minors, state whether you obtained consent from parents or guardians. If the need for consent was waived by the ethics committee, please include this information.

If you are reporting a retrospective study of medical records or archived samples, please ensure that you have discussed whether all data were fully anonymized before you accessed them and/or whether the IRB or ethics committee waived the requirement for informed consent. If patients provided informed written consent to have data from their medical records used in research, please include this information.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: Summary and Overall Evaluation

In this article, the authors examined the effect of several individual difference measures, among them, risk and ambiguity preferences, numeracy, empathy, optimism, and donation motivation. The main dependent variable is if and if then how much participants were willing to donate from their experimental endowment to one of two different types of charities, which the authors call sure-thing or probabilistic charities. Another dependent variable was the donation to an individual charity (either sure-thing or probabilistic) at a later stage of the experiment. In addition, the authors examined whether an intervention, namely advertising the EV-maximizing principle as a decision criterion, has an effect on donation choices and whether presenting anonymous and context-free donations rather than real-world charities with concrete description of their cause lead to different behavior than the main condition. As results, none of the individual difference measures significantly affected the choice of donating to a sure-thing or probabilistic charity, and only empathy and donation type affected overall donation in the choice to donate to an individual charity in some models.

The paper is clearly written and reports all methods, hypotheses and analyses in an understandable fashion. I appreciate that the study was pre-registered and I think the general research question is important and interesting. Moreover, the individual difference measures are all taken from previous research and seemed to be reasonable choices to measure the respective constructs. However, I am not fully convinced of the classification into sure-thing and probabilistic charity, of the theoretical underpinning of the hypotheses, and the power of the study. I will discuss each of these issues in more detail in the next section.

Major Comments

1. The main dependent variable is the choice between so-called sure-thing and probabilistic charities. In total there were six charities, three classified as sure-thing and three classified as probabilistic. This classification is rather ad-hoc and neither rooted in empirical evidence, nor based on principled theoretical arguments. Whereas I might agree from reading the description that this classification could be justified, the current study did not establish that all or even the majority of participants perceived the charities as assumed in the classification. Instead, there is a lot of information in the descriptions of the charities and participants might focus on different information than impact probability. Second, participants could perceive other information about the charity as varying in riskiness. For example, the direct money transfer could be perceived as risky with respect to the ultimate use of the donated money. Finally, the charities in the different conditions could also differ on other dimensions than riskiness. For example, the sure-outcome charities all have a direct impact on individual people, whereas the probabilistic charities have a more indirect impact on a large amount of people. To be fair, the authors mention concerns about the classification of charities into sure-thing and probabilistic in the discussion. However, the provided evidence in favor of their classification are not convincing as they do not directly speak towards a difference in perceived riskiness between the two types of charity. Instead, a straight-forward way to examine this question empirically would be to ask participants either from the old participant pool if available or a new one, how much they rate the respective charities on several dimensions including the riskiness of the project.

2. Since there were different charities administered in the stimulus material, it would make sense to estimate regressions with stimulus random effects, instead of OLS. Other models that cluster errors on the stimulus level might also be possible. Similar to the logistic, this could be done as a robustness test.

3. More effort should be spent on the theory of why the examined individual differences should be correlated with the choice between the sure-thing and the probabilistic charity. In particular, why should risk preference be related to this choice, if—as mentioned by the authors—previous literature (Vives & FeldmanHall, 2018) did not find a connection between risk-preference and prosocial behavior, that the exact probabilities of success of a charity are most likely ambiguous and that risk- and ambiguity preferences are not strongly correlated? Similarly, the connection between optimism or numeracy and the main DV is not build on theoretical considerations nor on previous findings that suggest such a connection. A strong foundation of the hypotheses would considerably strengthen the importance of the article.

4. There is some uncertainty about the power of the regression results as the pre-registered power analyses were based on the assumption that 2/3 donated, but the true number of donations was only 1/3. I appreciate that this is mentioned in the discussion, but I think this aspect deserves a bit more attention. In particular the EV-max intervention and the context-free manipulations with a much smaller sample size than the main dependent variable, might not have enough power to draw definite conclusions. Instead, it might make sense to put these analyses in an appendix and label these analyses as inconclusive. For the main DV, I appreciate that the authors calculate equivalent tests to examine the measurement uncertainty around the true effect. In my view, a better or at least complementary approach would be to calculate Bayes factors to evaluate the evidence for the Null (see Jarosz & Wiley, 2014; Rouder & Morey, 2012). Bayes Factors could clearly state whether enough evidence has been collected to conclude that there is no effect of an individual difference measure on the choice between the sure-thing or the probabilistic charity. In case that the evidence is inconclusive, it might make sense to collect more data.

Minor Comments

I would suggest to describe the hypotheses in the introduction in terms of the alternative hypothesis. Presenting hypotheses as the Null makes the text wordier and more complex than necessary.

On page 5 it is stated that:

“Our research builds on this literature but is importantly different, primarily because of our focus on donations to actual charities and not on pro-social behaviour in abstract games.”

And on page 6:

“Specifically, we focus on donation behaviour between real world charities that are made with an earned endowment, in contrast to abstract laboratory game pro-social decisions and hypothetical choices.”

These comments suggest that there is basically no research about charitable giving and all research about social preferences is only done in abstract experimental paradigms. I think this is a wrong impression and literature about charitable giving should be properly cited in these situations. Possible literature and review articles of which some are also cited in the text are Bekkers and Wiepking (2011), Karlan and List (2007), and Vesterlund and Sonnevi (2007).

References

Bekkers, R., & Wiepking, P. (2011). A literature review of empirical studies of philanthropy: Eight mechanisms that drive charitable giving. Nonprofit and voluntary sector quarterly, 40(5), 924-973.

Jarosz, A. F., & Wiley, J. (2014). What are the odds? A practical guide to computing and reporting Bayes factors. The Journal of Problem Solving, 7(1), 2.

Karlan, D., & List, J. A. (2007). Does price matter in charitable giving? Evidence from a large-scale natural field experiment. American Economic Review, 97(5), 1774-1793.

Rouder, J. N., & Morey, R. D. (2012). Default Bayes factors for model selection in regression. Multivariate Behavioral Research, 47(6), 877-903.

Vesterlund, L., & Sonnevi, G. (2006). 24. Why Do People Give?. In The nonprofit sector (pp. 568-588). Yale University Press.

Vives, M. L., & FeldmanHall, O. (2018). Tolerance to ambiguous uncertainty predicts prosocial behavior. Nature communications, 9(1), 1-9.

Reviewer #2: While I was reading the paper, I found the research questions interesting and important. I liked it a lot. However, when I read though the experiment design, I found, unfortunately, the design in the Main choice and context-free choice is not appropriate to answer the research questions. The concern is that the current design asks the participants to make two choices: which type of charity to donate to and how much to donate, while these two decisions are correlated with each other. Consequently, the experiment is not well controlled.

To make this more clear, I’ll first describe what is an ideal design and then point out the issues about the current design. To answer the research questions or test the hypotheses, we can either test whether they prefer to donate to sure-thing charity or probabilistic charity or test how much they would like to donate given sure-thing charity or probabilistic charity. In the former case, we need to control for the donation level, i.e. if they donate, they donate the same amount of money. In the latter case, we randomly control for the charity type. The latter case is of course the Final Choice design in the current paper which I fully agree is the correct way.

The issue with the current design is that, theoretically, there should be a point where one is indifferent between donate a certain amount of money under sure-thing charity and another amount under probabilistic charity. So what we watched in current experiment is just one of the two choices they are indifferent from. Therefore, it is hard to tell from their preference for the charity type.

To extend the above point a bit more and perhaps it helps to make it more clear, suppose there are some amount of people who would donate zero no matter what type of charity they are assigned or chose in the Main choice (this happens I guess for sure because there are only 35.8% of participants made a donation). For these people, economically, their choice of the charity type is invalid as the cost is zero (they will not donate anyway). This is an extreme case.

Now return to the paper, to save it, first, I think the Final Choice design is good, the authors may want to rely more on the data generate from the Final Choice. But at the same time take it in mind that this is last task which may suffer from order or experimenter demand effect. Second, if the authors really want to use the data from the Main Choice, they should at least control for the donate amount in the regression. Though I still think this is not valid enough to test the hypotheses.

In table 4, to test the hypothesis using the Final Choice data, they should add in interactive term between treatment variable (sure-thing or probabilistic) and the major explanatory variables to test the difference in difference, that is whether the major explanatory variables can explain the donation difference under sure-thing or probabilistic charity (the major question this paper intends to answer).

Minor points

The paper, especially the abstract can be shortened to make it more readable.

It is helpful to report R2 in the regression to show how much can be explained by the factors measured in this study.

Typos: by conducting an a priori power analysis

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Sep 22;17(9):e0273971. doi: 10.1371/journal.pone.0273971.r002

Author response to Decision Letter 0


5 Apr 2022

Please see the enclosed document 'Response to Reviewers' for a properly formatted response.

Dear Editor/Reviewers,

Thank you for the opportunity to revise this manuscript based on the comments outlined in detail below. In this letter, we provide all comments, our response to these comments, as well as an example of how we have implemented this change in our manuscript (if applicable), so that the time to go back and forth between this document and the revised manuscript is minimised. However, note that this only reflects a part of the changes made.

Both reviewers indicated that the research question and the methods were scientifically interesting and that the analyses were conducted well. However, both reviewers outlined some worries regarding interpretation, framing, and missing analyses. In this revised manuscript, we hope to have addressed all reviewer comments in full.

Overall, every section of the previous manuscript has been reworked to significant extent, while parts of the manuscript have been fully rewritten (see Revised Manuscript -Tracked Changes). The main changes that we have implemented are: (i) a fully reworked literature and hypotheses sections that better motivate our research question and the inclusion of all explanatory variables, (ii) additional Bayesian analyses to provide further evidence in favour of a null effect, (iii) a refocusing towards Final Choice as well as Main Choice as equally central tests of our hypothesis, (iv) a requested set of additional regressions, and (v) an additional auxiliary study.

Smaller changes have also been made throughout, focusing on readability and structure of the paper, correct contextualisation within the background literature, further justification of our design choices, as well as additional minor analyses.

We hope that our revisions are satisfactory and that our manuscript can now be considered for publication in PLOS One. However, should there been any further changes and amendment requested, we would gladly take them on board and revise our manuscript further. Thank you very much again for your time and effort in helping us improve this manuscript!

Reviewer #1

Summary Comment:

In this article, the authors examined the effect of several individual difference measures, among them, risk and ambiguity preferences, numeracy, empathy, optimism, and donation motivation. The main dependent variable is if and if then how much participants were willing to donate from their experimental endowment to one of two different types of charities, which the authors call sure-thing or probabilistic charities. Another dependent variable was the donation to an individual charity (either sure-thing or probabilistic) at a later stage of the experiment. In addition, the authors examined whether an intervention, namely advertising the EV-maximizing principle as a decision criterion, has an effect on donation choices and whether presenting anonymous and context-free donations rather than real-world charities with concrete description of their cause lead to different behavior than the main condition. As results, none of the individual difference measures significantly affected the choice of donating to a sure-thing or probabilistic charity, and only empathy and donation type affected overall donation in the choice to donate to an individual charity in some models.

The paper is clearly written and reports all methods, hypotheses and analyses in an understandable fashion. I appreciate that the study was pre-registered and I think the general research question is important and interesting. Moreover, the individual difference measures are all taken from previous research and seemed to be reasonable choices to measure the respective constructs. However, I am not fully convinced of the classification into sure-thing and probabilistic charity, of the theoretical underpinning of the hypotheses, and the power of the study. I will discuss each of these issues in more detail in the next section.

Comment 1:

The main dependent variable is the choice between so-called sure-thing and probabilistic charities. In total there were six charities, three classified as sure-thing and three classified as probabilistic. This classification is rather ad-hoc and neither rooted in empirical evidence, nor based on principled theoretical arguments. Whereas I might agree from reading the description that this classification could be justified, the current study did not establish that all or even the majority of participants perceived the charities as assumed in the classification. Instead, there is a lot of information in the descriptions of the charities and participants might focus on different information than impact probability. Second, participants could perceive other information about the charity as varying in riskiness. For example, the direct money transfer could be perceived as risky with respect to the ultimate use of the donated money. Finally, the charities in the different conditions could also differ on other dimensions than riskiness. For example, the sure-outcome charities all have a direct impact on individual people, whereas the probabilistic charities have a more indirect impact on a large amount of people. To be fair, the authors mention concerns about the classification of charities into sure-thing and probabilistic in the discussion. However, the provided evidence in favor of their classification are not convincing as they do not directly speak towards a difference in perceived riskiness between the two types of charity. Instead, a straight-forward way to examine this question empirically would be to ask participants either from the old participant pool if available or a new one, how much they rate the respective charities on several dimensions including the riskiness of the project.

Response 1: Thank you very much for raising this critique and for suggesting such an actionable step. As a first step, we have improved on our theoretical justification of this distinction: We have elaborated in much more detail than before for why one might think theoretically that these charities can be divided into these two groups. Furthermore, following your recommendation we have now conducted an auxiliary study (n=101) where participants rated the charities based on how risky their interventions were (riskiness) and how quantifiable they thought these interventions were (ambiguity). We found that probabilistic charities were rated as significantly more risky than sure-thing charities (p<.001) and that they were also rated as significantly more difficult to quantify (i.e. as more ambiguous) (p<.001), indicating that participants do in fact pick up on the underlying dimension of variation in probability and ambiguity in the predicted direction. Taken together, we believe that this distinction is now much better justified than in the original manuscript both theoretically and empirically. This, we argue, allows for a relatively straight interpretation of the results and we hope that this allays most of your concerns about this distinction. However, we will remain to call attention to this potential issue in our discussion section to ensure that all readers are made aware of the potential design weakness and the evidence presented here such that they can form an adequate picture of the data presented in this paper. However, recall also that moving the level of riskiness from easily controllable aspects like donation in standard lab settings to the actual charities themselves is one of the contributions of this paper.

Example (p. 34-35):

Appendix A – Auxiliary Study

We conducted a post-hoc auxiliary study to empirically confirm our categorisation into sure-thing and probabilistic charities. This is to ensure that this distinction is not only theoretically grounded but also perceived as intended by the general public. We recruited a total of 101 participants on Prolific that had not participated in the main study, none of which failed the attention check. Participants were paid £0.75 for their participation. They were presented with all six charities and were asked to rate them on a scale from 0-10 on the likelihood that the charity’s intervention succeeds (relating to uncertainty over its interventions) and on the quantifiability of the charity’s intervention (relating to ambiguity). We also asked participants to rate the charities on their moral deservingness to keep the objective of this study opaque.

We find strong support for the distinction between sure-thing charities and probabilistic charities on the basis of both uncertainty and ambiguity. See Appendix Table 1 for means, standard deviations, and medians of the uncertainty and ambiguity ratings for all six charities, with 0 indicating low probability that the charity’s intervention will succeed and a low level of quantifiability of its interventions, and 10 indicating a high probability and quantifiability. In other words, the higher the scores, the less risky and the less ambiguous the charity’s respective intervention is.

APPENDIX TABLE 1—PROBABILITY-RATINGS FOR ALL SIX CHARITIES

Uncertainty

Mean (SD) Median

Sure-Thing Charity 1 (SCI Foundation) 7.19 (2.08) 8

Sure-Thing Charity 2 (GiveDirectly) 5.92 (2.09) 6

Sure-Thing Charity 3 (Against Malaria Foundation)

7.36 (1.83) 8

Probabilistic Charity 1 (Machine Intelligence Research Institute)

4.15 (2.46) 4

Probabilistic Charity 2 (Nuclear Threat Initiative)

fd 4.09 (2.64) 4

Probabilistic Charity 3 (The Center for Health Security) 4.30 (2.50) 4

Ambiguity

Mean (SD) Median

Sure-Thing Charity 1 (SCI Foundation) 7.28 (1.97) 8

Sure-Thing Charity 2 (GiveDirectly) 5.42 (2.49) 5

Sure-Thing Charity 3 (Against Malaria Foundation)

7.24 (2.10) 8

Probabilistic Charity 1 (Machine Intelligence Research Institute)

3.63 (2.72) 3

Probabilistic Charity 2 (Nuclear Threat Initiative)

fd 4.03 (2.73) 4

Probabilistic Charity 3 (The Center for Health Security) 4.69 (2.40) 4

Notes: Mean, Standard Deviation, and Median of risk and ambiguity ratings for all six charities.

We find that the data behave as generally expected, with sure-thing charities receiving higher ratings about the likelihood that their interventions will succeed as well as higher ratings for the quantifiability of their interventions, and probabilistic charities receiving lower rating correspondingly. Adding a subject’s scores for the individual charities of each bucket, we find that sure-thing charities are rated as having significantly higher probability interventions (M=20.47, SD=4.52) than the probabilistic charities (M=13.96, SD=5.83). This difference, 6.51, 95% CI [5.19, 7.82] was highly statistically significant, t(100)=9.81, p<.001. The same picture emerges with regard to the quantifiability of the interventions, with the mean of the sum of the quantifiability scores of sure-thing charities (M=19.83, SD=4.60) being significantly higher than that of the probabilistic charities (M=12.85, SD=6.70), with the difference of 6.98, 95% CI [5.62, 8.34] also being statistically significant at t(100)=10.17, p<.001. The effect sizes of these two differences, in Cohen’s d, is d=.98 for the probability ratings and d=1.01 for the quantifiability rankings. This provides strong support for our theoretically based distinction between sure-thing and probabilistic charities.

Comment 2:

Since there were different charities administered in the stimulus material, it would make sense to estimate regressions with stimulus random effects, instead of OLS. Other models that cluster errors on the stimulus level might also be possible. Similar to the logistic, this could be done as a robustness test.

Response 2: Thank you very much for suggesting this analysis. We agree that this type of regression would be appropriate for our data structure. We now report one such model in the appendix (Appendix Table 3, Model 5), where we find the same pattern of results as in the main analyses, suggesting that our results are also robust to this modelling choice.

Example (p. 36-37):

We also report an additional robustness check of Model (1) in Appendix Table 3, Model (5). Specifically, we report a random effects model with the stimulus material being treated as a random effect. As we had three sure-thing and three probabilistic charities, there were nine charity pairs that participants could have been presented with. In Model (5), we treated the stimulus (i.e. the charity pair presented) as a random effect (with nine levels). Because significance levels are not unproblematic in mixed models like this, we also report 95% confidence intervals of the estimates. The results indicate that our null effect is robust to this model choice as well.

APPENDIX TABLE 3— REGRESSION RESULTS FOR MAIN CHOICE – RANDOM EFFECTS ROBUSTNESS CHECK

PREDICTING CHOICE BETWEEN SURE-THING AND PROBABILISTIC CHARITIES

(5)

Risk Attitude .002 (.005) [-.008, .012]

Ambiguity Aversion -.002 (.005) [-.010, .006)

Numeracy .003 (.023) [-.042, 0.47]

Empathy .001 (.002) [-.005, .004]

Optimism

-.006* (.003) [-.012, .001]

Donor Type

Warm-Glow -.007 (.047) [-.099, .086]

Pure Altruism -.040 (.059) [-.156, .077]

Age .001 (.002) [-.003, .004]

Gender -.024 (.047) [-.115, .068]

Education

Undergraduate degree .041 (.048) [-.053, .134]

Postgraduate/Professional degree -.023 (.051) [-.123, .078]

Religion

Protestantism -.043 (.062) [-.165, .080]

Catholicism .020 (.068) [-.114, .154]

Islam -.035 (.113) [-.246, .188]

Judaism -.006 (.354) [-.691, .703]

Buddhism -.060 (.265) [-.581, .462]

Hinduism .325 (.254) [-.175, .825]

Religious Participation -.004 (.083) [-.168, .159]

Marriage Status -.019 (.046) [-.071, .108]

Children .009 (.050) [-.090, .107]

Financial Wellbeing .009 (.022) [-.035, .053]

Employment

Part-time employment .058 (.054) [-.047, .164]

Full-time employment .015 (.049) [-.083, .112]

Sample size 307

Notes: Random effects model, coefficients, and standard errors, as well as 95% CIs. *p<.1, **p<.05, ***p<.01, ****p<.001

Comment 3:

More effort should be spent on the theory of why the examined individual differences should be correlated with the choice between the sure-thing and the probabilistic charity. In particular, why should risk preference be related to this choice, if—as mentioned by the authors—previous literature (Vives & FeldmanHall, 2018) did not find a connection between risk-preference and prosocial behavior, that the exact probabilities of success of a charity are most likely ambiguous and that risk- and ambiguity preferences are not strongly correlated? Similarly, the connection between optimism or numeracy and the main DV is not build on theoretical considerations nor on previous findings that suggest such a connection. A strong foundation of the hypotheses would considerably strengthen the importance of the article.

Response 3: Thank you very much for pressing us on this. We agree that our explanations were not thorough enough in the previous manuscript. We have now expanded upon our justification for all our individual difference measures throughout the revised manuscript, making clearer our reasoning for their inclusion. Specifically, we have now backed up the inclusion of all items with a wider number of references to the literature, showing how all constructs have been previously found to be associated with some measure of pro-social behaviour or charitable giving to directly motivate and justify their inclusion. Further, we have also updated our analytical reasoning to further improve the justifications of inclusion on both counts.

Example (p. 6-7):

First, we investigate whether donor choices can be explained by individual differences in risk and ambiguity attitudes. Previous work in domains such as stock market participation (Bianchi & Tallon, 2018) and health-related field behaviours (Sutter et al., 2013) has found that attitudes to risk and ambiguity can play significant roles. In context of pro-social behaviour in game environments the results show that ambiguity aversion may play a role (while risk aversion sometimes does not) (Vives & FeldmanHall, 2018), though risk aversion has also been found to be “predictive for giving” (Cettolin, Riedl, & Tran, 2017, 95). As such, we argue that given risk and ambiguity aversion have been shown to impact behaviour in many contexts including charitable giving, this makes it an a priori interesting relation to test. This hypothesis is also theoretically grounded, in that it might be the case that one’s preference not to give to charities that have a low chance of making an impact might be driven by an individual’s general risk aversion profile, or it might be that given the ambiguous nature of charitable interventions that it is only ambiguity aversion that impacts this choice.

[…]

For example, previous research has found that those lower in numeracy were more insensitive to proportions of donation recipients (Kleber, Dickert, Peters, & Florack, 2013) and that they showed higher susceptibility to changes in numeric presentation (Dickert, Kleber, Peters, & Slovic, 2011). It may as such be the case that one’s level of numeracy also meaningfully impacts behaviour in the context studied here as the probabilistic charities include interventions that have a small chance of making a large impact. Understanding these proportions plausibly requires a certain level of numeracy. Further, one may also think that a general proclivity to optimism may bias individuals towards overestimating the success of probabilistic charitable interventions, or conversely that higher pessimism may explain a preference for sure-thing charities as those promise to have a reliable impact even in the worst-case scenario. This is corroborated by previous research that draws on the German socioeconomic panel and finds that optimism predicts charitable giving in some of their models (Boenigk & Mayr, 2016).

Comment 4:

There is some uncertainty about the power of the regression results as the pre-registered power analyses were based on the assumption that 2/3 donated, but the true number of donations was only 1/3. I appreciate that this is mentioned in the discussion, but I think this aspect deserves a bit more attention. In particular the EV-max intervention and the context-free manipulations with a much smaller sample size than the main dependent variable, might not have enough power to draw definite conclusions. Instead, it might make sense to put these analyses in an appendix and label these analyses as inconclusive.

Response 4: Thank you for your comment and the actionable advice. We agree with you that the context-free manipulations (and in some respects the EV-information interventions) might have a too-low sample size to draw conclusions from justifiable. We have now followed your suggestion and labelled these as ‘inconclusive’ and have put them in the appendix. However, we do not share the same pessimism about Main Choice: Recall that our original power analysis was extremely strict, having been aimed to have .95 power to detect at f2=.02. While we understand the statistical challenges of calculating power post-hoc, moving from .95 power to .8 in the above power analysis would have reduced the sample size needed to nearly a fifth, making our collected sample size much more reasonable with this in mind. Also, our data from Final Choice does not share these concerns, and following Reviewer 2, we have now increased the prominence of our discussion of the results from Final Choice.

Example (p. 17):

However, because the number of people who made donations was unexpectedly small, both of these conditions did not have the power that we calculated prior to running this study to detect a meaningful effect. This means that results of these conditions are inconclusive. We still report the full pre-registered analyses in the appendix, see Appendix D, but do not discuss them in the main results and discussion sections.

Comment 5: For the main DV, I appreciate that the authors calculate equivalent tests to examine the measurement uncertainty around the true effect. In my view, a better or at least complementary approach would be to calculate Bayes factors to evaluate the evidence for the Null (see Jarosz & Wiley, 2014; Rouder & Morey, 2012). Bayes Factors could clearly state whether enough evidence has been collected to conclude that there is no effect of an individual difference measure on the choice between the sure-thing or the probabilistic charity. In case that the evidence is inconclusive, it might make sense to collect more data.

Response 5: We are very glad that you appreciate our use of equivalence tests. We have now expanded our usage of equivalence tests (now also for the results from Final Choice) to provide the reader with a wide range of plausible test ranges that may allow us to conclude a null effect (and to narrowly estimate the range of the equivalence bounds). Further, we have also followed your suggestion and included Bayesian analyses following Rouder & Morey (2012). The Bayesian results corroborate our equivalence test results in that they provide strong evidence in favour of a null effect.

Example (p. 24):

Following Rouder & Morey (2012), we compute exploratory Bayes factors to evaluate the evidence for the null. We report Bayes factors for linear models based on Liang et al. (2008), using a Jeffrey-Zeelner-Siow mixture of g-priors. Applying the default r scale of .353 for the results regarding Main Choice in Model (1), we calculate a JZS Bayes Factor of 2.49*107, suggesting very strong evidence in favour of the null hypothesis. The same analysis regarding Final Choice in Model (2) also provides evidence in favour of the null, though at a significantly smaller magnitude, at a JZS Bayes Factor of 22.15. This corroborates the results of the equivalence tests and suggests that our data provide strong evidence in favour of a robust null result.

Comment 6:

I would suggest to describe the hypotheses in the introduction in terms of the alternative hypothesis. Presenting hypotheses as the Null makes the text wordier and more complex than necessary.

Response 6: Thank you for your comment! While we agree that statements in terms of alternative hypotheses would aid comprehension in some respects, we will retain the phrasing in terms of null hypotheses for the following reasons:

1) We have pre-registered the null hypotheses this way and would like to stick as closely as possible to our pre-registration.

2) More importantly, because the results in this paper are primarily null results, retaining the null hypothesis structure enables us to more meaningfully discuss our inability to reject the null and to discuss evidence in favour of the null (with equivalence tests). As such, while alternative hypotheses might be easier to understand in the earlier sections of the paper, we believe that when it comes to the statistical results and discussions of them, null hypothesis phrasings are actually more straightforward.

Comment 7:

On page 5 it is stated that:

“Our research builds on this literature but is importantly different, primarily because of our focus on donations to actual charities and not on pro-social behaviour in abstract games.”

And on page 6:

“Specifically, we focus on donation behaviour between real world charities that are made with an earned endowment, in contrast to abstract laboratory game pro-social decisions and hypothetical choices.”

These comments suggest that there is basically no research about charitable giving and all research about social preferences is only done in abstract experimental paradigms. I think this is a wrong impression and literature about charitable giving should be properly cited in these situations. Possible literature and review articles of which some are also cited in the text are Bekkers and Wiepking (2011), Karlan and List (2007), and Vesterlund and Sonnevi (2007).

References

Bekkers, R., & Wiepking, P. (2011). A literature review of empirical studies of philanthropy: Eight mechanisms that drive charitable giving. Nonprofit and voluntary sector quarterly, 40(5), 924-973.

Jarosz, A. F., & Wiley, J. (2014). What are the odds? A practical guide to computing and reporting Bayes factors. The Journal of Problem Solving, 7(1), 2.

Karlan, D., & List, J. A. (2007). Does price matter in charitable giving? Evidence from a large-scale natural field experiment. American Economic Review, 97(5), 1774-1793.

Rouder, J. N., & Morey, R. D. (2012). Default Bayes factors for model selection in regression. Multivariate Behavioral Research, 47(6), 877-903.

Vesterlund, L., & Sonnevi, G. (2006). 24. Why Do People Give?. In The nonprofit sector (pp. 568-588). Yale University Press.

Vives, M. L., & FeldmanHall, O. (2018). Tolerance to ambiguous uncertainty predicts prosocial behavior. Nature communications, 9(1), 1-9.

Response 7: Thank you very much for pressing us on this! We of course did not mean to suggest that there was no research fitting these criteria. We have now adjusted our language throughout the literature review and cited all the literature outlined by you (except the ones that we had already cited before), also clearing up some additional issues in these sections: Specifically, we now make explicit that our contribution is primarily moving risk over donations in tightly controlled abstract games (as has been done before) to the risk/ambiguity over actual charities’ interventions, something that has both not been studied in detail and is much closer to the actual choice environment that potential donors find themselves in. Now we make clear that our claim to increase ecological validity is with regards to this aspect specifically, and of course not to charitable giving overall (though we agree that our previous manuscript did not make this sufficiently clear). We hope that our updated language now removes all ambiguity regarding this issue.

Example (p. 4-5):

Our research builds on the literature on risk that has so far mostly employed directly controllable levels of risk in the lab. For example, in abstract game scenarios, risk can be controlled and stated precisely, for example by imposing a 50% chance of one’s donation not going through, or by introducing a 5% chance that one’s donation is matched. Crucially though, our research is substantially different from the discussed literature primarily because we move the level of risk from directly calculable interventions in the lab (as outlined above) to the actual charities themselves. While this introduces several design challenges, we argue that this step leads to an increased level of ecological validity of any potential finding. However, note that there is already a large literature on charitable giving generally that has a similar or higher level of external validity (Bekkers & Wiepking, 2011; Karlan & List, 2007; Vesterlund & Sonevi, 2006). However, our paper’s main contribution is the moving of our focus on risk and ambiguity to actual organisations and their interventions and away from aspects that can be controlled in the lab. Having risk and ambiguity at the level of actual charities is the level at which risk and ambiguity typically enter people’s decision-making processes; rarely are we uncertain as to whether our donation will randomly increase when we donate (as in some experimental lab studies), but we are almost always acting under uncertainty about the charity’s interventions that we consider donating to.

Reviewer #2:

Summary Comment:

While I was reading the paper, I found the research questions interesting and important. I liked it a lot. However, when I read though the experiment design, I found, unfortunately, the design in the Main choice and context-free choice is not appropriate to answer the research questions. The concern is that the current design asks the participants to make two choices: which type of charity to donate to and how much to donate, while these two decisions are correlated with each other. Consequently, the experiment is not well controlled.

Comment 1:

To make this more clear, I’ll first describe what is an ideal design and then point out the issues about the current design. To answer the research questions or test the hypotheses, we can either test whether they prefer to donate to sure-thing charity or probabilistic charity or test how much they would like to donate given sure-thing charity or probabilistic charity. In the former case, we need to control for the donation level, i.e. if they donate, they donate the same amount of money. In the latter case, we randomly control for the charity type. The latter case is of course the Final Choice design in the current paper which I fully agree is the correct way.

The issue with the current design is that, theoretically, there should be a point where one is indifferent between donate a certain amount of money under sure-thing charity and another amount under probabilistic charity. So what we watched in current experiment is just one of the two choices they are indifferent from. Therefore, it is hard to tell from their preference for the charity type.

To extend the above point a bit more and perhaps it helps to make it more clear, suppose there are some amount of people who would donate zero no matter what type of charity they are assigned or chose in the Main choice (this happens I guess for sure because there are only 35.8% of participants made a donation). For these people, economically, their choice of the charity type is invalid as the cost is zero (they will not donate anyway). This is an extreme case.

Now return to the paper, to save it, first, I think the Final Choice design is good, the authors may want to rely more on the data generate from the Final Choice. But at the same time take it in mind that this is last task which may suffer from order or experimenter demand effect.

Response 1: Thank you very much for your thoughtful comments. We actually agree with you that the design for Final Choice is superior to that of Main Choice, primarily because Main Choice is a more substantive (and thus more confoundable) design, compared to the overall much cleaner design of Final Choice. As such, we have followed your recommendation and have prioritised the data from Final Choice. Specifically, we now report Main Choice and Final Choice results next to each other throughout the entire manuscript and discuss their comparative strengths and weaknesses in the discussion section. Further, on suggestion of Reviewer 1, we have also moved the results and discussion of the no-context condition as well as the expected-value treatment to the appendix, giving the Final Choice data even more space (effectively doubling its share of discussion space in the main manuscript). However, we will not solely rely on the data from Final Choice as we pre-registered our main hypotheses with both Main Choice and Final Choice in mind, and because we do believe that the data from Main Choice do meaningfully contribute to our understanding of the research question. We hope that our restructuring and comparative focus on Final Choice satisfactorily responds to your actionable request to rely more on Final Choice.

Lastly, we believe that given we have understood your worry correctly, we have some empirical evidence from our data that should go towards alleviating your worry to some extent. We believe that one testable prediction from your worry would be that those who donated to sure-thing charities before (i.e., in Main Choice) should donate less to probabilistic charities at Final Choice. We have investigated this in our data, looking at participants who donated to a sure-thing charity in Main Choice and correlated these donations with their donation amounts (including 0) when they were presented with a probabilistic charity. We do not find a statistically significant correlation at r(184)=.087, p=.238. We also do not find the predicted negative correlation in the reverse (with participants who donated to a probabilistic charity first and were then shown a sure-thing charity in Final Choice, though these results rest on a very low sample size and should probably be disregarded), with r(25)=.442, p=.027. These results suggest that your worry may not impact the data as much as one might have thought, making our solution of presenting both Main Choice and Final Choice as central pieces of the paper justified. Further, when you claim that one piece of your worry is that some people would “donate zero no matter what type of charity they are assigned or chose in the Main choice” we just want to quickly point out that those people would not be included in any of the analyses of Main Choice as only those making a donation are included in the regression models based on our pre-registered exclusion criteria.

Example (p. 6; 24-25):

In our experiment, each participant is first presented with a randomly selected pair of charities consisting of one charity of each type to control for accidental confounds relating the charity’s context as each are presented with substantial additional accurate information to increase the naturalness of the choice. Participant choices with respect to this randomly selected charity pair then allows us to isolate and capture the element of probability between the two charity types. In the second part of this experiment, we study participant behaviour when they are shown only one randomly selected charity (either sure-thing or probabilistic), which more narrowly captures the predictive value of individual differences on donation choices to charities of specific types. Overall, we find little to no evidence that individual differences in risk/ambiguity attitudes, numeracy, optimism, and donor type predict charitable giving behaviour. However, we do find that a purely altruistic donor type predicts donations to probabilistic charities. As such, we take this paper to be primarily reporting a null-result.

[…]

Overall, we find little or no evidence in favour of rejecting our null hypotheses, and as such this paper primarily reports a negative result: First, we do not find evidence that risk and ambiguity attitudes robustly predict charitable decision-making between sure-thing and probabilistic charities (null hypothesis #1), and we also do not find evidence that individual differences in numeracy, optimism, donor type, and empathy predict behaviour in this choice as well (null hypothesis #2). Further, we only find weak evidence predicting behaviour when participants are presented with either a sure-thing or a probabilistic charity, where we find that pure altruistic donor types significantly predict donation frequency to probabilistic charities (null hypothesis #5). Further, due to a lower than planned sample size, we are unable to conclusively evaluate the data about our no-context condition (null hypothesis #3) and with regard to the effect of the informational treatment introduced participants to expected-value reasoning (null hypothesis #4).

[…]

As such, it may be justified to put more interpretative emphasis on the results from Final Choice compared to Main Choice.

Comment 2:

Second, if the authors really want to use the data from the Main Choice, they should at least control for the donate amount in the regression. Though I still think this is not valid enough to test the hypotheses.

Response 2: Thank you for your comment. However, we have decided not to follow it because we are mindful of introducing endogeneity in the regression models by having two dimensions of the same choice in the same model as independent and dependent variables (choice of charity as Dep. V and size of that donation choice as Ind. V). If there is an opportunity for further revision, could you please elaborate for why you think we should include this control variable despite the challenges of endogeneity (and how we could circumvent these)? Then we’d be more than happy to follow your advice.

Comment 3:

In table 4, to test the hypothesis using the Final Choice data, they should add in interactive term between treatment variable (sure-thing or probabilistic) and the major explanatory variables to test the difference in difference, that is whether the major explanatory variables can explain the donation difference under sure-thing or probabilistic charity (the major question this paper intends to answer).

Response 3: Thank you very much for this comment! We agree that this would be a better approach and have moved our original (pre-registered) regressions to the appendix and now report two regressions with the interaction terms as proposed by you in the main text and rely on their results throughout the paper. We also use the interaction terms for our equivalence tests of the Final Choice coefficients.

Example (p. 21-22):

Second, we investigate general donation behaviour in Final Choice where participants were either presented with a sure-thing or a probabilistic charity. Here, we report two further regression models relevant to null hypothesis #5. These are not the pre-registered ones but instead include interaction terms that we did not pre-register. For specifications of the regression models as pre-registered, see Appendix C (Appendix Table 6). Model (2) explains frequency of donation, and Model (3) predicts size of donation. For Model (2), not making a donation is coded as 0 and making a donation is coded as 1. The central variables are interaction terms, where we interact the main individual difference measures with the condition. Specifically, they are coded with 0 = sure-thing charity and 1 = probabilistic charity for all our main explanatory variables. In these analyses, participants from all conditions are included and are split only by which type of charity they were presented with at Final Choice; recall that each participant here was only presented with one randomly selected charity.

TABLE 3—REGRESSION RESULTS FOR CHARITABLE GIVING BEHAVIOR IN FINAL CHOICE

PREDICTING FREQUENCY OF GIVING AND SIZE OF DONATION

(2) (3)

Risk Attitude .002 (.005) .124 (.120)

Ambiguity Aversion -.002 (.004) -.127 (.109)

Numeracy .015 (.017) -.193 (.453)

Empathy .007**** (.002) .212**** (.048)

Optimism

.003 (.003) .062 (.071)

Donor Type

Warm-Glow .061* (.047) 2.170* (1.237)

Pure Altruism .045 (.058) -.144 (1.525)

Condition X Risk Attitude .003 (.006) .019 (.161)

Condition X Ambiguity Aversion .002 (.005) .085 (.143)

Condition X Numeracy .006 (.018) .275 (.482)

Condition X Empathy -.002* (.002) -.086 (.056)

Condition X Optimism -.004 (.003) -.033 (.088)

Donor Type

Condition X Warm-Glow -.034 (.063) -.993 (1.649)

Condition X Pure Altruism .177** (.083) 2.955 (2.176)

Age .001 (.001) .047 (.030)

Gender -.050* (.030) -.747 (.788)

Education

Undergraduate degree -.050 (.031) -1.129 (.800)

Postgraduate/Professional degree .030 (.034) -.095 (.897)

Religion

Protestantism .071* (.043) .153 (1.129)

Catholicism .045 (.047) 1.468 (1.239)

Islam .173* (.142) 6.606*** (2.317)

Judaism .134 (.142) 2.636 (3.730)

Buddhism .159 (.160) 4.994 (4.207)

Hinduism .149 (.126) 7.049** (3.301)

Sikhism .046 (.260) -2.343 (6.819)

Religious Participation .015 (.054) -.265 (1.406)

Marriage Status -.025 (.031) -.469 (.819)

Children .032 (.033) .697 (.877)

Financial Wellbeing .023 (.015) .398 (.384)

Employment

Out of the workforce .017 (.055) -.820 (1.442)

Part-time employment -.019 (.047) -2.211* (1.243)

Full-time employment -.029 (.042) -1.436 (1.107)

R2 .076 .085

Sample size 1177 1177

Notes: OLS regressions reporting unstandardised coefficients and standard errors. Model (2) predicts frequency of donation and Model (3) predicts size of donation. Interaction terms interact the condition (0 = sure-thing charity, 1 = probabilistic charity) with the explanatory variables *p<.1, **p<.05, ***p<.01, ****p<.001

Here we find that while empathy predicts donation frequency and size overall, it is only the interaction term with pure altruism that statistically significantly predicts frequency of donation to probabilistic charities. This is in line with theoretical predictions that hold that pure altruists would be more likely to give to probabilistic charities as they primarily care about the impact on social welfare. Further, in our robustness check that reproduces Model (2) as a logit model (Appendix C, Appendix Table 5), we find the same result. All other pre-registered variables (interacting with condition) do not predict donor behaviour.

Comment 4:

The paper, especially the abstract can be shortened to make it more readable.

Response 4: Thank you for your comment. In this revision we have tried to make the paper shorter throughout (and focused especially on the abstract). Most of the robustness checks that were requested by the reviewers are now in the appendix to ensure that the paper’s length is not massively increased by responding to reviewer comments. The paper without references and appendix is now under 30 pages.

Comment 5:

It is helpful to report R2 in the regression to show how much can be explained by the factors measured in this study.

Response 5: Thank you very much for suggesting this. We now report R2 (or Cox and Snell R2) for all our regressions.

Comment 6:

Typos: by conducting an a priori power analysis

Response 6: Apologies, but we are unsure what the typo here is.

Attachment

Submitted filename: Response to Reviewers.docx

Decision Letter 1

Junhuan Zhang

16 May 2022

PONE-D-21-35635R1Sure-Thing vs. Probabilistic Charitable Giving: Experimental Evidence On the Role of Individual Differences in Risky and Ambiguous Charitable Decision-MakingPLOS ONE

Dear Dr. Schoenegger,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jun 30 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Junhuan Zhang, PhD

Academic Editor

PLOS ONE

Journal Requirements:

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: No

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: No

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: In general, the authors incorporated all my comments and improved the paper considerably. I would like to thank and congratulate them for their work and their constructive responses to the raised comments. I think the paper can be published with only minor revisions:

1. I very much like the additional study and the demonstration that participants indeed perceived the one class of charities as riskier and more ambiguous than the other. Given the importance of this result in judging the whole validity of the study with respect to the research question, I feel this study deserves much more space in the main manuscript. I think it should get a small methods and results section (maybe called study to validate the stimuli) and should also be mentioned already in the Methods of the main study when the different charities are introduced as stimuli. Importantly, while the additional study helps to support the claim that the charities differ in perceived riskiness, it does not exclude the possibility that the groups of charities also differed on other dimensions (e.g., individual vs. group as recipients etc.). I think this is an important limitation that should be mentioned in the discussion.

2. It is great that the authors conducted a Bayes Factor analysis. However, it was not exactly clear to me which models the authors compared. If it were the full model with all predictors vs. the null model with only an intercept this should be clearly reported and interpreted accordingly. Given the hypotheses, it would make sense to compare all individual models with one predictor from the five individual difference measures each against the null model with just an intercept. That way, the evidence for the individual null hypotheses could be unambiguously assessed.

3. Again, I think the authors did a good job in adding more literature and theoretical arguments to justify why the selected measures should correlate with donation behavior in their study. However, I also noticed that most of the studies they cite are based on different experimental designs. For example, Kleber et al. (2013) found an effect of numeracy on the effect of the number and share of helped people. This feature is not central to the current manipulation of interest. Similarly, in Cettolin et al. (2017) risk preferences are measures with certainty equivalents in decisions from description, whereas the current study uses an experience-based risk elicitation task with a different dependent variable. We already know that behavioral elicitation methods for risk preference correlate little with each other (see Frey et al., 2017). Thus, it is not straight-forward to assume the same underlying relation from different elicitation methods. Ultimately, I think this is not a weakness of the current study, but rather of the field as a whole. I would suggest the authors mention this problem in the discussion and call for more specific cognitive models and theories that build stronger connections between elicitation methods, constructs, and cognitive processes influencing behavior.

References:

Frey, R., Pedroni, A., Mata, R., Rieskamp, J., & Hertwig, R. (2017). Risk preference shares the psychometric structure of major psychological traits. Science Advances, 3(10), e1701381.

Reviewer #2: I appreciate the authors effort to give more weight to Final Choice. The paper has been improved a lot for sure. However, I’m not sure whether the authors understand my major concern well. The point is that you have two moving parts in your Main Choice design. Such a design flaw makes it unable to test your hypotheses and hence is totally invalid. The reasoning is simple: when one can choose both the type of the charity and the amount to donate, one can choose two different compositions that one is indifferent from. For example, if I’m indifferent between $1/sure thing charity and $0.5/probabilistic charity, I can choose randomly from the two. Overall, you would find roughly 50% choose sure thing and 50% choose probabilistic charity. This means you cannot test the preference for charity type under such a design. This is point one. Point two, suppose people are more likely to donate more under sure-thing charity, then when you excluded 0 donation as you preregistered, you exclude more people who chose probabilistic charity (you can test this with your data). Th reason is that there must be some people who are indifferent between $positive amount/sure thing charity and $0/probabilistic charity, and these people chose randomly between the two. In the observed results, you excluded those who chose probabilistic charity but not the other type. Point 3, because of the above issue, if you regress charity type preference on other variables, you have to control for the amount donated, as an ex-post control, because you did not control for the donate amount in the experiment. I’m not clear what endogeneity this may cause. But if there is any, I don’t believe the data from Main Choice can prove anything. There is no problem if you want to report the analyses as you preregistered. But a preregistration doesn’t mean your design and analysis (Main Choice) is not problematic.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Sebastian Olschewski

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Sep 22;17(9):e0273971. doi: 10.1371/journal.pone.0273971.r004

Author response to Decision Letter 1


23 Jun 2022

[PLEASE SEE ATTACHED FILE FOR A PROPERLY FORMATTED RESPONSE]

Dear Reviewers,

Thank you very much for evaluating our revised manuscript and for giving us the opportunity to revise and improve it further.

We are very happy that both of you agreed that our previous revision “improved the paper considerably” (R1) and “improved the paper a lot” (R2). However, we acknowledge that we have not yet fully addressed all of the comments raised to a satisfactory extent and we hope that in this revision, we will have done so.

Overall, we have made the following general improvements.

Following Reviewer 1’s recommendations, we have now moved our auxiliary study into the main text to give it more spotlight compared to the previous appendix placement. We have also updated our Bayesian analyses, now reporting comparisons of all individual models with the intercept-only model. Further, in part in response to Reviewer 2’s comment from the first round of revision, we have also provided model-averaged coefficients for the full Bayesian models of Final Choice to provide further analyses of the condition that suffers less design worries.

Following Reviewer 2’s recommendations, we now control for the amount donated in all analyses of Main Choice (and have updated our equivalence tests accordingly) such that Main Choice’s design downsides are at least in part addressed. In the discussion section, we now also discuss in much detail the major limitation that Reviewer 2 has identified with Main Choice. While we cannot drop these analyses from our paper as they were our primary pre-registered analyses, we now give significant space to this objection such that readers are fully informed about this design choice and its flaws. Further, we now report a regression in the appendix that mirrors Main Choice but aims to control for some of those worries.

Additionally, we have slightly reworked the literature section, updated the parts of the discussion section, and made further minor changes throughout the document.

We hope that these changes are sufficient for this paper to be considered for publication at PLOS One. However, if you have any further improvements that you’d like to see us make, we’d be more than happy to accommodate them too!

Reviewer #1

General Comment: In general, the authors incorporated all my comments and improved the paper considerably. I would like to thank and congratulate them for their work and their constructive responses to the raised comments. I think the paper can be published with only minor revisions:

General Response: Thank you very much! We hope that the changes below properly respond to your requested revisions.

Comment 1: I very much like the additional study and the demonstration that participants indeed perceived the one class of charities as riskier and more ambiguous than the other. Given the importance of this result in judging the whole validity of the study with respect to the research question, I feel this study deserves much more space in the main manuscript. I think it should get a small methods and results section (maybe called study to validate the stimuli) and should also be mentioned already in the Methods of the main study when the different charities are introduced as stimuli. Importantly, while the additional study helps to support the claim that the charities differ in perceived riskiness, it does not exclude the possibility that the groups of charities also differed on other dimensions (e.g., individual vs. group as recipients etc.). I think this is an important limitation that should be mentioned in the discussion.

Response 1: Glad to hear the addition of the auxiliary study has adequately responded to your previous request for revisions. We have now moved the study to the main text as you suggested. We have also mentioned the valid concern you have raised in the discussion section.

Example (p. 30): However, it is worth pointing out that this auxiliary study cannot help us rule out whether the two groups of charities also differed along other dimensions in the perceptions of participants.

Comment 2: It is great that the authors conducted a Bayes Factor analysis. However, it was not exactly clear to me which models the authors compared. If it were the full model with all predictors vs. the null model with only an intercept this should be clearly reported and interpreted accordingly. Given the hypotheses, it would make sense to compare all individual models with one predictor from the five individual difference measures each against the null model with just an intercept. That way, the evidence for the individual null hypotheses could be unambiguously assessed.

Response 2: Thank you for pointing out this shortcoming and giving a clear suggestion on how to improve it. We have now run a set of Bayesian linear regressions where we compare the individual models (with only one variable of interest) to the intercept-only model. For Final Choice, we do this for the interaction terms. In addition, in part in response to Reviewer 2’s suggestion to focus more on Final Choice, we also report model-averaged coefficients for Bayesian linear regression models with all variables (including controls) entered for the variables of interest, corroborating previous results.

Example (p. 26-27): Further, we conduct exploratory Bayesian analyses (Rouder & Morey 2012; Liang et al. 2008) where we use Bayesian linear regression analyses that draw on Bayesian model averaging (e.g., Hinne et al., 2019, Maier et al., 2022). First, in Table 7, we report results where we compare the individual models with one predictor each (the respective explanatory variable) against the null model with an intercept only. We report Bayes Factor Model Odds for the null for both Main Choice and Final Choice (frequency of donation as well as size of donation). For the analyses in Table 7, we use a uniform model prior.

Table 7—Bayes Factor Model Odds for Null Model

Main Choice Final Choice (Freq.) Final Choice (Size)

Risk Attitude 7.46

Ambiguity Aversion 7.81

Numeracy 7.44

Empathy 7.89

Optimism 3.50

Warm-Glow Type 7.94

Pure Altruistic Type 5.94

Condition X Risk Attitude .04 <.001

Condition X Ambiguity Aversion 13.41 7.57

Condition X Numeracy <.001 <.001

Condition X Empathy <.001 <.001

Condition X Optimism <.001 <.001

Condition X Warm Glow 3.82 6.23

Condition X Pure Altruism 1.46 15.32

Notes: Bayes Factor Model Odds for the Null Model with Uniform Prior for both the explanatory variables for Main Choice and the interaction terms for Final Choice.

The results in Table 7 indicate that for Main Choice, we have strong additional evidence in favour of a null effect of all variables of interest with Bayes factors of between 3.5 and 7.94. The picture is more complicated with regard to the interaction effects from Final Choice. There, we find that compared to models with ambiguity aversion or altruistic types individually, the null model is much more likely given the data. However, the other explanatory variables do not share this pattern. To further investigate the effect of our main variables on our outcomes of interest, we also report the model averaged coefficients that account for uncertainty over the estimates as well as uncertainty over model choice with all interaction effects and all control variables (age, gender, education, marital status, children, financial well-being, employment, religious affiliation, and religious participation). We report those coefficients as well as their 95% credible intervals that represent a weighted average that is weighted by the posterior probability of predictor inclusion. We use a uniform model prior and a JZS parameter prior with the default r scale of .354.

Table 8—Bayesian Linear Regression Coefficients and 95% Credible Intervals

Final Choices (Freq.) Final Choice (Size)

Condition X Risk Attitude .000 [-.001, .005] .02 [-.03, .24]

Condition X Ambiguity Aversion .000 [-.001, .005] -.003 [-.008, .10]

Condition X Numeracy -.01 [-.02, .000] -.49 [-.89, .000]

Condition X Empathy .000 [-.002, .001] .001 [-.07, .06]

Condition X Optimism -.001 [-.005, .000] -.000 [-.03, .08]

Condition X Warm Glow -.004 [-.01, .07] .19 [-.25, 2.19]

Condition X Pure Altruism .20 [.08, .32] .90 [.000, 4.66]

Notes: Model averaged coefficients and 95% credible intervals.

The results reported in Table 8 are in line with previous frequentist regression analyses, where we find that pure altruistic donor type predicts donations to probabilistic charities while most other interaction effects do not show any effects or only provide weak evidence in favour of them once all variables are entered into the model. As such, we conclude that the exploratory Bayesian analyses roughly corroborate our previous results.

Comment 3: Again, I think the authors did a good job in adding more literature and theoretical arguments to justify why the selected measures should correlate with donation behavior in their study. However, I also noticed that most of the studies they cite are based on different experimental designs. For example, Kleber et al. (2013) found an effect of numeracy on the effect of the number and share of helped people. This feature is not central to the current manipulation of interest. Similarly, in Cettolin et al. (2017) risk preferences are measures with certainty equivalents in decisions from description, whereas the current study uses an experience-based risk elicitation task with a different dependent variable. We already know that behavioral elicitation methods for risk preference correlate little with each other (see Frey et al., 2017). Thus, it is not straight-forward to assume the same underlying relation from different elicitation methods. Ultimately, I think this is not a weakness of the current study, but rather of the field as a whole. I would suggest the authors mention this problem in the discussion and call for more specific cognitive models and theories that build stronger connections between elicitation methods, constructs, and cognitive processes influencing behavior.

References:

Frey, R., Pedroni, A., Mata, R., Rieskamp, J., & Hertwig, R. (2017). Risk preference shares the psychometric structure of major psychological traits. Science Advances, 3(10), e1701381.

Response 3: We agree with your comment and have added this reference and made the recommendation you outlined in the discussion section.

Example (p.30-31): Additionally, it is important to point that much of the literature that this paper builds on is based on fundamentally different experimental designs. As Frey et al. (2017) have already shown, behavioural elicitation methods for risk preferences often show low correlations between each other, making it not as straight-forward as presented previously to assume a similar underlying relation of different elicitation methods. While we argue that this is not primarily a weakness of the currently, we see it as general issue with the field overall, suggesting that more specific cognitive models and theories that build stronger connections between elicitation methods, constructs, and cognitive processes influencing behaviour would be sorely needed.

Reviewer #2

General Comment: I appreciate the authors effort to give more weight to Final Choice. The paper has been improved a lot for sure. However, I’m not sure whether the authors understand my major concern well. The point is that you have two moving parts in your Main Choice design. Such a design flaw makes it unable to test your hypotheses and hence is totally invalid.

General Response: Thank you very much for giving us the opportunity to revise the paper further and for acknowledging that our previous revisions improved the paper a lot. We have now followed your recommendation in full and are controlling for the amount of donation in all analyses regarding Main Choice and have updated our equivalence test results and Bayesian analyses accordingly.

We would also like to thank you for spelling out your main objection to our Main Choice analysis in much more detail in this round of review. You were right that we had not fully grasped it at the last R&R stage, but we now believe that we have a much better understanding of your main objection to this design. While we hope that you can appreciate that we cannot remove the Main Choice analyses from our paper due to them being our pre-registered hypotheses, we have made four significant changes to our paper that we hope will help address your point. First, we have put even more weight on Final Choice, for example by conducting Bayesian linear regression analyses with, reporting model-averaged coefficients for it (but not for Main Choice). Second, we have followed your recommendation and are now controlling for amount donated in all Main Choice analyses. Third, we now outline your full objection in detail in the discussion section, where we give it substantive space, outline your examples, and detail these challenges for the interpretation of data collected from Main Choice. This is to ensure that all readers are informed about your concerns regarding our design, without compromising on our commitment to conduct, report, and analyse our experiment as pre-registered. Fourth, we have conducted two sets of regressions (one in the appendix, one with same results in this letter) aimed at controlling for your concerns, showing no difference in results. We hope that this goes towards addressing your main concern, though if you have any further specific recommendations, we are happy to follow them too.

Comment 1: The reasoning is simple: when one can choose both the type of the charity and the amount to donate, one can choose two different compositions that one is indifferent from. For example, if I’m indifferent between $1/sure thing charity and $0.5/probabilistic charity, I can choose randomly from the two. Overall, you would find roughly 50% choose sure thing and 50% choose probabilistic charity. This means you cannot test the preference for charity type under such a design. This is point one.

Response 1: Thank you very much for spelling out your main objection in more detail. We now believe that we better understand your main concern. Your concern is of course legitimate and true. In general, indifference is always a possibility (unless we make some assumptions in terms of the decision maker’s preferences). For example, even if we had given our subjects the choice between donating a fixed amount to the sure thing charity or to the probabilistic charity or donating nothing to either of them, a subject’s choice to donate to one of the two types of charities would not rule out the possibility that the decision maker was indifferent between donating the fixed amount to either of them. Of course, you are right that when the decision maker can choose not just between a sure thing charity and a probabilistic charity indifference but also the amount to be donated, then the indifference situation is more likely to arise than when the amount to be donated is fixed, and the only choice is to which charity to donate or to not donate at all. In the text, the new revised version now explicitly acknowledges the indifference issue you pointed out.

There is, however, a way to delve further into the data to get an idea of the concern you raised. That way involves considering a participant’s choices at Main Choice and Final Choice simultaneously to identify possible cases of indifference. We consider that those who: i) donated to a sure-thing charity in Main Choice but did not donate at all in Final Choice when they were shown a probabilistic charity, might have been indifferent between a strictly positive donation to a sure-thing charity and donating nothing to a probabilistic charity; ii) donated to a probabilistic charity in Main Choice but did not donate at all in Final Choice when they were shown a sure-thing charity, might have been indifferent between a strictly positive donation to a probabilistic charity and donating nothing to a sure-thing charity. Thus, in Appendix F we exclude such subjects from the regressions of Main Choice. The regression estimates replicate findings of the regressions in the main body of the text. See Example 1 below.

Of course, Appendix F’s regressions which are based on the above exclusions do not fully eliminate all the subjects who might have been indifferent between donating a strictly positive amount to one type of charity and a zero donation to the other type of charity. The reasons for this are that: i) subjects who donated to a sure-thing charity at Main Choice and were faced with the decision to donate a sure-thing charity at Final Choice might have been indifferent between donating a positive amount to a sure-thing charity and donating zero to a probabilistic charity at Main Choice, but since they were not faced with a probabilistic charity at Final Choice they did not have the chance to make choices to reveal that indifference; i) likewise, subjects who donated to a probabilistic charity at Main Choice and were faced with the decision to donate a probabilistic charity at Final Choice might have been indifferent between donating a positive amount to a probabilistic charity and donating zero to a sure-thing charity at Main Choice, but since they were not faced with a sure-thing charity at Final Choice they did not have the chance to make choices to reveal that indifference.

This led us to re-run the regressions focusing on different exclusion criteria. We ran the regressions on the subsample of subjects who faced a probabilistic charity at Final Choice, but among those excluded the ones who had donated a positive amount to a sure-thing charity. The excluded subjects made decisions consistent with the indifference concern you stated (i.e., donated a strictly positive amount to a sure-thing charity but made no donation when faced with a probabilistic charity). The non-excluded subjects made decisions that are not consistent with that possibility. The regression estimates broadly replicate findings of the regressions in the main body of the text for this selected sample of subjects. The results for this regression are below (but not in the paper) as the results are the same as in Appendix F.

Table X — Review Response Letter Regression Results for Charitable Giving Behaviour in Main Choice

Predicting Choice between Sure-Thing and Probabilistic Charities

(X)

Risk Attitude .496 (.980)

Ambiguity Aversion .008 (.015)

Numeracy -.002 (.015)

Empathy -.047 (.081)

Optimism

.004* (.007)

Donor Type

Warm-Glow -.014 (.008)

Pure Altruism .002 (.120)

Donation (amount) -.082 (.155)

Age -.001 (.002)

Gender .006 (.005)

Education

Undergraduate degree .066 (.124)

Postgraduate/Professional degree .215 (.129)

Religion

Protestantism .056** (.125)

Catholicism -.350 (.152)

Islam -.030 (.163)

Judaism -.115 (.339)

Buddhism -.257 (.538)

Hinduism .307 (.584)

Religious Participation .190 (.203)

Marriage Status .118 (.120)

Children -.067 (.133)

Financial Wellbeing .030 (.067)

Employment

Out of the workforce -.347 (.229)

Part-time employment .052 (.205)

Full-time employment -.139 (.198)

Adj. R2 .009

Sample size 93

Notes: OLS regression reporting unstandardised coefficients and standard errors. Outcome variable is charity choice (0 = sure-thing charity, 1 = probabilistic charity). *p<.1, **p<.05, ***p<.01, ****p<.001

We very much hope that these additional analyses and regressions make headway to address the concerns you raised.

Furthermore, as outlined in General Response, we will now present your main objection in significant detail in the discussion section such that readers are fully aware of this design limitation and its impact on the results. We also outline your criticism throughout the paper by pointing out repeatedly that Final Choice ought to be given more weight. See Example 2 and Example 3 below.

Example 1 (p. 48-49):

In Appendix Table 11, Model (17), we report the same OLS regression for Main Choice as in Model (1). The main change in this regression is that, in response to the worry that by excluding those that do not donate, we may be biasing our estimates as participants may be indifferent between donating a non-zero amount to one type of charity (say a sure-thing charity) and no donation to a probabilistic charity. As we had additional data on all participants from their choices in Final Choice, we excluded the following participants on top of our standard exclusion criteria. First, we excluded those who donated to a sure-thing charity in Main Choice but did not donate at all in Final Choice when they were shown a probabilistic charity (n=70). Second, we also excluded those who donated to a probabilistic charity in Main Choice but who did not donate at all in Final Choice when they were shown a sure-thing charity (n=5). Overall, we do not find a difference in results which suggests that that our data may go some way towards addressing this worry.

APPENDIX TABLE 11—REGRESSION RESULTS FOR CHARITABLE GIVING BEHAVIOUR IN MAIN CHOICE

PREDICTING CHOICE BETWEEN SURE-THING AND PROBABILISTIC CHARITIES

(17)

Risk Attitude .000 (.007)

Ambiguity Aversion -.004 (.006)

Numeracy -.005 (.030)

Empathy -.001 (.003)

Optimism

-.004 (.004)

Donor Type

Warm-Glow -.012 (.065)

Pure Altruism -.086 (.076)

Donation (amount) -.001 (.001)

Age .004 (.002)

Gender -.002 (.062)

Education

Undergraduate degree .068 (.064)

Postgraduate/Professional degree -.017 (.069)

Religion

Protestantism -.116 (.084)

Catholicism -.045 (.089)

Islam -.111 (.171)

Judaism -.198 (.412)

Buddhism -.151 (.318)

Hinduism .319 (.303)

Religious Participation .057 (.109)

Marriage Status .044 (.062)

Children .000 (.069)

Financial Wellbeing .007 (.030)

Employment

Out of the workforce -.213 (.122)

Part-time employment -.005 (.104)

Full-time employment -.099 (.096)

R2 .081

Sample size 232

Notes: OLS regression reporting unstandardised coefficients and standard errors. Outcome variable is charity choice (0 = sure-thing charity, 1 = probabilistic charity). *p<.1, **p<.05, ***p<.01, ****p<.001

Example 2 (p. 31-32): Moreover, there is an additional reason to significantly favour the results of Final Choice over those of Main Choice throughout the interpretation of this paper’s data. This is because the design of Main Choice has some inherent flaws that make interpretation of results difficult. This is because when participants make choices between two charities, they can make choices between compositions that they are indifferent between. For example, any one donor may be indifferent between a donation of £0.25 to a sure-thing charity and a donation of £0.10 to a probabilistic charity. In this scenario, the donor would choose randomly between the two. This in itself makes it difficult to cleanly identify a preference for one charity type or the other in the design that we have pre-registered and analysed in this paper. Based on this reason, we argue that the results from Final Choice should be given greater weight compared to Main Choice throughout this paper.

This concern is even more pointed if one considers that it may be the case that donors could be indifferent between donating a positive amount to a sure-thing charity and no donation to a probabilistic charity. As before, if they are indifferent between the two and choose randomly, this inhibits straightforward interpretation of our results. This is because we exclude all participants who choose not to donate (to understand their behaviour between these two options). Further, because it is quite plausible that, given the distribution of donation choices (most people who donate donate to a sure-thing charity), we may be excluding significantly more people favouring probabilistic charities but not sure-thing charities from those who are indifferent between a non-zero donation to a sure-thing charity and a zero donation to a probabilistic (and thus choose randomly between the two).

Some of these concerns have been addressed ex-post in this paper, for example by controlling for amount donated in the analyses of Main Choice that we did not initially pre-register. While this goes some way towards addressing this concern, we argue that some fundamental design constraints of the set-up that we chose remain. As such, we have put a higher emphasis on our analyses of Final Choice (which do not fall prey to the same structural challenges) throughout this paper and argue that one ought to be generally cautious in interpreting the results of Main Choice. However, given that it was pre-registered, we continue to report it fully (and where we deviate from the pre-registered protocol, we document this in detail and provide the original analyses in the appendix). We hope that this discussion, highlighting these issues in detail, properly contexualises the results for readers.

Example 3 (p. 15): This condition [Final Choice] controls for a number of potential confounds in Main Choice which allows for it to answer the paper’s central question more directly and cleanly, though it being the last task of the experiment, we cannot rule out potential order effects.

Comment 2: Point two, suppose people are more likely to donate more under sure-thing charity, then when you excluded 0 donation as you preregistered, you exclude more people who chose probabilistic charity (you can test this with your data). Th reason is that there must be some people who are indifferent between $positive amount/sure thing charity and $0/probabilistic charity, and these people chose randomly between the two. In the observed results, you excluded those who chose probabilistic charity but not the other type.

Response 2: Thank you again for further clarifications on this. As outlined above, we now discuss in detail your objection in the discussion section, including the examples you have provided here to ensure we do not misrepresent your main concern, while also offering regressions that hopefully go some way towards addressing this concern. Further, in order to make sure we are not missing something in our exchanges, we re-read our paper line by line. In doing so, we have realised that we have not made clear enough in the main paper that participant choices in Main Choice are between donating a non-zero amount to either of the two charities or not donating at all. Participants cannot choose a charity and then enter ‘0’. We have now made this clear in the paper at multiple points and apologise for this oversight in our previous revision.

Example (p. 5, 9, 14, 29): In attempting the construction of an externally valid donation choice that allows us to capture the difference between sure-thing charities and probabilistic charities in actual charities that participants can donate money to, our main outcome variables of interest are (i) the choice between two real charities, one sure-thing charity and one probabilistic charity and (ii) the choice about one of those charities that has been randomly selected. Both of these charitable decision-making scenarios have strengths and weaknesses from an experimental design perspective, but we hope that they jointly allow us to better understand the role of individual differences in charitable decision-making scenarios like these. We outline the main weakness of (i) in the discussion section and argue that, overall, (ii) is a cleaner design.

Lastly, one may also be interested in donation behaviour not between these two types of charities, but rather just in the context where potential donors are presented with one such charity. This may reduce the chance of additional confounds (like worrying that the design that presents two charities is artificial in its dichotomous presentation; after all, most naturalistic decisions are not decisions between two distinct choices). It also is overall a cleaner design that brings with it less drawbacks regarding interpretation of results.

In both cases, we collect data on whether they donate at all, if they do, which charity they choose in Main Choice (participants can choose up to one charity), and how much they donate. In other words, participants can donate a non-zero amount to either charity or not donate at all (i.e. one cannot select a charity and choose to donate ‘0’).

We have some evidence in favour of this worry as at least some of the pre-registered factors predict directionally as expected in Final Choice, where only one charity was presented to participants. As such, it may be justified to put more interpretative emphasis on the results from Final Choice compared to Main Choice.

Comment 3: Point 3, because of the above issue, if you regress charity type preference on other variables, you have to control for the amount donated, as an ex-post control, because you did not control for the donate amount in the experiment. I’m not clear what endogeneity this may cause. But if there is any, I don’t believe the data from Main Choice can prove anything. There is no problem if you want to report the analyses as you preregistered. But a preregistration doesn’t mean your design and analysis (Main Choice) is not problematic.

Response 3: Thank you very much for clarifying your comment from last time. As a result of your explanation of your main objection above, we have now followed your recommendation in full by controlling for the amount donated as an ex-post control for all analyses of Main Choice.

Example (p. 20-22): First, we investigate general donation behaviour relating to Main Choice. The results presented in Table 3 speak to the central null hypotheses #1, #2, and #3. Model (1) reports the results for the Main Condition. The outcome variable is the type of charity conditional on a donation being made, with 0 being coded as the sure-thing charity and 1 as the probabilistic charity. The gender variable is coded 1 for female, all other categorical variables are coded 1 for the affirmative. As outlined above, the risk attitude measure is a discrete variable of the number of boxes opened, the ambiguity aversion is the result of the subtraction of the reservation prices. All other scales are the sum of the (re-reversed) individual items. As specified in our pre-registration, we report main regression results for binary outcomes using an OLS model and also for the corresponding logit model as a robustness check in Appendix B (Appendix Table 2) to check for sensitivity to functional form choice, where we find no impact of model choice. We also report a robustness check in Appendix B (Appendix Table 3) where we report a regression with random effects at the stimulus level, finding that our null result is also robust to this model choice.

Table 3—Regression Results for Charitable Giving Behaviour in Main Choice

Predicting Choice between Sure-Thing and Probabilistic Charities

(1)

Risk Attitude .000 (.005)

Ambiguity Aversion -.003 (.005)

Numeracy .004 (.024)

Empathy <.001 (.003)

Optimism

-.004 (.003)

Donor Type

Warm-Glow -.014 (.051)

Pure Altruism -.062 (.064)

Donation (amount) -.001 (.001)

Age .003 (.002)

Gender .013 (.050)

Education

Undergraduate degree .044 (.051)

Postgraduate/Professional degree -.020 (.055)

Religion

Protestantism -.082 (.067)

Catholicism .012 (.073)

Islam -.101 (.122)

Judaism -.163 (.379)

Buddhism -.105 (.287)

Hinduism .378 (.276)

Religious Participation .041 (.089)

Marriage Status .045 (.050)

Children -.007 (.054)

Financial Wellbeing .005 (.024)

Employment

Out of the workforce -.154 (.097)

Part-time employment .003 (.085)

Full-time employment -.054 (.079)

R2 .064

Sample size 307

Notes: OLS regression reporting unstandardised coefficients and standard errors. Outcome variable is charity choice (0 = sure-thing charity, 1 = probabilistic charity). *p<.1, **p<.05, ***p<.01, ****p<.001

We find that none of the main variables nor the demographic control variables predict charity choice in the Main Condition. We can straightforwardly conclude from Model (1) that we do not have evidence to reject the null hypotheses #1 and #2 as none of the independent variables meaningfully predict donor behaviour. However, note that Model (1) includes the control variable of amount donated that was not pre-registered but suggested by a helpful anonymous reviewer. For the pre-registered regression model without this control with no difference in results, please see Appendix A.

Attachment

Submitted filename: Respopnse to Reviewers.docx

Decision Letter 2

Junhuan Zhang

16 Aug 2022

PONE-D-21-35635R2Sure-Thing vs. Probabilistic Charitable Giving: Experimental Evidence On the Role of Individual Differences in Risky and Ambiguous Charitable Decision-MakingPLOS ONE

Dear Dr. Schoenegger,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please explain the limitations about this kind of study using web-based sample. 

Please submit your revised manuscript by Sep 30 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Junhuan Zhang, PhD

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #2: (No Response)

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #2: (No Response)

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: (No Response)

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #2: (No Response)

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #2: I thank the authors for their excellent second round revisions which I believe make the paper strong and publishable. It is smart to connect the data from the Main and Final treatments. It is now clear and important to point out the flaws in the Main treatment and putting more weight to the Final treatment. Good work!

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2022 Sep 22;17(9):e0273971. doi: 10.1371/journal.pone.0273971.r006

Author response to Decision Letter 2


18 Aug 2022

Dear Editor,

Below we outline our response to your last comment.

Comment: Please explain the limitations about this kind of study using web-based sample.

Response: We agree that we have not properly stated this limitation before and now do so explicitly in the revised manuscript.

Example (p. 33): A second potential limitation is the use of a web-based sample from Prolific. While we have tried to counterbalance this concern by relying on a representative sample (at least along the dimensions of age, sex, and ethnicity), there may still be some dimension along which our sample is not representative of the population as a whole. This, in turn, means that that there may be some external validity worries inherent in using this sample, which may impact generalisability of our results.

Attachment

Submitted filename: Response to Editor.docx

Decision Letter 3

Junhuan Zhang

19 Aug 2022

Sure-Thing vs. Probabilistic Charitable Giving: Experimental Evidence On the Role of Individual Differences in Risky and Ambiguous Charitable Decision-Making

PONE-D-21-35635R3

Dear Dr. Schoenegger,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Junhuan Zhang, PhD

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Junhuan Zhang

4 Sep 2022

PONE-D-21-35635R3

Sure-Thing vs. Probabilistic Charitable Giving: Experimental Evidence on the Role of Individual Differences in Risky and Ambiguous Charitable Decision-Making

Dear Dr. Schoenegger:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Junhuan Zhang

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix. Preregistered model specification for model (1).

    (PDF)

    S2 Appendix. Robustness checks for main choice.

    (PDF)

    S3 Appendix. Robustness checks for final choice.

    (PDF)

    S4 Appendix. Additional hypotheses tests with inconclusive results.

    (PDF)

    S5 Appendix. Additional equivalence tests.

    (PDF)

    S6 Appendix. Additional regression for main choice.

    (PDF)

    S7 Appendix. Experimental texts.

    (PDF)

    S1 File

    (PDF)

    S2 File

    (DOCX)

    S1 Data

    (CSV)

    S2 Data

    (CSV)

    Attachment

    Submitted filename: Response to Reviewers.docx

    Attachment

    Submitted filename: Respopnse to Reviewers.docx

    Attachment

    Submitted filename: Response to Editor.docx

    Data Availability Statement

    The data are stored on the OSF repository and freely available here: https://osf.io/w9gfu/.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES