Abstract
Underlying preferences are often considered to be persistent, and are important inputs into economic models. We first conduct an extensive review of the disparate literature studying the stability of preferences measured in experiments. Then, we test the stability of individuals’ choices in panel data from rural Paraguay over almost a decade. Answers to social preference survey questions are quite stable. Experimental measures of risk, time, and social preferences do not exhibit much stability. Correlations between experimental measures of risk aversion are a more precisely estimated zero, whereas correlations for time and social preferences are larger and noisier. We also find no systematic evidence that real world shocks influence play in games. We suggest that in a developing country context researchers should explore designing simpler experiments and including survey questions in addition to experiments to measure preferences.
1 Introduction
Time, risk, and social preference parameters are crucial inputs into many economic models. They have important impacts on outcomes in models of technology adoption, migration, savings, and risk-sharing among others. Over the past decades, experimental economists have worked on perfecting methods for measuring these parameters. Preferences are assumed constant by theory, so the experiment and survey-based constructs which purport to measure them should also remain constant. If preferences do vary over time, theory would suggest that this is due to some shock in the environment the individuals are facing (e.g., a monetary windfall or a recently experienced theft).
We study the stability over time of risk and time preferences as measured by experiments, and of social preferences as measured by both survey questions and experiments. We also study whether these measures are affected by real world shocks or experiences in experiments in previous years. Because the literature on this topic is spread across many journals in many different disciplines including economics, psychology, management, and marketing, our first contribution is an extensive cross-disciplinary review of the literature regarding the stability of experimentally-measured preferences over time. Most of the existing literature is focused on developed countries and more educated populations, though that emphasis has been changing more recently. We compile the correlations of preferences over time found in each paper, which can serve as a baseline to compare the correlations found in our own data and those found in future studies.
Next, we make use of a unique dataset which follows households in rural Paraguay, running surveys and experiments with them in 2002, 2007, 2009, and 2010. This data contains experimental measures of social, risk, and time preferences as well as survey measures of social preferences. Our contribution compared to the previous work is that we look at all three types of preferences,1 over relatively long periods of time, and with a relatively large and diverse population in a developing country.
We should note that we do not observe preferences themselves, we observe choices made in experiments and answers given to survey questions. These choices will typically depend on both underlying preferences and the environment. If there are aggregate shocks (e.g., a recession) this should affect the average level of preferences, but an individual’s relative position in the distribution should not change. Our results account for the environment by controlling for idiosyncratic characteristics of the environment (e.g., household-level income and village fixed effects) when measuring the stability of choices.
We find that survey measures of social preferences are extremely stable over long periods of time in our data. None of the experimental measures approaches such a level of stability. Experimental measures of risk preferences are not stable over time. We find weak evidence of stability in experimental measures of time and social preferences, but the data seems too noisy to estimate this relationship with much precision.
One potential explanation for the fact that experimental measures of preferences are not stable over time in our data might be that households are facing shocks which cause their decisions or preferences to change. For rural Paraguayans who do not have access to crop or health insurance, such shocks may take a large toll on the household’s financial situation. We test whether income, health, and theft shocks are correlated with play in games or answers to survey questions. We find no evidence that these shocks have any impact on the choices people make in experiments or their answers to the survey preference questions. The only somewhat consistent effect is that households which experienced health shocks give less in experiments. We do find some suggestive evidence that experiences in earlier experiments impact play in future games; e.g., that being linked with a more generous partner in the past makes a player more generous himself in future games and more likely to punish ungenerous players. But, it seems that these effects have more to do with learning about the game rather than actual changes in preferences.
In sum, in our dataset survey measures of social preferences are quite consistent over time; on the other hand, experimental measures of social and time preferences are only weakly correlated over time. Experimental measures of risk preferences are not at all stable. Neither income, theft, nor health shocks impact answers to survey questions or choices made in experiments.
The greater stability of social preferences measured by surveys may be due to many reasons (also see Section 4). One reason is that many of the theoretically elegant experiments designed by experimental economists may be better suited for student populations found in university labs who tend to be highly educated and non-poor.
Recent research suggests that economic scarcity consumes attentional resources and interferes with cognitive function which may then lead to errors and biases in decision making (Shah et al., 2012). Schofield (2014) shows that nutritional scarcity also interferes with cognitive function. Benjamin et al. (2013) and Burks et al. (2009) give evidence that individuals with lower cognitive ability exhibit more behavioral biases in risk and time experiments and play less consistently in risk experiments. The scarcity experienced by poor individuals in rural areas of developing countries may lead the decisions they make in experiments to contain more errors than decisions made by individuals who do not experience such need. This may then explain why the decisions they make in experiments are not correlated over time.
In addition to the effect of poverty on cognitive function and decision making, the low levels of education in our rural Paraguayan sample may imply that the experiments are significantly harder for them to understand than are the survey questions. In our data the median level of education is only five years and the maximum is twelve. Using a representative sample in the Netherlands, Choi et al. (2014) find that less educated respondents play less consistently in risk and time preference games. Similarly, Charness and Viceisza (2015) compare rates of player inconsistency between experiments run in the developed and developing world, and find that a greater share of players in the developing world make inconsistent choices.
Using a sample of Canadians, Dave et al. (2010) find that simpler experiments measuring risk aversion work better among less numerate subjects, as more complex experiments lead to much noise in decision-making. Charness and Viceisza (2015) measure the risk preferences of a population of rural Senegalese individuals and find suggestive evidence that more complex tasks lead to lower levels of understanding and more noisy responses. Cook (2015) finds that data collected in risk aversion experiments in low income countries contains suggestive evidence that players are confused.
In terms of comparing survey measures to experimental measures, Dohmen et al. (2011) and Lönnqvist et al. (2014) find that survey measures of risk aversion perform better than experimental measures and we find the same for social preferences. On the other hand, Burks et al. (2012) find that a survey measure of time preference perform worse than experimental measures. Although economists have tended to dismiss survey measures of preferences, they may want to reconsider the usefulness of such measures.
This discussion highlights the difficulties involved in conducting economic experiments in developing countries among populations with high levels of poverty and low levels of education. More research on how to best design experiments for developing country populations, as well as on when survey questions might be more informative than experiments is still needed.
2 Previous Literature
While Loewenstein and Angner (2003) ponder whether underlying preferences are stable over time from a philosophical and theoretical perspective, we empirically examine whether preferences measured from surveys and experiments are stable. Below we review three related strands of the literature. The empirical literature on preference stability over time is most directly related and we review those papers first and most thoroughly. Our paper can also be related to the growing number of studies looking at how shocks such as illness, income shocks, civil wars, and natural disasters lead to changes in measures of preferences.2 Finally, we briefly review the literature on the stability of preferences as measured in different games but played on the same day, which is relevant since in many cases we do not play exactly the same game in each year.
2.1 Stability over Time of Preferences Measured in Experiments
First we review papers studying the stability of measures of risk, time, and social preferences over time. Compared with our data, much of this literature either uses a smaller sample, a shorter period of time, or both. These papers only focus on one type of preference with the large majority studying only risk preferences. More recently the number of papers looking at time preferences has been growing, while there are still very few that look at the stability of social preferences. Very few of these papers discuss attrition, although the sample sizes reported often suggest that many fewer individuals participate in later rounds compared to earlier rounds of the experiments.3 In addition many focus on student populations, which leads to worries about selection and external validity. The majority use better-educated samples from developed countries, although this has been changing more recently.
Most of the papers which report correlations find that measured preferences are significantly correlated over time. This may potentially be due to the file drawer effect - many individuals run similar experiments with the same population at two points in time but if the correlations are not significant then the results may not be published anywhere and are instead filed away in a drawer. There are two components to the file drawer effect: the fact that papers finding no correlation are more likely to be rejected; and the fact that researchers who do not find a significant correlation are less likely to write up the results.
Risk Preferences
The results regarding stability of risk preferences are summarized in Table 1. The evidence suggests that risk preferences are relatively stable over time, with reported correlations ranging from a low of −.38 to a high of 0.68. If one excludes the studies with fewer than 100 observations, the range is 0.13 to 0.55.4,5 Interestingly, there does not appear to be much systematic difference in the correlations reported by studies which measure risk preferences over shorter versus longer periods of time. Likewise, there doesn’t appear to be much difference in the the stability of hypothetical versus incentivized measures, or from student versus non-student populations.
Table 1.
Paper | Population | Time | Corr | Sig | Inc |
---|---|---|---|---|---|
Menkhoff and Sakha (2014) | 384 rural Thai | 5 years | ? | yes | inc |
Levin et al. (2007) | 124 US children/parents | 3 years | 0.20 – 0.38 | yes | inc |
Guiso et al. (2011) | 666 Italian investors | 2 years | 0.131 | yes | hyp |
Kimball et al. (2008) | 700 older Americans | 2 years | 0.27 | ? | hyp |
Love and Robison (1984) | 23 US farmers | 2 years | −0.38 – 0.231 | no2 | hyp |
Sahm (2012) | 12000 older Americans | multiple years | 0.18 | yes | hyp |
Beauchamp et al. (2012) | 489 Swedish twins | 1 year | 0.483 | yes | hyp |
Goldstein et al. (2008) | 75 Americans | 1 year | 0.43 | yes | hyp |
Lönnqvist et al. (2014) | 43 German students | 1 year | 0.21 | no | inc |
Smidts (1997) | 205 Dutch farmers | 1 year | 0.44 | yes | hyp |
Wehrung et al. (1984) | 84 N. American businessmen | 1 year | 0.36 | yes | hyp |
Andersen et al. (2008) | 97 Danes | 3– 17 months | ? | yes | inc |
Harrison et al. (2005) | 31 US students | 6 months | ? | yes | inc |
Vlaev et al. (2009) | 69 British students/adults | 3 months | 0.20–0.634 | yes | hyp |
Horowitz (1992) | 66 US students & 23 PTA | 2 months | ? | no | inc |
Wölbert and Riedl (2013) | 53 Dutch students | 5–10 weeks | 0.36–0.68 | yes | inc |
Schoemaker and Hershey (1992) | 109 US MBA students | 3 weeks | 0.55 | yes | hyp |
Hey (2001) | 53 British students | a few days | ? | yes | inc |
Corr - the correlation of risk preferences over time. Sig - whether risk preferences are significantly related over time. Inc - whether the experiment was incentivized rather than hypothetical.
Our own calculation, from the raw data reported in the original paper.
In fact, the negative correlation of −0.38 is significant at the 10% level.
This is a polychoric correlation, which may be larger because it suffers from less attenuation bias.
Interestingly, the one correlation which was not statistically significantly different than 0 (value of 0.20)was for the risk question which used the Multiple Price List (MPL) mechanism over gains. This is arguably the most common method for collecting risk preferences.
Time Preferences
The results regarding stability of experimentally measured time preferences can be found in Table 2. Of the papers looking at the stability of time preferences, the reported correlation coefficients range from 0.004 to 0.75. Most numbers seem to be slightly higher than the range of risk preference correlation coefficients mentioned above, although there are fewer observations. If one excludes the studies with less than 100 observations four studies remain, with correlations (for the comparisons with over 100 observations) ranging from 0.09 to 0.68. This is similar to the range of the correlations in the risk aversion studies.6
Table 2.
Paper | Population | Time | Corr | Sig | Inc |
---|---|---|---|---|---|
Meier and Sprenger (2015) | 250 US low-income | 2 years | 0.40 | yes | inc |
Krupka and Stephens Jr (2013) | 1194 Americans | 1 year | ? | ? | hyp |
Harrison et al. (2006) | 97 Danes | 3 – 17 months | ? | yes | inc |
Kirby et al. (2002) | 95–123 Bolivian Amerindians | 3 –12 months | 0.004–0.46 | yes | inc |
Kirby (2009) | 46–81 US students | 1 – 12 months | 0.57–0.75 | yes | inc |
Li et al. (2013) | 336–516 Americans | 1 week – 14 months | 0.33–0.68 | yes | hyp |
Wölbert and Riedl (2013) | 53 Dutch students | 5–10 weeks | 0.61–0.68 | yes | inc |
Dean and Sautmann (2014) | 961 peri-urban Malians | 1 week | 0.61–0.67 | yes | inc |
Corr - the correlation of time preferences over time. Sig - whether time preferences are significantly related over time. Inc - whether the experiment was incentivized rather than hypothetical.
Social Preferences
The results regarding stability of experimentally-measured social preferences can be found in Table 3. In this case, correlation coefficients range from −0.15 to 0.69, similar to the range for time and risk preferences. If one excludes the studies with less than 100 observations then only one study remains, with a range of 0.12 to 0.28.7
Table 3.
Paper | Population | Time | Corr | Sig | Inc | game / preference |
---|---|---|---|---|---|---|
Carlsson et al. (2014) | 196 Vietnamese | 6 years | 0.12 – 0.28 | yes | inc | public good |
Lönnqvist et al. (2014) | 22 German students | 1 year | 0.69 | yes | inc | trust |
Brosig et al. (2007) | 40 German students | 3 months | 0.09–0.481 | no/yes | inc | altruism |
Brosig et al. (2007) | 40 German students | 3 months | −0.15–0.561 | no/yes | inc | sequential PD |
Corr - the correlation of social preferences over time. Sig - whether social preferences are significantly related over time. Inc - whether the experiment was incentivized rather than hypothetical.
Our own calculation, from the raw data reported in the original paper.
Scanning all three tables, no obvious patterns appear. The results are similar for games played at longer and shorter intervals, for incentivized and hypothetical games, and for games played with student and non-student populations. While we see that there are many papers which study the stability of measures of preferences over time, our contribution is that we look at both experimental and survey measures of all three (risk, time, and social preferences) in a relatively large dataset over relatively long periods of time with a diverse population in a developing country, and that we also have data on other real-world outcomes for these same individuals.
2.2 Impact of Events on Preferences: Economic Shocks, Natural Disasters, and Conflict
Research studying how shocks affect preferences usually starts from the underlying implicit assumption that, in the absence of the shock, preferences would have changed less. Originally researchers were most interested in studying how job market shocks cause changes in preferences. Many papers find that changes in income, unemployment, health status, and family composition do not lead to changes in risk preferences (Brunnermeier and Nagel, 2008; Chiappori and Paiella, 2011; Sahm, 2012) or time preferences (Harrison et al., 2006; Meier and Sprenger, 2015).8
Specific examples of papers which find no evidence of changes in preferences after varying shocks in low income settings include Giné et al. (2014) who find no impact of household shocks such as a death in the family or shortfalls in expected income on time preferences in rural Malawi. Carvalho et al. (2014) look at low-income Americans and find that individuals are more present-biased regarding money before payday than after, but they hypothesize that this is most likely due to differences in liquidity constraints rather than differences in preferences since they behave similarly for inter-temporal decisions regarding non-monetary rewards and play no differently in risk experiments. Meier and Sprenger (2015) suggest that changes in measured preferences over time thus may either be noise, or may be orthogonal to socio-demographics.
Another group of papers disagrees with this assessment and does find that preferences are affected by shocks. Fisman et al. (2014) find that the Great Recession increased selfishness while Gerrans et al. (2015) and Necker and Ziegelmeyer (2015) find that it decreased risk tolerance. Krupka and Stephens Jr (2013) show that economic shocks are correlated with changes in time preferences.9
In a developing country context Dean and Sautmann (2014) find, using weekly data, that individuals become more patient when they have positive income or savings shocks, and more impatient when they face negative shocks. Menkhoff and Sakha (2014) find that agricultural shocks (drought and flood) and economic shocks (an increase in the price of inputs or the collapse of a business) cause rural Thai respondents to become more risk averse. Finally, using a randomized controlled trial, Carvalho et al. (2014) find that Nepali women who are randomly offered savings accounts are more risk-taking and more patient than those who are not. Kandasamy et al. (2014) provide a psychological mechanism to explain the effects of stress on risk aversion: they find that randomly raising the stress hormone cortisol increases risk aversion.
Education has also sometimes been found to have an impact on preferences. Jakiela et al. (2015) find that higher educational achievement among female primary school students in Kenya increased their generosity. Lührmann et al. (2015) find that randomly providing high school students with financial education decreases present bias. Booth et al. (2014) find that randomly assigning female college students to single-sex discussion sections causes them to make more risky decisions in games. On the other hand, Dasgupta et al. (2014) find that a randomized vocational training program among women residing in Indian slums does not effect risk preferences.
There is a new burgeoning literature looking at how preferences are impacted by extreme events such as civil wars or natural disasters. While this literature is fascinating, authors tend to face two main difficulties. First, data on preferences is usually only available after the event and not before. Second, it is difficult to construct a control group, since these events affect different populations differentially. Papers looking at the impact of natural disasters and civil wars find amazingly divergent results. The number of these papers has been growing at a rapid clip, with no signs of converging to one consistent set of results. This lack of consistency may suggest that there are nuances involved in the experience of disaster that we are currently not considering, or that experimental choices are filled with much noise. It would be helpful for future researchers to attempt to reconcile these results.
The research on natural disasters (including earthquakes, famines, floods, hurricanes, and tsunamis) suggests that such shocks increase risk aversion (Cameron and Shah, 2015; Cassar et al., 2011; Chantarat et al., 2015; Samphantharak and Chantarat, 2015; van den Berg et al., 2009), decrease risk aversion (Bchir and Willinger, 2013; Eckel et al., 2009; Hanaoka et al., 2014; Ingwersen, 2014; Page et al., 2014; Willinger et al., 2013), have no effect at all on risk preferences (Becchetti et al., 2012), or have no consistent effect on risk preferences (Said et al., 2015); increase impatience (Bchir and Willinger, 2013; Cassar et al., 2011; Sawada and Kuroishi, 2015), decrease impatience (Callen, 2011; Chantarat et al., 2015), or have no consistent effect on time preferences (Willinger et al., 2013); increase trust (Cassar et al., 2011), decrease trust (Chantarat et al., 2015), or have no effect on the level of trust (Andrabi and Das, 2010); decrease trustworthiness (Fleming et al., 2014); and increase altruism (Becchetti et al., 2012; Chantarat et al., 2015), decrease altruism (Samphantharak and Chantarat, 2015), or have no consistent effect on altruism (Afzal et al., 2015).10
The research on conflict (including civil wars and political violence) likewise shows contradictory results. Findings suggest that conflict may decrease risk aversion (Voors et al., 2012) or increase risk aversion (Callen et al., 2014; Kim and Lee, 2014; Moya, 2011); decrease patience (Voors et al., 2012); lower trust (Cassar et al., 2013; De Luca and Verpoorten, 2015) or have no effect on trust (Grosjean, 2014); increase initial trustworthiness but lower subsequent trustworthiness (Becchetti et al., 2014); increase altruism (Voors et al., 2012) or decrease altruism to out-groups with no effect within group (Silva and Mace, 2014); and increase egalitarianism (Bauer et al., 2014).
Finally, there is a bit of evidence regarding how experiences in one experiment may impact play in later experiments and real-world decisions. Conte et al. (2014) and Matthey and Regner (2013) find that individuals who have participated in experiments in the past behave more selfishly in subsequent experiments. Mengel et al. (2015) show that subjects become more risk averse after playing games involving ambiguity rather than games involving pure risk only, and He and Hong (2014) show that subjects become more risk averse after playing games involving higher levels of risk. Cai and Song (2014) find that experiencing more disaster in an insurance game increases uptake of a real insurance product but does not influence risk aversion as measured in a subsequent game. Jamison et al. (2008) find that being deceived in one experiment leads participants to behave more inconsistently (exhibit more multiple-switching behavior) in future experiments measuring risk aversion.
2.3 Stability of Preferences Measured in Different Games
There are many papers which look at the stability of preferences measured in different games and we only mention a few here. This subsection of the literature review is not as comprehensive as the subsection reviewing the literature on stability measured over time. Most research shows that risk preferences are not stable across different settings or different games (Anderson and Mellor, 2009; Berg et al., 2005; Binswanger, 1980; Dulleck et al., 2015; Eckel and Wilson, 2004; Isaac and James, 2000; Kruse and Thompson, 2003; Vlaev et al., 2009). On the other hand, Choi et al. (2007) find that risk preferences are stable across games when the games in question are quite similar to one another. Reynaud and Couture (2012) find that risk preferences are stable both across games and within the same game at different levels of stakes, while Dulleck et al. (2015) using a similar design only find stability across levels of stakes for the same game but not across different games.
Time preferences have been found to be correlated across goods (Reuben et al., 2010; Ubfal, 2014) and across lengths of time (Halevy, 2015; McLeish and Oxoby, 2007). Social preferences have been found to be rather stable both using variants of the same game (Andreoni and Miller, 2002; Fisman et al., 2007) and using very different games (Ackert et al., 2011; de Oliveira et al., 2012), although Blanco et al. (2011, Table 3) find a low correlation across most games in their sample.
Although our review of previous research suggests that risk preferences may be less stable across different games than time and social preferences are, our review in this subsection of the literature on stability across games is not exhaustive nor does it take into account publication bias. One message that seems to be supported by the majority of papers is that preferences tend to be more stable when measured in similar games or games which are variants of one another. When preferences are measured in very different games, they are less likely to be consistent. This conclusion was reached long ago by Slovic (1972) with regards to experiments on risk taking and seems to more generally continue to be verified.
3 Datasets
We have survey and experimental data from 2002, 2007, 2009, and 2010 for different subsets of the same sample. As the original purpose of this data collection was not to look at the stability of preferences over time, both the samples and the experiments varied across rounds. Summary statistics of all the outcome variables analyzed here can be found in Table 4.
Table 4.
Mean | Sd | Min | Max | # Obs. | |
---|---|---|---|---|---|
bet in 2002 | 3.43 | 2.05 | 0 | 8 | 189 |
roll of die in risk game in 2002 | 3.37 | 1.69 | 1 | 6 | 172 |
# risky choices in 2007 (hyp) | 2.09 | 1.77 | 0 | 5 | 449 |
# risky choices in 2009 (hyp) | 1.55 | 1.53 | 0 | 5 | 176 |
time preference in 2007 (hyp) | 199.35 | 561.22 | 75 | 5000 | 449 |
time preference in 2009 (hyp) | 111.99 | 116.61 | 75 | 1000 | 176 |
| |||||
sent as trustor in 2002 | 3.74 | 1.99 | 0 | 8 | 188 |
sent as dictator in anonymous game in 2007 | 5.08 | 2.69 | 0 | 14 | 371 |
sent as dictator in chosen non-revealed game in 2007 | 5.39 | 2.68 | 0 | 14 | 371 |
sent as dictator in revealed game in 2007 | 5.47 | 2.69 | 0 | 14 | 371 |
sent as dictator in chosen revealed game in 2007 | 5.93 | 2.84 | 0 | 14 | 371 |
sent as dictator in anonymous game in 2009 | 2.51 | 1.94 | 0 | 10 | 176 |
sent as dictator in chosen non-revealed game in 2009 | 2.72 | 2.02 | 0 | 13 | 129 |
sent as dictator in revealed game in 2009 | 3.01 | 2.18 | 0 | 12 | 176 |
sent as dictator in chosen revealed game in 2009 | 3.69 | 2.71 | 0 | 14 | 129 |
| |||||
share returned as trustee in 2002 | 0.44 | 0.20 | 0 | 1 | 188 |
amount received as trustee in 2002 | 11.23 | 5.97 | 0 | 24 | 188 |
proportion received back as trustor in 2002 | 1.30 | 0.61 | 0 | 3 | 175 |
| |||||
trust people in the world in 2002 (survey) | 2.40 | 0.64 | 1 | 5 | 214 |
trust people in the world in 2007 (survey) | 2.58 | 0.92 | 1 | 5 | 449 |
trust people in the world in 2009 (survey) | 2.84 | 1.15 | 1 | 5 | 176 |
| |||||
trust people in the village in 2002 (survey) | 3.14 | 1.13 | 1 | 5 | 214 |
trust people in the village in 2007 (survey) | 3.22 | 1.09 | 1 | 5 | 449 |
trust people in the village in 2009 (survey) | 3.40 | 1.22 | 1 | 5 | 176 |
trust people in the village in 2010 (survey) | 3.40 | 1.11 | 1 | 5 | 119 |
| |||||
trust closest neighbors in 2002 (survey) | 3.96 | 1.27 | 1 | 5 | 214 |
trust closest neighbors in 2007 (survey) | 3.97 | 1.30 | 1 | 5 | 449 |
trust closest neighbors in 2009 (survey) | 3.96 | 1.34 | 1 | 5 | 176 |
| |||||
would villagemates take advantage if had opportunity in 2002 (survey) | 2.90 | 1.15 | 1 | 5 | 214 |
would villagemates take advantage if had opportunity in 2007 (survey) | 2.44 | 1.09 | 1 | 5 | 449 |
would villagemates take advantage if had opportunity in 2009 (survey) | 2.56 | 1.14 | 1 | 5 | 176 |
| |||||
bad to buy something you know is stolen in 2002 (survey) | 0.90 | 0.30 | 0 | 1 | 214 |
bad to buy something you know is stolen in 2007 (survey) | 0.87 | 0.34 | 0 | 1 | 449 |
bad to buy something you know is stolen in 2009 (survey) | 0.85 | 0.36 | 0 | 1 | 176 |
| |||||
negative reciprocity in 2007 (survey) | 0.25 | 0.43 | 0 | 1 | 449 |
negative reciprocity in 2009 (survey) | 0.15 | 0.36 | 0 | 1 | 176 |
negative reciprocity in 2010 (survey) | 0.16 | 0.37 | 0 | 1 | 119 |
| |||||
positive reciprocity in 2010 | 0.71 | 0.46 | 0 | 1 | 119 |
negative reciprocity in 2010 | 0.18 | 0.39 | 0 | 1 | 119 |
Note: Hyp denotes that the game was hypothetical. All data is from experiments unless otherwise stated.
From looking at the summary statistics table, we note something which may hint at the results we will find below. The means of the survey questions (for example, questions such as what share of people in the world do you trust) appear to be much more stable over time than the experimental data (for example, the amounts sent in the dictator games). On the other hand, we do see a similar pattern in both the 2007 and 2009 dictator data, where the least is sent in the anonymous game; the most is sent in the chosen-revealed game; and the amounts sent in the revealed and chosen non-revealed games lie somewhere in between. A discussion of and explanation for this pattern in the 2007 data can be found in Ligon and Schechter (2012).
3.1 Sample Selection
In 1991, the Land Tenure Center at the University of Wisconsin in Madison and the Centro Paraguayo de Estudios Sociológicos in Asunción worked together in the design and implementation of a survey of 285 rural Paraguayan households in fifteen randomly chosen villages in three departments (comparable to states) across the country. The households were stratified by land-holdings and chosen randomly.11 The original survey was followed by subsequent rounds of survey data collection in 1994 and 1999. Subsequent rounds including both survey and experimental data were conducted in 2002, 2007, 2009, and 2010.
In 2002 the survey began to include questions measuring trust and economic experiments were conducted. By that point it was only possible to interview 214 of the original 285 households. Of the 214 households interviewed, 188 also sent a household member to participate in the economic experiments.
In 2007, new households were added to the survey in an effort to interview 30 households in each of the fifteen villages. (This meant adding between 6 and 24 new households in any village in addition to the original households.) Villages range in size from around 30 to 600 households. In total 449 households were interviewed, of which 371 sent a household member to participate in the economic experiments. We interviewed 195 of the previous 214 households and added 254 new households.
In 2009, we returned only to the two smallest villages. Of the 59 households interviewed in these two villages in 2007, we interviewed someone from 52 of those households in 2009. In this round, the experiments were conducted as part of the survey and so all surveyed households also participated in the experiments.
Finally, in 2010 we returned to the ten villages in the two more easily accessible departments, excluding the five villages in the more distant department. These ten villages included the two villages interviewed in 2009. In this round, we played a reciprocity experiment with 119 of the 299 individuals interviewed in 2007 in those ten villages. The households who participated were those chosen by political middlemen.
The numbers cited in the previous paragraphs refer to the number of households followed over time. As preferences are individual characteristics, rather than household characteristics, we also care about whether the same individual responded to the survey or participated in the experiments. In 2002 we told individuals that we preferred to run the surveys and games with the household heads but allowed other household members to participate if the head was not available. For the 2007 survey we tried to follow up with the 2002 game player if possible. The 2007 games were with the household head if possible. In 2009 we interviewed all adults in each household. In 2010 we followed up with the 2007 survey respondent. For more details on the 2002 dataset see Schechter (2007a), for the 2007 dataset see Ligon and Schechter (2012), and for the 2010 dataset see Finan and Schechter (2012).
In sum, while 214 households were interviewed in 2002, and the sample was expanded to cover 449 households in 2007, the number of observations we can compare across rounds is smaller. Table 5 presents the number of observations in each survey and the number of observations for the same person across rounds.
Table 5.
02 survey | 02 game | 07 survey | 07 game | 09 participate | 10 participate | |
---|---|---|---|---|---|---|
02 survey | 214 | 142 | 123 | 79 | 21 | 39 |
02 game | 188 | 139 | 103 | 17 | 43 | |
07 survey | 449 | 282 | 49 | 119 | ||
07 game | 371 | 41 | 81 | |||
09 participate | 176 | 23 | ||||
10 participate | 119 |
Note: Each cell contains the number of individuals participating in both the row and the column data.
The individuals we follow over time may be a rather select sample. They are still alive and haven’t migrated away. They also include those households in which we were consistently able to contact the same individual in each round. One could conduct a selection analysis if one had access to a variable which affected the likelihood of staying in our sample, but did not affect social, time, or risk preferences. Though we can think of no such variable, we believe that our sample consists of precisely those people whose preferences are most likely to be consistent over time. They are the most stable individuals: people who continue to live in the same village and be available to be interviewed over a period of many years. So, we believe that our estimates of the stability of preferences over time should overstate this stability. Given that we find that experimental measures of preferences are not very stable in our sample, one might expect that a sample without such attrition problems would exhibit even less stability.12
We test whether the people who remain in our survey are different from those who attrit. Table 6 shows average characteristics (as measured in 2002) for two mutually exclusive subsets of the individuals who participated in 2002: those how remained in later rounds and those who attrited. Table 7 shows average characteristics (as measured in 2007) for the individuals who participated in 2007 and did or did not attrit in later rounds. To distinguish how much selection there is at the village versus the individual level (since not all villages were visited in each survey round) we show unconditional tests for significant differences in characteristics and also show such tests conditional on village fixed effects.
Table 6.
2002 survey and 2007 survey | 2002 survey and 2010 survey | 2002 game and 2007 game | 2002 game and 2010 game | |||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
in both (1) |
attrit (2) |
diff1 (3) |
diff2 (4) |
in both (5) |
attrit (6) |
diff1 (7) |
diff2 (8) |
in both (9) |
attrit (10) |
diff1 (11) |
diff2 (12) |
in both (13) |
attrit (14) |
diff1 (15) |
diff2 (16) |
|
hhd income | 19,611 (47,977) | 18,830 (50,209) | 780 (6,811) [0.880] | 1,724 (6,065) [0.768] | 10,323 (8,840) | 21,275 (53,656) | −10,952** (4,299) [0.091]* | 787 (1,499) [0.590] | 13,821 (18,475) | 23,167 (69,961) | −9,346 (7,800) [0.102] | −8607 (7467) [0.113] | 10,348 (8,396) | 20,330 (55,512) | −9,981** (4,791) [0.229] | 1,472 (1,439) [0.293] |
log(hhd inc) | 9.25 (0.97) | 9.07 (0.98) | 0.18 (0.14) [0.162] | 0.24* (0.12) [0.070]* | 8.96 (0.80) | 9.22 (1.01) | −0.26* (0.15) [0.155] | 0.05 (0.15) [0.695] | 9.09 (0.85) | 9.14 (1.06) | −0.05 (0.14) [0.591] | −0.03 (0.13) [0.771] | 8.96 (0.77) | 9.16 (0.99) | −0.21 (0.14) [0.328] | 0.10 (0.15) [0.403] |
hhd size | 5.72 (2.51) | 5.48 (2.31) | 0.23 (0.33) [0.466] | 0.34 (0.31) [0.233] | 5.59 (2.09) | 5.62 (2.49) | −0.03 (0.38) [0.951] | 0.59 (0.40) [0.313] | 5.67 (2.51) | 5.48 (2.44) | 0.19 (0.36) [0.580] | 0.23 (0.35) [0.513] | 5.53 (2.13) | 5.60 (2.57) | −0.07 (0.39) [0.870] | 0.39 (0.41) [0.392] |
male | 0.76 (0.43) | 0.77 (0.42) | −0.01 (0.06) [0.934] | −0.01 (0.06) [0.849] | 0.79 (0.41) | 0.76 (0.43) | 0.03 (0.07) [0.665] | 0.13 (0.08) [0.194] | 0.73 (0.45) | 0.65 (0.48) | 0.08 (0.07) [0.276] | 0.10 (0.07) [0.181] | 0.74 (0.44) | 0.68 (0.47) | 0.07 (0.08) [0.381] | 0.22** (0.09) [0.030]** |
age | 52.89 (13.25) | 50.30 (16.60) | 2.59 (2.11) [0.262] | 2.39 (2.08) [0.375] | 53.85 (12.75) | 51.33 (15.20) | 2.52 (2.33) [0.337] | 0.61 (2.39) [0.799] | 50.74 (15.15) | 46.21 (20.05) | 4.53* (2.64) [0.201] | 4.03 (2.91) [0.198] | 52.56 (14.62) | 47.54 (18.32) | 5.01* (2.69) [0.114] | 4.09 (3.20) [0.229] |
ed | 4.54 (2.38) | 4.98 (2.38) | −0.43 (0.36) [0.236] | −0.36 (0.35) [0.323] | 4.38 (1.99) | 4.81 (2.66) | −0.42 (0.38) [0.197] | −0.51 (0.41) [0.186] | 4.55 (2.37) | 5.09 (2.78) | −0.54 (0.38) [0.275] | −0.45 (0.44) [0.392] | 4.26 (1.88) | 4.96 (2.73) | −0.70* (0.36) [0.058]* | −0.79* (0.43) [0.084]* |
bet | 3.57 (2.20) | 3.27 (1.85) | 0.30 (0.30) [0.423] | 0.28 (0.31) [0.453] | 3.77 (1.91) | 3.34 (2.08) | 0.43 (0.34) [0.183] | 0.81** (0.34) [0.026]** | ||||||||
sent as trustor | 3.73 (2.10) | 3.76 (1.86) | −0.04 (0.29) [0.900] | −0.19 (0.31) [0.533] | 3.95 (2.11) | 3.68 (1.95) | 0.27 (0.36) [0.588] | 0.70* (0.36) [0.289] | ||||||||
trust in village | 3.13 (1.15) | 3.15 (1.09) | −0.02 (0.15) [0.911] | −0.05 (0.16) [0.791] | 3.15 (1.09) | 3.14 (1.14) | 0.02 (0.19) [0.928] | −0.11 (0.22) [0.434] | 3.17 (1.16) | 3.19 (1.11) | −0.02 (0.17) [0.900] | −0.16 (0.18) [0.434] | 3.12 (1.10) | 3.19 (1.15) | −0.08 (0.19) [0.698] | −0.35* (0.21) [0.087]* |
# Obs. | 123 | 91 | 39 | 175 | 103 | 85 | 43 | 145 |
Note: Standard deviations are in parentheses for means while heteroskedasticity robust standard errors are in parentheses for the tests of differences. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: *** p<0.01,
p<0.05,
p<0.1.
Columns 1, 5, 9, and 13 show information for the non-attriters while Columns 2, 6, 10, and 14 show information for the attriters. Diff1 is coefficient in regression controlling only for participation at the later date. Diff2 is coefficient in regression aditionally controlling for village fixed effects. Income is in thousands of Guarani. All variables are measured in 2002.
Table 7.
2007 survey and 2009 survey | 2007 survey and 2010 survey | 2007 game and 2009 game | 2007 game and 2010 game | |||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
in both (1) |
attrit (2) |
diff1 (3) |
diff2 (4) |
in both (5) |
attrit (6) |
diff1 (7) |
diff2 (8) |
in both (9) |
attrit (10) |
diff1 (11) |
diff2 (12) |
in both (13) |
attrit (14) |
diff1 (15) |
diff2 (16) |
|
hhd income | 14,614 (11,564) | 24,545 (39,286) | −9,931*** (2,559) | −4,161 (3,946) | 17,510 (13,006) | 25,607 (42,735) | −8,097*** (2,638) [0.050]** | 243 (1,481) [0.904] | 14,743 (11,728) | 21,600 (31,981) | −6,858*** (2,529) | 1,634 (2,016) | 15,922 (8,752) | 22,216 (34,051) | −6,294*** (2,224) [0.036]** | 320 (1,290) [0.853] |
log(hhd inc) | 9.33 (0.75) | 9.70 (0.82) | −0.37*** (0.11) | −0.34 (0.21) | 9.56 (0.67) | 9.70 (0.86) | −0.14* (0.08) [0.279] | 0.03 (0.08) [0.758] | 9.35 (0.72) | 9.63 (0.78) | −0.28** (0.12) | −0.03 (0.16) | 9.50 (0.65) | 9.63 (0.81) | −0.12 (0.09) [0.259] | 0.06 (0.09) [0.628] |
hhd size | 4.88 (2.63) | 4.91 (2.36) | −0.03 (0.39) | −0.24 (0.66) | 4.98 (2.58) | 4.88 (2.32) | 0.10 (0.27) [0.750] | 0.29 (0.27) [0.355] | 4.78 (2.49) | 5.00 (2.37) | −0.22 (0.41) | −1.05 (0.93) | 4.75 (2.29) | 5.03 (2.41) | −0.28 (0.29) [0.197] | −0.08 (0.29) [0.818] |
male | 0.63 (0.49) | 0.68 (0.47) | −0.04 (0.07) | −0.04 (0.17) | 0.64 (0.48) | 0.68 (0.47) | −0.04 (0.05) [0.337] | 0.06 (0.06) [0.044]** | 0.49 (0.51) | 0.65 (0.50) | −0.16** (0.08) | −0.01 (0.20) | 0.65 (0.48) | 0.63 (0.48) | 0.03 (0.06) [0.612] | 0.13** ((0.07) [0.051]* |
age | 48.49 (17.56) | 50.15 (15.30) | −1.66 (2.60) | 7.71 (6.91) | 51.39 (15.62) | 49.46 (15.51) | 1.93 (1.67) [0.420] | 1.72 (1.88) [0.567] | 48.17 (18.80) | 47.95 (16.14) | 0.23 (3.04) | 16.93** (6.97) | 51.95 (15.09) | 46.86 (16.64) | 5.09*** (1.94) [0.035]** | 6.32*** (2.13) [0.022]** |
ed | 5.12 (2.50) | 5.05 (3.03) | 0.08 (0.39) | −1.94 (1.25) | 5.05 (2.63) | 5.05 (3.09) | −0.00 (0.29) [0.988] | −0.23 (0.34) [0.570] | 4.90 (2.36) | 5.21 (2.98) | −0.31 (0.40) | −1.91** (0.81) | 4.89 (2.10) | 5.26 (3.11) | −0.37 (0.30) [0.231] | −0.78** (0.36) [0.047]** |
risky choices | 1.71 (1.68) | 2.13 (1.78) | −0.42 (0.25) | −0.87 (0.56) | 2.04 (1.78) | 2.10 (1.77) | −0.06 (0.19) [0.756] | −0.21 (0.22) [0.349] | 1.90 (1.67) | 2.13 (1.79) | −0.23 (0.28) | −0.45 (0.67) | 2.15 (1.80) | 2.09 (1.77) | 0.06 (0.23) [0.822] | −0.13 (0.25) [0.646] |
time preference | 120 (67) | 209 (593) | −89*** (31) | 8 (18) | 156 (456) | 215 (594) | −58 (53) [0.446] | −23 (56) [0.593] | 117 (59) | 196 (570) | −79** (32) | 19 (18) | 182 (550) | 188 (536) | −7 (69) [0.942] | 59 (74) [0.626] |
sent as dictator | 4.59 (2.33) | 5.15 (2.73) | −0.56 (0.39) | −0.84 (0.75) | 4.79 (2.46) | 5.17 (2.75) | −0.38 (0.32) [0.468] | 0.17 (0.35) [0.717] | ||||||||
trust in village | 3.51 (1.19) | 3.18 (1.08) | 0.33* (0.18) | 0.61* (0.37) | 3.35 (1.15) | 3.17 (1.07) | 0.19 (0.12) [0.072]* | 0.12 (0.13) [0.190] | 3.71 (1.19) | 3.22 (1.07) | 0.49** (0.19) | 0.66** (0.31) | 3.47 (1.10) | 3.21 (1.09) | 0.26* (0.14) [0.051]* | 0.14 (0.15) [0.304] |
# Obs. | 49 | 400 | 119 | 330 | 41 | 330 | 81 | 290 |
Note: Standard deviations are in parentheses for means while heteroskedasticity robust standard errors are in parentheses for the tests of differences. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: *** p<0.01,
p<0.05,
p<0.1.
Columns 1, 5, 9, and 13 show information for the non-attriters while Columns 2, 6, 10, and 14 show information for the attriters. Diff1 is coefficient in regression controlling only for participation at the later date. Diff2 is coefficient in regression aditionally controlling for village fixed effects. Income is in thousands of Guarani. All variables are measured in 2002.
Looking across the two tables, the number of significant differences we find is higher than what we would have expected by chance. In 2009 and 2010 we followed up with the poorer villages, and left out the richer villages. Many of these differences decrease in size once controlling for village fixed effects and so seem to be due to selection in terms of which villages we re-surveyed rather than selective attrition within villages. Once controlling for village fixed effects, it still appears that those individuals who remain in our sample over time are slightly older and less educated; as well as slightly more trusting. In all tables which follow we control for village fixed effects, age, education, sex, log income, and household size to help account for some of this selection.
3.2 Survey Data
The data for which we have the most continuity across rounds is the survey trust data. In 2002 the survey asks what share of people in the world, people in the village, and close neighbors they trust. The survey also asks what share of their village-mates would take advantage of them if given the opportunity. Possible answers for both sets of questions are 5-all, 4-more than half, 3-half, 2-less than half, and 1-none. While the correct cardinality is approximately 1, .75, .50, .25, and 0, as this is just a linear transformation of the 1–5 scale we have left the variable in its original form for regression analysis. Respondents are also asked if they think it is bad if somebody buys something knowing it is stolen (The variable equals 1 if they say it is very bad and 0 if they say it is a little bad or not bad at all.)
These same questions all reappear in the 2007 survey, although for the final question it is made more specific asking if they think it is bad if somebody buys a radio knowing it is stolen. In this round a new negative reciprocity question was added asking if someone put them in a difficult position would they to do the same to that person (1 equals always or sometimes, 0 equals never). All of these questions were continued in the 2009 survey. The 2010 survey was shortened so that it only asks about trust in villagemates and negative reciprocity.
We include measures of trust (from both surveys and experiments) in our analysis of social preferences although one might argue that trust measures beliefs about others rather than an underlying social preference (Fehr, 2009; Sapienza et al., 2013). For example, if one sees that trust is going down over time, as measured either by a decrease in the amount sent in a trust game (described later) or an increase in the share of one’s village-mates one believes would take advantage if given the opportunity, this may be due to one of many different reasons. It may be because the respondent’s social preferences are changing, for example the respondent has become inherently more trusting; it may be because social preferences in the village are changing, for example trustworthiness in the village is going down causing the individual to be more trusting; or it may be because the respondent’s beliefs have changed, for example the respondent experienced a robbery and so now believes his village-mates to be less trustworthy. Given that some piece of trust is due to underlying preferences, and that we can control for village fixed effects which capture the environment of trustworthiness in the village, we include trust in our analysis.
3.3 Experimental Data
Because each round of data collection was conducted to answer different questions, rather than with the express purpose of looking at the stability of preferences over time, some of the experiments conducted in each round differ while some are repeated.
3.3.1 Risk Preferences
In 2002 we conducted an incentivized risk experiment in which players were given 8,000 Gs and chose how much of that to bet on the roll of a die (Schechter, 2007b). At that point in time the exchange rate was approximately 4,800 Gs to the dollar, and a day’s wages was 12,000 Gs. Participants could choose to bet 0, 2, 4, 6, or 8 thousand Gs, and different rolls of the dice led to losing all their bet, half their bet, keeping their bet, or earning an extra 50%, 100%, or 150% of their bet. The measure of risk preferences is the amount of money bet on the die.
The 2007 and 2009 surveys both contained hypothetical risk questions based on the question asked in the Mexican Family Life Survey (MxFLS). The respondent is asked if he prefers 50 thousand Gs for sure or a 50-50 chance of 50 or 100 thousand Gs. If he prefers the lottery, he is then asked if he prefers 50 thousand Gs to a 40/100 lottery. Respondents who still prefer the lottery are then offered 30/100, 20/100, and finally 10/100 lotteries. The measure of risk preferences is the number of risky choices preferred. The 2010 survey does not measure risk preferences.
3.3.2 Time Preferences
The 2007 and 2009 surveys ask hypothetical questions measuring time preferences, while the 2002 and 2010 surveys do not broach this topic. Respondents are asked whether they prefer 50,000 Gs today or 75,000 Gs a month from today. For those who prefer the money today, they are asked if they would prefer 50,000 Gs today or 100,000 Gs a month from today. If the person still prefers the money today, he is asked how much one would have to offer him to convince him to wait for a month. The measure of time preferences is the amount of money he would need one month later.
3.3.3 Social Preferences
Every round of survey collection conducted at least one incentivized experiment measuring social preferences, although some of the experiments differed across rounds.
2002 Trust Game
We conducted an incentivized trust game in which every participant played both the role of trustor and trustee (Berg et al., 1995). The trustor was given 8,000 Gs and decided how much of that to keep and how much to send to the trustee. The trustee was randomly chosen from the group of players and received the amount sent by the trustor, tripled. The trustee then decided how much of that to keep and how much to send back to the trustor, if any. These games were anonymous as neither player knew with whom he was matched. The amount sent by the trustor is often used to measure trust (although it also measures risk aversion and altruism), while the amount sent by the trustee is often used to measure trustworthiness or reciprocity (although it also measures altruism).
2007 Variants of Dictator Game
Participants were asked to play four distinct dictator games. For more details see Ligon and Schechter (2012). The games varied in whether or not the dictator was anonymous, and in whether the recipient was randomly selected from the set of households in the village, or was chosen by the dictator. In each of the games the dictator was given 14,000 Gs and decided how to divide it between herself and another household. The other household could be any household in the village, not just those participating as dictators in the experiment. We doubled any money shared by the dictator and also added a random component with mean 5,000 Gs before passing it on to the recipient at which point the game ended.
In the ‘Anonymous-Random’ (AR) game, the dictator decided how much to share with a randomly selected household in the village, and neither the dictator nor the recipient ever learned who the other was. This is the canonical dictator game and it measures undirected altruism or benevolence. In the ‘Revealed-Random’ (RR) game, the dictator once more chose how much to send to a randomly selected household. It was known that the identities of the dictator and recipient would subsequently be revealed to each other. The amount sent measures a combination of undirected altruism and fear of sanctions.
In the two ‘Chosen’ games the dictator chose a single, common recipient. In the ‘Anonymous-Chosen’ game, the recipient never learned the identity of the dictator. The amount sent measures a combination of undirected and directed altruism. In the ‘Revealed-Chosen’ game, the dictator’s identity is revealed to the recipient. The amount sent in this final game measures a combination of undirected and directed altruism, fear of sanctions, and reciprocity.
2009 Variants of Dictator Game
Participants played in games which were quite similar to those in 2007. The difference is that in 2009 no random component was added to the amount received by the recipient. And, some (but not all) participants played the two ‘Chosen’ games.
2010 Reciprocity Game
Respondents in 2010 were asked to play a reciprocity game similar to that designed by Andreoni et al. (2003) with the political middlemen who chose them. The respondent was told that the middleman was given 12,000 Gs and could choose to send 2, 4, 6, 8, or 12,000 Gs to the respondent. The respondent could take the money and choose to do nothing, or he could choose to reward or fine the middleman. For every 100 Gs the respondent put toward the reward or fine, the middleman received an extra 500 Gs, or had 500 Gs taken away from him. Neither player could earn less than 0 Gs total.
We used the strategy method, asking the respondent what he would do if he were to receive each potential amount. Here we define a negative reciprocal individual as one who would fine the middleman if he sent 2,000 Gs (the lowest possible amount to send) and a positive reciprocal individual as one who would reward him if he sent all 12,000 Gs.
3.3.4 Experiences while Playing
Not only will we look at the stability of preferences over time, but we will also look at whether and how experiences in the experiments in one year impact play in later years of the experiments. The different experiences we will look at include the following:
In the 2002 risk games the player’s winnings were determined by the roll of the die. The die was rolled in front of the player. The higher the roll of the die, the luckier the player was and the more he won.
In the 2002 trust game, the trustee received a certain amount from the trustor. Receiving more can change the trustee’s perception of the trust and generosity of his village-mates.
In the 2002 trust game, the trustor chose how much to send to the trustee. The trustee then decided how much of that, if any, to return. Receiving back a higher share of the amount sent can change the trustor’s perception of the generosity and reciprocity of his village-mates.
3.4 Shock Data
Finally, we study how preferences react to household experiences. We have data on income, health, and theft shocks which are arguably some of the most important shocks faced by agricultural households who do not have access to crop insurance or health insurance. Specifically, the 2002, 2007, and 2009 surveys contain information on household income and theft experienced in the past year. Those rounds also asked the days of school or work lost to illness in the past year.
4 Analysis and Results
Our hypothesis is that preferences as measured by experiments and surveys are significantly correlated over time. In the first stage of our analysis we look at the stability of preferences (as measured in experiments and surveys) over time. We show the correlation coefficient between the two variables of interest and evaluate whether it is statistically significantly different from 0. In addition, we run regressions of the later variable on the earlier variable with village fixed effects while also controlling for log income, sex, age, and education level as measured in the later year.13 In results not shown here, we re-run all of the analysis dividing our sample into those below and above the median in terms of age (50 years) or education (5 years) but do not find any consistent differences.
In all tables of results we show asterisks (*) to represent levels of significance according to traditional p-values which treat each test as a lone independent test. These p-values are based on standard errors which are robust to heteroskedasticity. Given that we are testing multiple hypotheses at the same time, these traditional p-values may over-reject the hypothesis of zero correlation of measures over time. To account for this, we additionally show plus signs (+) to represent levels of significance according to q-values for False Discovery Rates (FDR) to correct for multiple comparisons. We use the calculation designed by Benjamini and Hochberg (1995) and explored in detail by Anderson (2008).
In addition to simply correcting our standard errors for heteroskedasticity, we can also correct for potential correlations between observations within clusters (villages). Because we do not have a large number of clusters, we use the wild cluster bootstrap suggested by Cameron et al. (2008) to generate p-values. The 2002 and 2007 data sets have 15 villages; the 2010 data has 10 villages; and the 2009 data has only 2 villages. Because two villages is too few to account for clustering, we do not show clustered p-values in any analysis using data from 2009.
Theory would suggest that if the decisions made by the same individual at two points in time are not highly correlated, it is because the environment has changed. For example, think about an individual with exogenously determined income y. Assume this person has constant relative risk aversion (CRRA) utility, with coefficient of relative risk aversion γ. He must decide between a sure thing (with payoff s) and a lottery (winning payoff h with probability p and outcome l with probability (1 − p)). The individual chooses whether or not to gamble where g = 0 if he chooses the safe option and g = 1 if he chooses the gamble. His total consumption will be his exogenous income plus his earnings in the risk game.
The maximization problem is thus:
This equation makes clear that a person may make different decisions at different points in time either because his underlying risk preference γ has changed, or because his income y has changed. A monetary windfall may cause a person to choose the more risky option. A similar income confound arises for social and time preferences. And, one could imagine shocks other than income, such as health shocks, changing the optimal decision. The main point is that a person can have stable preferences but still behave differently in experiments at different points in time.
We account for this in our study in three ways. First, if everyone is getting poorer or wealthier across the board at the same rate, then we would still imagine that the people who were most risk averse in earlier years would continue to be the people who were most risk averse in later years. Second, our regressions control for income, education, sex, age, and village fixed effects and so our results are conditional on the environment. Finally, if the environment played a large role, then we would expect to find significant results in an analysis looking at how income, health, and theft shocks are related to play in games. But we do not find much correlation between shocks and preferences.
Overall, we find that experimental measures are less stable than survey measures. This may be because some of the experiments varied across years, while the survey questions remained the same in each survey. Semantically, when we compare play in the same game at different points in time we are running “test-retest” analysis, while when we compare play in different games purportedly measuring the same preferences at different points in time we are running a “construct validity” analysis. That said, even when we do have experimental measures for the exact same game at two points in time, for example the variants of the dictator games or the repeated risk game, the choices are not any more stable than measures of the same preferences from different games.
The decisions made in the experiments are not very correlated over time, nor are they correlated with shocks to the environment, suggesting that individuals’ decisions may contain substantial noise. This may be more likely in a developing country context where participants experience high levels of poverty and scarcity and have low levels of education.
With so much data from so many different experiments and survey questions, in so many different rounds of data collection, it would be possible to look at thousands of correlations. We limit ourselves by only showing results which fit the following criteria. First, given that we have much incentivized experimental social preference data, we do not analyze the data from the hypothetical social preference experiments.14 We do analyze the hypothetical risk and time preference data, since we have less data on these preferences.
Second, we do not include comparisons between the 2002 and 2009 data, since the sample size in those comparisons is only 21, nor do we include comparisons between the 2009 and 2010 data, since that sample size is only 23. The comparisons we do include are 2002 vs 2007, 2002 vs 2010, 2007 vs 2009, and 2007 vs 2010.
Third, we only compare variables which are measures of the same thing at different points in time. For example, we look at how survey measures of trust are correlated over time, and we look at how experimental measures of trust are correlated over time. But, we do not look at how survey measures of trust are correlated with experimental measures of trust over time, or at how experimental measures of trust are correlated with experimental measures of risk aversion over time. Although there are definitely many interesting comparisons to be made, we avoid data mining by setting these rules for ourself in advance.
4.1 Stability of Preferences
4.1.1 Risk and Time Experiments
Table 8 shows the results for risk and time preferences. The 2002 game was incentivized. The same hypothetical risk game was played in both 2007 and 2009. The research cited in Section 2.3 has shown that players do not have stable preferences across different risk games played at the same point in time. This makes it less likely that we will find risk preferences are correlated between 2002 and 2007 but this critique should not affect the 2007–2009 comparison. Still, the table shows no significant correlation between risk preferences measured at different points in time. If we run the analysis using coefficients of relative risk aversion rather than the number of risky choices, we again find insignificant results of similar magnitudes. The magnitudes of the correlations in measures of risk preferences over time we find in Table 8 are far below those magnitudes cited in Section 2.1.
Table 8.
Explanatory variable |
Dependent variable |
Correlation coefficient |
Regression coefficient |
# Obs. |
---|---|---|---|---|
bet in 2002 | # risky choices in 2007 (hyp) | 0.070 | 0.066 (0.087) [0.488] | 140 |
bet in 2002 | # risky choices in 2007 (hyp-d) | 0.043 | 0.045 (0.090) [0.586] | 112 |
# risky choices in 2007 (hyp) | # risky choices in 2009 (hyp) | −0.059 | −0.012 (0.102) | 49 |
# risky choices in 2007 (hyp-d) | # risky choices in 2009 (hyp-d) | −0.009 | 0.003 (0.112) | 32 |
Time preference in 2007 (hyp) | Time preference in 2009 (hyp) | 0.432*** ++ | 1.036 (0.646) | 49 |
Notes: Heteroskedasticity robust standard errors are in parentheses. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: ***p <0.01,
p <0.05,
p <0.10.
FDR q-values: +++ q <0.01,
q <0.05,
q <0.10 calculated for 5 hypotheses within table and column.
In regression column these are based on the heteroskedasticity robust standard errors. Controls in regressions include log income, sex, age, education, and village fixed effects. Hyp denotes that the game was hypothetical. Hyp-d additionally denotes that individuals choosing the strictly dominated optoin are excluded.
On the other hand, time preferences (measured from the answers to the same series of hypothetical questions in 2007 and 2009) are relatively stable over time. We cannot reject that the coefficient on time preferences in 2007 predicting time preferences in 2009 is 1. We also can not reject that it equals 0 at conventional levels of confidence, although it is statistically different from 0 at the 12% significance level. This is in stark contrast to the results for the hypothetical risk games in 2007 and 2009 where the coefficient is a much more precise 0. This may be because time preferences are more stable over time than risk preferences, or because people understood the time questions better and so they contain less noise.
One indication that the 2007/2009 risk preference data contains much noise is the fact that 19% of respondents in 2007 and 27% in 2009 preferred 50,000 Gs for sure to a 50/50 chance of winning 50,000 or 100,000 Gs, even after the game was explained to them a second time. If one assumed that an equal share did not understand the game but choose the better option by chance, this would imply that 38–54% of respondents do not understand the game. Such low levels of understanding wreak havoc on our ability to use these games to measure risk preferences.
While the share of individuals choosing dominated options may seem high, they are in fact quite similar to those found in other studies. In the MxFLS, upon which the questions we used in our survey were based, nearly a quarter of the individuals chose the dominated option even after having it explained a second time (Hamoudi, 2006). In a similar set of questions in the 2007 Indonesian Family Life Survey (IFLS), 41% of the almost 30,000 respondents chose the dominated option (Caruthers, 2013). Even in the US, Bhargava et al. (2015) find that over half of individuals choose dominated health insurance plans, and bad choices are even more common among the poor and the elderly.
This is reminiscent of the debate following Gneezy et al. (2006) who find that many individuals value a risky prospect less than its worst possible realization. Subsequent work argued that the effect is eliminated either when focussing on the subset of the subject pool which passes a comprehension test or when making the instructions more clear (Keren and Willemsen, 2009; Rydval et al., 2009). Choices made by individuals who do not fully understand the experiment may be filled with much noise, and thus these choices may not be significantly correlated over time.
This suggests that one reason we find such low correlations for measures of risk aversion compared to those found in the rest of the literature may be that our sample has a much lower level of education than most of the other experiments. Choi et al. (2014) find that younger, more educated, and wealthier respondents behave more consistently across multiple games measuring risk and time preferences played at the same point in time.
We explore this in two ways. First, we exclude those individuals who chose the strictly dominated option. We present those results in Table 8 and find that the magnitudes are no larger and the correlations no closer to being significant when excluding those individuals.
Second, we divide our sample into those below and above the median in terms of age or education. When doing that we find a similar lack of significance in all groups for risk preferences.15 Given that the median in our sample is 5 years of education and the maximum is 12, the more educated half of our sample still has much lower education levels than respondents in most other samples studied in the literature. For time preferences, the relatively strong correlation seems to be coming entirely from the younger and more educated halves of the population.
Looking forward, in the next subsection we show evidence that the more specific a trust question is, the more stable its answer. While some risk games conducted in developing countries are framed in terms of crop choice or other real-world investment decisions, our risk games were very abstract. It would be interesting for future research to study whether decisions made in framed experiments are more stable than those made in artefactual (neutrally framed) experiments.
4.1.2 Social Preference Surveys
Table 9 shows the results for the survey questions which measure social preferences and were repeated across years.16 Many of the variables show a surprising amount of stability over time. Almost all of the comparisons are significant. On the other hand, when looking at the regression coefficients we can also reject that the relationship is one-to-one.
Table 9.
Years | Variable | Correlation coefficient |
Regression coefficient |
# Obs. |
---|---|---|---|---|
2002 vs 2007 | trust people in the world | 0.064 | 0.068 (0.119) [0.585] | 123 |
2007 vs 2009 | trust people in the world | 0.284** + | 0.339* (0.176) | 49 |
| ||||
2002 vs 2007 | trust people in the village | 0.137 | 0.162* (0.096) [0.211] | 123 |
2002 vs 2010 | trust people in the village | 0.440*** ++ | 0.425*** (0.132)++ [0.011]** | 39 |
2007 vs 2009 | trust people in the village | 0.525*** +++ | 0.525*** (0.146)+++ | 49 |
2007 vs 2010 | trust people in the village | 0.254*** ++ | 0.206** (0.094)+ [0.052]* | 119 |
| ||||
2002 vs 2007 | trust closest neighbors | 0.273*** ++ | 0.275** (0.108)++ [0.013]** | 123 |
2007 vs 2009 | trust closest neighbors | 0.463*** +++ | 0.545*** (0.159)+++ | 49 |
| ||||
2002 vs 2007 | bad to buy something you know is stolen | 0.141 | 0.226* (0.123) [0.230] | 123 |
2007 vs 2009 | bad to buy something you know is stolen | 0.372*** ++ | 0.353** (0.150)+ | 49 |
| ||||
2002 vs 2007 | would villagemates take advantage if had opportunity | 0.251*** ++ | 0.167* (0.094) [0.108] | 123 |
2007 vs 2009 | would villagemates take advantage if had opportunity | 0.355** ++ | 0.420** (0.190)+ | 49 |
| ||||
2007 vs 2009 | negative reciprocity | 0.360** ++ | 0.236 (0.154) | 49 |
2007 vs 2010 | negative reciprocity | 0.245*** ++ | 0.176* (0.099) [0.176] | 119 |
Notes: Heteroskedasticity robust standard errors are in parentheses. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: ***p <0.01,
p <0.05,
p <0.10.
FDR q-values: +++ q <0.01,
q <0.05,
q <0.10 calculated for 14 hypotheses within table and column.
In regression column these are based on the heteroskedasticity robust standard errors. Controls in regressions include log income, sex, age, education, and village fixed effects.
Imagine a model in which observed trust is equal to an underlying trust parameter plus two additional pieces: recent experiences and measurement error. If most of the variation around the underlying trust parameter is due to recent experiences, we might expect the correlation to go down as more time passed. On the other hand, if most of the variation around the underlying trust parameter were due to measurement error, then we might expect the size of the correlations to be similar over time.
The correlation and regression coefficients are almost always highest for the 2007 vs 2009 comparison, which is also the comparison over the shortest time span. Unfortunately, this is also the only comparison which involves the 2009 data. Because the 2009 data only comes from the two smallest villages, there is very little overlap between either 2002 or 2010 and 2009. Thus, we cannot test whether the high correlation between 2007 and 2009 is due to the select sample in 2009 or if it is due to the shorter time span.17
Given the large magnitude of the coefficients for the one comparison shown between 2002 and 2010 (trust people in village), it suggests that it is not the case that the correlation goes down significantly over long periods of time. In fact, we can replicate the results in Table 9, but only for those 39 individuals who are surveyed in 2002, 2007, and 2010. When we do so (results not shown), we continue to find no significant difference in the coefficients for the 2002–2007 2007–2010, and 2002–2010 comparisons for trust in village, with 2002–2007 continuing to be the smallest coefficient. This may imply that most variation is due to measurement error rather than the influence of recent experiences.
These questions may measure innate preferences which are stable and not strongly affected by shocks. On the other hand, the stability of answers to these questions may indicate that individuals have a well-developed self-identity or feel a strong social norm regarding how to answer the questions which may or may not be what researchers have in mind when modeling social preferences.
There also seems to be some suggestive evidence that as the questions become more specific, the responses exhibit more stability. For example, the correlation and regression coefficients regarding trusting people in the world are rather low and often not significant. The coefficients regarding trusting people in the village are higher, while the coefficients regarding trusting close neighbors are usually even higher, although most of these differences are not significant.18
4.1.3 Social Preference Experiments
Table 10 shows the results for the experimental measures of social preferences. The results are loosely divided up by social preference into altruism, trust, and reciprocity. Many of the decisions made in the games measure multiple preferences at the same time. For example, the amount sent by the trustor in the trust game is a combination of both trust and altruism. The amount returned by the trustee in the trust game combines altruism, trust, and reciprocity. The amounts sent by the dictator in the two non-revealed games measure altruism. The amounts sent in the two revealed games additionally measure trust since a person who trusts his village-mates more will send more in the hopes that the recipient will return more to the dictator once the recipient learns the dictator’s identity. Finally, the reciprocity experiments measure reciprocity.
Table 10.
Explanatory variable |
Dependent variable |
Correlation coefficient |
Regression coefficient |
# Obs. |
---|---|---|---|---|
ALTRUISM | ||||
sent as trustor in 2002 | sent as dictator in anonymous game in 2007 | 0.297*** ++ | 0.298* (0.173) [0.221] | 103 |
share returned as trustee in 2002 | sent as dictator in anonymous game in 2007 | 0.132 | 1.171 (1.939) [0.628] | 103 |
sent as dictator in anonymous game in 2007 | sent as dictator in anonymous game in 2009 | −0.107 | −0.180 (0.190) | 41 |
sent as dictator in chosen non-revealed game in 2007 | sent as dictator in chosen non-revealed game in 2009 | 0.138 | 0.126 (0.091) | 33 |
| ||||
TRUST | ||||
sent as trustor in 2002 | sent as dictator in revealed game in 2007 | 0.354*** +++ | 0.513*** (0.174)++ [0.037]** | 103 |
share returned as trustee in 2002 | sent as dictator in revealed game in 2007 | 0.283*** ++ | 4.335* (2.189) [0.118] | 103 |
sent as dictator in revealed game in 2007 | sent as dictator in revealed game in 2009 | 0.049 | −0.050 (0.145) | 41 |
sent as dictator in chosen revealed game in 2007 | sent as dictator in chosen revealed game in 2009 | −0.118 | −0.236 (0.229) | 33 |
| ||||
RECIPROCITY | ||||
share returned as trustee in 2002 | positive reciprocity in 2010 | 0.009 | 0.473 (0.376) [0.309] | 43 |
share returned as trustee in 2002 | negative reciprocity in 2010 | 0.123 | −0.430 (0.334) [0.216] | 43 |
Notes: Heteroskedasticity robust standard errors are in parentheses. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: ***p <0.01,
p <0.05,
p <0.10.
FDR q-values: +++ q <0.01,
q <0.05,
q <0.10 calculated for 10 hypotheses within table and column.
In regression column these are based on the heteroskedasticity robust standard errors. Controls in regressions include log income, sex, age, education, and village fixed effects.
When looking at the stability of different social preferences over time we look at choices made in different games, some of which are more pure measures of the preference than others. For example, in 2007 and 2009 we have measures of altruism from anonymous dictator games which should arguably be the best measures of altruism. In 2002, we only have the amounts sent as trustor and trustee in the trust game. Although altruism is sure to influence those decisions, trust and reciprocity will influence them as well. We look at whether these decisions are correlated over time not because we believe that there is a single latent trait which drives behavior across all the experiments. We look at them because we believe that all three of those decisions involve altruism, and if altruism is a stable preference over time then those decisions should be correlated.
With regards to altruism, we find that the amount sent as trustor in 2002 and the amount sent in the anonymous game in 2007 are positively correlated. We find a strong correlation between the two measures of trust in 2002 and 2007. Individuals who give generously in one game at one point in time tend to be more generous in a different game at a different point in time.
We find no correlation between the share returned in 2002 and later measures of reciprocity or later measures of altruism, but we do find correlation between the share returned in 2002 and our 2007 measure of trust. Although this correlation is relatively strong, interpreting the share returned as a measure of trust (rather than of trustworthiness) is a bit tenuous.
Finally, within the trust and altruism sections, we look at whether the amounts sent in the various dictator games played in 2007 and 2009 are correlated. Given the high degree of similarity in the games, it is surprising that not a single one of the four comparisons is significant.
According to our loose categorization of preferences measured by different games, we see the most stability in trust preferences, weak stability in altruistic preferences, and no stability in reciprocity. What is perhaps more noticeable than the stability of experimental measures of social preferences, may actually be their lack of stability. In comparing Table 9 with Table 10, the higher relative stability of survey-based measures of social preferences compared with experiment-based measures is evident.
4.1.4 Graphical Representation of Standardized Results
It is difficult to compare the magnitudes of the stability of preferences shown in Tables 8, 9, and 10 as they are spread across multiple tables and the variables are not standardized. Thus, we condense all of the previous results in Figure 1. In this figure we run similar regressions to those in our other tables (controlling for log income, sex, age, education, and village fixed effects) and we show whisker plots of the coefficients and 90% confidence intervals.
The difference between the results in the figure and those in the tables is that in this figure we standardize the preference variables on both the right and left-hand side so that they have a mean of 0 and standard deviation of 1. One thing to keep in mind when looking at this figure is that the explanatory and dependent variables were standardized separately. For example, in Table 8 the regression coefficient on the time variable was approximately 1. But, in the figure we see that it is closer to 0.4. This is because the mean and standard deviation of that variable is much higher in 2007 compared to 2009.
What becomes clear from this figure is that the survey measures of trust are consistently quite strongly correlated over time. On the other hand, it also becomes clear that this relationship is far from being one-to-one. When looking at the experimental measures of social preferences, they do tend to be positively correlated over time, but the confidence intervals on the experimental measures are much wider than those on the survey measures. Finally, the confidence intervals around the measures of risk preferences are rather narrow, and that relationship seems like a more precisely estimated zero.
4.2 Impact of Outcomes in Previous Games on Later Games
4.2.1 Impact of Experience in the Risk Game
In the 2002 risk experiment the players’ payoffs depended on the roll of a die. For those most risk averse people who didn’t bet anything, we didn’t roll the die, but we have an observation for all others. Appendix Table A-1 explores whether the roll of the die has an impact on play in the later risk experiment. We additionally control for the baseline level of risk aversion as measured in the earlier experiment.
We find weak evidence that having a good roll of the die in 2002 leads people to choose more risky choices in 2007. The magnitude of this effect is such that a roll of the die that is higher by 1, leads to 0.2 extra risky decisions being made out of 5 in 2007. If we look instead at coefficients of relative risk aversion (CRRA), we find that a roll of the die that is higher by 1 leads to a decrease of 0.16 in the individual’s estimated CRRA in 2007.
A good roll of the die might lead players to think they will be more lucky in games played with us in the future. In results not shown here we do test if the roll of the die had an impact on subsequent risk-related decisions made in the real-world, such as the number of different crops planted, and we find no impact. This suggests that the impact only involves learning about the experimental setting, and does not change the players’ underlying risk preferences.
4.2.2 Impact of Experience in the Trust Game
In 2002, the payoffs of trustors and trustees depended both on the choices they themselves made, as well as on the actions of their partners. The payoffs of a trustor who sent a positive amount depended on how much (if anything) the trustee with whom he was paired decided to return.19 This experience could give a player new information regarding how trustworthy and how altruistic his village-mates were within the experimental setting, and could then potentially impact how he played in games in later years. This impact could either be due to a change in how he expected others to behave, or due to a change in how he thought he himself ought to behave given his new signal regarding local experimental norms.
Appendix Table A-2 contains the results of this analysis. Because it might be the case that trustees return a different share depending on how much they received from the trustor, when looking at impacts on generosity in 2007 we also control for the amount sent by the respondent as trustor in 2002 in all regressions in this table. This also helps to control for baseline levels of social preferences, and we additionally control for the average share the respondent returned as trustee in 2002.
We find that receiving a higher share back as trustor in 2002 leads to an increase in altruism in 2007. The impact of going from receiving back all your money in the trust game, to doubling your money in the trust game is to increase the amount sent in 2007 by almost 1,000 Gs (out of a potential total of 14,000 Gs).20 This is in line with Matthey and Regner (2013) who find that individuals who have participated in more experiments keep more money for themselves, with suggestive evidence that this is due to them receiving less than they had expected in the previous experiment.
We see a similar magnitude increase in the amount sent in all versions of the 2007 dictator games (anonymous and revealed, and randomly assigned and explicitly chosen). Thus, it seems likely that this impact is due to changes in the individual’s sense of what the norm he should be following is, rather than a change in how he expects his village-mates will respond. Finally, receiving more as trustor in 2002 leads individuals to be significantly more negative reciprocal in 2010. If they receive more money from their partner in 2002, they are more likely to punish bad behavior from partners in 2010, enforcing norms of good behavior. This last result is the only one which stands up when considering that we are testing multiple hypotheses at once.
In 2002, trustees also receive different amounts of money depending on the play of their partner. Similar to the discussion above, this could change the individual’s perception of how generous his village-mates are, and could also change his perception of the social norm in the village. Appendix Table A-3 looks at the impacts the amount received as trustee in 2002 may have on play in later games. Similarly to the previous table, we see that individuals who receive more as trustee in 2002 send more as dictator in all four versions of the dictator game. So, again, we do see that altruism in 2007 is increased if people receive more money from their village-mates in 2002.
We think that these changes are due to learning about the within-game sharing norms. We have tested whether different experiences as trustor or trustee lead to differences in real-world outcomes such as future gift-giving, and find no impact. This suggests that these experiences do not change underlying preferences but only change behavior within the experimental setting.
4.3 Impact of Real-World Shocks on Preferences
Finally we study the correlations between preferences and changes in income, changes in theft, and sickness in the past year. In the population we are studying, these are the most common and most relevant shocks. Income shocks are not exogenous, and so such an analysis should be taken with a grain of salt. Theft and health shocks are more plausibly exogenous than income shocks. Because we did not measure income, theft, or sickness in 2010, this analysis will only look at 2002, 2007, and 2009 data. Results can be found in Appendix B.
For income and theft, we regress levels of preferences on changes in income or changes in theft experienced while controlling for levels of preferences in the previous round.21 One might prefer to run regressions of changes in preferences on changes in income or theft. But, since our preference variables are often not exactly the same in each survey round, it is problematic to take their difference. Instead, of putting the difference in preferences on the left hand side, we control for past preferences on the right hand side.
Appendix Tables B-1 and B-2 basically show no correlation between preferences and income or theft shocks. Although the results are not very significant, the results regularly show that survey measures of trust are negatively correlated with increases in theft. The results for the effect of health shocks on social preferences shown in Tables B-4 and B-5 are mostly inconsistent and insignificant with one exception. Table B-5 suggests that health shocks induce individuals to be less generous in experiments. This result retains significance when taking into account multiple hypothesis testing.22
In contrast to our results, Dean and Sautmann (2014) do find an impact of income and expenditure shocks (which includes, for example, expenditure on family illness) on time preferences in a developing country context - peri-urban Mali. One difference between their setting and ours is that their shock and preference data was taken weekly over a period of 3 weeks. Family illness may influence choices made in experiments in that very week, but months later that effect may wear off. The use of yearly data, as we do here, may make it less likely to find impacts of shocks on preferences. Additionally, Dean and Sautmann (2014) find no impact of labor income on preferences, and hypothesize that this is due to its endogeneity. They only find an impact of non-labor earnings which are more likely to be exogenous. In our rural Paraguayan sample there are almost no sources of non-labor income, so we are not able to look at this type of income. Giné et al. (2014) find no impact of shocks on time preferences among rural Malawians over six to eight weeks, but the households in their sample suffer very few shocks over this time period, so their tests suffer from low power. Overall, the general message of these tables is that shocks of the size measured in this paper do not significantly impact our measures of risk, time, or social preferences.
5 Conclusion
The results presented here suggest that social preferences measured by simple survey questions are quite stable over time, while social and time preferences measured by experiments are much less stable. Risk preferences exhibit a relatively tight 0 correlation over time. This may suggest that some underlying preferences, for example risk preferences, are not stable. Or, underlying preferences may be stable, but the measures we get from experiments may contain significant amounts of noise.
Given that our measures of preferences are not very stable over time, we then also explore whether they are affected by observable shocks. There is some evidence that being matched with a generous partner in a trust game in 2002 leads individuals to send more in dictator games in 2007. We surmise that this may be due to a change in the players’ perceptions of social norms within the experimental setting in their village. On the other hand, we find no evidence that income, theft, or health shocks are correlated with changes in preferences other than the finding that health shocks cause individuals to be less generous in experiments.
One potentially fruitful avenue for future research would be to run experiments which purposely allow respondents to make inconsistent or dominated choices. Of our experiments, only the risk experiments in 2007 and 2009 allowed individuals to make a dominated choice, and none gave individuals the opportunity to make inconsistent choices. Jacobson and Petrie (2009) study inconsistent choices in Rwanda and find that risk preferences are only correlated with real world outcomes when simultaneously controlling for having made an inconsistent choice. Keren and Willemsen (2009) and Rydval et al. (2009) find that increasing respondent comprehension, or excluding respondents who don’t pass a comprehension test, leads to results which are more consistent at one point in time. It is an open question as to whether this would also lead to results which are more stable across time.
The benefit of designing experiments which allow respondents to make inconsistent choices is that the researcher then has a measure of inconsistency and he can either drop individuals who make inconsistent choices or control for this measure. On the other hand, experiments which involve many choices and allow for inconsistencies may also be more complicated and harder for populations with low levels of education and nutrition to understand.
A different avenue for future research would be exploring whether even more simple experiments could lead to better measures. There is a trade-off between getting narrowly pinpointed measures of social preferences and experiment simplicity. Dave et al. (2010) find that simple risk aversion experiments which give broad ranges of risk parameters perform better in low-educated populations in the US. Charness and Viceisza (2015) conducts a similar exercise in rural Senegal. Both of these papers focus on risk aversion experiments.
Another possibility which some researchers have posited is that framing experiments can aid in comprehension by developing country populations (Charness and Viceisza, 2015). A study comparing results for abstract risk aversion games and games framed as seed choice could lead to new insights.
A final fruitful avenue of research could be for researchers to explore more deeply whether survey measures of preferences are robustly more stable than experimental measures, especially in low-education settings. This would be in accord with the findings of Dohmen et al. (2011) and Lönnqvist et al. (2014) that survey measures of risk aversion are more correlated with real world outcomes and more stable over time than experimental measures of risk aversion. On the other hand, Burks et al. (2012) find that a survey measure of time preference does not perform as well as experimental measures of time preferences. This is potentially in accord with what is found in this paper. Risk preference experiments tend to be complicated and hard to understand, and so it would make sense that survey measures do a better job. Time preference experiments on the other hand tend to be simpler, and so in that domain experiments may do better. None of the above-mentioned papers were conducted in a developing country setting, and our data does not include survey information on risk or time preferences.
Overall, these results are thought-provoking for researches who use experiments to measure both preferences and their impact on real-world outcomes, especially in a developing country setting with low levels of education and high levels of poverty. Similarly to Meier and Sprenger (2015), our results suggest that variability in preference measures may be mostly due to noise. The fact that the experimental measures show such low levels of stability may encourage researchers working in developing countries both to design simpler experiments, and to include survey questions measuring preferences in addition to the experiments.
Supplementary Material
A Appendix: Impact of Previous Games on Play
Table A-1.
Dependent variable |
Correlation coefficient |
Regression coefficient |
# Obs. |
---|---|---|---|
# risky choices in 2007 (hyp) | 0.119 | 0.184* (0.103) [0.044]** | 126 |
Notes: Heteroskedasticity robust standard errors are in parentheses. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: ***p <0.01,
p <0.05,
p <0.10.
Controls in regressions include log income, sex, age, education, bet in 2002, and village fixed effects. Hyp denotes that the game was hypothetical.
Table A-2.
Dependent variable |
Correlation coefficient |
Regression coefficient |
# Obs. |
---|---|---|---|
ALTRUISM | |||
sent as dictator in anonymous game in 2007 | 0.207** | 0.916 (0.557) [0.167] | 95 |
sent as dictator in chosen non-revealed game in 2007 | 0.143 | 0.684 (0.593) [0.200] | 95 |
| |||
TRUST | |||
sent as dictator in revealed game in 2007 | 0.159 | 1.000** (0.467) [0.056]* | 95 |
sent as dictator in chosen revealed game in 2007 | 0.146 | 0.800 (0.607) [0.267] | 95 |
| |||
RECIPROCITY | |||
positive reciprocity in 2010 | 0.175 | 0.126 (0.150) [0.538] | 42 |
negative reciprocity in 2010 | 0.421*** ++ | 0.286** (0.105)+ [0.051]* | 42 |
Notes: Heteroskedasticity robust standard errors are in parentheses. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: ***p <0.01,
p <0.05,
p <0.10.
FDR q-values: +++ q <0.01,
q <0.05,
q <0.10 calculated for 6 hypotheses within table and column.
In regression column these are based on the heteroskedasticity robust standard errors. Controls in regressions include log income, sex, age, education, amount respondent sent as trustor in 2002, share returned by respondent as trustee in 2002, and village fixed effects.
Table A-3.
Dependent variable |
Correlation coefficient |
Regression coefficient |
# Obs. |
---|---|---|---|
ALTRUISM | |||
sent as dictator in anonymous game in 2007 | 0.204** | 0.098 (0.059) [0.085]* | 103 |
sent as dictator in chosen non-revealed game in 2007 | 0.162 | 0.123** (0.060) [0.214] | 103 |
| |||
TRUST | |||
sent as dictator in revealed game in 2007 | 0.179* | 0.116* (0.067) [0.068]* | 103 |
sent as dictator in chosen revealed game in 2007 | 0.179* | 0.144** (0.061) [0.123] | 103 |
| |||
RECIPROCITY | |||
positive reciprocity in 2010 | −0.094 | 0.003 (0.014) [0.875] | 43 |
negative reciprocity in 2010 | 0.032 | −0.002 (0.014) [0.902] | 43 |
Notes: Heteroskedasticity robust standard errors are in parentheses. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: ***p <0.01,
p <0.05,
p <0.10.
FDR q-values: +++ q <0.01,
q <0.05,
q <0.10 calculated for 6 hypotheses within table and column.
In regression column these are based on the heteroskedasticity robust standard errors. Controls in regressions include log income, sex, age, education, amount respondent sent as trustor in 2002, share returned by respondent as trustee in 2002, and village fixed effects.
B Appendix: Impact of Real-World Shocks on Play
Table B-1.
Explanatory variable |
Dependent variable |
Correlation coefficient |
Regression coefficient |
# Obs. |
---|---|---|---|---|
Δ log income 2002 to 2007 | # risky choices in 2007 (hyp) | −0.003 | −0.150 (0.231) [0.578] | 140 |
Δ log income 2007 to 2009 | # risky choices in 2009 (hyp) | −0.027 | 0.013 (0.399) | 49 |
Δ log theft 2002 to 2007 | # risky choices in 2007 (hyp) | 0.095 | 0.091** (0.043) [0.127] | 140 |
Δ log theft 2007 to 2009 | # risky choices in 2009 (hyp) | −0.193 | −0.067 (0.072) | 49 |
days sick 2007 | # risky choices in 2007 (hyp) | −0.150* | −0.006* (0.004) [0.175] | 140 |
days sick 2009 | # risky choices in 2009 (hyp) | 0.292** | 0.013* (0.007) | 49 |
| ||||
Δlog income 2007 to 2009 | time preference in 2009 (hyp) | 0.001 | 1.502 (31.95) | 49 |
Δ log theft 2007 to 2009 | time preference in 2009 (hyp) | −0.014 | −2.384 (4.663) | 49 |
days sick 2009 | time preference in 2009 (hyp) | −0.100 | −0.164 (0.414) | 49 |
Notes: Heteroskedasticity robust standard errors are in parentheses. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: ***p <0.01,
p <0.05,
p <0.10.
FDR q-values: +++ q <0.01,
q <0.05,
q <0.10 calculated for 9 hypotheses within table and column.
In regression column these are based on the heteroskedasticity robust standard errors. Hyp denotes that the game was hypothetical. Controls in regressions include sex, age, education, and village fixed effects. Additionally the risk regressions control for risk aversion in the previous round, while the time regressions control for time preferences in the previous round. Days sick - number of days non-disabled people in the household aged 11–74 couldn’t work due to illness. Days sick regressions additionally control for number of non-disabled people aged 11–74. Days sick and theft regressions additionally control for log income.
Table B-2.
Explanatory variable |
Dependent variable |
Correlation coefficient |
Regression coefficient |
# Obs. |
---|---|---|---|---|
Δ log income 2002 to 2007 | Outcome in 2007 | |||
sent as dictator in anonymous game | −0.086 | −0.403 (0.325) [0.352] | 103 | |
sent as dictator in chosen non-revealed game | 0.033 | −0.087 (0.314) [0.789] | 103 | |
sent as dictator in revealed game | −0.100 | −0.573 (0.387) [0.275] | 103 | |
sent as dictator in chosen revealed game | −0.075 | −0.566 (0.369) [0.261] | 103 | |
trust people in the world | −0.097 | −0.140 (0.114) [0.180] | 122 | |
trust people in the village | −0.016 | −0.056 (0.139) [0.736] | 122 | |
trust closest neighbor | −0.047 | −0.069 (0.144) [0.546] | 122 | |
bad to buy something you know is stolen | 0.001 | 0.019 (0.035) [0.507] | 122 | |
would villagemates take advantage if had opportunity | 0.045 | 0.104 (0.108) [0.301] | 122 | |
| ||||
Δ log income 2007 to 2009 | Outcome in 2009 | |||
sent as dictator in anonymous game | −0.204 | −0.409 (0.456) | 41 | |
sent as dictator in chosen non-revealed game | −0.151 | −0.277 (0.430) | 33 | |
sent as dictator in revealed game | −0.100 | −0.123 (0.458) | 41 | |
sent as dictator in chosen revealed game | −0.054 | 0.364 (0.903) | 33 | |
trust people in the world | −0.102 | −0.125 (0.302) | 49 | |
trust people in the village | −0.005 | 0.030 (0.280) | 49 | |
trust closest neighbor | −0.076 | −0.270 (0.300) | 49 | |
bad to buy something you know is stolen | −0.113 | −0.030 (0.088) | 49 | |
would villagemates take advantage if had opportunity | −0.004 | 0.027 (0.229) | 49 | |
negative reciprocity | 0.239* | 0.103 (0.089) | 49 |
Notes: Heteroskedasticity robust standard errors are in parentheses. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: ***p <0.01,
p <0.05,
p <0.10.
FDR q-values: +++ q <0.01,
q <0.05,
q <0.10 calculated for 9 or 10 hypotheses within table and column.
In regression column these are based on the heteroskedasticity robust standard errors. Controls in regressions include sex, age, education, and village fixed effects. Each regression additionally controls for social preferences in the previous round. In the top panel, the first four rows control for the amount sent as trustor and the average share returned as trustee in 2002. In the remaining rows of the first panel, the 2002 version of the 2007 outcome variable is included as a control variable. In all rows of the second panel, the 2007 version of the 2009 outcome is included as a control variable.
Table B-3.
Explanatory variable |
Dependent variable |
Correlation coefficient |
Regression coefficient |
# Obs. |
---|---|---|---|---|
Δ log theft 2002 to 2007 | Outcome in 2007 | |||
sent as dictator in anonymous game | −0.030 | −0.024 (0.076) [0.708] | 103 | |
sent as dictator in chosen non-revealed game | −0.027 | 0.013 (0.105) [0.919] | 103 | |
sent as dictator in revealed game | 0.055 | 0.073 (0.085) [0.367] | 103 | |
sent as dictator in chosen revealed game | 0.001 | 0.083 (0.104) [0.550] | 103 | |
trust people in the world | −0.116 | −0.059** (0.024)+ [0.024]** | 123 | |
trust people in the village | −0.079 | −0.064** (0.027)+ [0.121] | 123 | |
trust closest neighbor | 0.037 | 0.007 (0.030) [0.852] | 123 | |
bad to buy something you know is stolen | 0.057 | 0.008 (0.007) [0.426] | 123 | |
would villagemates take advantage if had opportunity | −0.075 | −0.007 (0.025) [0.739] | 123 | |
| ||||
Δ log theft 2007 to 2009 | Outcome in 2009 | |||
sent as dictator in anonymous game | 0.117 | 0.080 (0.106) | 41 | |
sent as dictator in chosen non-revealed game | 0.078 | −0.015 (0.095) | 33 | |
sent as dictator in revealed game | 0.156 | 0.070 (0.086) | 41 | |
sent as dictator in chosen revealed game | 0.122 | 0.068 (0.117) | 33 | |
trust people in the world | −0.025 | −0.001 (0.049) | 49 | |
trust people in the village | 0.035 | −0.013 (0.068) | 49 | |
trust closest neighbor | −0.077 | −0.083 (0.079) | 49 | |
bad to buy something you know is stolen | 0.022 | 0.005 (0.009) | 49 | |
would villagemates take advantage if had opportunity | 0.064 | 0.066 (0.058) | 49 | |
negative reciprocity | −0.008 | −0.002 (0.021) | 49 |
Notes: Heteroskedasticity robust standard errors are in parentheses. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: ***p <0.01,
p <0.05,
p <0.10.
FDR q-values: +++ q <0.01,
q <0.05,
q <0.10 calculated for 9 or 10 hypotheses within table and column and panel.
In regression column these are based on the heteroskedasticity robust standard errors. Controls in regressions include log income, sex, age, education, and village fixed effects. Each regression additionally controls for social preferences in the previous round. In the top panel, the first four rows control for the amount sent as trustor and the average share returned as trustee in 2002. In the remaining rows of the first panel, the 2002 version of the 2007 outcome variable is included as a control variable. In all rows of the second panel, the 2007 version of the 2009 outcome is included as a control variable.
Table B-4.
Explanatory variable |
Dependent variable |
Correlation coefficient |
Regression coefficient |
# Obs. |
---|---|---|---|---|
Outcome in 2007 | ||||
days sick 2007 | trust people in the world | −0.056 | −0.002 (0.002) [0.390] | 123 |
trust people in the village | 0.072 | 0.004 (0.003) [0.114] | 123 | |
trust closest neighbor | −0.061 | −0.001 (0.004) [0.796] | 123 | |
bad to buy something you know is stolen | −0.015 | −0.000 (0.001) [0.756] | 123 | |
would villagemates take advantage if had opportunity | 0.049 | 0.002 (0.002) [0.261] | 123 | |
| ||||
Outcome in 2009 | ||||
days sick 2009 | trust people in the world | −0.111 | −0.006 (0.007) | 49 |
trust people in the village | 0.173 | 0.007 (0.005) | 49 | |
trust closest neighbor | 0.109 | 0.004 (0.004) | 49 | |
bad to buy something you know is stolen | 0.217 | 0.002 (0.002) | 49 | |
would villagemates take advantage if had opportunity | −0.219 | −0.009** (0.004) | 49 | |
negative reciprocity | −0.004 | −0.000 (0.002) | 49 |
Notes: Heteroskedasticity robust standard errors are in parentheses. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: ***p <0.01,
p <0.05,
p <0.10.
FDR q-values: +++ q <0.01,
q <0.05,
q <0.10 calculated for 5 or 6 hypotheses within table and column and panel.
In regression column these are based on the heteroskedasticity robust standard errors. Controls in regressions include log income, sex, age, education, number of non-disabled people in the household aged 11–74, and village fixed effects. Days sick - number of days non-disabled people in the household aged 11–74 couldn’t work due to illness. Each regression additionally controls for social preferences in the previous round. In the first panel, the 2002 version of the 2007 outcome variable is included as a control variable. In the second panel, the 2007 version of the 2009 outcome is included as a control variable
Table B-5.
Explanatory variable |
Dependent variable |
Correlation coefficient |
Regression coefficient |
# Obs. |
---|---|---|---|---|
Outcome in 2007 | ||||
days sick 2007 | sent as dictator in anonymous game | −0.052 | −0.001 (0.007) [0.958] | 103 |
sent as dictator in chosen non-revealed game | −0.035 | 0.004 (0.009) [0.751] | 103 | |
sent as dictator in revealed game | −0.242** | −0.015** (0.006)+ [0.028]** | 103 | |
sent as dictator in chosen revealed game | −0.110 | 0.000 (0.009) [0.976] | 103 | |
| ||||
Outcome in 2009 | ||||
days sick 2009 | sent as dictator in anonymous game | −0.126 | −0.009 (0.006) | 41 |
sent as dictator in chosen non-revealed game | −0.296* | −0.006** (0.003)+ | 33 | |
sent as dictator in revealed game | −0.236 | −0.009** (0.004)+ | 41 | |
sent as dictator in chosen revealed game | −0.250 | −0.013** (0.005)+ | 33 |
Notes: Heteroskedasticity robust standard errors are in parentheses. Wild cluster bootrap p-values are in brackets.
Per-comparison significance: ***p <0.01,
p <0.05,
p <0.10.
FDR q-values: +++ q <0.01,
q <0.05,
q <0.10 calculated for 8 hypotheses within table and column.
In regression column these are based on the heteroskedasticity robust standard errors. Controls in regressions include log income, sex, age, education, number of non-disabled people in houseshold aged 11–74, and village fixed effects. Days sick - number of days non-disabled people in household aged 11–74 couldn’t work due to illness. Each regression additionally controls for social preferences in the previous round. In the top panel the amount sent as trustor and the average share returned as trustee in 2002 are included as control variables. In the second panel, the 2007 version of the 2009 outcome is included as a control variable.
Footnotes
We thank Jennifer Alix-Garcia, Glenn Harrison, Jessica Hoel, Dean Karlan, Ethan Ligon, John Liu, Hendrik Wolff, and seminar participants at IFPRI, MIEDC, UIUC, University of Washington, and UW Madison for helpful comments. Funding for data collection came from the Russell Sage Foundation, UC Berkeley’s IBER, and the UW Graduate School.
We do not look at correlations of preferences across domains. Nevertheless, an advantage of looking at all three domains (risk, time, and social preferences) with one subject pool is that it slightly lessens the potential for data-mining, picking and choosing the subject pools or the preferences which are or are not significantly correlated.
Jamison et al. (2012) conduct a nice, more general, overview of the literature on the measurement of risk and time preferences, including experimental design, incentivized versus hypothetical experiments, correlations with real-world outcomes, and stability over time.
Some welcome exceptions which do discuss attrition include Andersen et al. (2008), Beauchamp et al. (2012), Meier and Sprenger (2015), and Wölbert and Riedl (2013).
One of the authors has collected data from 17 Wisconsin farmers over 2 years and finds a significant correlation of 0.46 when looking at number of risky choices but an insignificant correlation of 0.11 when looking at coefficients of relative risk aversion. See Barham et al. (2014) for information about the first round of data collected.
There are also papers which look at the stability of prospect theory risk preferences including Baucells and Villasís (2010) (141 students over three months, correlation of 0.32), Glöckner and Pachur (2012) (64 students over one week, correlations between 0.22 and 0.59), and Zeisberger et al. (2012) (73 German undergraduates over one month).
McLeish and Oxoby (2007) (86 Canadian students over seven weeks) and Halevy (2015) (117 Canadian students over four weeks) seem to have the data available to test stability of time preferences, but do not do so.
Two additional papers, Volk et al. (2012), and Sass and Weimann (2012), study the aggregate decay of cooperation over time and/or categorize individuals into different types based on their play and look at the stability of their type over time. It is not clear whether or not these two papers find evidence of stable decisions in the same game repeated over time at the individual level.
Guiso et al. (2011) do find their entire sample to become more risk averse after the financial crisis, but find no time-varying variables at the individual level which impact levels of risk aversion.
Alesina and La Ferrara (2002) and Malmendier and Nagel (2011) use survey rather than experimental data and find that individuals who have experienced economic shocks are less trusting and less willing to take risks respectively.
Castillo and Carter (2011) find non-linear effects of weather shocks on trust and trustworthiness, with small shocks increasing both but larger shocks decreasing both.
There was a sixteenth community which was not chosen randomly and which consisted of Japanese farmers. After the 2002 survey these households were dropped and their data is not analyzed here.
Of course, even if one could measure preferences over time without attrition in a university lab, there is selection in terms of what type of person signs up to participate in university experiments. In this case, we start out with a more representative sample of the Paraguayan countryside.
The only exception is that we do not have a measure of income in 2010 and so we use 2007 income.
We have analyzed this data, and most of the time the hypothetical social preference data are not stable over time.
Of course, dividing the sample in half can lead to lower precision. For the 2002–2007 risk aversion comparison, the heteroskedasticity robust standard error on the regression coefficient goes up from 0.09 overall to 0.13 for the more educated individuals and 0.15 for the less educated. For the 2007–2009 risk aversion comparison, the standard error goes from 0.10 overall to 0.15 and 0.14 for the more and less educated.
Not all questions were asked in all years.
There are only 14 individuals who participated in the 2002, 2007, and 2009 surveys.
This difference in stated trust in people in general compared to people in the village cannot be caused by the experiment, since the surveys took place one-on-one in people’s homes while the experiment took place at a central location a few days later.
These games are a fairly unusual and memorable experience for the participants. One author particularly remembers a participant who gave everything in the trust game in 2002 and was matched with a partner who took all of it. When we returned in 2007 he approached the author and told her that he remembered how badly things had turned out last time and that he had learned his lesson.
Since we tripled the money sent by the trustor, he could potentially receive back anywhere between none of the money he sent and three times the money he sent.
In a previous draft of the paper we ran this analysis without controlling for preferences in the previous round and the results are similarly unexciting. The advantage of not controlling for previous measures of preferences is that sample sizes are larger because we can then include any individual whose household appeared in the previous round, no matter which person responded in that previous round. We have also tried regressing changes in preferences on changes in income or theft, and similarly find no strong results.
The use of the change in the amount stolen (rather than the level) may allay some concerns regarding the endogeneity of theft. Some people tend to experience more theft than others. Here we explore whether experiencing a particularly good or particularly bad year might impact their preferences. We have also tried using the change in a binary indicator for whether or not the household experienced theft, as well as including binary indicators for experiencing more or less theft in the later year, and the results are similar. Likewise, we have tried using a dummy for missing at least one day of work rather than the number of days missed. The results are similar with slightly more evidence that household health shocks lead individuals to state they trust their village-mates and neighbors more and less evidence that these shocks lead them to give less in experiments.
Contributor Information
Yating Chuang, UW Madison.
Laura Schechter, UW Madison.
References
- Ackert LF, Gillette AB, Martinez-Vazquez J, Rider M. Are benevolent dictators altruistic in groups? A within-subject design. Experimental Economics. 2011;14(3):307–321. [Google Scholar]
- Afzal U, Turner G, Said F. Natural disasters and social capital: Evidence from a field experiment in Pakistan. 2015 Unpublished Manuscript. [Google Scholar]
- Alesina A, La Ferrara E. Who trusts others? Journal of Public Economics. 2002;85(2):207–234. [Google Scholar]
- Andersen S, Harrison G, Lau M, Rutström E. Lost in state space: Are preferences stable? International Economic Review. 2008;49(3):1091–1112. [Google Scholar]
- Anderson L, Mellor J. Are risk preferences stable? Comparing an experimental measure with a validated survey-based measure. Journal of Risk and Uncertainty. 2009;39(2):137–160. [Google Scholar]
- Anderson ML. Multiple inference and gender differences in the effects of early intervention: A reevalaution of the Abecedarian, Perry Preschool, and Early Training Projects. Journal of the American Statistical Association. 2008;103(484):1481–1495. [Google Scholar]
- Andrabi T, Das J. In aid we trust: Hearts and minds and the Pakistan earthquake of 2005. 2010 Unpublished Manuscript. [Google Scholar]
- Andreoni J, Harbaugh W, Vesterlund L. The carrot or the stick: Rewards, punishments, and cooperation. American Economic Review. 2003;93(2):893–902. [Google Scholar]
- Andreoni J, Miller J. Giving according to GARP: An experimental test of the consistency of preferences for altruism. Econometrica. 2002;70(2):737–753. [Google Scholar]
- Barham BL, Chavas J-P, Fitz D, Ríos Salas V, Schechter L. The roles of risk and ambiguity in technology adoption. Journal of Economic Behavior & Organization. 2014;97:204–218. [Google Scholar]
- Baucells M, Villasís A. Stability of risk preferences and the reflection effect of prospect theory. Theory and Decision. 2010;68(1):193–211. [Google Scholar]
- Bauer M, Cassar A, Chytilová J, Henrich J. War’s enduring effects on the development of egalitarian motivations and in-group biases. Psychological Science. 2014;25(1):47–57. doi: 10.1177/0956797613493444. [DOI] [PubMed] [Google Scholar]
- Bchir MA, Willinger M. Does the exposure to natural hazards affect risk and time preferences? Some insights from a field experiment in Perú. 2013 Unpublished Manuscript. [Google Scholar]
- Beauchamp J, Cesarini D, Johannesson M. The psychometric properties of measures of economic risk preferences. 2012 Unpublished Manuscript. [Google Scholar]
- Becchetti L, Castriota S, Conzo P. Calamity, aid and indirect reciprocity: The long run impact of tsunami on altruism. 2012 Unpublished Manuscript. [Google Scholar]
- Becchetti L, Conzo P, Romeo A. Violence, trust, and trustworthiness: Evidence from a Nairobi slum. Oxford Economic Papers. 2014;66(1):283–305. [Google Scholar]
- Benjamin DJ, Brown SA, Shapiro JM. Who is ‘behavioral’? Cognitive ability and anomalous preferences. Journal of the European Economic Association. 2013;11(6):1231–1255. doi: 10.1111/jeea.12055. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Benjamini Y, Hochberg Y. Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society, Series B. 1995;57(1):289300. [Google Scholar]
- Berg J, Dickhaut J, McCabe K. Trust, reciprocity, and social history. Games and Economic Behavior. 1995;10:122–142. [Google Scholar]
- Berg J, Dickhaut J, McCabe K. Risk preference instability across institutions: A dilemma. Proceedings of the National Academy of Sciences of the United States of America. 2005;102(11):4209–4214. doi: 10.1073/pnas.0500333102. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bhargava S, Loewenstein G, Sydnor J. Do individuals make sensible health insurance decisions? Evidence from a menu with domintated options. 2015 Unpublished Manuscript. [Google Scholar]
- Binswanger HP. Attitudes toward risk: Experimental measurement in rural India. American Journal of Agricultural Economics. 1980;62(3):395–407. [Google Scholar]
- Blanco M, Engelmann D, Normann HT. A within-subject analysis of other-regarding preferences. Games and Economic Behavior. 2011;72(2):321–338. [Google Scholar]
- Booth A, Cardona-Sosa L, Nolen P. Gender differences in risk aversion: Do single-sex environments affect their development? Journal of Economic Behavior and Organization. 2014;99:126–154. [Google Scholar]
- Brosig J, Riechmann T, Weimann J. Selfish in the end? An investigation of consistency and stability of individual behavior. 2007 Unpublished Manuscript. [Google Scholar]
- Brunnermeier M, Nagel S. Do wealth fluctuations generate time-varying risk aversion? Micro-evidence on individuals’ asset allocation. American Economic Review. 2008;98(3):713–36. [Google Scholar]
- Burks S, Carpenter J, Götte L, Rustichini A. Which measures of time preference best predict outcomes? Journal of Economic Behavior & Organization. 2012;84(1):308–320. [Google Scholar]
- Burks SV, Carpenter Jp, Götte L, Rustichini A. Cognitive skills affect economic preferences, strategic behavior, and job attachment. Proceedings of the National Academy of Sciences of the United States of America. 2009;106(19):7745–7750. doi: 10.1073/pnas.0812360106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Cai J, Song C. Do hypothetical experiences affect real financial decisions? Evidence from insurance take-up. 2014 Unpublished Manuscript. [Google Scholar]
- Callen M. Catastrophes and time preference: Evidence from the Indian Ocean earthquake. 2011 Unpublished Manuscript. [Google Scholar]
- Callen M, Isaqzadeh M, Long JD, Sprenger C. Violent trauma and risk preference: Experimental evidence from Afghanistan. American Economic Review. 2014;104(1):123–148. [Google Scholar]
- Cameron AC, Gelbach JB, Miller DL. Bootstrap-based improvements for inference with clustered errors. Review of Economics and Statistics. 2008;90(3):414–427. [Google Scholar]
- Cameron L, Shah M. Risk-taking behavior in the wake of natural disasters. Journal of Human Resources. 2015 Forthcoming. [Google Scholar]
- Carlsson F, Johansson-Stenman O, Nam PK. Social preferences are stable over long periods of time. Journal of Public Economics. 2014;117:104–114. [Google Scholar]
- Caruthers H. Risk management and rural to urban migration decisions in Indonesia. 2013 Unpublished Manuscript. [Google Scholar]
- Carvalho LS, Meier S, Wang SW. Poverty and economic decision-making: Evidence from changes in financial resources at payday. 2014 doi: 10.1257/aer.20140481. Unpublished Manuscript. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Carvalho LS, Prina S, Sydnor J. The effect of saving on risk attitudes and intertemporal choices. 2014 Unpublished Manuscript. [Google Scholar]
- Cassar A, Grosjean P, Whitt S. Legacies of violence: Trust and market development. Journal of Economic Geography. 2013;18(3):285–318. [Google Scholar]
- Cassar A, Healy A, von Kessler C. Trust, risk and time preferences after a natural disaster: Experimental evidence from Thailand. 2011 Unpublished Manuscript. [Google Scholar]
- Castillo M, Carter M. Behavioral responses to natural disasters. 2011 Unpublished Manuscript. [Google Scholar]
- Chantarat S, Chheng K, Minea K, Oum S, Samphantharak K, Sann V. The effects of natural disasters on households’ preferences and behaviours: Evidence from Cambodian rice farmers after the 2011 mega flood. In: Sawada Y, Oum S, editors. Disaster Risks, Social Preferences, and Policy Effects: Field Experiments in Selected ASEAN and East Asian Coutries. Chapter 4. 2015. pp. 85–130. ERIA Research Project Report 2013–34. [Google Scholar]
- Charness G, Viceisza A. Three risk-elicitation methods in the field: Evidence from rural Senegal. 2015 Unpublished Manuscript. [Google Scholar]
- Chiappori P-A, Paiella M. Relative risk aversion is constant: Evidence from panel data. Journal of the European Economics Association. 2011;9(6):1021–1052. [Google Scholar]
- Choi S, Fisman R, Gale D, Kariv S. Consistency and heterogeneity of individual behavior under uncertainty. American Economic Review. 2007;97(5):1921–1938. [Google Scholar]
- Choi S, Kariv S, Müller W, Silverman D. Who is (more) rational? American Economic Review. 2014;104(6):1518–1550. [Google Scholar]
- Conte A, Levati MV, Montinari N. Experience in public goods experiments. 2014 Unpublished Manuscript. [Google Scholar]
- Cook J. Confusion in risk aversion experiments in low-income countries. 2015 Unpublished Manuscript. [Google Scholar]
- Dasgupta U, Gangadharan L, Maitra P, Mani S. De gustibus non est disputandum: An experimental investigation. 2014 Unpublished Manuscript. [Google Scholar]
- Dave C, Eckel C, Johnson C, Rojas C. Eliciting risk preferences: When is simple better? Journal of Risk and Uncertainty. 2010;41(3):219–243. [Google Scholar]
- De Luca G, Verpoorten M. Civil war, social capital and resilience in Uganda. Oxford Economic Papers. 2015;67(3):661–686. [Google Scholar]
- de Oliveira A, Croson R, Eckel C. The stability of social preferences in a low-income neighborhood. Southern Economic Journal. 2012;79(1):15–45. [Google Scholar]
- Dean M, Sautmann A. Credit constraints and the measurement of time preferences. 2014 Unpublished Manuscript. [Google Scholar]
- Dohmen T, Falk A, Huffman D, Sunde U, Schupp J, Wagner G. Individual risk attitudes: Measurement, determinants, and behavioral consequences. Journal of the European Economic Association. 2011;9(3):522–550. [Google Scholar]
- Dulleck U, Fooken J, Fell J. Within-subject intra- and inter-method consistency of two experimental risk attitude elicitation methods. German Economic Review. 2015 Forthcoming. [Google Scholar]
- Eckel C, El-Gamal M, Wilson R. Risk loving after the storm: A Bayesian-network study of Hurricane Katrina evacuees. Journal of Economic Behavior & Organization. 2009;69(2):110–124. [Google Scholar]
- Eckel C, Wilson R. Is trust a risky decision? Journal of Economic Behavior & Organization. 2004;55(4):447–465. [Google Scholar]
- Fehr E. On the economics and biology of trust. Journal of the European Economic Association. 2009;7(2–3):235266. [Google Scholar]
- Finan F, Schechter L. Vote-buying and reciprocity. Econometrica. 2012;80(2):863–881. [Google Scholar]
- Fisman R, Jakiela P, Kariv S. How did distributional preferences change during the Great Recession? 2014 Unpublished Manuscript. [Google Scholar]
- Fisman R, Kariv S, Markovits D. Individual preferences for giving. American Economic Review. 2007;97(5):1858–1876. [Google Scholar]
- Fleming DA, Chong A, Bejarano H. Trust and reciprocity in the aftermath of natural disasters. Journal of Development Studies. 2014;50(11):1482–1493. [Google Scholar]
- Gerrans P, Faff R, Hartnett N. Individual financial risk tolerance and the global financial crisis. Accounting & Finance. 2015;55(1):165–185. [Google Scholar]
- Giné X, Goldberg J, Silverman D, Yang D. Revising commitments: Field evidence on the adjustment of prior choices. 2014 Unpublished Manuscript. [Google Scholar]
- Glöckner A, Pachur T. Cognitive models of risky choice: Parameter stability and predictive accuracy of prospect theory. Cognition. 2012;123(1):21–32. doi: 10.1016/j.cognition.2011.12.002. [DOI] [PubMed] [Google Scholar]
- Gneezy U, List JA, Wu G. The uncertainty effect: When a risky prospect is valued less than its worst possible outcome. Quarterly Journal of Economics. 2006;121(4):1283–1309. [Google Scholar]
- Goldstein DG, Johnson EJ, Sharpe WF. Choosing outcomes versus choosing products: Consumer-focused retirement investment advice. Journal of Consumer Research. 2008;35(3):440–456. [Google Scholar]
- Grosjean P. Conflict and social and political preferences: Evidence from World War II and civil conflict in 35 European countries. Comparative Economic Studies. 2014;56(3):424–451. [Google Scholar]
- Guiso L, Sapienza P, Zingales L. Time varying risk aversion. 2011 Unpublished Manuscript. [Google Scholar]
- Halevy Y. Time consistency: Stationarity and time invariance. Econometrica. 2015 Forthcoming. [Google Scholar]
- Hamoudi A. Risk preferences in households and families. 2006 Unpublished Manuscript. [Google Scholar]
- Hanaoka C, Shigeoka H, Watanabe Y. Do risk preferences change? Evidence from panel data before and after the Great East Japan Earthquake. 2014 Unpublished Manuscript. [Google Scholar]
- Harrison GW, Johnson E, McInnes MM, Rutström EE. Temporal stability of estimates of risk aversion. Applied Financial Economics Letters. 2005;1(1):31–35. [Google Scholar]
- Harrison GW, Lau MI, Rutström EE. Dynamic consistency in Denmark: A longitudinal field experiment. 2006 Unpublished Manuscript. [Google Scholar]
- He T-S, Hong F. Exposure to risk and risk aversion: A laboratory experiment. 2014 Unpublished Manuscript. [Google Scholar]
- Hey J. Does repetition improve consistency? Experimental Economics. 2001;4(1):5–54. [Google Scholar]
- Horowitz J. A test of intertemporal consistency. Journal of Economic Behavior & Organization. 1992;17(1):171–182. [Google Scholar]
- Ingwersen N. Impact of a natural disaster on observed risk aversion. 2014 Unpublished Manuscript. [Google Scholar]
- Isaac R, James D. Just who are you calling risk averse? Journal of Risk and Uncertainty. 2000;20(2):177–187. [Google Scholar]
- Jacobson S, Petrie R. Learning from mistakes: What do inconsistent choices over risk tell us? Journal of Risk and Uncertainty. 2009;38:143–158. [Google Scholar]
- Jakiela P, Miguel E, te Velde VL. You’ve earned it: Estimating the impact of human capital on social preferences. Experimental Economics. 2015 Forthcoming. [Google Scholar]
- Jamison J, Karlan D, Schechter L. To deceive or not to deceive: The effect of deception on behavior in future laboratory experiments. Journal of Economic Behavior & Organization. 2008;68(3–4):477–488. [Google Scholar]
- Jamison J, Karlan D, Zinman J. Measuring risk and time preferences and their connections with behavior. 2012 Unpublished Manuscript. [Google Scholar]
- Kandasamy N, Hardy B, Page L, Schaffner M, Graggaber J, Powlson AS, Fletcher PC, Gurnell M, Coates J. Cortisol shifts financial risk preferences. Proceedings of the National Academy of Sciences. 2014;111(9):3608–3613. doi: 10.1073/pnas.1317908111. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Keren G, Willemsen MC. Decision anomalies, experimenter assumptions, and participants’ comprehension: Revaluating the uncertainty effect. Journal of Behavioral Decision Making. 2009;22(3):301–317. [Google Scholar]
- Kim Y-I, Lee J. The long-run impact of a traumatic experience on risk aversion. Journal of Economic Behavior and Organization. 2014;108:174–186. [Google Scholar]
- Kimball MS, Sahm CR, Shapiro MD. Imputing risk tolerance from survey responses. Journal of the American Statistical Association. 2008;103(483):1028–1038. doi: 10.1198/016214508000000139. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kirby KN. One-year temporal stability of delay-discount rates. Psychonomic Bulletin & Review. 2009;16(3):457–462. doi: 10.3758/PBR.16.3.457. [DOI] [PubMed] [Google Scholar]
- Kirby KN, Godoy R, Reyes-García V, Byron E, Apaza L, Leonard W, Pérez E, Vadez V, Wilkie D. Correlates of delay-discount rates: Evidence from Tsimane’ Amerindians of the Bolivian rain forest. Journal of Economic Psychology. 2002;23:291–316. [Google Scholar]
- Krupka EL, Stephens M., Jr The stability of measured time preferences. Journal of Economic Behavior & Organization. 2013;85:11–19. [Google Scholar]
- Kruse J, Thompson M. Valuing low probability risk: Survey and experimental evidence. Journal of Economic Behavior & Organization. 2003;50(4):495–505. [Google Scholar]
- Levin IP, Hart SS, Weller JA, Harshman LA. Stability of choices in a risky decision-making task: A 3-year longitudinal study with children and adults. Journal of Behavioral Decision Making. 2007;30:241–252. [Google Scholar]
- Li Y, Baldassi M, Johnson EJ, Weber EU. Complementary cognitive capabilities, economic decision making, and aging. Psychology and Aging. 2013;28(3):595–613. doi: 10.1037/a0034172. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ligon E, Schechter L. Motives for sharing in social networks. Journal of Development Economics. 2012;99(1):13–26. [Google Scholar]
- Loewenstein G, Angner E. Predicting and indulging changing preferences. In: Loewenstein G, Read D, Baumeister RF, editors. Time and Decision: Economic and Psychological Perspectives on Intertemporal Choice. Chapter 12. New York: Russell Sage Foundation Press; 2003. pp. 351–391. [Google Scholar]
- Lönnqvist J-E, Verkasalo M, Walkowitz G, Wichardt PC. Measuring individual risk attitudes in the lab: Task or ask? An empirical comparison. 2014 Unpublished Manuscript. [Google Scholar]
- Love RO, Robison J. An empirical analysis of the intertemporal stability of risk preference. Southern Journal of Agricultural Economics. 1984;16(1):159–66. [Google Scholar]
- Lührmann M, Serra-Garcia M, Winter JK. The impact of financial education on adolescents’ intertemporal choices. 2015 Unpublished Manuscript. [Google Scholar]
- Malmendier U, Nagel S. Depression babies: Do macroeconomic experiences affect risk taking? The Quarterly Journal of Economics. 2011;126(1):373–416. [Google Scholar]
- Matthey A, Regner T. On the independence of history: Experience spill-overs between experiments. Theory and Decision. 2013;75:403–419. [Google Scholar]
- McLeish KN, Oxoby RJ. Gender, affect and intertemporal consistency: An experimental approach. 2007 Unpublished Manuscript. [Google Scholar]
- Meier S, Sprenger CD. Temporal stability of time preferences. Review of Economics and Statistics. 2015 Forthcoming. [Google Scholar]
- Mengel F, Tsakas E, Vostroknutov A. Past experience of uncertainty affects risk aversion. Experimental Economics. 2015 Forthcoming. [Google Scholar]
- Menkhoff L, Sakha S. Risk aversion over time: Experimental evidence in rural Thailand. 2014 Unpublished Manuscript. [Google Scholar]
- Moya A. Violence, mental trauma, and induced changes in risk attitudes among the internally displaced population in Colombia. 2011 Unpublished Manuscript. [Google Scholar]
- Necker S, Ziegelmeyer M. Household risk taking after the financial crisis. The Quarterly Review of Economics and Finance. 2015 Forthcoming. [Google Scholar]
- Page L, Savage DA, Torgler B. Variation in risk seeking behaviour following large losses: A natural experiment. European Economic Review. 2014;71:121–131. [Google Scholar]
- Reuben E, Sapienza P, Zingales L. Time discounting for primary and monetary rewards. Economics Letters. 2010;106(2):125–127. [Google Scholar]
- Reynaud A, Couture S. Stability of risk preference measures: Results from a field experiment on French farmers. Theory and Decision. 2012;73(2):203–221. [Google Scholar]
- Rydval O, Ortmann A, Prokosheva S, Hertwig R. How certain is the uncertainty effect? Experimental Economics. 2009;12(4):473487. [Google Scholar]
- Sahm C. How much does risk tolerance change? Quarterly Journal of Finance. 2012;2(4) doi: 10.1142/S2010139212500206. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Said F, Afzal U, Turner G. Risk taking and risk learning after a rare event: Evidence from a field experiment in Pakistan. Journal of Economic Behavior & Organization. 2015 Forthcoming. [Google Scholar]
- Samphantharak K, Chantarat S. The effects of natural disasters on households’ preferences and behaviours: Evidence from Thai farmers during and after the 2011 mega flood. In: Sawada Y, Oum S, editors. Disaster Risks, Social Preferences, and Policy Effects: Field Experiments in Selected ASEAN and East Asian Coutries. Chapter 2. 2015. pp. 57–84. ERIA Research Project Report 2013–34. [Google Scholar]
- Sapienza P, Toldra A, Zingales L. Understanding trust. Economic Journal. 2013;123(573):13131332. [Google Scholar]
- Sass M, Weimann J. The dynamics of individual preferences in repeated public good experiments. 2012 Unpublished Manuscript. [Google Scholar]
- Sawada Y, Kuroishi Y. How does a natural disaster affect people’s preference? The case of a large scale flood in the Philippines using the convex time budget experiments. In: Sawada Y, Oum S, editors. Disaster Risks, Social Preferences, and Policy Effects: Field Experiments in Selected ASEAN and East Asian Coutries. Chapter 3. 2015. pp. 57–84. ERIA Research Project Report 2013–34. [Google Scholar]
- Schechter L. Theft, gift-giving, and trustworthiness: Honesty is its own reward in rural Paraguay. American Economic Review. 2007a;97(5):1560–1582. [Google Scholar]
- Schechter L. Traditional trust measurement and the risk confound: An experiment in rural Paraguay. Journal of Economic Behavior & Organization. 2007b;62(2):272–292. [Google Scholar]
- Schoemaker PJH, Hershey JC. Utility measurement: Signal, noise, and bias. Organizational Behavior and Human Decision Processes. 1992;52:397–424. [Google Scholar]
- Schofield H. The economic costs of low caloric intake: Evidence from India. 2014 Unpublished Manuscript. [Google Scholar]
- Shah AK, Mullainathan S, Shafir E. Some consequences of having too little. Science. 2012;338(6107):682–685. doi: 10.1126/science.1222426. [DOI] [PubMed] [Google Scholar]
- Silva AS, Mace R. Cooperation and conflict: Field experiments in Northern Ireland. Proceedings of the Royal Society of London B: Biological Sciences. 2014;281(1792) doi: 10.1098/rspb.2014.1435. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Slovic P. Information processing, situation specificity, and the generality of risk-taking behavior. Journal of Personality and Social Psychology. 1972;22(1):128–134. [Google Scholar]
- Smidts A. The relationship between risk attitude and strength of preference: A test of intrinsic risk attitude. Management Science. 1997;43(3):357–370. [Google Scholar]
- Ubfal D. How general are time preferences? Eliciting good-specic discount rates. 2014 Unpublished Manuscript. [Google Scholar]
- van den Berg M, Fort R, Burger K. Natural hazards and risk aversion: Experimental evidence from Latin America. 2009 Unpublished Manuscript. [Google Scholar]
- Vlaev I, Chater N, Stewart N. Dimensionality of risk perception: Factors affecting consumer understanding and evaluation of financial risk. Journal of Behavioral Finance. 2009;10(3):158–181. [Google Scholar]
- Volk S, Thöni C, Ruigrok W. Temporal stability and psychological foundations of cooperation preferences. Journal of Economic Behavior & Organization. 2012;81(2):664–676. [Google Scholar]
- Voors M, Nillesen E, Verwimp P, Bulte E, Lensink R, van Soest D. Violent conflict and behavior: A field experiment in Burundi. American Economic Review. 2012;102(2):941–964. [Google Scholar]
- Wehrung D, MacCrimmon K, Brothers K. Utility assessment: Domains, stability, and equivalence procedures. INFOR. Information Systems and Operational Research. 1984;22(2):98–115. [Google Scholar]
- Willinger M, Bchir MA, Heitz C. Risk and time preferences under the threat of background risk: A case-study of lahars risk in central Java. 2013 Unpublished Manuscript. [Google Scholar]
- Wölbert E, Riedl A. Measuring time and risk preferences: Reliability, stability, domain specificity. 2013 Unpublished Manuscript. [Google Scholar]
- Zeisberger S, Vrecko D, Langer T. Measuring the time stability of prospect theory preferences. Theory and Decision. 2012;72(3):359–386. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.