Skip to main content
Frontiers in Psychology logoLink to Frontiers in Psychology
. 2019 Apr 5;10:342. doi: 10.3389/fpsyg.2019.00342

How to Measure Behavioral Spillovers: A Methodological Review and Checklist

Matteo M Galizzi 1,*, Lorraine Whitmarsh 2
PMCID: PMC6460990  PMID: 31024368

Abstract

A growing stream of literature at the interface between economics and psychology is currently investigating ‘behavioral spillovers’ in (and across) different domains, including health, environmental, and pro-social behaviors. A variety of empirical methods have been used to measure behavioral spillovers to date, from qualitative self-reports to statistical/econometric analyses, from online and lab experiments to field experiments. The aim of this paper is to critically review the main experimental and non-experimental methods to measure behavioral spillovers to date, and to discuss their methodological strengths and weaknesses. A consensus mixed-method approach is then discussed which uses between-subjects randomization and behavioral observations together with qualitative self-reports in a longitudinal design in order to follow up subjects over time. In particular, participants to an experiment are randomly assigned to a treatment group where a behavioral intervention takes place to target behavior 1, or to a control group where behavior 1 takes place absent any behavioral intervention. A behavioral spillover is empirically identified as the effect of the behavioral intervention in the treatment group on a subsequent, not targeted, behavior 2, compared to the corresponding change in behavior 2 in the control group. Unexpected spillovers and additional insights (e.g., drivers, barriers, mechanisms) are elicited through analysis of qualitative data. In the spirit of the pre-analysis plan, a systematic checklist is finally proposed to guide researchers and policy-makers through the main stages and features of the study design in order to rigorously test and identify behavioral spillovers, and to favor transparency, replicability, and meta-analysis of studies.

Keywords: spillovers, mixed-methods, experimental design, lab-field experiments, behavioral spillovers

Introduction

What Does Spillover Offer?

Academic and policy interest in ‘behavioral spillover’ has grown considerably in recent years (e.g., Austin et al., 2011; Truelove et al., 2014; Nilsson et al., 2016). Spillover is where the adoption of one behavior causes the adoption of additional, related behaviors. As we discuss below, we assume that the initial behavior change is due to an intervention, although other definitions of behavioral spillovers do not assume this (Nash et al., 2017). From a policy or practitioner perspective, the notion of behavioral spillover is attractive because it appears to hold the promise of changing a suite of behaviors in a cost-effective manner with little regulation which might be politically unpopular. For many pressing social issues, such as climate change or obesity, spillover is thus a promising method of achieving the scale of lifestyle change required to address these, in contrast to the typically small-scale behavioral changes achieved from most individually focussed interventions (Capstick et al., 2014). From an academic perspective, spillover is intriguing because it sheds new light on the process of lifestyle change: rather than examining behavior change from the perspective of individual behaviors in isolation, spillover draws attention to the holistic relationships between behaviors within and between contexts, and hence refocus the researchers’ perspective on the complex behavioral ecologies that represent lifestyles (Geller, 2001; Schatzki, 2010).

A variety of empirical methods have been used to measure behavioral spillovers to date, from qualitative self-reports to statistical/econometric analyses, from online and lab experiments to field experiments. Detecting spillover has often proved challenging, and there is a need for both conceptual and methodological clarity in order to move the field forward. The aim of this paper is to critically review the main experimental and non-experimental methods to measure behavioral spillovers to date, and to discuss their methodological strengths and weaknesses. A consensus mixed-method approach is then discussed which uses between-subjects randomization and behavioral observations together with qualitative self-reports in a longitudinal design in order to follow up subjects over time. We conclude by proposing a systematic checklist to guide researchers and policy-makers through the main stages and features of the study design in order to rigorously test and identify behavioral spillovers, and to favor transparency, replicability, and meta-analysis of studies.

Definition of Behavioral Spillover

The term ‘spillover’ has been applied to a wide variety of phenomena, including the spread of knowledge, attitudes, roles/identities, or behaviors from a given domain (e.g., health, environment, care-giving), group, or location, to a different domain, group or location (e.g., Geller, 2001; Poortinga et al., 2013; Littleford et al., 2014; Rodriguez-Muñoz et al., 2014; Poroli and Huang, 2018). The main appeal of such broad definition of behavioral spillover is that it encompasses a rich variety of spillover effects at both a micro and a macro level which are of key interest for policy and practice purposes, such as cross-domains, inter-personal, and cross-regional spillover effects of phenomena and interventions. However, the processes underpinning these diverse effects are highly heterogeneous, ranging from cognition (e.g., learning, problem-solving) and self-regulation, through interpersonal effects (e.g., modeling, contagion) to individual behavior change, and there is little these processes have in common besides the idea of (often unanticipated) diffusion of some effect.

In what follows, we assume a narrower and more specific definition of behavioral spillover that matches more closely the methodological approach that we have in mind. In particular, behavioral spillover can be defined as the observable and causal effect that a change in one behavior (behavior 1) has on a different, subsequent behavior (behavior 2). More specifically, to constitute behavioral spillover, the two behaviors must be different (i.e., not related components of a single behavior), sequential (i.e., behavior 2 follows behavior 1), and sharing, at a conscious or unconscious way, an underlying motive (i.e., an overarching goal or a ‘deep preference,’ such as, for example, pro-environmentalism or a healthy life) (Dolan and Galizzi, 2015; Nash et al., 2017). This concept of spillover has been examined in relation to different domains (safety, environment, health, finances, etc.) for some decades, although these effects have previously been labeled in diverse ways, including ‘response generalization’ (Ludwig and Geller, 1997; Geller, 2001), ‘the foot in the door effect’ (Freedman and Fraser, 1966; Beaman et al., 1983), and ‘moral licensing’ (Blanken et al., 2015; Mullen and Monin, 2016). We have conducted a systematic review of the literature (see Appendix for full details) and found that a total of 106 studies to date have used the above, more specific, definition of behavioral spillovers.1

Behavioral spillovers can be categorized as ‘promoting,’ ‘permitting,’ ‘purging,’ or ‘precipitating,’ as illustrated in Table 1.

Table 1.

Types of behavioral spillovers (adapted from Dolan and Galizzi, 2015: no copyright permissions are required for the reproduction of this table): examples from health behavior.

Behavior 2
Behavior 1 Eat healthily Eat less healthily
A run after work Promoting Permitting
I ran an hour, let’s keep up the good work I ran an hour, I deserve a big slice of cake
Sofa-sitting after work Purging Precipitating
I’ve been lazy today, best not eat so much tonight I’ve been lazy today, so, what the heck, let’s have a big slice of cake

Other real world examples from environmental behavior are whether a behavioral intervention to monetarily incentivize household waste separation has a significant effect not just on waste separation (behavior 1), but also on green shopping, traveling, and support to environmental policies (behavior 2), for instance (Xu et al., 2018a); or whether an intervention to restrict irrigation has a significant impact not just on water conservation (behavior 1), but also on recycling behavior (behavior 2), for example (Sintov et al., 2019).

The mechanisms thought to explain promoting or positive spillovers vary by discipline and theoretical framework. Psychological approaches have focussed particularly on two mechanisms: (a) self-perception, identity, or preference for consistency (behavior 1 changes how one sees oneself and the desire to act consistently with that self-image leads to behavior 2) and (b) self-efficacy, knowledge, or self-motivation/empowerment (satisfactorily undertaking behavior 1 increases confidence and perceived efficacy of action, motivating change in behavior 2; Nash et al., 2017). Permitting or negative spillovers have been typically explained in terms of moral licensing, whereby a virtuous initial behavior licenses or ‘permits’ a second indulgent or morally questionable behavior, or by a contribution ethic whereby an initial behavior justifies subsequent inaction (e.g., Thøgersen, 1999; Karmarkar and Bollinger, 2015). Rebound effects are a related phenomenon, studied more from an economic than psychological perspective, and describe increased energy consumption due to technical efficiency gains, thereby offsetting energy savings achieved (e.g., Sorrell et al., 2009).

Evidence for spillover remains somewhat mixed, with some studies finding effects under certain conditions that are not replicated in other studies (Nash et al., 2017). Conceptually, spillover remains defined and explained in a variety of ways, and there remain considerable gaps in understanding (e.g., the role of social processes, such as norms, in spillover; Nash et al., 2017). Methodologically, there is also no coherent approach to researching spillover, which may in part explain the mixed and inconsistent empirical results, and critically highlights a need to improve the rigor and transparency of spillover research.

Overview of Spillover Research Methods and Measurement

A growing stream of the literature at the interface between economics and psychology is currently investigating ‘behavioral spillovers’ in (and across) different domains, including health, environmental, and pro-social behaviors. To date, there have been a variety of methods applied to studying spillover (see Table 2). These range from qualitative retrospective self-reports using biographical interviews (e.g., Nash et al., 2019) to controlled laboratory experiments with randomization to condition (e.g., Van der Werff et al., 2014a,b). Each approach offers different strengths and weaknesses. For example, qualitative approaches are able to elucidate unexpected spillovers and additional insights (e.g., drivers, barriers, mechanisms) not anticipated or measured in quantitative approaches. On the other hand, quantitative approaches allow for more measurement standardization and potentially for generalization, as well as affording insights into factors shaping behavior that individuals may be unable or unwilling to reflect on consciously through self-report.

Table 2.

Overview of methods used to research behavioral spillover: examples from environmental behavior.

Methodological approach Data collection and analysis methods Examples from environmental behavior Strengths Weaknesses
Qualitative • Interviews or open-ended survey questions Austin et al., 2011; Boström et al., 2015; Nash et al., 2019; Uzzell and Räthzel, 2018 • Expose unexpected spillovers • Risk of presentational bias
• Thematic, content, discourse (or similar) analysis • Shed light on spillover mechanisms, drivers and barriers • Partial or selective recollection • No measurement standardization
• Self-reports or other (e.g., practitioner) accounts
• Biographical (retrospective) or evaluative (during/immediately after intervention)
Quantitative (cross-sectional) • Survey, card sort or secondary data analysis (e.g., retail data) Thøgersen, 1999; Barr et al., 2010; Whitmarsh and O’Neill, 2010; Austin et al., 2011; Gabe-Thomas et al., 2016 • Quantify strength of relationships between measured behaviors • No causal relationships identified
• Cluster or factor analysis • Measurement standardization • Limited to expected spillovers
• Correlational analysis
• Regression analysis
Quantitative (longitudinal) • Surveys at 2+ timepoints Thøgersen and Noblet, 2012; Kaida and Kaida, 2015; Poortinga et al., 2013; Thomas et al., 2016. • Quantify strength of relationships between measured behaviors • No causal relationships identified
• Repeated measures analysis or multi-level modeling • Measurement standardization • Limited to expected spillovers
• Correlational analysis
• Regression analysis (including time series, panel data, and difference-in-difference models)
Quantitative (experimental) • Online, laboratory, or field experiments Van der Werff et al., 2014a,b; Juhl et al., 2017. • Causal relationships and mechanisms identified • Limited to expected spillovers
• Self-reported or observed behavior • Measurement standardization
• Randomization to behavioral intervention
• Analysis of variance
• Regression analysis
Mixed-methods • Combination of qualitative and quantitative methods (e.g., experiment and interviews) Verfuerth, in preparation; Lede, in preparation. As above As above

Measurement of spillover has been undertaken in a variety of ways that reflect the range of methods used. Qualitative approaches tend to rely on self-reported accounts of behavior change; whereas quantitative approaches may use self-reports or observations of behavior. A key weakness in the literature to date, has been a reliance on self-reported behavior, which is known to be only weakly correlated with actual behavior (e.g., Kormos and Gifford, 2014). Furthermore, several studies claiming to find spillover have found change in behavioral intentions or attitudes following an initial behavior change, which is not strictly spillover (Van der Werff et al., 2014a). Few studies also conduct follow-up measurements, so the durability of any immediate spillover effects is unknown. There has also been a reliance on correlational or longitudinal designs which are unable to shed light on causal processes; and within longitudinal designs approaches differ in how to detect spillover (Capstick et al., submitted). Finally, there have also been few attempts to bring together quantitative and qualitative approaches, thus providing complementary insights and addressing respective weaknesses in approaches (Creswell, 2014). In the following section, we describe how spillover should be measured in experimental and non-experimental approaches that seeks to build on this literature and address limitations in the methods used to date.

Measuring Spillover

We now turn from our observations of previous spillover research to a discussion of how we propose spillover research should ideally be conducted in order to reliably detect any spillover effects and expose mechanisms through which they may operate. Drawing on best practice in research design and reflecting principles of transparency and validity (e.g., Open Science Collaboration, 2015), we first discuss experimental studies, which elucidate causal mechanisms, and then non-experimental approaches, which afford other insights into spillover, as discussed above.

How to Measure Behavioral Spillover: Experimental Studies

Rigorously designing and implementing randomized controlled experiments allows the researchers to obtain an unbiased estimate of the average treatment effect of a behavioral intervention (e.g., a ‘nudge,’ a monetary or non-monetary incentive, a ‘boost’ or ‘prime’). Because of sample selection bias, it is only by randomly assigning subjects to a treatment or to a control group that the researchers can identify the causal effect of a behavioral intervention on an observed outcome (Heckman, 1979; Burtless, 1995; Angrist and Pischke, 2009; List, 2011; Gerber and Green, 2012).

In practice, a variety of different randomized controlled experiments is available to researchers interested in testing behavioral spillovers. It is useful to refer here to the influential taxonomy of experiments in social sciences originally proposed by Harrison and List (2004): conventional lab experiments involve student subjects, abstract framing, a lab context, and a set of imposed rules; artefactual field experiments depart from conventional lab experiments in that they involve non-student samples; framed field experiments add to artefactual field experiments a field context in the commodity, stakes, task or information; and, finally, natural field experiments depart from framed field experiments in that subjects undertake the tasks in their natural environment, and subjects do not know that they take part into an experiment. The main idea behind natural field experiments is that the mere act of observation and measurement necessarily alters what is being observed and measured. In key areas of interest for behavioral spillovers, such as health, the environment or pro-social behavior, for instance, there are potential experimenter demand effects (i.e., participants change behavior due to cues about what represents ‘appropriate’ behavior for the experimenter: Bardsley, 2005; Levitt and List, 2007a,b; Zizzo, 2010); Hawthorne effects (i.e., simply knowing they are part of a study makes participants feel important and improves their effort and performance: Franke and Kaul, 1978; Adair, 1984; Jones, 1992; Levitt and List, 2011); and John Henry effects (i.e., participants who perceive that they are in the control group exert greater effort because they treat the experiment like a competitive contest and they want to overcome the disadvantage of being in the control group: Campbell and Stanley, 1963; Cook and Campbell, 1979).

Other, more recent, typologies of randomized controlled experiments are online experiments (Horton et al., 2011) conducted, for instance, using Amazon’s Mechanical Turk (MTurk) (Paolacci et al., 2010; Horton et al., 2011; Paolacci and Chandler, 2014); and lab-field experiments that consist of a first-stage intervention under controlled conditions (in the lab) linked to a naturalistic situation (in the field) where subjects are not aware that their behavior is actually observed. Lab-field experiments have been used to look at the unintended spillover effects of behavioral interventions in health (Dolan and Galizzi, 2014, 2015; Dolan et al., 2015), as well as at the spillover effects in terms of external validity of lab-based behavioral economics games of pro-social behavior (Galizzi and Navarro-Martinez, 2018).

Investigating experimentally the occurrence of behavioral spillover requires a mixed, longitudinal experimental design combining elements of between- and within-subjects design. Participants in an experiment are randomly allocated by the researcher either to a control group, or to (at least) one behavioral intervention group. In the control group (C), subjects are observed while they engage in a first behavior (behavior 1) and then in a different, subsequent, behavior (behavior 2). Each of the two subsequent behaviors is operationally captured and reflected into (at least) one corresponding outcome variable: B1 and B2. In practice, the choice of behavior 1 and behavior 2, as well as the choice of the corresponding outcome variables B1 and B2, is often based on theoretical expectations, previous literature, or qualitative evidence. It is also based on other, more pragmatic, considerations related, for example, to the ease of observing some specific positive or negative spillovers in the lab or the field, and to the ethical and logistical acceptability of changing some behaviors in an experimental setting. In what follows, we illustrate the measurement of behavioral spillovers in the simplest possible case of one single behavioral intervention group, and one single outcome variable for both B1 and B2. The extension to more complex cases is straightforward.

In the treatment group (T), a behavioral intervention (e.g., a ‘nudge,’ a monetary or non-monetary incentive, a ‘boost’ or ‘prime’) is introduced to directly target behavior 1, thus affecting the outcome variable B1. The between-subjects design naturally allows the researcher to test the effects of the behavioral intervention on the targeted behavior 1, by directly comparing B1 across the control and the treatment groups, that is, by comparing B1C versus B1T.

The between-subjects design, together with the longitudinal dimension of the experiment, also allows the researcher to check if the behavioral intervention has a ramification effect on the non-targeted behavior 2, thus affecting the outcome variable B2. In particular, the outcome of behavior 2 in the control group (B2C) serves as the baseline level for the extent to which behavior 2 is affected by behavior 1 in the absence of any behavioral intervention targeting behavior 1 (B1C) (see Table 3).

Table 3.

Experimental design and variables to test behavioral spillovers.

Behavior 1 Behavior 2
Control group (C) B1C B2C
Treatment group (T) B1T B2T
Difference ΔB1 ΔB2

In contrast, the outcome of behavior 2 in the treatment group (B2T) captures the extent to which behavior 2 is affected by the ‘perturbed’ level of behavior 1 as a consequence of the introduction of the behavioral intervention (B1T).

Therefore, by directly comparing B2T and B2C, the difference ΔB2 = B2T – B2C captures the positive or negative change in the outcome variable for behavior 2 which is directly attributable to the change in the outcome variable for behavior 1, ΔB1 = B1T – B1C, which, in turn, is causally affected by the introduction of the behavioral intervention. That is, ΔB2 = B2T – B2C captures the ‘knock on’ behavioral spillover effect of the behavioral intervention targeting behavior 1 on the non-targeted, subsequent behavior 2.

In terms of sizes and statistical significance, such spillover effects may not be significantly different from zero (ΔB2 = 0), may be significantly and positively different from zero (i.e., ΔB2 > 0), or, finally, may be significantly and negatively different from zero (i.e., ΔB2 < 0). If the two behaviors share one common underlying ‘motive’ (in the sense of Dolan and Galizzi, 2015, of some overarching goal or deep preference such as ‘being healthy,’ ‘being pro-environmental,’ or ‘being pro-social’) then the experimental findings may thus be interpreted as evidence of no behavioral spillovers (ΔB2 = 0), evidence of originating ‘promoting’ or ‘precipitating’ behavioral spillover (ΔB2 > 0) or, finally, evidence of ‘permitting’ or ‘purging’ behavioral spillover (ΔB2 < 0).

Such an experimental design also allows the researchers to estimate not only the sign and the statistical significance of the behavioral spillover effects, but also their size. In particular, by comparing the relative changes in the outcome variables for behavior 1 and 2 as effects of the introduction of the behavioral intervention, the ratio between the proportional change (ΔB2/B2C) and the proportional change (ΔB1/B1C) allows the researcher to estimate the ‘elasticity’ of the behavioral spillovers: in analogy with standard price elasticity concepts, the elasticity is defined as the percentage change in behavior 2 per unitary percentage change in behavior 1, that is 𝜀BS = (ΔB2/B2C)/(ΔB1/B1C).

This, in turn, allows the researcher to conclude whether a behavioral intervention causes behavioral ramifications which are small or large compared to the directly targeted change in behavior. In case of permitting or purging behavioral spillovers (i.e., ΔB1 and ΔB2 having opposite signs), and provided that B1 and B2 share the same metrics (or provided that they feed into the underlying motive in a way that the relative sizes of their changes ΔB1 and ΔB2 are conceptually comparable), this can provide further evidence on whether the permitting or purging spillovers are compensating each other completely or only partially (e.g., ‘backfire’ or ‘rebound’ effects).

Two further considerations are in order here. First, the above described definition and framework to measure behavioral spillovers in an experimental setting is sufficiently general and comprehensive to nest as a special case the situation where the behavioral intervention consists of behavior 1 itself. For example, in the ‘question-behavior’ and ‘survey’ promoting spillover effects discussed in Dolan and Galizzi (2015), the behavioral intervention consists of randomly assigning subjects to a brief survey or questionnaire eliciting past health, environmental, or purchasing behavior (e.g., Fitzsimons and Shiv, 2001; Zwane et al., 2011; Van der Werff et al., 2014a). In such a case, in fact, the behavioral intervention in the treatment group merely consists of exposing subjects to behavior 1 (e.g., a survey) before behavior 2 takes place. In the control group, on the other hand subjects go through behavior 2 without being previously exposed to behavior 1. Also in this, simpler, special case, behavioral spillover is measured as ΔB2 = B2T – B2C, but in this case the behavioral spillover captures the positive or negative change in the outcome variable for behavior 2 which is directly attributable to the mere exposure of subjects to behavior 1 in the treatment group (which, in this case, coincides with the behavioral intervention).

Second, the decision about the timeframe is crucial for the measurement of behavioral spillovers. Following subjects over longer timeframes implies, naturally, that it is more likely that spillover effects are effectively detected (Poortinga et al., 2013). Considering substantially long timeframe (ideally a few weeks or even months after the end of the intervention) is desirable in order to be able to assess the durability of spillover effects. Considering even longer timeframes (ideally over 3 or 6 months after the end of the intervention) is particularly important to be able to detect the formation of new habits sustained over time (Lally et al., 2010), rather than a behavioral change that is only transient. In any case, in order to favor transparency and replicability of experimental results, it is crucial that the researchers pre-specify in advance the timeframe over which subjects are followed up over time. The timeframe, in fact, is a key point of the checklist that we propose below.

How to Measure Behavioral Spillover: Non-experimental Quantitative Studies

An analogous strategy can be used in non-experimental settings along the line of the difference-in-difference empirical approach (e.g., Card, 1992, 1996; Card and Krueger, 1994, 2000; see more below). In particular, the researcher can exploit the variation occurring naturally in the field outside their control and can use some ‘natural experiment’ as an exogenous ‘intervention’ in order to identify the likely effect of such an exogenous change on the variables of interest, despite the fact that participants are not randomly assigned to a proper experimental intervention.

The exogenous variation occurring naturally in the field can be a change in policy, a natural ‘shock’ (e.g., a health shock, a natural disaster, a political shock, an economic shock), a life event (e.g., birth of a child, death of a relative, divorce, unemployment), a technological advance, a discontinuity in the availability or in the access of a resource or an infrastructure. The source of the exogenous variation can also be ‘cognitive’ or ‘behavioral,’ such as an exogenous change in attention or awareness, provided that there are convincing reasons to argue that such a source of variation is exogenous (rather than endogenous) to the occurrence of behavioral spillovers.

In the standard difference-in-difference approach, two areas (e.g., two regions, two countries, two schools, two hospitals), are compared before and after the occurrence of a natural event (e.g., a policy, a shock) affecting one area (T) but not the other one (C). Typically, the change of the outcome of behavior 1 before (t = 0) and after (t = 1) the natural event in the ‘control’ area B1Ct = 1 – B1Ct = 0 is compared over time to the analogous change in the ‘treatment’ area B1Tt = 1 – B1Tt = 0, in order to see whether the trends show any significant difference in differences across the two areas (i.e., if B1Tt = 1 – B1Tt = 0, is statistically significantly different from B1Ct = 1 – B1Ct = 0).

In principle, an analogous comparison can be made considering the outcome variable of behavior 2 (B2, instead of B1), to see whether the natural event also has ramifications on a different, subsequent behavior, far and beyond the initial change on behavior 1. Therefore, the researcher can compare the change over time of the outcome variable for behavior 2 before (t = 0) and after (t = 1) the natural event in the ‘control’ area B2Ct = 1 – B2Ct = 0 to the analogous change in the outcome variable for behavior 2 in the ‘treatment’ area B2Tt = 1 – B2Tt = 0, in order to see whether the trends show any significant difference in differences across the two areas (i.e., whether B2Tt = 1 – B2Tt = 0, is statistically significantly different from B2Ct = 1 – B2Ct = 0). Analogous considerations to the ones described above can be made here concerning the sign, significance, and size of the behavioral spillovers in a non-experimental setting (e.g., Claes and Miliute-Plepiene, 2018).

As mentioned above, our framework is sufficiently general and comprehensive to nest, as a special case, the situation where the ‘intervention’ in an experimental setting, or the ‘shock’ or exogenous variation in a non-experimental setting, consists of behavior 1 itself. In such a case, the difference-in-difference approach described above reduces to the comparison of the change in the outcome variable for behavior 2 in the ‘treatment’ area that has been exposed to behavior 1 (B2Tt = 1 – B2Tt = 0) with the analogous change in the ‘control’ area which has not been exposed to behavior 1 (B2Ct = 1 – B2Ct = 0).

The empirical strategy described above has been illustrated having in mind our specific definition of behavioral spillover proposed in section “Definition of Behavioral Spillover,” that is, the observable and causal effect that a change in one behavior (behavior 1) has on a different, subsequent behavior (behavior 2). Nonetheless, a corresponding strategy can be adapted to some of the instances encompassed by the broader definition of spillover reported at the beginning of section “Definition of Behavioral Spillover,” that is the impact that an intervention in a given domain (e.g., health, the environment), group, or location, has on a different domain, group or location. In principle, two locations (e.g., two countries), can be compared before and after the occurrence of a natural event (e.g., a natural phenomenon, an intervention) affecting one domain (e.g., the environment) in one area (T) but not in the other one (C). The researcher can compare not only the change over time of the outcome variable for the domain directly involved in the phenomenon or originally targeted by the intervention (e.g., the environment), but also the change over time of the outcome variable for a different domain (e.g., health). Considering the knock-on effects of the phenomenon or intervention on different groups or regions is also possible in principle, although in practice the empirical analysis would need to account for other underlying intra-groups or intra-regional differences between the ‘control’ and the ‘treatment’ areas.

How to Study Behavioral Spillover: Qualitative and Mixed-Methods Studies

A different, but potentially complementary, approach to studying spillover involves using qualitative methods, such as interviews analyzed thematically (e.g., Boström et al., 2015; Dittmer and Blazejewski, 2016; Nash et al., 2017; Uzzell and Räthzel, 2018; Thomas et al., 2019). As noted, such approaches have the advantage over quantitative approaches of exposing unexpected spillovers, as well as the shedding light on the drivers, barriers and mechanisms of spillover, and on participants’ experience and meanings associated with spillover. For example, Uzzell and Räthzel (2018) used life history interviews to examine how equivalent practices (as well as identities and meanings) develop over time and may be transferred between work and home; using diachronic and synchronic analyses allowed them to identify drivers and barriers to consistency of actions across time, as well as across contexts. Verfuerth et al. (2018) used depth interviews to explore the impacts of a workplace meat reduction intervention, and found unanticipated spillover across behaviors (e.g., to avoiding food waste) and contexts (to home); while Schütte and Gregory-Smith’s (2015) semi-structured interviews exposed cognitive and emotional barriers to pro-environmental spillover between home and holiday.

As such, qualitative methods provide valuable insight in their own right into spillover phenomena, but can also be combined with quantitative approaches in mixed-methods designs to address quantitative limitations (Verfuerth and Gregory-Smith, 2018). Various approaches can be used to ensure the quality of qualitative data, such as member validation (i.e., asking participants to check researcher interpretations), inter-rater reliability of coded data (i.e., using multiple coders and resolving any disagreement in interpretation), and reflexivity (i.e., fully documenting the processes used to collect data and the role and background of the researcher; Breakwell et al., 2012). Others have noted that the diversity of qualitative methods requires a range of criteria for assessing quality and validity (Reicher, 2000); but most agree at least that transparency and consistency are key (Braun and Clarke, 2006). The importance of being systematic is therefore a criterion of quality shared by both quantitative and qualitative methods.

A growing literature advocates the use of mixed-methods approaches in order to triangulate and provide complementary insights. Despite associations of qualitative and quantitative methods with divergent epistemological and ontological paradigms (Blaikie, 1991), this should not imply that qualitative and quantitative methods are essentially incommensurate (Bryman, 1988). Rather, the distinction between particular qualitative and quantitative methods can be understood as primarily technical, and not necessarily philosophical. Qualitative and quantitative methods offer different insights into spillover and each is better suited to answering different types of research question (e.g., What are the range of effects of an intervention? How is the development of identity and practices experienced over time and contexts? What causes and mediates spillover?). Thus, the rationale for combining methods stems from “the basic and plausible assertion that life is multifaceted and is best approached by the use of techniques that have a specialized relevance” (Fielding and Fielding, 1986, p. 34). Furthermore, using multiple methods allows interesting lines of inquiry exposed through one method to be explored further through another (Whitmarsh, 2009). At the same time, however, it is not assumed that aggregating data sources can provide a complete or ‘true’ picture of the social world (Silverman, 2001). Indeed, “the differences between types of data can be as illuminating as their points of coherence” (Fielding and Fielding, 1986, p. 31), for example leading to a re-examination of conceptual frameworks or assumptions (Tashakkori and Teddlie, 2003). The distinct challenges of researching spillover imply both qualitative and quantitative approaches are warranted to address different facets of the problem.

Mixed-methods designs may be sequential or concurrent, or both (Creswell, 2014). In the case of spillover studies, a mixed methods design might start with an initial qualitative and/or correlational phase to identify clusters of co-occurring behaviors which may indicate spillover, for which candidate behaviors (B1, B2, etc.) and the causal pathways connecting them can be examined in a subsequent experimental design, as outlined above. In addition, qualitative methods can be used alongside quantitative behavioral measures within the intervention phase to explore the experience, perceptions, and subjective wellbeing implications of the intervention, and to expose potentially unexpected spillover effects, as well as possible drivers, barriers, mechanisms, and mediating/moderating factors for any spillover. This might take the form of interviews with a sub-sample of experimental participants, or one or more open-ended questions in a post-intervention survey. Where spillover is detected through quantitative experimental methods, qualitative data may help explain why this effect has occurred, and how this has been subjectively perceived and experienced. In the event that spillover is not detected via the experimental methods outlined above, qualitative methods may explain why not, or they may expose other, unquantified spillover effects. Qualitative, quantitative, and experimental methods should thus be seen as complementary, rather than substitute, empirical methods to explore and assess behavioral spillovers. So far, there exist few mixed-methods studies of spillover, but those that have been undertaken appear to demonstrate that a mixed methodology can elucidate multiple aspects of spillover processes and experiences (Barr et al., 2010; Verfuerth et al., 2018; Thomas et al., 2019).

A Practical Checklist

Exploring and detecting behavioral spillovers is a research and policy task which should be undertaken using a systematic and transparent approach, in the same spirit of, and closely in line with, the recent best practices favoring and advocating systematization and transparency in psychological and behavioral sciences (Ioannidis, 2005; Higgins and Green, 2011; Simmons et al., 2011; Miguel et al., 2014; Simonsohn et al., 2014; Open Science Collaboration, 2015; Munafò et al., 2017). In the previous section, we outlined how this might be achieved using different research designs.

Abstracting from these exemplar designs, here we propose a checklist of points which should be explicitly stated and addressed by the researcher prior to undertaking of experimental and empirical analysis. The 20-item checklist is in line with, and in the same spirit of, other checklists designed to systematically assess the methodological quality of prospective studies, for example by the Cochrane Collaboration (Higgins and Green, 2011). The checklist is also in line with, and in the same spirit of, other more general checklists guiding researchers through pre-registration of studies and pre-analysis plans (e.g., the Open Science Framework2). Once filled in, the checklist for a prospective study should be deposited in a dedicated website which is going to be launched with the publication of this special issue, and which will be available at: https://osf.io/9cqjf/. The website will also include a data template where data from deposited studies could be shared, collated, and combined in order to conduct collaborative systematic reviews and meta-analyses of the literature.

The 20 questions of the checklist are below. In what follows we briefly illustrate each question with a real case study, the recent study by Xu et al. (2018a) on household waste separation:

  • simple 1.

    What are the setting and population of interest?

  • simple

          •Four geographically adjacent communities in the Yuhang District of Hangzhou, Zhejiang Province, China.

  • simple 2.

    Is this an experimental or a non-experimental study?

  • simple

          •An experimental study (a framed field experiment).

  • simple 3.

    If this is a non-experimental quantitative study, what is the empirical identification strategy (e.g., difference-in-difference)?

  • simple

          •N/A.

  • simple 4.

    If this is a quantitative study, what is the control group?

  • simple

          •The control group were participants in each community who were not exposed to any formal promotion of waste separation.

  • simple 5.

    How have the behaviors been selected (e.g., existing literature, qualitative evidence)?

  • simple

          •Based on previous findings and on the literature.

  • simple 6.

    What is the targeted behavior 1?

  • simple

          •Sorting daily garbage and bringing it to waste collection sites.

  • simple 7.

    What are the outcome variables for behavior 1 (i.e., how will you measure behavior 1)? (Please list them and briefly describe each outcome variable, indicating whether this is directly observed or self-reported behavior.)

  • simple

          •Difference in self-reported household waste collection before and after the interventions.

  • simple 8.

    How many intervention groups there are?

  • simple

          •Originally there were three intervention groups, but one condition (‘mixed condition’) was then excluded (see footnote 1 in page 28).

  • simple 9.

    What are the behavioral interventions targeting behavior 1? (Please list them and briefly describe each of them.)

  • simple

          •In the Environmental Appeal (EA) condition participants were given 3 monthly 30-min presentations where they were informed about the environmental benefits of waste separation. In the Monetary Incentive (MI) condition participants were given 3 monthly 30-min presentations where they were informed that they could earn ‘green scores’ from a recycling firm if they sorted their daily garbage and brought it to waste collection sites. In the ‘mixed condition’ participants were given 3 monthly 30-min presentations where they were informed of both EA and MI (this condition was later excluded from the analysis).

  • simple 10.

    What is the non-targeted behavior 2?

  • simple

          •A set of 25 self-reported environmental behaviors or self-reported willingness to engage in environmental behaviors, including both ‘private-sphere’ behaviors (e.g., green shopping, traveling) and ‘public-sphere’ behaviors (e.g., support to environmental policies, environmental citizenship actions).

  • simple 11.

    What are the outcome variables for behavior 2 (i.e., how will you measure behavior 2)? (Please list them and briefly describe each outcome variable, indicating whether this is directly observed or self-reported behavior.). If there are multiple outcome variables for behavior 2, does the study correct for multiple hypotheses testing? (Please describe which correction is used.)

  • simple

          •All the outcome variables for the 25 environmental behaviors or willingness to engage in environmental behaviors are self-reported, and are collected by a monthly survey. There is no explicit correction for multiple hypotheses testing.

  • simple 12.

    What is the expected underlying motive linking behavior 1 and behavior 2?

  • simple

          •Pro-environmental identity (page 28).

  • simple 13.

    What are the expected mechanisms moderating and/or mediating the changes in the outcome variables for behavior 2?

  • simple

          •The expected mechanisms are both promoting/positive behavioral spillovers such as the activation of a stronger pro-environmental identity, and permitting/negative behavioral spillovers such as moral licensing (page 28). Pro-environmental identity and environmental concern are expected to mediate promoting/positive spillovers. Relief of guilt is expected to mediate permitting/negative spillovers.

  • simple 14.

    What is the expected time frame during which behavioral spillovers will be tested, and during which the durability of spillover and habit formation will be assessed?

  • simple

          •The expected time frame is not explicitly mentioned, but participants are followed up for 3 months.

  • simple 15.

    What is the expected participant attrition between behavior 1 and behavior 2?

  • simple

          •There is no explicit discussion of expected attrition. However, attrition was not only high, but it was asymmetric across different conditions. At the end of the experiment (3 months after), only 195 out of the 400 participants originally recruited remained in the study: 80 (out of 100) in the EA group, 36 (out of 100) in the MI group, and 79 (out of 100) in the control group (all the 100 participants in the mixed condition group were excluded).

  • simple 16.

    What is the expected direction of the changes in the outcome variables for behaviors 1 and 2 between the intervention groups and the control group (i.e., are positive or negative spillovers expected)?

  • simple

          •Both promoting/positive and permitting/negative spillovers were expected (page 28).

  • simple 17.

    What are the expected sizes and standard errors of the changes in the outcome variables for behaviors 1 and 2 between the intervention groups and the control group?

  • simple

          •There is no explicit discussion of the expected effect size or standard errors of the changes in the outcome variables for behaviors 1 and 2.

  • simple 18.

    What is the minimum expected sample size to test and detect the occurrence of behavioral spillover?

  • simple

          •The study recruits n = 100 participants in each of the four groups, but there is no explicit justification of the minimum expected sample size to test and detect the occurrence of behavioral spillovers.

  • simple 19.

    If collecting qualitative data, how will the quality of this data be ensured and assessed (e.g., reflexivity, consistency)?

  • simple

          •A number of psychological constructs were collected (including four items to measure personal identification with environmental protection; three items to measure personal concern for the environment, ecology, and the earth; three items to measure feelings of disappointment, guilt, and regret for past environmentally unfriendly behaviors) and used in exploratory factor analysis, but no further qualitative data was collected.

  • simple 20.

    If using mixed-methods approaches, how will insights from different methods be combined?

  • simple

          •N/A.

Conclusion

We have critically reviewed the main methods to measure behavioral spillovers to date, and discussed their methodological strengths and weaknesses. We have proposed a consensus mixed-method approach which uses a longitudinal between-subject design together with qualitative self-reports: participants are randomly assigned to a treatment group where a behavioral intervention takes place to target behavior 1, or to a control group where behavior 1 takes place absent any behavioral intervention. A behavioral spillover is empirically identified as the effect of the behavioral intervention in the treatment group on a subsequent, not targeted, behavior 2, compared to the corresponding change in behavior 2 in the control group.

In the spirit of the pre-analysis plan, we have also proposed a systematic checklist to guide researchers and policy-makers through the main stages and features of the study design in order to rigorously test and identify behavioral spillovers, and to ensure transparency, reproducibility, and meta-analysis of studies.

While ours is arguably the first methodological note on how to measure behavioral spillovers, it has of course limitations. The main limitation is that our experimental and empirical identification strategy relies on our specific definition of behavioral spillover – i.e., the observable and causal effect that a change in one behavior (behavior 1) has on a different, subsequent behavior (behavior 2). As mentioned in section “Definition of Behavioral Spillover,” broader definitions of spillover exist that can encompass attitudinal change, learning, interpersonal influences, and other disparate processes. While we have suggested here that a similar approach to ours (i.e., longitudinal mixed-methodology) might apply in these cases, there may be also be methodological considerations specific to each type of spillover that warrants its own methodological checklist. Even applying our more specific definition of behavioral spillover, it would be possible to define alternative methodological checklists that, for example, apply solely quantitative or qualitative methods (cf. Uzzell and Räthzel, 2018). However, as we have argued, we believe there is benefit in combining methods as they can offer different insights or address different research questions relating to spillover.

We would like to conclude by briefly mentioning a few other directions where we envisage promising methodological developments in the years to come. First, the current technological landscape naturally lends itself to a systematic measurement of behavioral spillovers in a variety of research and policy domains. Today an unprecedented richness of longitudinal data are routinely collected at an individual level in terms of online surveys, apps, smart phones, internet of things (IoT) and mobile devices, smart cards and scan data, electronic administrative records, biomarkers, and other longitudinal panels. This is creating, for the first time in history, an immense potential for following up individuals across different contexts and domains, and over time, for months, years, and even decades. This new technological landscape is also creating previously unexplored opportunities for ‘behavioral data linking,’ that is, for the linkage of behavioral experiments with other sources of longitudinal data (Galizzi, 2017; Galizzi et al., 2017; Galizzi and Wiesen, 2018; Krpan et al., 2019). On the one hand, the scope for systematically testing the occurrence of behavioral spillovers using rigorous empirical and experimental methods is therefore enormous. On the other hand, the endless wealth of research hypotheses, outcome variables, and data points makes even more important for researchers to embrace the best practices discussed above in order to ensure transparency, openness, and reproducibility of science.

Second, a promising methodological line of research about behavioral spillover concerns the rigorous investigation of the factors mediating and moderating the occurrence of behavioral spillover, for example in terms of accessibility (Sintov et al., 2019). Further work in this direction is likely to develop also thanks to the triangulation of different sources of data enabled by the above described shift in the technological landscape.

All these future developments reinstate the importance of developing a collective discussion about clear and transparent methodological guidelines to measure behavioral spillovers. We hope that with the present article we have contributed to at least start such a discussion. The time is ripe to foster a collaborative endeavor to systematically test behavioral spillovers across all research and policy domains, contexts, and settings.

Author Contributions

MG initiated and led the paper writing. LW contributed to paper writing.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Appendix: Methodology of Systematic Review of the Literature

In conducting and reporting our systematic review of the literature, we followed as closely as possible the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) checklist and guidelines (Moher et al., 2009), as explained below.

Search Strategy and Key Terms

Google Scholar was searched in December 2018 using the following combinations of exact phrases in the advanced search settings:

  • simple (1)

    “behavioral spillover” (field TX all text) OR

  • simple (2)

    “behavioral spillover” (field TX all text).

Selection and Exclusion Criteria

The authors reviewed and assessed all the references systematically, following a two-stage strategy. In the first stage, the inclusion criteria were applied to the title, the keywords, and the abstract; in the second stage, the criteria were applied to the abstract and the full text. All the papers were independently assessed for inclusion by each of the authors. Differences in opinions between the authors were solved through discussion.

The two stages worked as follows. In the first stage, a study was included only if it satisfied the following three criteria:

  • simple (1)

    The study was available (no broken link).

  • simple (2)

    The study was written in English.

  • simple (3)

    The study presented new scientific material, in terms of: new empirical evidence or original experimental analysis of behavioral spillover; new theoretical definitions or conceptual frameworks for behavioral spillovers; systematic reviews or meta-analyses of existing studies on behavioral spillovers. This criterion excluded non-systematic reviews, commentaries, editorials, letters, or similar items.

  • simple

          Each article was sequentially evaluated against the three criteria, starting with criterion one and ending on criterion three. Whenever a criterion was not met, the article was excluded.

  • simple

          In the second stage, the abstract and the full text of the studies shortlisted in the first stage were screened, evaluated, and finally included according to two further criteria:

  • simple (4)

    The study considered human behavior.

  • simple (5)

    The study used a definition of behavioral spillover substantially in line with our operational definition in section “Definition of Behavioral Spillover,” that is, the observable and causal effect that a change in one behavior (behavior 1) has on a different, subsequent behavior (behavior 2).

We included both published and unpublished studies, for example studies in working paper or in dissertation form. If both published and unpublished versions of the study were available, we considered the published version. If different dates of the unpublished versions were available, we considered the most recent one.

To ensure that the set of studies retrieved was exhaustive and comprehensive, for each included study, we also back-tracked and screened all the references cited in the article, applying the same inclusion criteria explained above.

Search Results

The initial Google Scholar search resulted in a total number of n = 529 entries on December 17th, 2018 (n = 305 for “behavioral spillovers” and n = 224 for “behavioral spillovers.” After n = 51 duplicates were removed, the resulting number of studies was n = 478. We then excluded the papers that were not accessible (n = 16), were not written in English (n = 11), or did not present new scientific material (n = 97). A total of n = 354 studies met all three criteria in this first stage of our selection strategy.

The abstract and the full text of the n = 354 studies shortlisted were then screened and evaluated. We then excluded the studies that did not focus on human behavior (n = 17), and the studies whose definitions of behavioral spillovers was substantially different from our operational definition – or which did not define behavioral spillovers at all (n = 240). A total of n = 97 studies matched all the inclusion criteria in this second stage.

Back-tracking, screening, and evaluating the references cited in these n = 97 articles against the same inclusion criteria retrieved further n = 9 studies. So, at the end of the whole process, the systematic review resulted in a total of n = 106 selected studies.

Of the n = 106 selected studies, n = 12 are Doctoral theses, n = 5 are Master theses, and n = 12 are still unpublished works, all which shows the growing interest on behavioral spillovers.

The selection process and the number of papers excluded and included in each stage are summarized in the PRISMA flow chart in Figure A1.

Figure A1.

Figure A1

PRISMA flow diagram of systematic review.

Funding. Funding for LW was received from the European Research Council, CASPI Starting Grant (336665).

1

These 106 studies are: Bratt, 1999, Thøgersen, 1999, Hertwich, 2005, Karremans et al., 2005, Cornelissen et al., 2008, Hecht and Boies, 2009, Sorrell et al., 2009, Zimmerman, 2009, Savikhin, 2010, Sheremeta et al., 2010, Dickinson and Oxoby, 2011, Nolan, 2011, Bednar et al., 2012, 2015, Cason et al., 2012, Thøgersen and Noblet, 2012, Xanthopoulou and Papagiannidis, 2012, Alpizar et al., 2013a,b, Baca-Motes et al., 2013, Cason and Gangadharan, 2013, Falk et al., 2013, Godoy et al., 2013, Juvina et al., 2013, Norden, 2013, Poortinga et al., 2013, Savikhin and Sheremeta, 2013, Swim and Bloodhart, 2013, Tiefenbeck et al., 2013, Bech-Larsen and Kazbare, 2014, Dolan and Galizzi, 2014, 2015, Lanzini and Thøgersen, 2014, Littleford et al., 2014, Spence et al., 2014, Tiefenbeck, 2014, Truelove et al., 2014, 2016, Van der Werff et al., 2014a,b, Goswami and Urminsky, 2015, Kaida and Kaida, 2015, Karmarkar and Bollinger, 2015, Lacasse, 2015, 2016, 2017, Schütte and Gregory-Smith, 2015, Steinhorst et al., 2015, Zawadzki, 2015, Banerjee, 2016, Dittmer and Blazejewski, 2016, Eby, 2016, Gholamzadehmir, 2016, Ha and Kwon, 2016, Lauren et al., 2016, 2017, Margetts and Kashima, 2016, Nilsson et al., 2016, Polizzi di Sorrentino et al., 2016, Steinhorst and Matthies, 2016, Suffolk, 2016, Suffolk and Poortinga, 2016, Thomas et al., 2016, 2019, Carpenter and Lawler, 2017, Carrico et al., 2017, Crookes, 2017, Fenger, 2017, Galbiati et al., 2017, Hedrick et al., 2017, Jessoe et al., 2017, Juhl et al., 2017, Kesternich et al., 2017, Klein, 2017, Krieg and Samek, 2017, McCoy and Lyons, 2017, Nash et al., 2017, 2019, Werfel, 2017, Xie et al., 2017, Angelovski et al., 2018, Bednar and Page, 2018, Chatelain et al., 2018, Claes and Miliute-Plepiene, 2018, Dutschke et al., 2018, Ghesla et al., 2018, Lawler, 2018, Liu et al., 2018, Panos, 2018, Peters et al., 2018, Santarius and Soland, 2018, Schmitz, 2018, Seebauer, 2018, Shreedhar, 2018, Shreedhar and Mourato, 2018, Tippet, 2018, Vasan, 2018, Verfuerth and Gregory-Smith, 2018, Vincent and Koessler, 2018, Whitmarsh et al., 2018, Xu et al., 2018a,b, Capstick et al., 2019, Fanghella et al., 2019, Krpan et al., 2019, Sintov et al., 2019.

References

  1. Adair G. (1984). The hawthorne effect: a reconsideration of the methodological artefact. J. Appl. Psychol. 69 334–345. 10.1037/0021-9010.69.2.334 [DOI] [Google Scholar]
  2. Alpizar F., Norden A., Pfaff A., Robalino J. (2013a). Behavioral spillovers from targeted incentives: losses from excluded individuals can counter gains from those selected. Duke Environmental and Energy Economics Working Paper EE 13-07, Turrialba. [Google Scholar]
  3. Alpizar F., Norden A., Pfaff A., Robalino J. (2013b). Effects of exclusion from a conservation policy: negative behavioral spillovers from targeted incentives. Duke Environmental and Energy Economics Working Paper EE 13-16Turrialba. [Google Scholar]
  4. Angelovski A., Di Cagno D., Guth W., Marazzi F., Panaccione L. (2018). Behavioral spillovers in local public good provision: an experimental study. J. Econ. Psychol. 67 116–134. 10.1016/j.joep.2018.05.003 [DOI] [Google Scholar]
  5. Angrist J. D., Pischke J. S. (2009). Mostly Harmless Econometrics: An Empiricist’s Companion. Princeton, NJ: Princeton University Press. [Google Scholar]
  6. Austin A., Cox J., Barnett J., Thomas C. (2011). Exploring Catalyst Behaviours: Full Report: A Report to The Department for Environment, Food and Rural Affairs. London: Department for Environment, Food and Rural Affairs. [Google Scholar]
  7. Baca-Motes K., Brown A., Gneezy A., Keenan E. A., Nelson L. D. (2013). Commitment and behavior change: evidence from the field. J. Consum. Res. 39 1070–1084. 10.1086/667226 [DOI] [Google Scholar]
  8. Banerjee R. (2016). Corruption, norm violation and decay in social capital. J. Public Econ. 137 14–27. 10.1016/j.jpubeco.2016.03.007 [DOI] [Google Scholar]
  9. Bardsley N. (2005). Experimental economics and the artificiality of alteration. J. Econ. Methodol. 12 239–251. 10.1080/13501780500086115 [DOI] [Google Scholar]
  10. Barr S., Shaw G., Coles T., Prillwitz J. (2010). ‘A holiday is a holiday’: practicing sustainability, home and away. J. Transp. Geogr. 18 474–481. 10.1016/j.jtrangeo.2009.08.007 [DOI] [Google Scholar]
  11. Beaman A. L., Cole C. M., Preston M., Klentz B., Steblay N. M. (1983). Fifteen years of foot-in-the door research: a meta-analysis. Personal. Soc. Psychol. Bull. 9 181–196. 10.1177/0146167283092002 [DOI] [Google Scholar]
  12. Bech-Larsen T., Kazbare L. (2014). Spillover of diet changes on intentions to approach healthy food and avoid unhealthy food. Health Educ. 114 367–377. 10.1108/HE-04-2013-0014 [DOI] [Google Scholar]
  13. Bednar J., Chen Y., Liu T. X., Page S. (2012). Behavioral spillover and cognitive load in multiple games: experimental results. Games Econ. Behav. 74 12–31. 10.1016/j.geb.2011.06.009 [DOI] [Google Scholar]
  14. Bednar J., Joones-Rooy A., Page S. E. (2015). Choosing a future based on the past: institutions, behaviour, and path dependence. Eur. J. Polit. Econ. 40 312–332. 10.1016/j.ejpoleco.2015.09.004 [DOI] [Google Scholar]
  15. Bednar J., Page S. E. (2018). When order affects performance: culture, behavioral spillovers, and institutional path dependence. Am. Polit. Sci. Rev. 112 82–98. 10.1017/S0003055417000466 [DOI] [Google Scholar]
  16. Blaikie N. W. H. (1991). A critique of the use of triangulation in social research. Qual. Quan. 25 115–136. 10.1007/BF00145701 [DOI] [Google Scholar]
  17. Blanken I., van de Ven N., Zeelenberg M. (2015). A meta-analytic review of moral licensing. Personal. Soc. Psychol. Bull. 41 540–558. 10.1177/0146167215572134 [DOI] [PubMed] [Google Scholar]
  18. Boström M., Gilek M., Hedenström E., Jönsson A. M. (2015). How to achieve sustainable procurement for “peripheral” products with significant environmental impacts. Sustain. Sci. Pract. Policy 11 21–31. 10.1080/15487733.2015.11908136 [DOI] [Google Scholar]
  19. Bratt C. (1999). Comsumers’ environmental behaviour: generalized, sector-based, or compensatory? Environ. Behav. 31 28–44. 10.1177/00139169921971985 [DOI] [Google Scholar]
  20. Braun V., Clarke V. (2006). Using thematic analysis in psychology. Qual. Res. Psychol. 3 77–101. 10.1191/1478088706qp063oa [DOI] [Google Scholar]
  21. Breakwell G., Smith J. A., Wright D. B. (2012). Research Methods in Psychology, 4th Edn. London: Sage. [Google Scholar]
  22. Bryman A. (1988). Quantity and Quality in Social Research. London: Unwin Hyman. [Google Scholar]
  23. Burtless G. (1995). The case for randomised field trials in economic and policy research. J. Econ. Perspect. 9 63–84. 10.1257/jep.9.2.63 [DOI] [Google Scholar]
  24. Campbell D. T., Stanley J. C. (1963). Experimental and Quasi-Experimental Designs for Research. Chicago, IL: Rand McNally. [Google Scholar]
  25. Capstick S., Lorenzoni I., Corner A., Whitmarsh L. (2014). Prospects for radical emissions reduction through behavior and lifestyle change. Carbon Manage. 5 429–445. 10.1080/17583004.2015.1020011 [DOI] [Google Scholar]
  26. Capstick S. B., Whitmarsh L., Nash N. (2019). Behavioural spillover in the UK and Brazil: findings from two large-scale longitudinal surveys. J. Environ. Psychol. (in press). [Google Scholar]
  27. Card D. (1992). Using regional variation to measure the effect of the federal minimum wage. Ind. Labor Relat. Rev. 46 22–37. 10.1177/001979399204600103 [DOI] [Google Scholar]
  28. Card D. (1996). The effect of unions on the structure of wages: a longitudinal analysis. Econometrica 64 957–979. 10.2307/2171852 [DOI] [Google Scholar]
  29. Card D., Krueger A. (1994). Minimum wages and employment: a case study of the fast-food industry in New Jersey and Pennsylvania. Am. Econ. Rev. 84 772–784. [Google Scholar]
  30. Card D., Krueger A. (2000). Minimum wages and employment: a case study of the fast-food industry in New Jersey and Pennsylvania: reply. Am. Econ. Rev. 90 1397–1420. 10.1257/aer.90.5.1397 [DOI] [Google Scholar]
  31. Carpenter C. S., Lawler E. C. (2017). Direct and spillover effects of middle school vaccination requirements. Am. Econ. J. Econ. Policy 11 92–125. 10.3386/w23107 [DOI] [Google Scholar]
  32. Carrico A. R., Raimi K. T., Truelove H. B., Eby B. (2017). Putting your money where your mouth is: an experimental test of pro-environmental spillover from reducing meat consumption to monetary donations. Environ. Behav. 50 723–748. 10.1177/0013916517713067 [DOI] [Google Scholar]
  33. Cason T., Savikhin A. C., Sheremeta R. M. (2012). Behavioral spillovers in coordination games. Eur. Econ. Rev. 56 233–245. 10.1016/j.euroecorev.2011.09.001 [DOI] [Google Scholar]
  34. Cason T. M., Gangadharan L. (2013). Cooperation spillovers and price competition in experimental markets. Econ. Inq. 51 1715–1730. 10.1111/j.1465-7295.2012.00486.x [DOI] [Google Scholar]
  35. Chatelain G., Hille S. L., Sander D., Patel M., Hahnel U. J. J., Brosch T. (2018). Feel good, stay green: positive affect promotes pro-environmental behaviors and mitigates compensatory “mental bookkeeping” effects. J. Environ. Psychol. 56 3–11. 10.1016/j.jenvp.2018.02.002 [DOI] [Google Scholar]
  36. Claes E., Miliute-Plepiene J. (2018). Behavioral spillovers from food-waste collection in Swedish municipalities. J. Environ. Econ. Manage. 89 168–186. 10.1016/j.jeem.2018.01.004 [DOI] [Google Scholar]
  37. Cook T. D., Campbell D. T. (1979). Quasi-Experimentation: Design and Analysis for Field Settings. Chicago, IL: Rand McNally. [Google Scholar]
  38. Cornelissen G., Pandelaere M., Warlop L., Dewitte S. (2008). Positive cueing: promoting sustainable consumer behaviour by cueing common environmental behaviors as environmental. Int. J. Res. Market. 25 46–55. 10.1016/j.ijresmar.2007.06.002 [DOI] [Google Scholar]
  39. Creswell J. (2014). A Concise Introduction to Mixed Methods Research. London: Sage Publications. [Google Scholar]
  40. Crookes R. D. (2017). Exploring the Underlying Processes and the Long Term Effects of Choice Architecture. Doctoral Thesis, Columbia University, New York, NY. [Google Scholar]
  41. Dickinson D. L., Oxoby R. J. (2011). Cognitive dissonance, pessimism, and behavioral spillover effects. J. Econ. Psychol. 32 295–306. 10.1016/j.joep.2010.12.004 [DOI] [Google Scholar]
  42. Dittmer F., Blazejewski S. (2016). “Sustainable at home – Sustainable at work? The impact of pro-environmental life-work spillover effects on sustainable intra- or entrepreneurship,” in Sustainable Entrepreneurship and Social Innovation, eds Nicolopoulou K., Karatas-Ozkan M., Janssen F., Jermier J. M. (Milton Park: Taylor & Francis; ), 73–100. [Google Scholar]
  43. Dolan P., Galizzi M. M. (2014). Because I’m worth it. A lab-field experiment on spillover effects of incentives in health. LSE CEP Discussion Paper CEPDP 1286 London. [Google Scholar]
  44. Dolan P., Galizzi M. M. (2015). Like ripples on a pond: behavioral spillovers and their consequences for research and policy. J. Econ. Psychol. 47 1–16. 10.1016/j.joep.2014.12.003 [DOI] [Google Scholar]
  45. Dolan P., Galizzi M. M., Navarro-Martinez D. (2015). Paying people to eat or not to eat? Carryover effects of monetary incentives on eating behavior. Soc. Sci. Med. 133 153–158. 10.1016/j.socscimed.2015.04.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Dutschke E., Frondel M., Schleich J., Vance C. (2018). Moral licensing: another source of rebound? Front. Energy Res. 6:38 10.3389/fenrg.2018.00038 [DOI] [Google Scholar]
  47. Eby B. D. (2016). Color Me Geeen: The Influence of Environmental Identity Labelling on Spillovers in Pro-Environmental Behaviors. Master Thesis, University of Colorado, Boulder. [Google Scholar]
  48. Falk A., Fischbacher U., Gachter S. (2013). Living in two neighborhoods – Social interaction effects in the laboratory. Econ. Inq. 51 563–578. 10.1111/j.1465-7295.2010.00332.x [DOI] [Google Scholar]
  49. Fanghella V., d’Adda G., Tavoni M. (2019). On the use of nudges to affect spillovers in environmental behaviors. Front. Psychol. 10:61. 10.3389/fpsyg.2019.00061 [DOI] [PMC free article] [PubMed] [Google Scholar]
  50. Fenger M. H. J. (2017). Essays on the Dynamics of Consumer Behavior. Doctoral Thesis, Aarhus University, Aarhus. [Google Scholar]
  51. Fielding N. G., Fielding J. L. (1986). Linking Data. Newbury Park, CA: Sage Publications; 10.4135/9781412984775 [DOI] [Google Scholar]
  52. Fitzsimons G. J., Shiv B. (2001). Non-conscious and contaminative effects of hypothetical questions on subsequent decision-making. J. Consum. Res. 334 782–791. [Google Scholar]
  53. Franke R. H., Kaul J. D. (1978). The Hawthorne experiments: first statistical interpretation. Am. Sociol. Rev. 43 623–643. 10.2307/2094540 [DOI] [Google Scholar]
  54. Freedman J. L., Fraser S. C. (1966). Compliance without pressure: the foot-in-the-door technique. J. Personal. Soc. Psychol. 4 195–202. 10.1037/h0023552 [DOI] [PubMed] [Google Scholar]
  55. Gabe-Thomas E., Walker I., Verplanken B., Shaddick G. (2016). Householders’ mental models of domestic energy consumption: using a sort-and-cluster method to identify shared concepts of appliance similarity. PLoS One 11:e0158949. 10.1371/journal.pone.0158949 [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Galbiati R., Henry E., Jacquemet N. (2017). Spillovers, persistence, and learning: institutions and the dynamic of cooperation. CEPR Discussion Paper No. DP12128 Paris: 10.2139/ssrn.3010782 [DOI] [Google Scholar]
  57. Galizzi M. M. (2017). “Behavioral aspects of policy formulation: experiments, behavioral insights, nudges,” in Handbook of Policy Formulation. Handbooks of Research on Public Policy, eds Howlett M., Mukherjee I. (Cheltenham: Edward Elgar Publishing; ), 410–429. 10.4337/9781784719326.00034 [DOI] [Google Scholar]
  58. Galizzi M. M., Harrison G. W., Miraldo M. (2017). “Experimental methods and behavioural insights in health economics: estimating risk and time preferences in health,” in Health Econometrics in Contributions to Economic Analysis, eds Baltagi B., Moscone F. (Bingley: Emerald Publishing; ). [Google Scholar]
  59. Galizzi M. M., Navarro-Martinez D. (2018). On the external validity of social preferences games: a systematic lab-field study. Manage. Sci. (in press). 10.1287/mnsc.2017.2908 [DOI] [Google Scholar]
  60. Galizzi M. M., Wiesen D. (2018). “Behavioral experiments in health economics,” in Oxford Research Encyclopedia of Economics and Finance, eds Hamilton J. H., Dixit A., Edwards S., Judd K. (Oxford: Oxford Research Encyclopedias, Oxford University Press; ). [Google Scholar]
  61. Geller E. S. (2001). From ecological behaviorism to response generalization: where should we make discriminations? J. Organ. Behav. Manage. 21 55–73. 10.1300/J075v21n04_05 [DOI] [Google Scholar]
  62. Gerber A. S., Green D. P. (2012). Field Experiments: Design, Analysis, and Interpretation, New York, NY: W.W. Norton & Company. [Google Scholar]
  63. Ghesla C., Grieder M., Schmitz J. (2018). Nudge for good? Choice defaults and spillover effects. Front. Psychol. 10:178. 10.3389/fpsyg.2019.00178 [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Gholamzadehmir M. (2016). The Impact of Moral Action and Moral Values on Moral Judgement and Moral Behaviour. Doctoral Thesis, University of Sussex, Falmer. [Google Scholar]
  65. Godoy S., Morales A. J., Rodero J. (2013). Competition lessens competition: an experimental investigation of simultaneous participation in a public good game and a lottery contest game with shared endowment. Econ. Lett. 120 419–423. 10.1016/j.econlet.2013.05.021 [DOI] [Google Scholar]
  66. Goswami I., Urminsky O. (2015). The nature and extent of post-reward crowding-out: the ‘effort balancing’ account. SSRN Working Paper 2733335 Chicago, IL. 10.2139/ssrn.2733335 [DOI] [Google Scholar]
  67. Ha S., Kwon S. Y. (2016). Spillover from past recycling to green apparel shopping behaviour: the role of environmental concerns and anticipated guilt. Fashion Text. 3:16 10.1186/s40691-016-0068-7 [DOI] [Google Scholar]
  68. Harrison G. W., List J. A. (2004). Field experiments. J. Econ. Literat. 42 1009–1055. 10.1257/0022051043004577 [DOI] [Google Scholar]
  69. Hecht T. D., Boies K. (2009). Structure and correlates of spillover from nonwork to work: an examination of nonwork activities, well-being, and work outcomes. J. Occupat. Health Psychol. 14 414–426. 10.1037/a0015981 [DOI] [PubMed] [Google Scholar]
  70. Heckman J. J. (1979). Sample selection bias as a specification error. Econometrica 47 153–162. 10.2307/1912352 [DOI] [Google Scholar]
  71. Hedrick V. E., Davy B. M., You W., Porter K. J., Estabrooks P. A., Zoellner J. M. (2017). Dietary quality changes in response to a sugar-sweetened beverage-reduction intervention: results from the Talking Heath randomized controlled clinical trial. Am. J. Clin. Nutr. 105 824–833. 10.3945/ajcn.116.144543 [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Hertwich E. G. (2005). Consumption and the rebound effect: an industrial ecology perspective. J. Ind. Ecol. 9 85–98. 10.1162/1088198054084635 [DOI] [Google Scholar]
  73. Higgins J. P. T., Green S. (2011). “Cochrane handbook for systematic reviews of interventions,” in Cochrane Collaboration, eds Higgins J. P. T., Green Sally. (Hoboken, NJ: John Wiley & Sons; ). [Google Scholar]
  74. Horton J., Rand D., Zeckhauser R. (2011). The online laboratory: conducting experiments in a real labor market. Exp. Econ. 14 399–425. 10.1007/s10683-011-9273-9 [DOI] [Google Scholar]
  75. Ioannidis J. P. (2005). Why most published research findings are false. PLoS Med. 2:e124. 10.1371/journal.pmed.0020124 [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Jessoe K., Lade G. E., Loge F., Spang E. (2017). Spillover from behavioral interventions: experimental evidence from water and energy use. E2e Working Paper 033, Sacramento, CA. [Google Scholar]
  77. Jones S. R. G. (1992). Was there a hawthorne effect? Am. J. Sociol. 98 451–148. 10.1086/230046 [DOI] [Google Scholar]
  78. Juhl H. J., Fenger M. H. J., Thøgersen J. (2017). Will the consistent organic food consumer step forward? An empirical analysis. J. Consum. Res. 44 519–535. 10.1093/jcr/ucx052 [DOI] [Google Scholar]
  79. Juvina I., Saleem M., Martin J. M., Gonzalez C., Lebiere C. (2013). Reciprocal trust mediates deep transfer of learning between games of strategic interaction. Organ. Behav. Hum. Decis. Process. 120 206–215. 10.1016/j.obhdp.2012.09.004 [DOI] [Google Scholar]
  80. Kaida N., Kaida K. (2015). Spillover effect of congestion charging on pro-environmental behavior. Environ. Dev. Sustain. 7 409–421. 10.1007/s10668-014-9550-9 [DOI] [Google Scholar]
  81. Karmarkar U. R., Bollinger B. (2015). BYOB: how bringing your own shopping bags leads to treating yourself and the environment. J. Market. 79 1–5. 10.1509/jm.13.0228 [DOI] [Google Scholar]
  82. Karremans J. C., Van Lange P. A. M., Holland R. W. (2005). Forgiveness and its associations with prosocial thinking, feeling, and doing beyond the relationship with the offender. Pers. Soc. Psychol. Bull. 31 1315–1326. 10.1177/0146167205274892 [DOI] [PubMed] [Google Scholar]
  83. Kesternich M., Roemer D., Flues F. (2017). The power of active choice: field experimental evidence on repeated contribution decisions to a carbon offsetting program. Euro. Econ. Rev. 114 76–91. 10.1016/j.euroecorev.2019.02.001 [DOI] [Google Scholar]
  84. Klein F. (2017). A Drop in the Ocean? Behavioral Spillover Effects and Travel Mode Choice. Master Thesis, Swedish University of Agricultural Sciences, Uppsala. [Google Scholar]
  85. Kormos C., Gifford R. (2014). The validity of self-report measures of pro-environmental behavior: a meta-analytic review. J. Environ. Psychol. 40 359–371. 10.1016/j.jenvp.2014.09.003 [DOI] [Google Scholar]
  86. Krieg J., Samek A. (2017). When charities compete: a laboratory experiment with simultaneous public goods. J. Behav. Exp. Econ. 66 40–57. 10.1016/j.socec.2016.04.009 [DOI] [Google Scholar]
  87. Krpan D., Galizzi M. M., Dolan P. (2019). Looking at behavioural spillovers at the mirror: the ubiquitous case for behavioural spillunders (in press). [DOI] [PMC free article] [PubMed] [Google Scholar]
  88. Lacasse K. (2015). Addressing the “go green” debate: policies that encourage small green behaviors and their political spillover effects. J. Sci. Policy Governance 3 1–33. [Google Scholar]
  89. Lacasse K. (2016). Don’t be satisfied, identify! Strengthening positive spillover by connecting pro-environmental behaviors to an “environmentalist” label. J. Environ. Psychol. 48 149–158. 10.1016/j.jenvp.2016.09.006 [DOI] [Google Scholar]
  90. Lacasse K. (2017). Can’t hurt, might help: examining the spillover effects from purposefully adopting a new pro-environmental behaviour. Environ. Behav. 51 259–287. 10.1177/0013916517748164 [DOI] [Google Scholar]
  91. Lally P., van Jaarsveld C., Potts H., Wardle J. (2010). How habits are formed: modelling habit formation in the real world. Eur. J. Soc. Psychol. 40 998–1009. 10.1002/ejsp.674 [DOI] [Google Scholar]
  92. Lanzini P., Thøgersen J. (2014). Behavioral spillover in environmental domain: an intervention study. J. Environ. Psychol. 40 381–390. 10.1016/j.jenvp.2014.09.006 [DOI] [Google Scholar]
  93. Lauren N., Fielding K. S., Smith L., Louis W. R. (2016). You did, so you can and you will: self-efficacy as a mediator of spillover from easy to more difficult pro-environmental behavior. J. Environ. Psychol. 48 191–199. 10.1016/j.jenvp.2016.10.004 [DOI] [Google Scholar]
  94. Lauren N., Smith L. D., Louis W. R., Dean A. J. (2017). Promoting spillover: how past behaviors increase environmental intentions by cueing self-perceptions. Environ. Behav. 51 235–258. [Google Scholar]
  95. Lawler E. C. (2018). Three Essays in Health Economics: Evidence From U.S. Vaccination Policy. Doctoral Thesis, Vanderbilt University, Nashville. [Google Scholar]
  96. Levitt S. D., List J. A. (2007a). Viewpoint: on the generalizability of lab behaviour to the field. Can. J. Econ. 40 347–370. 10.1111/j.1365-2966.2007.00412.x [DOI] [Google Scholar]
  97. Levitt S. D., List J. A. (2007b). What do laboratory experiments measuring social preferences reveal about the real world? J. Econ. Perspect. 21 153–174. 10.1257/jep.21.2.153 [DOI] [Google Scholar]
  98. Levitt S. D., List J. A. (2011). Was there really a Hawthorne effect at the Hawthorne Plant? An analysis of the original illumination experiments. Am. Econ. J. Appl. Econ. 3 224–238. 10.1257/app.3.1.224 [DOI] [Google Scholar]
  99. List J. A. (2011). Why economists should conduct field experiments and 14 tips for pulling one off. J. Econ. Perspect. 25 3–16. 10.1257/jep.25.3.321595323 [DOI] [Google Scholar]
  100. Littleford C., Ryley T. J., Firth S. K. (2014). Context, control, and the spillover of energy use behaviours between office and home settings. J. Environ. Psychol. 40 157–166. 10.1016/j.jenvp.2014.06.002 [DOI] [Google Scholar]
  101. Liu T. X., Bednar J., Chen Y., Page S. (2018). Directional behavioural spillover and cognitive load effects in multiple repeated games. Exp. Econ. 1–30. [Google Scholar]
  102. Ludwig T. D., Geller E. S. (1997). Assigned versus participative goal setting and response generalization: managing injury control among professional pizza deliverers. J. Appl. Psychol. 82:253. 10.1037/0021-9010.82.2.253 [DOI] [PubMed] [Google Scholar]
  103. Margetts E. A., Kashima Y. (2016). Spillover between pro-environmental behaviors: the role of resources and perceived similarity. J. Environ. Psychol. 49 30–42. 10.1016/j.jenvp.2016.07.005 [DOI] [Google Scholar]
  104. McCoy D., Lyons S. (2017). Unintended outcomes of electricity smart-metering: trading-off consumption and investment behaviour. Energy Effic. 10 299–318. 10.1007/s12053-016-9452-9 [DOI] [Google Scholar]
  105. Miguel E., Camerer C. F., Casey K., Cohen J., Esterling K. M., Gerber A., et al. (2014). Promoting transparency in social science research. Science 343 30–31. 10.1126/science.1245317 [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Moher D., Liberati A., Tetzlaff J., Altman D. G. the PRISMA Group (2009). Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. Ann. Internal Med. 151 264–269. 10.7326/0003-4819-151-4-200908180-00135 [DOI] [PubMed] [Google Scholar]
  107. Mullen E., Monin B. (2016). Consistency versus licensing effects of past moral behavior. Ann. Rev. Psychol. 67 363–385. 10.1146/annurev-psych-010213-115120 [DOI] [PubMed] [Google Scholar]
  108. Munafò M. R., Nosek B., Bishop D. V. M., Button K. S., Chambers C. D., Percieet al. (2017). A manifesto for reproducible science. Nat. Hum. Behav. 1:0021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  109. Nash N., Capstick S., Whitmarsh L. (2019). Behavioural spillover in context: negotiating environmentally-responsible lifestyles in Brazil, China and Denmark. Front. Psychol. (in press). [DOI] [PMC free article] [PubMed] [Google Scholar]
  110. Nash N., Whitmarsh L., Capstick S., Hargreaves T., Poortinga W., Thomas G., et al. (2017). Climate-relevant behavioural spillover and the potential contribution of social practice theory. WIREs Clim. Change 8:e481. [Google Scholar]
  111. Nilsson A., Bergquist M., Schultz W. P. (2016). Spillover effects in environmental behaviors, across time and context: a review and research agenda. Environ. Educ. Res. 3 1–7. [Google Scholar]
  112. Nolan J. M. (2011). The cognitive ripple of social norms communications. Group Process. Intergroup Relat. 14 689–702. 10.1177/1368430210392398 [DOI] [Google Scholar]
  113. Norden A. (2013). Essays on Behavioral Economics and Policies for Provision Of Ecosystem Services. Doctoral Thesis, Goteborg University, Goteborg. [Google Scholar]
  114. Open Science Collaboration (2015). Estimating the reproducibility of psychological science. Science 349:6251. [DOI] [PubMed] [Google Scholar]
  115. Panos M. (2018). Treat Yourself or Promote Your Health: A Presentation and Examination of the Mechanisms Behind Health Behavior Spillover. Doctoral Thesis, University of Minnesota, Minneapolis. [Google Scholar]
  116. Paolacci G., Chandler J. (2014). Inside the turk: understanding mechanical turk as a participant pool. Curr. Direct. Psychol. Sci. 23 184–188. 10.1177/0963721414531598 [DOI] [Google Scholar]
  117. Paolacci G., Chandler J., Ipeirotis P. G. (2010). Running experiments on Amazon mechanical turk. Judgment Decis. Making 5 411–419. [Google Scholar]
  118. Peters A. M., van der Werff E., Steg L. (2018). Beyond purchasing: electric vehicle adoption motivation and consistent sustainable energy behaviour in the Netherlands. Energy Res. Soc. Sci. 39 234–247. 10.1016/j.erss.2017.10.008 [DOI] [Google Scholar]
  119. Polizzi di Sorrentino E., Woelbert E., Sala S. (2016). Consumers and their behavior: state of the art in behavioral science supporting use phase modeling in LCA and ecodesign. Int. J. Life Cycle Assess. 21 237–251. 10.1007/s11367-015-1016-2 [DOI] [Google Scholar]
  120. Poortinga W., Whitmarsh L., Suffolk C. (2013). The introduction of a single-use carrier bag charge in Wales: attitude change and behavioural spillover effects. J. Environ. Psychol. 36 240–247. 10.1016/j.jenvp.2013.09.001 [DOI] [Google Scholar]
  121. Poroli A., Huang L. V. (2018). Spillover effects of a university crisis: a qualitative investigation using situational theory of problem solving. J. Mass Commun. Quart. 95 1128–1149. 10.1177/1077699018783955 [DOI] [Google Scholar]
  122. Reicher S. (2000). Against methodolatry: some comments on Elliott. Fischer, and Rennie. Br. J. Clin. Psychol. 39 1–6. 10.1348/014466500163031 [DOI] [PubMed] [Google Scholar]
  123. Rodriguez-Muñoz A., Sanz-Vergel A. I., Demerouti E., Bakker A. B. (2014). Engaged at work and happy at home: a spillover-crossover model. J. Happin. Stud. 15 271–283. 10.1007/s10902-013-9421-3 [DOI] [Google Scholar]
  124. Santarius T., Soland M. (2018). How technological efficiency improvements change consumer preferences: towards a psychological theory of rebound effects. Ecol. Econ. 146 414–424. 10.1016/j.ecolecon.2017.12.009 [DOI] [Google Scholar]
  125. Savikhin A. (2010). Essays on Experimental and Behavioral Economics. Doctoral Thesis, Purdue University, West Lafayette. [Google Scholar]
  126. Savikhin A. C., Sheremeta R. M. (2013). Simultaneous decision-making in competitive and cooperative environments. Econ. Inq. 51 1311–1323. 10.1111/j.1465-7295.2012.00474.x [DOI] [Google Scholar]
  127. Schatzki T. (2010). The Timespace of Human Activity. Lexington: Lexington Books. [Google Scholar]
  128. Schmitz J. (2018). Temporal dynamics of pro-social behaviour – An experimental analysis. Exp. Econ. 22 1–23. 10.1007/s10683-018-9583-2 29327424 [DOI] [Google Scholar]
  129. Schütte L., Gregory-Smith D. (2015). Neutralisation and mental accounting in ethical consumption: the case of sustainable holidays. Sustainability 7 1–14. 10.3390/su7067959 [DOI] [Google Scholar]
  130. Seebauer S. (2018). The psychology of rebound effects: explaining energy efficiency rebound behaviours with electric vehicles and building insulation in Austria. Energy Res. Soc. Sci. 4 6 311–320. 10.1016/j.erss.2018.08.006 [DOI] [Google Scholar]
  131. Sheremeta R. M., Cason T. N., Samek A. (2010). Behavioral spillovers in coordination games. Euro. Econ. Rev. 56 233–245. 10.2139/ssrn.1755609 [DOI] [Google Scholar]
  132. Shreedhar G. (2018). Experiments in Behavioural Environmental Economics. Doctoral Thesis, London School of Economics and Political Science, London. [Google Scholar]
  133. Shreedhar G., Mourato S. (2018). Do biodiversity conservation videos cause pro-environmental spillover effects? Grantham Research Institute on Climate Change and the Environment Working Paper No. 302, London. [Google Scholar]
  134. Silverman D. (2001). Interpreting Qualitative Data: Methods for Analysing Talk, Text and Interaction, 2nd Edn London: Sage Publications. [Google Scholar]
  135. Simmons J. P., Nelson L. D., Simonsohn U. (2011). False-positive psychology: undisclosed flexibility in data collection and analysis allows presenting anything as significant. Psychol. Sci. 22 1359–1366. 10.1177/0956797611417632 [DOI] [PubMed] [Google Scholar]
  136. Simonsohn U., Nelson L. D., Simmons J. P. (2014). P-curve and effect size: correcting for publication bias using only significant results. Perspect. Psychol. Sci. 9 666–681. 10.1177/1745691614553988 [DOI] [PubMed] [Google Scholar]
  137. Sintov N., Geislar S., White L. V. (2019). Cognitive accessibility as a new factor in pro-environmental spillover: results from a field study of household food waste management. Environ. Behav. 51 50–80. 10.1177/0013916517735638 [DOI] [Google Scholar]
  138. Sorrell S., Dimitropoulos J., Sommerville M. (2009). Empirical estimates of the direct rebound effect: a review. Energy Policy 37 1356–1371. 10.1016/j.enpol.2008.11.026 [DOI] [Google Scholar]
  139. Spence A., Leygue C., Bedwell B., O’Malley C. (2014). Engaging with energy reduction: does a climate change frame have the potential for achieving broader sustainable behaviour? J. Environ. Psychol. 38 17–28. 10.1016/j.jenvp.2013.12.006 [DOI] [Google Scholar]
  140. Steinhorst J., Klockner C. A., Matthies E. (2015). Saving electricity – For the money or the environment? Risks of limiting pro-environmental spillover when using monetary framing. J. Environ. Psychol. 43 125–135. 10.1016/j.jenvp.2015.05.012 [DOI] [Google Scholar]
  141. Steinhorst J., Matthies E. (2016). Monetary or environmental appeals for saving electricity? Potentials for spillover on low carbon policy acceptability. Energy Policy 93 335–344. 10.1016/j.enpol.2016.03.020 [DOI] [Google Scholar]
  142. Suffolk C. (2016). Rebound and Spillover Effects: Occupant Behaviour After Energy Efficiency Improvements are Carried Out. Doctoral Thesis, Cardiff University, Cardiff. [Google Scholar]
  143. Suffolk C., Poortinga W. (2016). “Behavioural changes after energy efficiency improvements in residential properties,” in Rethinking Climate and Energy Policies, eds Santarius T., Walnum H., •l C. (Berlin: Springer; ), 121–142. [Google Scholar]
  144. Swim J. K., Bloodhart B. (2013). Admonishment and praise: interpersonal mechanisms for promoting proenvironmental behaviour. Ecopsychology 5 20–31. 10.1089/eco.2012.0065 [DOI] [Google Scholar]
  145. Tashakkori A., Teddlie C. (2003). Handbook of Mixed Methods in Social and Behavioral Research. Thousand Oaks, CA: Sage Publications. [Google Scholar]
  146. Thøgersen J. (1999). Spillover processes in the development of a sustainable consumption pattern. J. Econ. Psychol. 20 53–81. 10.1016/S0167-4870(98)00043-9 [DOI] [Google Scholar]
  147. Thøgersen J., Noblet C. (2012). Does green consumerism increase the acceptance of wind power? Energy Policy 51 854–862. 10.1016/j.enpol.2012.09.044 [DOI] [Google Scholar]
  148. Thomas G. O., Poortinga W., Sautkina E. (2016). The welsh single-use carrier bag charge and behavioural spillover. J. Environ. Psychol. 47 126–135. 10.1016/j.jenvp.2016.05.008 [DOI] [Google Scholar]
  149. Thomas G. O., Sautkina E., Poortinga W., Wolstenholme E., Whitmarsh L. (2019). The English plastic bag charge changed behaviour and increased support for other charges to reduce plastic waste. Front. Psychol. 10:266. 10.3389/fpsyg.2019.00266 [DOI] [PMC free article] [PubMed] [Google Scholar]
  150. Tiefenbeck V. (2014). Behavioral Interventions to Reduce Residential Energy and Water Consumption: Impact, Mechanisms, and Side Effects. Doctoral Thesis, ETH Zurich, Zurich. [Google Scholar]
  151. Tiefenbeck V., Staake T., Roth K., Sachs O. (2013). For better or for worse? Empirical evidence of moral licensing in a behavioral energy conservation campaign. Energy Policy 57 160–171. 10.1016/j.enpol.2013.01.021 [DOI] [Google Scholar]
  152. Tippet E. (2018). Behavioral Spillover Across Health and Environmental Domains. Master Thesis, Ohio State University, Ohio. [Google Scholar]
  153. Truelove H. B., Carrico A. R., Weber E. U., Raimi K. T., Vandenbergh M. P. (2014). Positive and negative spillover of pro-environmental behavior: an integrative review and theoretical framework. Glob. Environ. Change 29 127–138. 10.1016/j.gloenvcha.2014.09.004 [DOI] [Google Scholar]
  154. Truelove H. B., Yeung K. L., Carrico A. R., Gillis A. J., Raimi K. T. (2016). From plastic bottle recycling to policy support: an experimental test of pro-environmental spillover. J. Environ. Psychol. 46 55–66. 10.1016/j.jenvp.2016.03.004 [DOI] [Google Scholar]
  155. Uzzell D., Räthzel N. (2018). Border crossing and the logics of space: a case study in pro-environmental practices. Front. Psychol. 9:2096. 10.3389/fpsyg.2018.02096 [DOI] [PMC free article] [PubMed] [Google Scholar]
  156. Van der Werff E., Steg L., Keizer K. (2014a). Follow the signal: when past pro-environmental actions signal who you are. J. Environ. Psychol. 40 273–282. 10.1016/j.jenvp.2014.07.004 [DOI] [Google Scholar]
  157. Van der Werff E., Steg L., Keizer K. (2014b). I am what I am, by looking past the present the influence of biospheric values and past behavior on environmental self-identity. Environ. Behav. 46 626–657. 10.1177/0013916512475209 [DOI] [Google Scholar]
  158. Vasan A. A. (2018). Affect as a Model of Pro-Environmental Spillover. Master Thesis, Vanderbilt University, Nashville. [Google Scholar]
  159. Verfuerth C., Gregory-Smith D. (2018). “Spillover of pro-environmental behaviour,” in Research Handbook on Employee Pro-environmental Behaviour, eds Wells V. K., Gregory-Smith D., Manika D. (Cheltenham: Edward Elgar; ). [Google Scholar]
  160. Verfuerth C., Jones C., Gregory-Smith D., Oates C. (2018). Spillover Effects Through the Lens of Identity Process Theory – A Qualitative Approach. Rome: International Association of People-Environment Studies. [Google Scholar]
  161. Vincent M., Koessler A. K. (2018). Mapping Moral Pluralism in Behavioural Spillovers: A Cross-Disciplinary Account of the Multiple Ways in Which We Engage in Moral Valuing. Hamburg: ZWB – Leibniz Information Centre for Economics Discussion Paper. [Google Scholar]
  162. Werfel S. H. (2017). Household behaviour crowds out support for climate change policy when sufficient progress is perceived. Nat. Clim. Change 7:512 10.1038/nclimate3316 [DOI] [Google Scholar]
  163. Whitmarsh L. (2009). What’s in a name? Commonalities and differences in public understanding of “climate change” and “global warming”. Public Understand. Sci. 18 401–420. 10.1177/0963662506073088 [DOI] [Google Scholar]
  164. Whitmarsh L., O’Neill S. (2010). Green identity, green living? The role of pro-environmental self-identity in determining consistency across diverse pro-environmental behaviours. J. Environ. Psychol. 30 305–314. 10.1016/j.jenvp.2010.01.003 [DOI] [Google Scholar]
  165. Whitmarsh L. E., Haggar P., Thomas M. (2018). Waste reduction behaviors at home, at work, and on holiday: what influences behavioral consistency across contexts? Front. Psychol. 9:2447. 10.3389/fpsyg.2018.02447 [DOI] [PMC free article] [PubMed] [Google Scholar]
  166. Xanthopoulou D., Papagiannidis S. (2012). Play online, work better? Examining the spillover of active learning and transformational leadership. Technol. Forecast. Soc. Change 79 1328–1339. 10.1016/j.techfore.2012.03.006 [DOI] [Google Scholar]
  167. Xie X., Lu Y., Gou Z. (2017). Green building pro-environment behaviors: are green users also green buyers? Sustainability 9:1703 10.3390/su9101703 [DOI] [Google Scholar]
  168. Xu L., Zhang X., Ling M. (2018a). Pro-environmental spillover under environmental appeals and monetary incentives: evidence from an intervention study on household waste separation. J. Environ. Psychol. 60 27–33. 10.1016/j.jenvp.2018.10.003 [DOI] [Google Scholar]
  169. Xu L., Zhang X., Ling M. (2018b). Spillover effects of household waste separation policy on electricity consumption: evidence from Hangzhou, China. Resour. Conserv. Recycl. 129 219–231. 10.1016/j.resconrec.2017.10.028 [DOI] [Google Scholar]
  170. Zawadzki S. J. (2015). The Interdependent Nature of Environmental Behaviors: Testing a Conceptual Framework for Behavioral Spillover. Master Thesis, Pennsylvania State University, State College. [Google Scholar]
  171. Zimmerman K. L. (2009). Operationalizing the Antecedents of Work-Family Positive Spillover: A Longitudinal Study. Doctoral Thesis, Portland State University, Portland. [Google Scholar]
  172. Zizzo D. J. (2010). Experimenter demand effects in economic experiments. Exp. Econ. 13 75–98. 10.1007/s10683-009-9230-z [DOI] [Google Scholar]
  173. Zwane A. P., Zinman J., Van Dusen E., Pariente W., Null C., Miguel E., et al. (2011). Being surveyed can change later behaviour and related parameter estimates. Proc. Natl. Acad. Sci. U.S.A. 108 1821–1826. 10.1073/pnas.1000776108 [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Frontiers in Psychology are provided here courtesy of Frontiers Media SA

RESOURCES