Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2019 Jun 25;9:9271. doi: 10.1038/s41598-019-45725-8

Neurocognitive mechanisms of reactions to second- and third-party justice violations

Claudia Civai 1,, Inge Huijsmans 2,3, Alan G Sanfey 2,3
PMCID: PMC6592880  PMID: 31239496

Abstract

The aim of the current study was threefold: (i) understand people’s willingness to engage in either punishment of the perpetrator or compensation of the victim in order to counteract injustice; (ii) look into the differences between victims of and witnesses to injustice; (iii) investigate the different role played by social preference and affective experience in determining these choices. The sample tested here showed an equal preference for punishment and compensation; neuroimaging findings suggested that compensation, as opposed to punishment, was related to Theory of Mind. Partially supporting previous literature, choosing how to react to an injustice as victims, rather than witnesses, triggered a stronger affective response (striatal and prefrontal activation). Moreover, results supported the idea that deciding whether or not to react to an injustice and then how severely to react are two distinct decisional stages underpinned by different neurocognitive mechanisms, i.e., sensitivity to unfairness (anterior insula) and negative affectivity (amygdala). These findings provide a fine-grained description of the psychological mechanisms underlying important aspects of social norm compliance.

Subject terms: Social behaviour, Decision

Introduction

Norms such as fairness and cooperation are fundamental principles of society, and in recent years there has been a dramatic increase in investigations of the psychological underpinnings of these social norms. Many experimental investigations have demonstrated people’s willingness to spend personal resources, in sometimes significant amounts, to promote norm compliance. For example, people often reject an unfair division of resources by punishing the proposer of an unfair offer, even when this act is personally financially costly13. Importantly, these rejections happen even when people are not the target of injustice themselves, but are instead deciding on behalf of a third-party4,5. This behaviour has typically been termed ‘costly altruistic punishment’, denoting the idea that people are often willing to override their own immediate self-interest in order to promote compliance with social norms, both when directly (i.e. Second Party (SP) punishment2,3) and non-directly (i.e. Third Party (TP) punishment5) affected by norm violations. Using one’s own resources to punish offenders is one strategy to enforce compliance - for example, most of us pay taxes which in turn finances the prison system. However, an alternative approach to emphasizing cooperative norms is that of compensating the victims of injustice, where fairness can be re-established by assisting the victim, offering aid to increase their depleted resources even though this aid comes at a personal cost611.

Although both of these approaches are reactions to a social transgression, these strategies differ substantially in terms of output: punishment discourages social norm violation by harming offenders and leaving them worse off, whereas compensation focuses on victims, signalling altruism and prosociality12. Despite these differences, and the obvious societal impact of these two methods of norm enforcement, relatively few comparative studies have been conducted exploring how these putative mechanisms of norm compliance may either overlap or differ in terms of their neural processes.

Therefore, the first goal of the current study was to investigate the neural processing of punishment and compensation. Recent research has begun to shed some light on the neural mechanisms underlying these processes, though several questions are still outstanding. In one recent study, Stallen and colleagues developed a novel task, termed the Justice Game, and found that when confronted with an obvious norm violation and then given the opportunity to take action either towards perpetrator or the victim, people typically preferred to punish the perpetrator than to compensate the victim11. This behavioural preference correlated with increased activation of the ventral striatum, one of the main components of the brain reward system13, consistent with previous research highlighting the rewarding nature of altruistic punishment3,14. However, in a separate study, Hu and colleagues found no behavioural difference between punishment and compensation rates, and also found common activation of the bilateral striatum across conditions, suggesting that reacting to injustice may always be perceived as ‘rewarding’8; on the other hand, Theory of Mind (ToM) areas such as temporo-parietal junction (TPJ)15 were associated more with compensation than punishment. This, and other findings9,10, suggests that compensation may be driven more by prosocial and other-regarding motivations.

However, though the tasks employed across these studies were superficially similar, there are some important differences to take into account: Stallen and colleagues offered participants certain trials in which they could punish and others in which they could compensate, allowing for separate assessment of the propensity to carry out each action, but which makes direct comparison between the two more difficult11; conversely, Hu and colleagues required their participants to always choose between either punishment and compensation8. Here, we developed a modified version of the aforementioned Justice Game, in which participants sometimes directly chose between punishment and compensation, and other times made a decision with only one of the two options available. This novel design enabled us to determine the preference for punishment or compensation when directly comparing the two (as in)8, as well as the general willingness to react to injustice when only one action is available (as in)11. This allowed for the categorising of participants based on their subjective preference, and the investigation of the mechanisms involved when their preferred option is, or is not, available. Neural data can then help to determine the extent to which motivations of reward (striatum) and prosociality (ToM areas) are involved in the choice. We hypothesised that reward areas were associated with a preference for punishment, whereas ToM areas were associated with a preference for compensation.

Costly altruistic punishment has been studied using laboratory-based Game Theory paradigms and hypothetical crime scenarios16. Economic games have been used to investigate both SP punishment1,3,17, i.e., when the victim of a social norm violation is given the chance to punish the transgressor, and TP punishment4,5, i.e., when the witness of a violation is given the chance to punish the transgressor, whereas crime scenarios have been mostly employed to investigate TP punishment1821. Recent meta-analyses of the neural basis of altruistic punishment suggested that TP punishment is an extension of SP punishment, and that the two share the same cognitive processes and neural underpinnings: sensitivity to norm violation and norm compliance (anterior insula (AIns) and dorsolateral prefrontal cortex (DLPFC)), negative affect (amygdala) and integration between self-referential (medial prefrontal cortex (MPFC) and posterior cingulate cortex (PCC)) and other-regarding (TPJ) processes16,18. Nevertheless, previous neuroimaging research directly comparing SP and TP punishment has shown both similarities and differences between the two: in terms of neural activity, stronger sensitivity to fairness norm violations and norm compliance, observed as increased AIns and DLPFC activation, can explain the rejection of unfair actions of another, as well as punishment, both in SP and in TP16,17,2225. However, people tend to exhibit different behaviours, with higher punishment rates observed when they themselves are victims (SP), as compared to when they witness (TP)5,11 an injustice; moreover, ventral striatum has been found to be more active for SP as compared to TP, suggesting that the rewarding sensation associated with punishment is heightened with personal involvement11,14. The second goal of this study was, therefore, to compare Second Party (SP) and Third Party (TP) punishment. Exploring how neural processing may be similar (and different) across these two conditions can help in disentangling mechanisms associated with general punishment processes and ones that are tied to personal involvement. We aimed to collect further evidence to investigate whether two different mechanisms, i.e., sensitivity to unfairness and reward, contribute to explaining SP and TP punishment differences: our task allowed for direct comparisons between SP and TP conditions when evaluating the extent of the injustice (when participants see the actions of others) as well as when making a decision (when they decide whether or not to act upon injustice). Based on previous studies, we expected that participants would punish more in SP, and that reward areas would be more active when choosing in the SP compared to the TP condition11,14. On the other hand, we hypothesised no difference between SP and TP in the activation of areas associated with sensitivity to unfairness and norm compliance (AIns, DLPFC) when the division is evaluated22,23.

Finally, the third objective of this study was to distinguish between two potentially different motives underpinning why we react to unfairness and injustice: social preference and affective experience. As proposed by11, the decision to react to injustice is underpinned by social motivations, such as inequality aversion17,26, whereas the severity of the reaction, i.e., how much people are willing to punish, might rather be linked to an affective experience, such as anger and frustration associated with unfairness27. Adapting11, the novel paradigm we used here allowed for the assessment of two psychologically distinct decision stages, one where participants indicated whether or not they want to act upon the injustice (response selection), and a second when they decided the severity of the punishment/compensation (how much they want to spend). We hypothesized that, if different motives act in distinct stages of the decision process, unfairness-related areas such as AIns17,22,23 would correlate with the decision to react to injustice, whereas areas associated with negative affective experience, such as amygdala28, would correlate with the amount of money spent to punish or compensate.

Results

Behavioral results

The target (SP or TP), the type of reaction to injustice made available (punish in SP; punish, compensate or both in TP), and the level of the injustice (0, 25, 50, 75, 100 coins could be taken from either the participant or the third-party) were manipulated. We measured the frequency of each reaction and the amount spent. No participant was excluded from the behavioural analysis. Four participants never reacted to injustice and always chose to Leave; one participant never reacted in the TP condition, and two never reacted in the SP condition; four participants never punished, neither in SP nor in TP, but did compensate; two participants never punished in TP, but did compensate; three participants never compensated, but did punish in TP. The remaining 22 participants chose to compensate and punish in SP and TP at least once. Analyses were performed using SPSS 2529. A repeated-measure Multivariate Analysis of Variance (MANOVA) was run considering two factors: Task and Injustice. Task had 3 levels: Second Party punishment (SPpun), Third Party punishment (TPpun), Third Party compensation (TPcomp). The analysis was limited to TP trials where only 2 options were available, i.e., the choice between punishment and no reaction (Leave or Take (LT) trials), or between compensation and no reaction (Leave or Give (LG) trials). Injustice had 5 levels: 0, 25, 50, 75, 100 out of 200. These were the amount of chips that player A could steal from either the participant (in SP condition) or player B (in TP condition). Frequency of reaction to injustice and amount spent to react were the dependent variables. Order of presentation was considered a between-participants factor. A significant main effect of Injustice was found both for frequency and for amount (frequency: F(4, 144) = 63.18, p < 0.001, partial η2 = 0.63; amount: F(4, 144) = 40.08, p < 0.001, partial η2 = 0.53), indicating that participants are more willing to act as a function of the level of injustice, and spend more as the injustice becomes greater. A significant main effect of Task was found for amount (F(2, 72) = 8.44, p < 0.001, partial η2 = 0.19), but not for frequency (Fig. 1a,b). Because of our interest in comparing SP and TP punishment, and TP punishment and compensation, planned comparisons were performed on both amount and frequency; the results of these comparisons, Bonferroni-corrected, indicated a significant difference between both SPpun and TPpun amounts (t(37) = 2.68, p < 0.05, Cohen’s d = 0.44), and SPpun and TPpun frequencies (t(37) = 2.91, p < 0.05 Cohen’s d = 0.47), across injustice levels (Fig. 1c,d). A post hoc comparison for the amount also revealed a significant difference between SPpun and TPcomp amounts (t(37) = 3.6, p < 0.01, Cohen’s d = 0.58). No difference between TPpun and TPcomp was found, neither for frequency nor for amount. Two t-tests were performed to analyse the difference between TP punishment and compensation frequencies in relation to the number of options (2-option or 3-option conditions): punishment in LT trials (i.e., only punishment option) was compared with compensation in LG trials (i.e., only compensation option), and punishment and compensation were compared in Leave-Take-Give (LTG) trials (i.e., both options available). No significant difference was found, suggesting an equal preference for the two strategies in our sample. No effect of order of presentation was found. Given the deceptive nature of the task, participants were asked at the end if they had any doubt about the plausibility of the final outcome and, whilst some were sceptical, most of them did not report any doubt.

Figure 1.

Figure 1

Behavioral results: (a) frequency of reaction and (b) amount spent for Task (Second Party (SP) punishment, Third Party (TP) punishment, TP compensation) and Injustice level represented in terms of amount of chips owned by both players A and B after player A’s decision (0 chips taken, or 200:200 = fair division; 100 chips taken, or 300:100 = extremely unfair division); (c,d) frequencies and amounts for each Task (SPpun; TPpun; TPcomp) for unfair trials (injustice levels collapsed across unfair divisions; NS = non-significant difference, *p < 0.05, **p < 0.01).

fMRI results

In addition to whole-brain analysis, given the clear brain location hypotheses based on previously discussed literature results, small volume correction (svc) was applied to the contrasts of interest. The following anatomical masks, built with Neuromorphometrics Atlas (MRI scans originated by the OASIS project, and labelled data provided by Neuromorphometrics, Inc), were used to determine the small volumes of interest: (1) AIns and Middle Frontal Gyrus (DLPFC) (unfairness sensitivity and norm compliance) for contrasts determining unfairness-related activation, i.e., main effect of injustice; (2) nucleus accumbens, caudate, putament (i.e., striatum, reward-related area) and angular gyrus and supramarginal gyrus (i.e., TPJ, ToM area) for contrasts looking at differences between punishment and compensation, for both the Option and the Choice model (see Methods section); (3) AIns, Middle Frontal Gyrus (DLPFC) and striatum, for contrasts comparing SP and TP, both when evaluating unfairness (Division screen) and when responding to it (Response screen), for both the Option and the Choice model; (4) AIns and amygdala (affective response) for contrasts looking at the differences between motives underlying reaction to injustice, i.e., Response selection and Amount selection. All results are reported using a common cluster-corrected threshold at p < 0.05 FWE on the basis of a whole-brain voxel-wise threshold of p < 0.001, unless otherwise specified in the tables (svc, cluster-corrected threshold at p < 0.05 FWE).

The reason why we have selected the current areas is that the aim was to focus on specific cognitive processes hypothesised to be involved in the behaviours of interest: reward and prosociality in punishment and compensation, sensitivity to injustice and reward in second-party versus third-party condition, and sensitivity to injustice and negative affect in response selection versus severity of the response. We have selected the brain areas associated with these processes based on previous studies8,11 using the same or a very similar paradigm. Nevertheless, the results of whole-brain analyses are reported, and the involvement of other relevant areas reported in recent meta-analyses16,18, such as MPFC and PCC, is discussed.

Comparison between punishment and compensation in response selection

No significant cluster was found when contrasting activations associated with only-punishment and only-compensation option (RespTP_LT-RespTP_LG contrast, or its reverse, Option model). When considering participants’ choices (Choice model), no significant cluster was found for punishing vs compensating choices (RespTPpun-RespTPcomp); the reverse contrast, i.e., compensating vs punishing choices (RespTPcomp- RespTPpun), returned activation in a small cluster of left TPJ, which, however, did not survive svc.

The full factorial analysis showed significant clusters in the precuneus, right TPJ and right middle temporal gyrus (MidTG) for Compensators compared to Punishers, supporting the idea that ToM areas are more involved in compensation (Fig. 2a, Table 1a). No significant cluster was found for the reverse main effect of Preference (Punishers vs Compensators). When the preferred option was available (Compensators could compensate (LG trials) and Punishers could punish (LT trials)), one cluster was active in ventromedial prefrontal cortex (VMPFC) (Fig. 2b, Table 1b), which survived svc (anatomical medial prefrontal gyrus used as small volume). Although VMPFC was not in our original list of regions of interest, this finding is worth of mention given the widely reported involvement of this area in subjective value-based choices and affect computations30. For the reverse contrast, i.e., when Compensators could only punish and Punishers could only compensate, no cluster survived neither whole-brain correction nor svc.

Figure 2.

Figure 2

Areas associated with preference for punishment and compensation: (a) Activation of TPJ, MidTG and precuneus when Compensators, as opposed to Punishers, selected a response; (b) Activation of VMPFC associated with response selection when the preferred option is the only one available. Activation maps are superimposed on an MNI template.

Table 1.

Significant clusters correlating with Punishment vs Compensation (peak MNI coordinates are reported; svc: small volume correction, based on a-priori defined anatomical areas).

Brain region Lateral x y z K (cluster size) Z pfwe-corr
(a) Compensators-Punishers (Option model, full factorial)
Precuneus R 4 −70 35 154 4.90 <0.001
MidTG R 50 −46 −4 135 4.50 0.002
Angular gyrus (TPJ) R 43 −60 24 30 3.72 0.039 (svc)
(b) Preferred choice exclusively available - (Compensators_LG + Punishers_LT)-rest (Option model, full factorial)
VMPFC R 4 46 −21 20 3.89 0.024 (svc)

Comparison between SP and TP

Injustice perception. The contrast between unfair vs fair divisions (DivUnfair-DivFair) was associated with significant clusters in the DLPFC and AIns (Table 2a). A significant whole-brain activation of the fusiform gyrus and the superior parietal lobule was found when comparing SP to TP. No suprathreshold activation was found in the selected ROIs, and no difference was found between SP and TP when processing unfair divisions (DivSP(Unfair-Fair)-DivTP(Unfair-Fair)).

Table 2.

Significant clusters correlating with second-party (SP) and third-party (TP) in Division and Response screen (peak MNI coordinates are reported; svc: small volume correction, based on a-priori defined areas).

Brain region Lateral x y z K (cluster size) Z pfwe-corr
(a) Main effect of Injustice - DivUnfair-DivFair (Option model)
Superior Parietal Lobule R 29 −63 38 82 Inf 0.013
Superior Parietal Lobule L −30 60 46 92 Inf 0.009
Thalamus R 22 −28 0 151 4.71 0.001
Pallidum L −10 7 4 162 4.66 0.001
DLPFC – Middle Frontal Gyrus L −41 7 32 27 4.84 0.041 (svc)
AIns L −34 18 4 23 5.85 0.05 (svc)
AIns R 32 21 7 24 5.47 (svc)
(b) Main effect of Target – DivSP-DivTP (Option model)
Fusiform Gyrus L −38 −52 −18 269 4.96 <0.001
Fusiform Gyrus R 40 −56 −18 120 4.31 <0.001
Superior Parietal Lobule R 29 −66 32 79 4.15 0.008
Inferior Occipital Gyrus R 32 −91 0 72 3.97 0.012
(c) Deciding in SP vs TP - RespSP-RespTP_LT (Option model)
MPFC/ACC L −6 38 7 198 4.94 <0.001
Middle Occipital Gyrus L −48 −77 7 113 6.14 <0.001
Precuneus L 6 −56 14 78 4.63 0.013
Posterior cingulate gyrus L −2 32 46 150 4.83 0.001
Planum Temporale R 54 −32 21 758 5.54 <0.001
Superior Temporal Gyrus L −62 −10 4 719 5.47 <0.001
Striatum -Putamen R 29 −14 7 62 4.49 0.006 (svc)
Striatum -Putamen L −24 4 −10 45 4.41 0.016 (svc)
(d) Deciding in TP vs SP – RespTP_LT-RespSP (Option model)
Calcarine Cortex L −13 −91 −7 723 6.90 <0.001
Superior Parietal Lobule R 29 −63 38 80 4.38 <0.001
(e) Deciding to punish in SP vs TP – RespSPpun-RespTPpun (Choice model)
ACC/White matter R 22 38 10 89 4.60 0.005

Response selection. When comparing the response time-window in SP and TP (Option model, contrast RespSP-RespTP_LT), clusters in the PCC, the dorsolateral striatum, and the MPFC/anterior cingulate cortex (ACC) were more active in SP compared to TP (Fig. 3 and Table 2b). Significant clusters in the occipital and parietal areas were found both for RespSP-RespTP_LT and for the reverse contrast (Table 2c). These activations are likely due to the difference in the visual appearance of the stimuli (words) in SP and TP. A significant cluster in a region encompassing the ACC and the right medial and dorsal PFC, as well as white matter, was found when participants decided to punish in SP compared to TP (Choice model, contrast RespSPpun-RespTPpun; Table 2d). No significant activation was found for the reverse contrast, i.e. when participants decided to punish in TP compared to SP.

Figure 3.

Figure 3

Areas associated with responding in Second Party (SP) as opposed to Third Party (TP): activation of MPFC and striatum. Activation map is superimposed on an MNI template.

Differences between Response Selection and Amount

A marginally significant activation in the left AIns was associated with the reaction to injustice (Response selection) as opposed to the severity of reaction (Amount selection) (Fig. 4a, Table 3a). Significant activation in the occipital cortex, most likely associated with the difference in the visual stimulation between the two separate time windows, was also found. The reverse contrast, indicating areas that increase their activation with the severity of the reaction and are more involved in this process compared to response selection, showed significant activations of the MPFC, bilateral central operculum/insula, and the amygdala (Fig. 4b, Table 3b).

Figure 4.

Figure 4

Areas associated with the willingness to react to injustice and the severity of reaction: (a) AIns associated with willingness to react; (b) ACC/VMPFC and amygdala associated with severity of reaction. Activation maps are superimposed on an MNI template.

Table 3.

Significant clusters correlating with the difference between Response Selection (willingness to react to injustice) and Amount (severity of reaction) (peak MNI coordinates are reported; svc: small volume correction, based on a-priori defined anatomical areas).

Brain region Lateral x y z K (cluster size) Z pfwe-corr
(a) Response vs Amount - [Resp_TP(pun + comp) + Resp_SP] − [AmTP(pun + comp)param + AmSPparam] (Choice model)
Calcarine Cortex R 15 −88 −4 113 5.44 0.003
Occipital Pole L −16 −94 −4 204 3.15 <0.001
AIns L −30 21 7 5 3.77 0.06 (svc)
(b) Amount vs Response - [AmTP (pun + comp)param + AmSPparam] − [Resp_TP(pun + comp) + Resp_SP] (Choice model)
Lingual Gyrus R 1 −74 0 256 4.47 <0.001
Inferior Occipital Gyrus R 46 −70 7 91 4.22 <0.001
MPFC/ACC L −10 46 −10 541 4.51 <0.001
Operculum/PIns L −58 −7 7 190 4.86 <0.001
Operculum/PIns R 54 −4 7 635 5.09 <0.001
Amygdala R 15 −4 −18 11 4.12 0.036 (svc)

Discussion

People typically have strong responses when they perceive injustice, and can react in different ways, such as by punishing offenders or compensating victims. These responses have different social consequences, i.e., a decrease in the offenders’ resources or an increase in the victims’ ones, and therefore investigating these decision processes in detail is crucial to better understand the psychosocial elements of justice and equality. In the current study, we addressed three issues: the preference to either punish a perpetrator or compensate a victim; the comparison between a willingness to punish when one is the victim as compared to when one is an unaffected witness; and the process differences between the initial reaction to injustice and the severity of that reaction.

Our participants demonstrated an equal preference for both punishment and compensation. Previous literature is inconclusive on whether there is a dominant strategy to enforce norm compliance. Some evidence points towards a preference for punishment over compensation11, whereas other studies have found the opposite pattern7. Further, some research shows no overall preference6,8, but suggests that individual differences in empathy and perspective taking may explain why some people prefer to compensate. Other personal traits may explain preference for punishment, but further research is needed in this direction. It has been shown that compensation is a more effective act than punishment for reputation enhancement, and thus may be the preferred choice in situations where there is no anonymity in the social exchange7,12, that is, where reputation is meaningful. In laboratory studies such as this one, reputation is likely less salient given the anonymous interactions. It is worth noting that this does not diminish findings from lab-based social decision-making studies: on the contrary, a more socially sterile environment may help elicit behaviours and motivations that are otherwise hidden, potentially mimicking the conditions created by internet and social media where people observe and engage in social interactions from an often solitary and isolated space. We acknowledge that some scepticism reported by few participants over the plausibility of the outcomes may have influenced the results; nevertheless, results of experiments using deception are usually in line with experiments where deception is not used31.

Based on previous literature, it was hypothesised that punishment would involve the reward system3,11,14, whereas compensation would involve perspective taking and ToM810. The current findings only partially support these hypotheses. We found no evidence that the reward system was specifically associated with punishment as opposed to compensation, which could be explained by the fact that participants did not show a preference for punishing over compensating behaviour in the third-party condition. On the other hand, TPJ and MidTG (ToM areas) were associated with an individual preference for compensation. A recent meta-analysis15 found TPJ as the only area involved in all aspects of ToM (from reading emotions from facial expressions to inferring others’ beliefs), whereas MidTG was specifically involved in goal directed tasks, i.e., tasks in which participants need to identify the specific goal to follow an action. Therefore, the involvement of MidTG in compensation may suggest that, in this type of scenario, compensating victims might be regarded as the most suitable action to perform. Importantly, these results suggest that ToM regions are not necessarily associated with the specific choice to compensate versus punish, but rather with the decisional stage in Compensators (individuals who, overall, prefer compensation); this suggests that Compensators engaged ToM abilities more than Punishers when choosing how to react to an injustice, irrespective of the specific choice made. VMPFC activation was correlated with the availability of the preferred option (when Compensators could compensate, and when Punishers could punish): it has been shown that VMPFC activity increases when comparing subjectively valuable options13,30,32; hence, this suggests that punishment is not necessarily more rewarding than compensation, or vice-versa, as the two do not have a distinguishable absolute value, but rather a subjective value that depends on individual preferences.

Based on previous literature on second- and third-party punishment, we expected people to punish more, and more harshly, in SP compared to TP condition5,11,14. We confirmed this hypothesis, in that we found that participants were willing to punish more and spend more to punish when they were directly damaged by injustice (SP) compared to when they witnessed injustice (TP). Nevertheless, neuroimaging findings support the idea that sensitivity to injustice does not differ between SP and TP22,23, since DLPFC and AIns, both of which are associated with unfairness sensitivity17,18,33,34, were more active for unfair compared to fair divisions in both SP and TP, with no difference found between the two conditions. The differences between the two conditions lay in activation of the dorsolateral striatum, the PCC and the MPFC/ACC. The striatum was more active when selecting a response in SP compared to TP; however, this activation did not specifically relate to punishing choices. Arguably, the greater involvement of this area in second-party scenarios is associated with decisions made in personally salient social situations35, which trigger affective responses irrespective of the value of the specific decision. According to the accounts put forward by [16, 18], the PCC and MPFC, in particular the medial portion, are part of the default mode network modulated by self-referential processes in altruistic punishment. These findings suggest that these processes are more involved in second- versus third-party punishment, possibly also reflecting the increased personal saliency of the SP condition. On the other hand, the ACC and dorsomedial portion of the prefrontal cortex were more active when punishing in SP compared to TP, supporting the idea that this area signals the affective reaction to unfairness and injustice in the presence of personal involvement36. Nevertheless, this interpretation remains speculative, given that the activation encompassed a large portion of white matter.

The current design breaks the decision down into multiple steps. One limitation of this is that, in the SP condition, participants could potentially make their decision very early, even at the Division screen stage, since the choice of options was fixed (Take or Leave). We reasoned that if participants did adopt this strategy in SP, decision activity would have increased in the Division screen as compared to the TP condition. Moreover, it would have decreased decision activity in the Response selection screen compared to TP, since participants in SP would have already made their choice by this time. A significant cluster in the superior parietal lobule was found when comparing SP and TP during the Division screen; according to [16, 18], this suggests that, in SP, participants were starting to decide on a punishment as soon as the division was presented. However, no difference between SP and TP was found when considering only the unfair divisions, hence making this interpretation less plausible. On the other hand, neural activity in the Response screen was larger for SP compared to TP in areas which have been associated with self-referential processes in decision-making16,18,36. Therefore, we assume that timing of the decisions was not completely shifted forward in SP.

The final goal of this study was to compare two different aspects of the decision process in response to injustice: deciding whether or not to react and, if so, how harshly. The computational model developed by Stallen and colleagues allowed for distinguishing the punishment choice itself from the actual severity of punishment; when applied to neuroimaging data, it was found that AIns correlated with the willingness to punish, whereas amygdala correlated with the severity of punishment. Our results replicate these findings in a different setting, supporting and extending these conclusions. Comparing two different time windows (Response selection and Amount spent), and thus two psychologically distinct decisional stages, we found that AIns was involved in deciding whether or not to react to injustice, whereas amygdala was involved in deciding how much to spend in order to react to injustice, both in SP and in TP. MPFC was also involved in determining the severity of the reaction, possibly reflecting the value computation, and the related affective valence, of the choice30. These results support the idea that the willingness to react to an injustice is driven by prosocial motivations such as a preference for equality17,22,26, whilst the severity of the reaction, whether it be punishing the offender or compensating the victim, reflects the affective component of the response to social norm violations27,37,38.

In summary, participants here demonstrated an equal preference for both punishment and compensation. ToM and perspective taking processes are more engaged in compensation, but neither option seems to be intrinsically rewarding per se; what is valuable is having the chance to react in the preferred way. Furthermore, we found support for two distinct psychological processes involved in the decision: unfairness perception, required to respond to an injustice, and the affective response to unfairness, associated with the severity of the reaction. Overall, the current study contributes to the developing body of research that aims to investigate the neuro-cognitive mechanisms underlying social behaviour, and, in particular, social norm compliance, a foundational aspect of a well-functioning society.

Methods

Participants

Forty right-handed participants (N(Males) = 7; Mage = 23,4 years; age range: 19–27) were recruited; two participants dropped out before undergoing (f)MRI, hence 38 completed the full task. Participants were paid €10 per hour, plus a potential small bonus depending on their choices. Exclusion criteria involved history of mental health problems and non-MRI compatible characteristics. The study was approved by the local regional ethics board (Commissie Mensgebonden Onderzoek Regio Arnhem-Nijmegen, CMO 2014/288 “Imaging Human Cognition”), in accordance with the Declaration of Helsinki regarding human experimentation. All participants provided written informed consent.

Task

A modified version of the Justice Game, presented to participants as “Chip Division Game” to avoid potential biases, was developed using Matlab Psychtoolbox3941. All players started with 200 monetary units (Mus; 1MU = 1 eurocent). In one condition (Second Party, SP), participants faced an opponent (Player A) that could steal MUs from them; in a second condition (Third Party, TP), participants observed two players, one of whom (player A) could steal MUs from the other (player B). Half of the trials were fair (when player A took 0, with both A and B possessing 200) and half were unfair (when player A took MUs from the participant/player B, so player A had more MUs than the other player), with a varying level of injustice (25, 50, 75 or 100 MUs taken from player B). At this point, participants could decide to either do nothing (“Leave”) or spend some of their endowment (200 MUs) to change the payoffs. In SP, 40 trials were administered, and the choice was always between “Leave” or “Take from A”, which meant to reduce player A’s payoff. In TP, 120 trials were administered: 40 trials presented all three options, i.e., “Leave”, “Take from A”, “Give to B” (LTG trials), and 80 trials presented only two, either “Leave” and “Take from A” (40 LT trials), or “Leave” and “Give to B” (40 LG trials). As in SP, “Take from A” meant reducing Player A’s MUs, whereas “Give to B” meant increasing player B’s MUs. The options were presented for 3 seconds before a green sign “Go!” appeared on the screen, prompting participants to enter their decision via a response box. If the choice was “Take from A” or “Give to B”, participants were required to indicate how much they wanted to spend to punish or compensate. Consistent with previous studies6,8,11, the punishment (compensation) ratio was 1:3, that is, if participants spent 10 MUs to punish (compensate) player A (B), player A (B) would lose (gain) 30 MUs. Figure 5 represents an example of an unfair trial. SP and TP conditions were presented to all participants in two separate blocks, with a counterbalanced order. Participants were told players A and B were real and had played in previous sessions, although all trials were pre-programmed. Participants were also told to treat each trial independent from the others, as each one represented a different pair of players (multiple one-shot trials). At the end, two trials were randomly selected to determine participants’ final payoff, as per the conversion rate described above.

Figure 5.

Figure 5

Justice Game task. Example of two Justice Game trials, in Second-Party and Third-Party conditions, where Player A took chips from the other player, and the participant decided to react to injustice. In the event of a “Leave” choice, the Amount screen would have been replaced by a screen asking to wait 3 seconds for the next trial. The time-windows of interest for the fMRI analysis are framed in red (Division, Response and Amount). The duration of the first two windows is fixed (3 seconds), while the Go and the Amount screen are self-paced (maximum duration of 2 and 6 seconds, respectively). The inter-stimulus interval between the Division and the Response screen and the inter-trial interval are jittered between 2 and 4 seconds.

Procedure

Upon arrival at the Donders Centre for Cognitive Neuroimaging’s fMRI facilities, participants were given both the informed consent and the MRI-compatibility checklist to fill out and sign. Then, they were taken in an experimental room for instructions and practice trials, before being placed into the MRI scanner. Depending on the presentation order of the conditions, participants were trained on either SP or TP condition outside the scanner, where they read the instructions and then answered some questions to ensure the understanding of the rules. For example, for the SP condition, participants were asked: “Player A took 50 chips from you. If you chose to spend 50 chips to take from player A, what will the final outcome (in chips) for A and yourself be?”. Participants would answer and be given the chance to discuss with the experimenter any doubts or uncertainties that may have arisen. After that, they would do 5 training trials, to become familiar with the response mode. The condition they were trained on outside the scanner was the condition with which the fMRI session would begin. The second condition was only explained inside the scanner immediately before playing it, and once the other condition was completed, to decrease demand characteristics and the risk of potentially biasing participants’ choices. A total of 160 trials was administered, for a maximum task time length of 55 minutes. A brief break was provided after 20 minutes. A structural image was acquired at the end. Overall, participants spent approximately one hour and 15 minutes inside the MRI scanner.

fMRI analysis

fMRI data acquisition and pre-processing

A 3T Siemens Magnetom Prisma MRI scanner collected functional and structural brain images. Functional images were collected using a multi echo T2*-weighted sequence. Thirty-five functional images per volume were collected in an ascending, interleaved manner (flip angle 90°, TR 2250 ms, TE 9.4 ms, 8.9 ms 8.4 ms and 7.9 ms, field of view: 224 mm, voxel size 3 × 3 × 3 mm and 0.5 mm slice gap), covering the entire cortex. High-resolution T1-weighted structural images (MPRAGE, 192 slices, TR 2300 ms, TE 3.03 ms, voxel size 1 × 1 × 1 mm) were also acquired.

Images were processed and analysed with SPM 1242. Images were time corrected and realigned to the first volume in order to correct for motion artefacts. The T1 image was co-registered to the mean functional image and segmented into tissue types. Parameters from segmentation were used to normalize all images to the MNI 152 brains template (12-parameter affine transformation); images were then spatially normalized into a standard space (MNI template), resliced to 3 mm isotropic voxels and smoothed with a 6 mm FWHM Gaussian kernel.

fMRI GLMs specifications

Two general linear models were specified at the first level, including run (one SP block, two TP blocks) as a fixed effect, 6 realignment parameters as regressors of no interests, and time derivatives for each parameter. Parameters were convolved with a double-gamma canonical hemodynamic response function (HRF). All contrasts of interests were estimated at the first-level (subject level), and then fed into a second-level analysis using a mixed-effect general linear model.

The Options model focused on the effects of the availability of different options to react to injustice, as well as on the correlates of unfairness perception. Blood Oxygenated Level Dependent (BOLD) signal was estimated for each participant, for each run and for each voxel as follows:

BOLD=β0+β1DivSPfair+β2DivTPfair+β3DivSPunfair++β4DivSPunfairInjusticeparam+β5DivTPunfair+β6DivTPunfairInjusticeparam+β7RespSP+β8RespTP_LT+β9RespTP_LG+β10RespTP_LTG+β11Go+β12Wait/Am

where Divfair and Divunfair (SP or TP) indicate a fair or unfair division screen; injusticeparam a parametric modulator of injustice (chips Player A took from the other player); Resp a response screen, with TP_LT and TP_LG referring to trials in which participants only had two options, i.e., leave or take from A (TP_LT), and leave or give to B (TP_LG), and TP_LTG referring to trials in which participants had all the three options available; Go the Go screen, and Wait/Am the screen following a decision (i.e., wait for the next trial after a Leave decision, or select an amount after a Take or Give decision). All 38 participants were included in this analysis.

The contrasts of interests estimated for the Options model were: DivUnfair-DivFair (main effect of Injustice, grouping SP and TP) to determine unfairness-related areas; DivSP-DivTP (main effect of Target, grouping unfair and fair divisions) to determine target-related areas; DivSP(Unfair-Fair)-DivTP(Unfair-fair) (interaction) to determine whether unfairness-related areas were more involved in SP compared to TP; RespSP-RespTP_LT, to determine areas that are more involved in SP responses compared to TP – note that for TP, the analysis was restricted to LT trials (only-punishment option) in order to keep it comparable to SP; RespTP_LT-RespTP_LG, and the reverse, to determine areas that are more involved in deciding whether to punish or leave compared to deciding whether to compensate or leave.

In order to compare individual preferences for punishment and compensation, participants were divided into Compensators and Punishers; those who chose compensation more often than punishment in TP_LTG trials (i.e., when all options were available), would be classified as Compensators (N = 19); vice versa, those who preferred punishment to compensation would be classified as Punishers (N = 14).

A full-factorial second-level design was run on the Response time-window, considering Preference (Compensators and Punishers) as a between-subject factor and Option (TP_LTG, TP_LT, TP_LG) as a within-subject factor. Contrasts of interests considered were (Punishers – Compensators) and reverse, (Compensators_LG + Punishers_LT) − rest, and (Compensators_LT + Punishers_LG) − rest. These contrasts tested the distinct neural processes involved when participants were given the opportunity to choose their preferred option, and when they could only choose their non-preferred option in order to react to injustice.

The Choice model focused on participants’ choices, i.e., punish or compensate, and the amount spent to do so. BOLD signal was estimated for each participant, for each run and for each voxel as follows:

BOLD=β0+β1DivSP_fair+β2DivSP_unfair+β3DivTP+β4DivTPfair+β5DivTPunfair+β6RespSPleave+β7RespTPleave+β8RespSPpun+β9RespTPpun+β10RespTPcomp+β11Go+β12Wait+β13AmSPpun+β14AmSPpunSPpunparam+β15AmTPpun+β16AmTPpunTPpunparam+β17AmTPcomp+β18AmTPcompTPcompparam

Where RespSPleave and RespSPpun indicate the choice to leave and to punish in SP; RespTPleave, RespTPpun and RespTPcomp the decisions in TP; Wait the screen after a leave decision; AmSP, AmTPpun and AmTPcomp the amount screen; SPpunparam, TPpunparam and TPcompparam the parametric modulators indicating the amount spent for SP punishment, TP punishment and TP compensation respectively. Only participants who had at least 3 trials for each type of response were included in this analysis8, and thus the sample was reduced to N = 20.

The contrasts estimated for the Choice model were: RespTPpun-RespTPcomp, and the reverse, to determine the simple effect of punishment vs compensation and viceversa; RespSPpun-RespTPpun, and the reverse, to determine the areas involved in SP vs TP punishment, and viceversa. The contrasts [Resp_TP(pun + comp) + Resp_SP] − [AmTP(pun + comp)param + AmSPparam], and reverse, comparing the Response screen and the Amount screen, parametrically modulated by the actual amount spent, were run in order to analyse the areas involved in response selection and areas involved in the severity of the reaction. TP and SP were collapsed because the focus here was on the reaction itself rather than the target.

For both models, all parameters were modelled as 3-sec boxcars, except the Go and the Wait and Amount screens, which were modelled as stick functions.

The datasets generated are available from the corresponding author on any reasonable request.

Acknowledgements

This work was supported by a Radboud Excellence Initiative Fellowship awarded to CC, and by a European Research Council grant (ERC StG 313454) to A.G.S. The authors would like to thank Dr Mirre Stallen for her helpful feedback on the latest version of the manuscript.

Author Contributions

C.C. and A.S. designed the study, I.H. and C.C. programmed the task, collect and analysed the data, C.C. and A.S. wrote the manuscript, all authors reviewed the manuscript

Competing Interests

The authors declare no competing interests.

Footnotes

Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Güth W, Schmittberger R, Schwarze B. An experimental analysis of ultimatum bargaining. J. Econ. Behav. Organ. 1982;3:367–388. doi: 10.1016/0167-2681(82)90011-7. [DOI] [Google Scholar]
  • 2.Civai C, Corradi-Dell’Acqua C, Gamer M, Rumiati RI. Are irrational reactions to unfairness truly emotionally-driven? Dissociated behavioural and emotional responses in the Ultimatum Game task. Cognition. 2010;114:89–95. doi: 10.1016/j.cognition.2009.09.001. [DOI] [PubMed] [Google Scholar]
  • 3.Fehr E, Gächter S. Altruistic punishment in humans. Nature. 2002;415:137. doi: 10.1038/415137a. [DOI] [PubMed] [Google Scholar]
  • 4.De Quervain DJ, Fischbacher U, Treyer V, Schellhammer M. The neural basis of altruistic punishment. Science. 2004;305:1254. doi: 10.1126/science.1100735. [DOI] [PubMed] [Google Scholar]
  • 5.Fehr E, Fischbacher U. Third-party punishment and social norms. Evol. Hum. Behav. 2004;25:63–87. doi: 10.1016/S1090-5138(04)00005-4. [DOI] [Google Scholar]
  • 6.Leliveld MC, van Dijk E, van Beest I. Punishing and compensating others at your own expense: The role of empathic concern on reactions to distributive injustice. Eur. J. Soc. Psychol. 2012;42:135–140. doi: 10.1002/ejsp.872. [DOI] [Google Scholar]
  • 7.Chavez AK, Bicchieri C. Third-party sanctioning and compensation behavior: Findings from the ultimatum game. J. Econ. Psychol. 2013;39:268–277. doi: 10.1016/j.joep.2013.09.004. [DOI] [Google Scholar]
  • 8.Hu Y, Strang S, Weber B. Helping or punishing strangers: neural correlates of altruistic decisions as third-party and of its relation to empathic concern. Front. Behav. Neurosci. 2015;9:24. doi: 10.3389/fnbeh.2015.00024. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Hu Y, et al. The effect of oxytocin on third-party altruistic decisions in unfair situations: An fMRI study. Sci. Rep. 2016;6:20236. doi: 10.1038/srep20236. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.David B, Hu Y, Krüger F, Weber B. Other-regarding attention focus modulates third-party altruistic choice: an fMRI study. Sci. Rep. 2017;7:43024. doi: 10.1038/srep43024. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Stallen, M. et al. Neurobiological Mechanisms of Responding to Injustice. J. Neurosci., 1242–17 (2018). [DOI] [PMC free article] [PubMed]
  • 12.Patil, I., Dhaliwal, N. & Cushman, F. A. Reputational and cooperative benefits of third-party compensation. PsyArXiv, 23, 10.31234/osf.io/c3bsj (2018).
  • 13.Bartra O, McGuire JT, Kable JW. The valuation system: A coordinate-based meta-analysis of BOLD fMRI experiments examining neural correlates of subjective value. NeuroImage. 2013;76:412–427. doi: 10.1016/j.neuroimage.2013.02.063. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Strobel A, et al. Beyond revenge: Neural and genetic bases of altruistic punishment. NeuroImage. 2011;54:671–680. doi: 10.1016/j.neuroimage.2010.07.051. [DOI] [PubMed] [Google Scholar]
  • 15.Schurz M, Radua J, Aichhorn M, Richlan F, Perner J. Fractionating theory of mind: A meta-analysis of functional brain imaging studies. Neurosci. Biobehav. Rev. 2014;42:9–34. doi: 10.1016/j.neubiorev.2014.01.009. [DOI] [PubMed] [Google Scholar]
  • 16.Krueger F, Hoffman M. The Emerging Neuroscience of Third-Party Punishment. Trends Neurosci. 2016;39:499–501. doi: 10.1016/j.tins.2016.06.004. [DOI] [PubMed] [Google Scholar]
  • 17.Sanfey AG, Rilling JK, Aronson JA, Nystrom LE, Cohen JD. The neural basis of economic decision-making in the ultimatum game. Science. 2003;300:1755–1758. doi: 10.1126/science.1082976. [DOI] [PubMed] [Google Scholar]
  • 18.Buckholtz JW, Marois R. The roots of modern justice: cognitive and neural foundations of social norms and their enforcement. Nat. Neurosci. 2012;15:655–661. doi: 10.1038/nn.3087. [DOI] [PubMed] [Google Scholar]
  • 19.Buckholtz JW, et al. The neural correlates of third-party punishment. Neuron. 2008;60:930–940. doi: 10.1016/j.neuron.2008.10.016. [DOI] [PubMed] [Google Scholar]
  • 20.Glass L, Moody L, Grafman J, Krueger F. Neural signatures of third-party punishment: evidence from penetrating traumatic brain injury. Soc. Cog. Affect. Neur. 2015;13:253–62. doi: 10.1093/scan/nsv105. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Krueger F, et al. Oxytocin selectively increases perceptions of harm for victims but not the desire to punish offenders of criminal offenses. Soc. Cog. Affect. Neur. 2012;2:494–8. doi: 10.1093/scan/nss026. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Civai C, Crescentini C, Rustichini A, Rumiati RI. Equality versus self-interest in the brain: Differential roles of anterior insula and medial prefrontal cortex. NeuroImage. 2012;62(1):102–112. doi: 10.1016/j.neuroimage.2012.04.037. [DOI] [PubMed] [Google Scholar]
  • 23.Corradi-Dell’Acqua C, Civai C, Rumiati RI, Fink GR. Disentangling self- and fairness-related neural mechanisms involved in the ultimatum game: an fMRI study. Soc. Cog. Affect. Neur. 2013;8:424–431. doi: 10.1093/scan/nss014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Zhong S, Chark R, Hsu M, Chew SH. Computational substrates of social norm enforcement by unaffected third parties. NeuroImage. 2016;129:95–104. doi: 10.1016/j.neuroimage.2016.01.040. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Buckholtz JW, et al. From blame to punishment: disrupting prefrontal cortex activity reveals norm enforcement mechanisms. Neuron. 2015;87:1369–1380. doi: 10.1016/j.neuron.2015.08.023. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Fehr E, Schmidt KM. A Theory of Fairness, Competition, and Cooperation. Q. J. Econ. 1999;114:817–868. doi: 10.1162/003355399556151. [DOI] [Google Scholar]
  • 27.Pillutla MM, Murnighan JK. Unfairness, Anger, and Spite: Emotional Rejections of Ultimatum Offers. Organ. Behav. Hum. Dec. 1996;68:208–224. doi: 10.1006/obhd.1996.0100. [DOI] [Google Scholar]
  • 28.Treadway MT, et al. Corticolimbic gating of emotion-driven punishment. Nat. Neurosci. 2014;17(9):1270–1275. doi: 10.1038/nn.3781. [DOI] [PubMed] [Google Scholar]
  • 29.IBM Corp. Released. IBM SPSS Statistics for Windows, Version 21.0. Armonk, NY: IBM Corp (2012)
  • 30.Delgado MR, et al. Viewpoints: Dialogues on the functional role of the ventromedial prefrontal cortex. Nat. Neurosci. 2016;19:1545–1552. doi: 10.1038/nn.4438. [DOI] [PubMed] [Google Scholar]
  • 31.Bonetti S. Experimental economics and deception. J Econ Psychol. 1998;19:377–95. doi: 10.1016/S0167-4870(98)00012-9. [DOI] [Google Scholar]
  • 32.Hare TA, Camerer CF, Rangel A. Self-control in decision-making involves modulation of the vmPFC valuation system. Science. 2009;324:646–648. doi: 10.1126/science.1168450. [DOI] [PubMed] [Google Scholar]
  • 33.Gabay AS, Radua J, Kempton MJ, Mehta MA. The Ultimatum Game and the brain: A meta-analysis of neuroimaging studies. Neurosci. Biobehav. Rev. 2014;47:549–558. doi: 10.1016/j.neubiorev.2014.10.014. [DOI] [PubMed] [Google Scholar]
  • 34.Feng C, Luo YJ, Krueger F. Neural signatures of fairness‐related normative decision making in the ultimatum game: A coordinate‐based meta‐analysis. Hum. Brain. Mapp. 2015;36:591–602. doi: 10.1002/hbm.22649. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Bhanji JP, Delgado MR. The social brain and reward: social information processing in the human striatum. Wiley Interdiscip. Rev. Cogn. Sci. 2014;5:61–73. doi: 10.1002/wcs.1266. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Civai C, Miniussi C, Rumiati RI. Medial prefrontal cortex reacts to unfairness if this damages the self: a tDCS study. Soc. Cog. Affect. Neur. 2014;10:1054–1060. doi: 10.1093/scan/nsu154. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Rabin M. Incorporating Fairness into Game Theory and Economics. Am. Econ. Rev. 1993;83:1281–1302. [Google Scholar]
  • 38.Pedersen EJ, Kurzban R, McCullough ME. Do humans really punish altruistically? A closer look. Proc. R. Soc. Lond. B. 2013;280:20122723. doi: 10.1098/rspb.2012.2723. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Brainard DH, Vision S. The psychophysics toolbox. Spatial vision. 1997;10:433–436. doi: 10.1163/156856897X00357. [DOI] [PubMed] [Google Scholar]
  • 40.Pelli DG. The VideoToolbox software for visual psychophysics. Spatial Vision. 1997;10:437–442. doi: 10.1163/156856897X00366. [DOI] [PubMed] [Google Scholar]
  • 41.Kleiner M, et al. What’s new in Psychtoolbox-3. Perception. 2007;36:1. [Google Scholar]
  • 42.The Wellcome Trust Center for Neuroimaging, London, UK, http://www.fil.ion.ucl.ac.uk/spm/.

Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES