Skip to main content
Springer logoLink to Springer
. 2012 Jan 24;217(4):783–796. doi: 10.1007/s00429-012-0380-y

Parsing the neural correlates of moral cognition: ALE meta-analysis on morality, theory of mind, and empathy

Danilo Bzdok 1,2,3, Leonhard Schilbach 4,5, Kai Vogeley 5,6, Karla Schneider 1,3, Angela R Laird 7, Robert Langner 1,2,3, Simon B Eickhoff 1,2,3,8,9,
PMCID: PMC3445793  PMID: 22270812

Abstract

Morally judicious behavior forms the fabric of human sociality. Here, we sought to investigate neural activity associated with different facets of moral thought. Previous research suggests that the cognitive and emotional sources of moral decisions might be closely related to theory of mind, an abstract-cognitive skill, and empathy, a rapid-emotional skill. That is, moral decisions are thought to crucially refer to other persons’ representation of intentions and behavioral outcomes as well as (vicariously experienced) emotional states. We thus hypothesized that moral decisions might be implemented in brain areas engaged in ‘theory of mind’ and empathy. This assumption was tested by conducting a large-scale activation likelihood estimation (ALE) meta-analysis of neuroimaging studies, which assessed 2,607 peak coordinates from 247 experiments in 1,790 participants. The brain areas that were consistently involved in moral decisions showed more convergence with the ALE analysis targeting theory of mind versus empathy. More specifically, the neurotopographical overlap between morality and empathy disfavors a role of affective sharing during moral decisions. Ultimately, our results provide evidence that the neural network underlying moral decisions is probably domain-global and might be dissociable into cognitive and affective sub-systems.

Electronic supplementary material

The online version of this article (doi:10.1007/s00429-012-0380-y) contains supplementary material, which is available to authorized users.

Keywords: Moral cognition, Theory of mind (ToM), Empathy, Social cognition, Meta-analysis, ALE

Introduction

Moral behavior is a building block of human societies and has classically been thought to be based on rational consideration. Aristotle (fourth century BC/1985), for example, argued that being a “good person” requires reasoning about virtues. Kant’s (1785/1993) famous categorical imperative was similarly rational, demanding that one should act according to principles that could also become a general law. More recently, Kohlberg et al. (1983) and Kohlberg (1969) advanced a six-stage developmental model acknowledging children’s increasing ability for abstraction and role-taking capacities in moral decisions. In contrast to rational models, the role of emotion in facilitating moral behavior has been less often emphasized (Haidt 2001). Hume (1777/2006) provides an early notable exception, as he believed in a key role of intuition for recognizing morally good and bad decisions, not requiring willful abstract reasoning. Charles Darwin (1874/1997) further argued that moral decisions are mainly influenced by emotional drives, which are rooted in socio-emotional instincts already present in non-human primates.

Notions of rationality and emotionality also serve as explanations in the contemporary neuroscientific literature on the psychological processes underlying moral decisions (henceforth: moral cognition). Results from functional magnetic resonance imaging (fMRI) studies by Greene et al. (2004) and Greene et al. (2001) were interpreted as revealing a dissociation between fast emotional responses and subsequent cognitive modulations in moral cognition. FMRI findings by Moll and Schulkin (2009), Moll et al. (2005a) and Moll et al. (2006), however, were interpreted as revealing a dissociation between group-oriented (i.e., pro-social) and self-oriented (i.e., egoistic) affective drives in moral cognition. Taken together, either abstract-inferential or rapid-emotional processing has been emphasized by most previous accounts explaining moral behavior. Rational explanations assumed that moral behavior arises from a conscious weighing of different rules, norms and situational factors, while emotional explanations emphasized the influence of uncontrolled emotional states rapidly evoked by a given situation (cf. Krebs 2008).

Previous evidence suggests that, the rational and emotional facets of moral cognition are likely related to other social skills: theory of mind (ToM) and empathy. ToM refers to the ability to contemplate other’s thoughts, desires, and behavioral dispositions by abstractinference (Premack and Woodruff 1978; Frith and Frith 2003). Indeed, accumulating evidence indicates that moral cognition is influenced by whether or not an agent’s action is perceived as intentional versus accidental (Knobe 2005; Cushman 2008; Killen et al. 2011; Moran et al. 2011). Empathy, on the other hand, refers to automatically adopting somebody’s emotionalstate while maintaining the self–other distinction (Singer and Lamm 2009; Decety and Jackson 2004). In moral decisions, experiencing empathy was shown to alleviate harmful actions towards others (Feshbach and Feshbach 1969; Zahn-Waxler et al. 1992; Eisenberger 2000). Conversely, the deficient empathy skills in psychopathic populations are believed to contribute to morally inappropriate behavior (Hare 2003; Blair 2007; Soderstrom 2003). Taken together, moral cognition is thought to involve the representation of intentions and outcomes as well as (vicariously experienced) emotional states (Decety et al. 2012; Leslie et al. 2006; Pizarro and Bloom 2003). This assumption is further supported by the observation that ToM and empathic skills precede mature moral reflection in primate evolution (Greene and Haidt 2002) and in ontogeny (Tomasello 2001; Kohlberg et al. 1983; Frith and Frith 2003; Piaget 1932) given that natural evolution tends to modify existing biological systems rather than create new ones from scratch (Jacob 1977; Krebs 2008).

Notably, theoretical accounts as well as empirical evidence suggest that ToM and empathy are partially overlapping psychological constructs. In particular, it has been proposed that embodied representations of affect, which should be relevant for empathic processing, may be further integrated into meta-representational or inferential processing (Keysers and Gazzola 2007; Spengler et al. 2009; De Lange et al. 2008; Mitchell 2005). That is, embodied representation and meta-representation might not constitute two mutually exclusive processes. In other words, more automatic, bottom-up driven mapping and awareness of others emotional states in the context of self–other distinction (i.e., empathy) might be modulated by more controlled, top-down processes involved in attributing mind states (i.e., ToM) (Leiberg and Anders 2006; Pizarro and Bloom 2003; Singer and Lamm 2009). Importantly, ToM and empathy differ in the representational content (mental states versus affect), yet both might be similarly brought about by interaction between top-down and bottom-up processes (Singer and Lamm 2009; Lamm et al. 2007; Spengler et al. 2009; Cheng et al. 2007). In line with the theoretical arguments for a partial overlap between ToM and empathy, ventromedial prefrontal cortex lesions associated with ToM impairments were empirically shown to debilitate elaborate forms of empathic processing, while lateral inferior frontal cortex lesions, which leave ToM skills intact, lead to an impairment of basic forms of empathic processing (Shamay-Tsoory et al. 2009). In sum, converging earlier evidence thus suggests that moral cognition is subserved by partially interrelated ToM and empathy processes.

From a neurobiological perspective, we therefore hypothesized that moral cognition might be subserved by brain areas also related to ToM and empathy, that is, brain networks associated with abstract-inferential and rapid-emotional processing, respectively. Furthermore, we sought to formally investigate to what extent the neurobiological correlates of ToM versus empathy overlap with the neural network associated with moral cognition. Moreover, we tested whether a common set of brain areas might be implicated in all three of these elaborate social-cognitive skills. We addressed these questions by means of a quantitative meta-analysis on peak coordinates reported in functional neuroimaging studies on moral cognition, ToM, and empathy.

Methods

Data used for the meta-analysis

We searched the PubMed database (http://www.pubmed.org) for fMRI and PET studies that investigated the neural correlates of moral cognition, ToM and empathy. Relevant papers were found by keyword search (search strings: “moral”, “morality”, “norm”, “transgression”, “violation”, “theory of mind”, “mentalizing”, “false belief”, “perspective taking”, “empathy”, “empathic”, “fMRI”, “PET”). Further studies were identified through review articles and reference tracing from the retrieved papers. Please note that in the context of ALE, the term “experiment” usually refers to any single (contrast) analysis on imaging data yielding localization information, while the term “study” usually refers to a scientific publication reporting one or more “experiments” (Laird et al. 2011; Eickhoff and Bzdok 2012). The inclusion criteria comprised full brain coverage as well as absence of pharmacological manipulations, brain lesions or mental/neurological disorders. Additionally, studies were only considered, if they reported results of whole-brain group analyses as coordinates corresponding to a standard reference space (Talairach/Tournoux, MNI). That is, experiments assessing neural effects in a priori defined regions of interest were excluded. We included all eligible neuroimaging studies published up to and including the year 2010. The exhaustive literature search yielded in the moral cognition category a total of 67 experiments reporting 507 activation foci; in the ToM category a total of 68 experiments reporting 724 activation foci, and in the empathy category a total of 112 experiments reporting 1,376 activation foci.

Methodologically equivalent to earlier ALE meta-analyses (Spreng et al. 2009; Lamm et al. 2011; Bzdok et al. 2011; Fusar-Poli et al. 2009; Mar 2011), study selection was grounded on the objective measure whether or not the authors claimed to have isolated brain activity that relates to moral cognition, ToM, or empathy. More specifically, we only included those neuroimaging studies into the “morality” category that required participants to make appropriateness judgments on actions of one individual towards others. In these studies, participants passively viewed or explicitly evaluated mainly textual, sometimes pictorial social scenarios with moral violations/dilemmas. The target conditions were frequently contrasted with neutral or unpleasant social scenarios (see Supplementary Table 1 for detailed study descriptions). Furthermore, we only included those neuroimaging studies into the “ToM” category that required participants to adopt an intentional stance towards others, that is, predict their thoughts, intentions, and future actions. These studies mostly presented cartoons and short narratives that necessitated understanding the beliefs of the acting characters. The target conditions were usually contrasted with non-social physical stories, which did not necessitate social perspective-taking. Finally, we only included those neuroimaging studies into the “empathy” category that aimed at eliciting the conscious and isomorphic experience of somebody else’s affective state. Put differently, in these studies participants were supposed to know and “feel into” what another person was feeling. These studies employed mostly visual, sometimes textual or auditory stimuli that conveyed affect-laden social situations which participants watched passively or evaluated on various dimensions.

Please note that we disregarded studies on empathy for pain because pain, although possessing an affective dimension, is not considered a classic emotion (Izard 1971; Ekman 1982). Rather, it is a bodily sensation mediated by distinct sensory receptors (Craig 2002; Saper 2000), and watching painful scenes does not induce isomorphic vicarious experiences (cf. Singer et al. 2004; Danziger et al. 2009). In particular, looking at a happy person usually elicits a sensation of happiness in the observer, yet watching a person in pain usually does not likewise evoke the physical experience of pain. Furthermore, we disregarded studies in which participants were presented with emotion recognition tasks using static pictures of emotional facial expressions. That is because such tasks are probably too simple to reliably entail sharing others’ emotions and maintaining a self–other distinction, both widely regarded as hallmarks of empathy (Singer and Lamm 2009).

Methodological foundation of activation likelihood estimation

The reported coordinates were analyzed for topographic convergence using the revised ALE algorithm for coordinate-based meta-analysis of neuroimaging results (Eickhoff et al. 2009; Turkeltaub et al. 2002; Laird et al. 2009a). The goal of coordinate-based meta-analysis of neuroimaging data is to identify brain areas where the reported foci of activation converge across published experiments. To this end, the meta-analysis determines if the clustering is significantly higher than expected under the null distribution of a random spatial association of results from the considered experiments while acknowledging the spatial uncertainty associated with neuroimaging foci.

As the first step, reported foci were interpreted as centers for 3D Gaussian probability distributions that capture the spatial uncertainty associated with each focus. This uncertainty is mostly a function of between-template (attributable to different normalization strategies and templates across laboratories) and between-subject (due to small sample sizes) variance. In fact, the between-template and between-subject variability are acknowledged based on empirical estimates, the latter being additionally gauged by individual sample size (Eickhoff et al. 2009).

In a second step, the probabilities of all activation foci in a certain experiment were combined for each voxel, yielding a modeled activation (MA) map (Turkeltaub et al. 2011). Voxel-wise ALE scores resulted from the union across these MA maps and quantified the convergence across experiments at each particular location in the brain.

The third and last step distinguished between random and ‘true’ convergence by comparing the ensuing ALE scores against an empirical null distribution reflecting a random spatial association between the experiments’ MA maps (Eickhoff et al. 2012). The within-experiment distribution of foci, however, was regarded to be fixed (Eickhoff et al. 2009). Thus, a random-effects inference was invoked, focusing on the above-chance convergence across different experiments (Eickhoff et al. 2009; Caspers et al. 2010; Kurth et al. 2010). The resulting ALE scores were tested against the earlier calculated ‘true’ ALE scores and cut off at a cluster-level-corrected threshold of p < 0.05. For cluster-level correction, the statistical image of uncorrected voxel-wise p values was first cut off by the cluster-forming threshold. Then, the size of the supra-threshold clusters was compared against a null distribution of cluster sizes derived from simulating 1,000 datasets with the same properties (number of foci, uncertainty, etc.) as the original experiments but random location of foci. The p value associated with each cluster was then given by the chance of observing a cluster of the given size in any particular simulation.

Additional conjunction and difference analyses were conducted to explore how different meta-analyses relate to each other. Conjunction-analyses testing for convergence between different meta-analyses employed inference by the minimum statistic, i.e., computing the intersection of the thresholded Z-maps (Caspers et al. 2010). That is, any voxel determined to be significant by the conjunction analysis constitutes a location in the brain which survived inference corrected for multiple comparisons in each of the individual meta-analyses. Difference analyses calculated the difference between corresponding voxels’ ALE scores for two sets of experiments. Then, the experiments contributing to either analysis were pooled and randomly divided into two analogous sets of experiments (Eickhoff et al. 2011). Voxel-wise ALE scores for these two sets were calculated and subtracted from each other. Repeating this process 10,000 times yielded a null distribution of recorded differences in ALE scores between two sets of experiments. The ‘true’ difference in ALE scores was then tested against these differences obtained under the null distribution yielding voxel-wise p values for the difference. These resulting non-parametric p values were thresholded at p < 0.001. Unfortunately, a statistical method to correct for multiple comparisons when assessing the differences between ALE maps has not yet been established. It should be mentioned, however, that the randomization procedure employed to compute the contrast between ALE-analyses is in itself highly conservative as it estimates the probability for a true difference between the two datasets.

Functional characterization

The converging activation patterns of tasks requiring moral cognition, theory of mind, or empathy were first determined by ALE meta-analysis. The conjunction across all three individual ALE analyses then yielded a computationally derived seed region for functional characterisation through quantitative correspondence with cognitive and experimental descriptions of the BrainMap taxonomy. In fact, BrainMap metadata describe the category of mental processes isolated by the statistical contrast of each experiment stored in the database (http://www.brainmap.org; Laird et al. 2009a). More specifically, behavioral domains (BD) include the main categories cognition, action, perception, emotion, interoception, as well as their subcategories. The respective paradigm classes (PC) categorize the specific task employed (a complete list of BDs and PCs can be found at http://www.brainmap.org/scribe/). We analyzed the behavioral domain and paradigm class metadata of BrainMap experiments associated with seed voxels to determine the frequency of domain ‘hits’ relative to its likelihood across the entire database. Using a binomial test, Bonferroni-corrected for multiple comparisons, the functional roles of the convergent network as a whole were identified by determining those BDs and PCs that showed a significant over-representation in experiments activating within the seed regions, relative to the entire BrainMap database (Laird et al. 2009b; Eickhoff et al. 2011).

Results

Individual meta-analyses of moral cognition, theory of mind, and empathy

All areas resulting from the ALE meta-analyses were anatomically labeled by reference to probabilistic cytoarchitectonic maps of the human brain using the SPM Anatomy Toolbox (see Tables 1, 2; Eickhoff et al. 2007; Eickhoff et al. 2005). Meta-analysis of neuroimaging studies related to moral cognition yielded convergent activation in the bilateral ventromedial/frontopolar/dorsomedial prefrontal cortex (vmPFC/FP/dmPFC), precuneus (Prec), temporo-parietal junction (TPJ), posterior cingulate cortex (PCC), as well as the right temporal pole (TP), right middle temporal gyrus (MTG), and left amygdala (AM) (Fig. 1; Table 1). Meta-analysis of studies related to ToM revealed convergence in the bilateral vmPFC/FP/dmPFC, Prec, TPJ, TP, MTG, posterior superior temporal sulcus (pSTS), and inferior frontal gyrus (IFG), as well as the right MT/V5. Finally, meta-analysis of studies related to empathy yielded convergence in the bilateral dmPFC, supplementary motor area (SMA), rostral anterior cingulate cortex (rACC), anterior mid-cingulate cortex (aMCC; cf. Vogt 2005), PCC, anterior insula (AI), IFG, midbrain, and TPJ, as well as the anterior thalamus on the left, further, the AM, MTG, pSTS, posterior thalamus, hippocampus, and pallidum on the right.

Table 1.

Peaks of activations for the brain areas consistently engaged in fMRI studies on moral cognition, theory of mind, and empathy as revealed by ALE meta-analysis

Macroanatomical location MNI coordinates
x y z
ALE meta-analysis of moral cognition
Ventromedial prefrontal cortex 4 58 −8
Ventromedial prefrontal cortex −10 42 −18
Frontopolar cortex 0 62 10
Frontopolar cortex −6 52 18
Dorsomedial prefrontal cortex 0 54 36
Precuneus 0 −56 34
Right temporo-parietal junction (PGa, PGp) 62 −54 16
Left temporo-parietal junction (PGa, PGp) −48 −58 22
Right temporal pole 54 8 −28
Right middle temporal gyrus 54 −8 −16
Left amygdala −22 −2 −24
Posterior cingulate cortex −4 −26 34
ALE meta-analysis of theory of mind
Ventromedial prefrontal cortex 0 52 −12
Frontopolar cortex 2 58 12
Dorsomedial prefrontal cortex −8 56 30
Dorsomedial prefrontal cortex 4 58 25
Precuneus 2 −56 30
Right temporo-parietal junction (PGa, PGp) 56 −50 18
Left temporo-parietal junction (PGa, PGp) −48 −56 24
Right temporal pole 54 −2 −20
Left temporal pole −54 −2 −24
Right middle temporal gyrus 52 −18 −12
Left middle temporal gyrus −54 −28 −4
Left middle temporal gyrus −58 −12 −12
Right posterior superior temporal sulcus 50 −34 0
Left posterior superior temporal sulcus −58 −44 4
Right inferior frontal gyrus (Area 45) 54 28 6
Left inferior frontal gyrus −48 30 −12
Right MT/V5 48 −72 8
ALE meta-analysis of empathy
Dorsomedial prefrontal cortex 2 56 18
Dorsomedial prefrontal cortex −8 54 34
Right anterior insula 36 22 −8
Left anterior insula −30 20 4
Right inferior frontal gyrus 50 12 −8
Right inferior frontal gyrus (Area 44) 54 16 20
Right inferior frontal gyrus (Area 45) 50 30 4
Left inferior frontal gyrus −44 24 −6
Supplementary motor area (Area 6) −4 18 50
Anterior mid-cingulate cortex −2 28 20
Rostral anterior cingulate cortex −4 42 18
Posterior cingulate cortex −2 −32 28
Right temporo-parietal junction (PGp) 52 −58 22
Left temporo-parietal junction (PGa) −56 −58 22
Right amygdala 22 −2 −16
Right middle temporal gyrus 54 −8 −16
Right posterior superior temporal sulcus 52 −36 2
Left anterior thalamus −12 −4 12
Right posterior thalamus 6 −32 2
Right hippocampus (SUB) 26 −26 −12
Midbrain 2 −20 −12
Right pallidum 14 4 0

All peaks are assigned to the most probable brain areas as revealed by the SPM Anatomy Toolbox (Eickhoff et al. 2007; Amunts et al. 2005; Geyer 2004; Caspers et al. 2006; Amunts et al. 1999; Eickhoff et al. 2005)

Table 2.

Conjunction analyses that test for topographical convergence between the individual ALE meta-analyses of moral cognition, theory of mind, and empathy

Macroanatomical location MNI coordinates
x y z
Morality ∩ (theory of mindempathy)
Ventromedial prefrontal cortex 2 54 −12
Frontopolar cortex 4 60 10
Dorsomedial prefrontal cortex 0 54 30
Right temporo-parietal junction (PGa) 62 −54 14
Left temporo-parietal junction (PGa, PGp) −50 −58 22
Right middle temporal gyrus 54 −16 −16
Right temporal pole 54 2 −24
Morality ∩ (empathytheory of mind)
Dorsomedial prefrontal cortex −4 50 20
Morality ∩ theory of mind ∩ empathy
Dorsomedial prefrontal cortex −5 54 34
Right temporo-parietal junction (PGp) 52 −58 20
Left temporo-parietal junction (PGa) −54 −58 22
Right middle temporal gyrus 54 −8 −16

All peaks are assigned to the most probable brain areas as revealed by the SPM Anatomy Toolbox (Eickhoff et al. 2007; Caspers et al. 2006; Eickhoff et al. 2005)

Fig. 1.

Fig. 1

ALE meta-analysis of neuroimaging studies on moral cognition, theory of mind, and empathy. Significant meta-analysis results displayed on frontal, right, and left surface view as well as sagittal, coronal, and axial sections of the MNI single-subject template. Coordinates in MNI space. All results were significant at a cluster-forming threshold of p < 0.05 and an extent threshold of k = 10 voxels (to exclude presumably incidental results)

Conjunction analyses

We conducted three conjunction analyses to examine where regions consistently involved in moral cognition converged with regions consistently involved in ToM, empathy, or both (Fig. 2; Table 2). In particular, we computed the conjunction across the individual meta-analysis of moral cognition and the difference between the analyses of ToM and empathy. Employing the difference excluded brain activity shared by both ToM and empathy. The conjunction across the neural network linked to moral cognition and the activation more robustly linked to ToM versus empathy revealed an overlap in the bilateral vmPFC/FP/dmPFC and TPJ, as well as the right MTG and TP. Conversely, the conjunction across the neural network linked to moral cognition and the activation more robustly linked to empathy versus ToM revealed an overlap in the dmPFC. Besides results from this conjunction analysis, it is noteworthy that the left AM was found in the analysis of moral cognition-related brain activity, while the right AM was found in the analysis of empathy-related brain activity. Moreover, the PCC showed significant convergence in both these individual meta-analyses at adjacent, yet non-overlapping locations. Specifically, convergence in the PCC was located slightly more rostro-dorsally in the ALE on moral cognition compared to the ALE on empathy (Fig. 1; Table 1). Finally, the conjunction analysis across the results of all three individual meta-analyses on moral cognition, ToM, and empathy yielded convergence in the dmPFC, right MTG, and bilateral TPJ. Additionally, we provide a list of those papers included in the present meta-analysis, which gave rise to the four converging clusters (Supplementary Table 2).

Fig. 2.

Fig. 2

Conjunction analyses for topographical convergence across brain activity related to moral cognition and theory of mind (ToM) or empathy. Left panel overlapping activation patterns between the meta-analysis on moral cognition and the difference analysis between ToM and empathy (cluster-forming threshold: p < 0.05). Right bottom panel overlapping activation patterns between the meta-analysis on moral cognition and the difference analysis between empathy and ToM (cluster-forming threshold: p < 0.05). Right top panel sagittal and coronal slices of juxtaposed results from the meta-analyses on moral cognition (green) and empathy (red) to highlight similar convergence in the posterior cingulate cortex and amygdala (extent threshold: k = 10 voxels to exclude presumably incidental results). Coordinates in MNI space

Functional characterization

Using the metadata of the BrainMap database for functional characterization as described above, we found that the common network observed across all three task domains was selectively associated with tasks tapping ToM, semantic processing, imagination, and social cognition, as well as episodic and explicit memory (Fig. 3). The network thus appears to be exclusively involved in higher-level cognitive processing.

Fig. 3.

Fig. 3

Functional characterization of the core-network implicated in moral cognition, theory of mind (ToM), and empathy. Left neural network consistently activated across individual meta-analyses on moral cognition, ToM, and empathy (extent threshold: k = 10 voxels to exclude presumably incidental results). Images were rendered using Mango (multi-image analysis GUI, Research Imaging Institute, San Antonio, Texas, USA; http://ric.uthscsa.edu/mango/). Right functional characterization of the convergent network across all three tasks by BrainMap metadata. The purple bars denote the number of foci for that particular metadata class within the seed network. The grey bars represent the number of foci that would be expected to hit the particular seed network if all foci with the respective class were randomly distributed throughout the cerebral cortex. That is, the grey bars denote the by-chance frequency of that particular label given the size of the cluster. All shown taxonomic classes reached significance according to a binomial test (p < 0.05). Asterisks denote classes that survived the Bonferroni correction for multiple comparisons

Discussion

Classic accounts of the psychology of moral cognition can be broadly divided into views that emphasize the involvement of either ‘rational’ or ‘emotional’ processes. In this study, we revisited this distinction by determining the consistent overlap between brain activation patterns reported in the neuroscientific literature on moral cognition with those of ToM (assumed to be more rational) and empathy (assumed to be more emotional) using coordinate-based ALE meta-analysis. By doing so, we demonstrated that moral cognition, indeed, recruits brain areas that are also involved in abstract-inferential (ToM) and rapid-emotional (empathy) social cognition. Furthermore, TPJ, mPFC, and MTG emerged as potential nodes of a network common to moral cognition, ToM, and empathy.

Individual analyses on the neural correlates of moral cognition, theory of mind, and empathy

To our knowledge, we here conducted the first quantitative meta-analytic assessment of the neural network engaged in moral cognition, that is, reflection of the social appropriateness of people’s actions. The obtained pattern of converged brain activation is in very good agreement with qualitative reviews of fMRI studies on moral cognition (Moll et al. 2005b; Greene and Haidt 2002). Concurrently, dysfunction in prefrontal, temporal and amygdalar regions is discussed as linked to psychopathy and anti-social behavior based on neuropsychological, lesion, and fMRI studies (Blair 2007; Anderson et al. 1999; Mendez et al. 2005).

The results of the meta-analysis of ToM were consistent with earlier quantitative analyses (Spreng et al. 2009; Mar 2011) of neuroimaging studies, in which participants attributed mental states to others to predict or explain their behavior. Put differently, the obtained network is likely to be implicated in the recognition and processing of others’ mental states (Amodio and Frith 2006; Gallagher and Frith 2003; Schilbach et al. 2010).

The meta-analysis on empathy across various affective modalities excluding pain was based on neuroimaging studies in which participants understood and vicariously shared the emotional experience of others. This analysis revealed the aMCC extending into the SMA and the AI extending into the IFG as the most prominent points of convergence. This is in line with a recent image-based meta-analysis on empathy for pain (Lamm et al. 2011). The present analysis on non-pain empathy, however, additionally revealed activation in the AM, rACC, and PCC. Consequently, empathy for pain- and non-pain-related affect appears to be implemented by an overlapping network that might recruit supplementary areas/networks depending on the specific affective modality with which participants are to empathize. In other words, the observed network is likely to be involved in vicariously experiencing others’ affective states (Singer et al. 2004; Wicker et al. 2003).

It is noteworthy that all brain areas revealed by the meta-analysis on moral cognition also converged significantly in the analyses on either ToM or empathy. It is thus tempting to speculate that moral cognition might rely on remodeling mental states and processing affective states of other people. Conversely, some of the brain areas, which show significant convergence in the ALE on ToM and empathy, were not significant in the ALE on moral cognition. This suggests that moral cognition might be realized by specific subsets, rather than the entirety of the neural correlates of ToM and empathy.

Converging neural correlates across moral cognition and theory of mind

Brain activity during moral cognition and ToM overlapped in the bilateral vmPFC/FP/dmPFC and TPJ, as well as the right TP and MTG. This extensive convergence indicates that moral cognition and ToM engage a highly similar neural network, which, in turn, entices to speculate about a close relationship between these two psychological processes. More specifically, increased activity along the dorso-ventral axis of the medial prefrontal cortex is heterogeneously discussed to reflect cognitive versus affective processes (Shamay-Tsoory et al. 2006), controlled/explicit versus automatic/implicit social cognition (Lieberman 2007; Forbes and Grafman 2010), goal versus outcome pathways (Krueger et al. 2009), dissimilarity versus similarity to self (Mitchell et al. 2006) as well as other-focus versus self-focus (Van Overwalle 2009). Apart from that, the spatially largest convergence across moral cognition- and ToM-related brain activity in the bilateral posterior temporal lobe/angular gyrus might be surprising given the divergent interpretation in the literature. That is, activation in this cortical region is conventionally interpreted as “posterior superior temporal sulcus” in the morality literature (Moll et al. 2005b; Greene and Haidt 2002) and as “temporo-parietal junction” in the ToM literature (Decety and Lamm 2007; Van Overwalle 2009; Saxe and Kanwisher 2003). This convergent activation in the ALE on moral cognition is, however, more accurately located to the TPJ, rather than to the pSTS (cf. Raine and Yang 2006; Binder et al. 2009). The potentially inconsistent neuroanatomical labeling might have disadvantageously affected discussion of this brain area in previous neuroimaging studies on moral cognition.

The engagement of ToM-associated areas during moral cognition has also been addressed in several recent fMRI studies by Young, Saxe, and colleagues. These authors proposed that during moral cognition, the dmPFC might process belief valence, while the TPJ and Prec might encode and integrate beliefs with other relevant features (Young and Saxe 2008). Especially, brain activity in the right TPJ was advocated to reflect belief processing during moral cognition. This argument was based on interaction effects with moral reasoning (Young et al. 2007), correlation with the participants’ self-reported tendency for acknowledging belief information (Young and Saxe 2009), and a significantly reduced impact of intentions after transient TPJ disruption (Young et al. 2010b). In line with our results, these fMRI studies suggest that moral cognition might involve reconstructing attributes and intentions that we apply to others and vice versa.

However, earlier neuroimaging evidence concerning the likely implication of the ToM network in moral cognition has two weak points. First, the interpretations of relevant fMRI studies were largely driven by reverse inference, i.e., identifying psychological processes from mere topography of brain activity (Poldrack 2006). That is, rather than investigating moral cognition tasks alone, the neural networks of moral cognition and ToM should have been compared directly by independent conditions involving either task. Second, the relevant fMRI studies crucially hinge on the repeatedly criticized functional localizer technique, which rests on strong a priori hypotheses (Friston et al. 2006; Mitchell 2008). The present ALE meta-analysis overcomes these two limitations by a largely hypothesis-free assessment of two independent pools of numerous whole-brain neuroimaging studies, providing strong evidence for the high convergence across the neural networks associated with moral cognition and ToM.

Converging neural correlates across moral cognition and empathy

Brain activity related to both moral cognition and non-pain empathy converged significantly in an area of the dmPFC, which was not revealed by the ALE on ToM. Converging activation of the dmPFC may suggest an implication of this highly associative cortical area in more complex social-emotional processing. In line with this interpretation, a recent fMRI study identified a similar brain location as highly selective for processing guilt (Wagner et al. 2012), an emotion closely related to moral and social transgression (Tangney et al. 2007). Moreover, the dmPFC has consistently been related to the (possibly interwoven) reflection of own and simulation of others’ mind states (Lamm et al. 2007; Jenkins and Mitchell 2010; Mitchell et al. 2006; D’Argembeau et al. 2007). Nevertheless, we feel that it might be currently unwarranted to confer precise functions to circumscribed parts of the dmPFC (cf. above), given the danger of reverse inference when deducing mental functions or states from regional activation patterns (Poldrack 2006). We therefore cautiously conclude that the observed convergence in the dmPFC probably reflects an unidentified, yet to be characterized, higher-level neural process that is related to affective and social processing.

It is interesting to note that the AM was significantly involved in the individual ALEs on morality and non-pain empathy, although in contralateral hemispheres. This brain region is thought to automatically extract biological significance from the environment (Ball et al. 2007; Sander et al. 2003; Bzdok et al. 2011; Müller et al. 2011). In particular, AM activity typically increases in the left hemisphere in elaborate social-cognitive processes and in the right hemisphere in automatic, basic emotional processes (Phelps et al. 2001; Bzdok et al. 2012; Markowitsch 1998; Gläscher and Adolphs 2003). This lateralization pattern thus seems well in line with the consistent engagement of the left AM in moral cognition and right AM in non-pain empathy. Moreover, activity in the PCC was found in adjacent, yet non-overlapping, locations during moral cognition and non-pain empathy. The PCC is thought important for the modality-independent retrieval of autobiographical memories and their integration with current emotional states (Fink et al. 1996; Maddock et al. 2001; Maddock 1999; Schilbach et al. 2008a). It might therefore be speculated that moral judgments and empathic processing could both rely on the integration on past experiences. Given that both PCC clusters do not overlap, however, it remains to be investigated whether (a) moral cognition and non-pain empathy engage distinct regions in the PCC, (b) the observed topographic pattern is purely incidental given the limited spatial resolution of meta-analyses, or (c) the differences in PPC activation reflect differences in stimulus-material and hence autobiographic associations.

Furthermore, it is important to note that the AI/IFG and aMCC/SMA were revealed as the most significant points of convergence in the ALE on non-pain empathy but did not show any overlap with the ALE on morality. In particular, the AI/IFG and aMCC/SMA form a network that is widely believed to represent one’s own and others’ emotional states regardless of the actual affective or sensory modality (Lamm et al. 2011; Wicker et al. 2003; Singer et al. 2004; Lamm and Singer 2010; Fan et al. 2011). Moreover, this network, especially the anterior insula, is not only implicated in meta-representation of emotional states but also in interoceptive awareness (Craig 2002, 2009; Kurth et al. 2010). Interoception-related meta-representation of emotion has thus been suggested to underlie the concomitant involvement of the aMCC/SMA and AI/IFG in neuroimaging studies on empathy as reflecting affective sharing (Singer and Lamm 2009; Fan et al. 2011).

In summary, we draw three conclusions from these observations. First, affective sharing, one core aspect of empathy, is unlikely to be involved in moral judgments, given the lack of consistent involvement of the AI/IFG or aMCC/SMA in paradigms probing the latter. Second, general affective processes might play a role in both moral judgments and empathy, given that the respective meta-analyses individually revealed the dmPFC (direct overlap) and the amygdala (oppositely lateralized). Third, the neural correlates of moral judgments are much closer related to the neural correlates of ToM than to those of empathy, as evidenced by the quantity of overlap in the conjunction analyses.

Converging neural correlates across all individual analyses

Moral cognition-, ToM-, and empathy-related brain activity converged in the bilateral TPJ, dmPFC, and right MTG, which therefore form a common network potentially involved in social-cognitive processes. In line with this, metadata profiling demonstrates solid associations of this network with neuroimaging studies related to ToM, explicit memory retrieval, language, and imagination of objects/scenes. Intriguingly, these four seemingly disparate psychological categories summarize what sets humans probably apart from non-human primates (Tomasello 2001; Frith and Frith 2010). They might functionally converge in the reciprocal relationship between the allocentric and egocentric perspective, instructed by self-reflection, social knowledge, and memories of past experiences (of social interactions). In particular, autobiographical memory supplies numerous building blocks of social semantic knowledge (Bar 2007; Binder et al. 2009). These isolated conceptual scripts may be reassembled to enable the forecasting of future events (Tulving 1983, 1985; Schacter et al. 2007). Similar brain mechanisms in remembering past episodes and envisioning future circumstances is supported by their engagement of identical brain areas, as evidenced by a quantitative meta-analysis (Spreng et al. 2009). Moreover, retrograde amnesic patients were reported to be impaired not only in prospection but also in imagining novel experience (Hassabis et al. 2007). These findings suggest a single neural network for mentally constructing plausible semantic scenarios of detached situations regardless of temporal orientation (Buckner and Carroll 2007; Hassabis and Maguire 2007). Indeed, construction of detached probabilistic scenes has been argued to influence ongoing decision making by estimating outcomes of behavioral choices (Boyer 2008; Suddendorf and Corballis 2007; Schilbach et al. 2008b). Taken together, moral cognition, ToM, and empathy jointly engage a network that might be involved in the automated prediction of social events that modulate behavior.

Relation to clinical research

Consistent with the demonstrated functional dissociation between cognitive and affective subsystems of the neural network related to moral cognition observed in our study, frontotemporal dementia has been reported to impair personal but not impersonal moral reasoning (Mendez et al. 2005). A cognitive-affective dissociation of moral cognition is also supported clinically by the psychopathic population’s immoral behavior in everyday life despite excellent moral reasoning skills (Cleckley 1941; Hare 1993). Apart from that, neither Greene’s nor Moll’s concept can exhaustively explain why vmPFC patients demonstrated a rationally biased approach to solving moral dilemmas (Koenigs et al. 2007; Moretto et al. 2010), yet, an emotionally biased approach to moral cognition in an economic game (Koenigs and Tranel 2007). Given the considerable neural commonalities of moral cognition and ToM tasks as revealed by the present analysis, contradictory findings of vmPFC patients dealing with moral dilemmas and economic games (Greene 2007; Koenigs and Tranel 2007; Koenigs et al. 2007) might resolve when meticulously probing ToM capabilities in future lesion studies.

Limitations

Several limitations of our study should be addressed. Meta-analyses are necessarily based on the available literature and hence may be affected by the potential publication-bias disfavoring null results (Rosenthal 1979). This caveat especially applies to the functional characterization of converging activation foci using the BrainMap database, as this database only contains about 21% of the published neuroimaging studies and its content therefore does not constitute a strictly representative sample. Furthermore, a part of the included neuroimaging studies on moral cognition might suffer from limited ecological validity. That is, the experimental tasks used might only partially involve the mental processes guiding real-life moral behavior (cf. Cima et al. 2010; Knutson et al. 2010; Young et al. 2010a). In particular, employing overly artificial moral scenarios (e.g., trolley dilemma), on top of the inherent limitations of neuroimaging paradigms, could have systematically overestimated cognitive versus emotional processes. This caveat might have contributed to the similar results in the meta-analyses on moral cognition and on ToM (rather than on empathy), given that emotion processing is thought to play a paramount role in real-world moral cognition (Krebs 2008; Tangney et al. 2007; Haidt 2001). Future neuroimaging studies should therefore strive for using more realistic moral scenarios to minimize the risk of investigating the neural correlates of “in vitro moral cognition” (cf. Schilbach et al. 2012; Schilbach 2010).

Conclusions

It is a topic of intense debate whether social cognition is subserved by a unitary specialized module or by a set of general-purpose mental operations (Mitchell 2006; Spreng et al. 2009; Bzdok et al. 2012; Van Overwalle 2011). The present large-scale meta-analysis provides evidence for a domain-global view of moral cognition, rather than for a distinct moral module (Hauser 2006), by showing its functional integration of distributed brain networks. More specifically, we parsed the neural correlates of moral cognition by reference to a socio-cognitive framework, exemplified by ToM cognition, and a socio-affective framework, exemplified by empathy. Ultimately, our results support the notion that moral reasoning is related to both seeing things from other persons’ points of view and to grasping others’ feelings (Piaget 1932; Tomasello 2001; Decety et al. 2012).

Electronic supplementary material

Acknowledgments

This study was supported by the German Research Council (DFG, IRTG 1328; KZ, SBE, DB), the Human Brain Project (R01-MH074457-01A1; ARL, SBE), and the Helmholtz Initiative on Systems Biology (Human Brain Model; KZ, SBE). The authors declare no conflict of interest.

References

  1. Amodio DM, Frith CD. Meeting of minds: the medial frontal cortex and social cognition. Nat Rev Neurosci. 2006;7(4):268–277. doi: 10.1038/nrn1884. [DOI] [PubMed] [Google Scholar]
  2. Amunts K, Schleicher A, Burgel U, Mohlberg H, Uylings HB, Zilles K. Broca’s region revisited: cytoarchitecture and intersubject variability. J Comp Neurol. 1999;412(2):319–341. doi: 10.1002/(sici)1096-9861(19990920)412:2<319::aid-cne10>3.0.co;2-7. [DOI] [PubMed] [Google Scholar]
  3. Amunts K, Kedo O, Kindler M, Pieperhoff P, Mohlberg H, Shah NJ, Habel U, Schneider F, Zilles K. Cytoarchitectonic mapping of the human amygdala, hippocampal region and entorhinal cortex: intersubject variability and probability maps. Anat Embryol (Berl) 2005;210(5–6):343–352. doi: 10.1007/s00429-005-0025-5. [DOI] [PubMed] [Google Scholar]
  4. Anderson SW, Bechara A, Damasio H, Tranel D, Damasio AR. Impairment of social and moral behavior related to early damage in human prefrontal cortex. Nat Neurosci. 1999;2(11):1032–1037. doi: 10.1038/14833. [DOI] [PubMed] [Google Scholar]
  5. Aristotle (1985) Nicomachean Ethics. Hackett, Indianapolis
  6. Ball T, Rahm B, Eickhoff SB, Schulze-Bonhage A, Speck O, Mutschler I. Response properties of human amygdala subregions: evidence based on functional MRI combined with probabilistic anatomical maps. PLoS One. 2007;2(3):e307. doi: 10.1371/journal.pone.0000307. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Bar M. The proactive brain: using analogies and associations to generate predictions. Trends Cogn Sci. 2007;11(7):280–289. doi: 10.1016/j.tics.2007.05.005. [DOI] [PubMed] [Google Scholar]
  8. Binder JR, Desai RH, Graves WW, Conant LL. Where is the semantic system? A critical review and meta-analysis of 120 functional neuroimaging studies. Cereb Cortex. 2009;19(12):2767–2796. doi: 10.1093/cercor/bhp055. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Blair RJ. The amygdala and ventromedial prefrontal cortex in morality and psychopathy. Trends Cogn Sci. 2007;11(9):387–392. doi: 10.1016/j.tics.2007.07.003. [DOI] [PubMed] [Google Scholar]
  10. Boyer P. Evolutionary economics of mental time travel? Trends Cogn Sci. 2008;12(6):219–224. doi: 10.1016/j.tics.2008.03.003. [DOI] [PubMed] [Google Scholar]
  11. Buckner RL, Carroll DC. Self-projection and the brain. Trends Cogn Sci. 2007;11(2):49–57. doi: 10.1016/j.tics.2006.11.004. [DOI] [PubMed] [Google Scholar]
  12. Bzdok D, Langner R, Caspers S, Kurth F, Habel U, Zilles K, Laird A, Eickhoff SB. ALE meta-analysis on facial judgments of trustworthiness and attractiveness. Brain Struct Funct. 2011;215(3–4):209–223. doi: 10.1007/s00429-010-0287-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Bzdok D, Langner R, Hoffstaedter F, Turetsky B, Zilles K, Eickhoff SB (2012) The modular neuroarchitecture of social judgments on faces. Cereb Cortex (in press) [DOI] [PMC free article] [PubMed]
  14. Caspers S, Geyer S, Schleicher A, Mohlberg H, Amunts K, Zilles K. The human inferior parietal cortex: cytoarchitectonic parcellation and interindividual variability. Neuroimage. 2006;33:430–448. doi: 10.1016/j.neuroimage.2006.06.054. [DOI] [PubMed] [Google Scholar]
  15. Caspers S, Zilles K, Laird AR, Eickhoff SB. ALE meta-analysis of action observation and imitation in the human brain. Neuroimage. 2010;50(3):1148–1167. doi: 10.1016/j.neuroimage.2009.12.112. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Cheng Y, Lin CP, Liu HL, Hsu YY, Lim KE, Hung D, Decety J. Expertise modulates the perception of pain in others. Curr Biol. 2007;17(19):1708–1713. doi: 10.1016/j.cub.2007.09.020. [DOI] [PubMed] [Google Scholar]
  17. Cima M, Tonnaer F, Hauser MD. Psychopaths know right from wrong but don’t care. Soc Cogn Affect Neurosci. 2010;5:59–67. doi: 10.1093/scan/nsp051. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Cleckley H. The mask of sanity: an attempt to clarify some issues about the so-called psychopathic personality. St. Louis: CV Mosby; 1941. [Google Scholar]
  19. Craig AD. How do you feel? Interoception: the sense of the physiological condition of the body. Nat Rev Neurosci. 2002;3(8):655–666. doi: 10.1038/nrn894. [DOI] [PubMed] [Google Scholar]
  20. Craig AD. How do you feel—now? The anterior insula and human awareness. Nat Rev Neurosci. 2009;10(1):59–70. doi: 10.1038/nrn2555. [DOI] [PubMed] [Google Scholar]
  21. Cushman F. Crime and punishment: distinguishing the roles of causal and intentional analyses in moral judgment. Cognition. 2008;108:353–380. doi: 10.1016/j.cognition.2008.03.006. [DOI] [PubMed] [Google Scholar]
  22. D’Argembeau A, Ruby P, Collette F, Degueldre C, Balteau E, Luxen A, Maquet P, Salmon E. Distinct regions of the medial prefrontal cortex are associated with self-referential processing and perspective taking. J Cogn Neurosci. 2007;19(6):935–944. doi: 10.1162/jocn.2007.19.6.935. [DOI] [PubMed] [Google Scholar]
  23. Danziger N, Faillenot I, Peyron R. Can we share a pain we never felt? Neural correlates of empathy in patients with congenital insensitivity to pain. Neuron. 2009;61(2):203–212. doi: 10.1016/j.neuron.2008.11.023. [DOI] [PubMed] [Google Scholar]
  24. Darwin C (1874/1997) The descent of man. Prometheus Books, New York
  25. Lange FP, Spronk M, Willems RM, Toni I, Bekkering H. Complementary systems for understanding action intentions. Curr Biol. 2008;18:454–457. doi: 10.1016/j.cub.2008.02.057. [DOI] [PubMed] [Google Scholar]
  26. Decety J, Jackson PL. The functional architecture of human empathy. Behav Cogn Neurosci Rev. 2004;3(2):71–100. doi: 10.1177/1534582304267187. [DOI] [PubMed] [Google Scholar]
  27. Decety J, Lamm C. The role of the right temporoparietal junction in social interaction: how low-level computational processes contribute to meta-cognition. Neuroscientist. 2007;13(6):580–593. doi: 10.1177/1073858407304654. [DOI] [PubMed] [Google Scholar]
  28. Decety J, Michalska KJ, Kinzler KD (2012) The contribution of emotion and cognition to moral sensitivity: a neurodevelopmental study. Cereb Cortex. doi:10.1093/cercor/bhr111 [DOI] [PubMed]
  29. Eickhoff SB, Bzdok D (2012) Meta-analyses in basic and clinical neuroscience: State of the art and perspective. In: Ulmer S, Jansen O (eds) fMRI—Basics and Clinical Applications, 2nd edn. Springer, Berlin (in press)
  30. Eickhoff SB, Stephan KE, Mohlberg H, Grefkes C, Fink GR, Amunts K, Zilles K. A new SPM toolbox for combining probabilistic cytoarchitectonic maps and functional imaging data. Neuroimage. 2005;25(4):1325–1335. doi: 10.1016/j.neuroimage.2004.12.034. [DOI] [PubMed] [Google Scholar]
  31. Eickhoff SB, Paus T, Caspers S, Grosbras MH, Evans AC, Zilles K, Amunts K. Assignment of functional activations to probabilistic cytoarchitectonic areas revisited. Neuroimage. 2007;36(3):511–521. doi: 10.1016/j.neuroimage.2007.03.060. [DOI] [PubMed] [Google Scholar]
  32. Eickhoff SB, Laird AR, Grefkes C, Wang LE, Zilles K, Fox PT. Coordinate-based activation likelihood estimation meta-analysis of neuroimaging data: a random-effects approach based on empirical estimates of spatial uncertainty. Hum Brain Mapp. 2009;30(9):2907–2926. doi: 10.1002/hbm.20718. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Eickhoff SB, Bzdok D, Laird AR, Roski C, Caspers S, Zilles K, Fox PT. Co-activation patterns distinguish cortical modules, their connectivity and functional differentiation. Neuroimage. 2011;57(3):938–949. doi: 10.1016/j.neuroimage.2011.05.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Eickhoff SB, Bzdok D, Laird AR, Kurth F, Fox PT (2012) Activation likelihood estimation meta-analysis revisited. Neuroimage. doi:10.1016/j.neuroimage.2011.09.017 [DOI] [PMC free article] [PubMed]
  35. Eisenberger N. Emotion, regulation, and moral development. Ann Rev Psychol. 2000;51:665–697. doi: 10.1146/annurev.psych.51.1.665. [DOI] [PubMed] [Google Scholar]
  36. Ekman P. Emotion in the human face. New York: Cambridge University Press; 1982. [Google Scholar]
  37. Fan Y, Duncan NW, Greck M, Northoff G. Is there a core neural network in empathy? An fMRI based quantitative meta-analysis. Neurosci Biobehav Rev. 2011;35(3):903–911. doi: 10.1016/j.neubiorev.2010.10.009. [DOI] [PubMed] [Google Scholar]
  38. Feshbach ND, Feshbach S. The relationship between empathy and aggression in two age groups. Dev Psychol. 1969;1:102–107. [Google Scholar]
  39. Fink GR, Markowitsch HJ, Reinkemeier M, Bruckbauer T, Kessler J, Heiss WD. Cerebral representation of one’s own past: neural networks involved in autobiographical memory. J Neurosci. 1996;16(13):4275–4282. doi: 10.1523/JNEUROSCI.16-13-04275.1996. [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Forbes CE, Grafman J. The role of the human prefrontal cortex in social cognition and moral judgment. Annu Rev Neurosci. 2010;33:299–324. doi: 10.1146/annurev-neuro-060909-153230. [DOI] [PubMed] [Google Scholar]
  41. Friston KJ, Rotshtein P, Geng JJ, Sterzer P, Henson RN. A critique of functional localisers. Neuroimage. 2006;30(4):1077–1087. doi: 10.1016/j.neuroimage.2005.08.012. [DOI] [PubMed] [Google Scholar]
  42. Frith U, Frith CD. Development and neurophysiology of mentalizing. Philos Trans R Soc Lond B Biol Sci. 2003;358(1431):459–473. doi: 10.1098/rstb.2002.1218. [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Frith U, Frith C. The social brain: allowing humans to boldly go where no other species has been. Philos Trans R Soc Lond B Biol Sci. 2010;365(1537):165–176. doi: 10.1098/rstb.2009.0160. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Fusar-Poli P, Placentino A, Carletti F, Landi P, Allen P, Surguladze S, Benedetti F, Abbamonte M, Gasparotti R, Barale F, Perez J, McGuire P, Politi P. Functional atlas of emotional faces processing: a voxel-based meta-analysis of 105 functional magnetic resonance imaging studies. J Psychiatry Neurosci. 2009;34(6):418–432. [PMC free article] [PubMed] [Google Scholar]
  45. Gallagher HL, Frith CD. Functional imaging of ‘theory of mind’. Trends Cogn Sci. 2003;7(2):77–83. doi: 10.1016/s1364-6613(02)00025-6. [DOI] [PubMed] [Google Scholar]
  46. Geyer S. The microstructural border between the motor and the cognitive domain in the human cerebral cortex. Adv Anat Embryol Cell Biol. 2004;174(I–VIII):1–89. doi: 10.1007/978-3-642-18910-4. [DOI] [PubMed] [Google Scholar]
  47. Gläscher J, Adolphs R. Processing of the arousal of subliminal and supraliminal emotional stimuli by the human amygdala. J Neurosci. 2003;23(32):10274–10282. doi: 10.1523/JNEUROSCI.23-32-10274.2003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Greene JD (2007) Why are VMPFC patients more utilitarian? A dual-process theory of moral judgment explains. Trends Cogn Sci 11(8):322–323. doi:10.1016/j.tics.2007.06.004 (author reply 323–324) [DOI] [PubMed]
  49. Greene JD, Haidt J. How (and where) does moral judgment work? Trends Cogn Sci. 2002;6(12):517–523. doi: 10.1016/s1364-6613(02)02011-9. [DOI] [PubMed] [Google Scholar]
  50. Greene JD, Sommerville RB, Nystrom LE, Darley JM, Cohen JD. An fMRI investigation of emotional engagement in moral judgment. Science. 2001;293(5537):2105–2108. doi: 10.1126/science.1062872. [DOI] [PubMed] [Google Scholar]
  51. Greene JD, Nystrom LE, Engell AD, Darley JM, Cohen JD. The neural bases of cognitive conflict and control in moral judgment. Neuron. 2004;44(2):389–400. doi: 10.1016/j.neuron.2004.09.027. [DOI] [PubMed] [Google Scholar]
  52. Haidt J. The emotional dog and its rational tail: a social intuitionist approach to moral judgment. Psychol Rev. 2001;108(4):814–834. doi: 10.1037/0033-295x.108.4.814. [DOI] [PubMed] [Google Scholar]
  53. Hare RD (1993) Without conscience: the disturbing world of the psychopath among us. Pocket Books, New York
  54. Hare RD (2003) The hare psychopathy checklist—revised, 2nd edn. Multi-Health Systems, Toronto
  55. Hassabis D, Maguire EA. Deconstructing episodic memory with construction. Trends Cogn Sci. 2007;11(7):299–306. doi: 10.1016/j.tics.2007.05.001. [DOI] [PubMed] [Google Scholar]
  56. Hassabis D, Kumaran D, Vann SD, Maguire EA. Patients with hippocampal amnesia cannot imagine new experiences. Proc Natl Acad Sci USA. 2007;104(5):1726–1731. doi: 10.1073/pnas.0610561104. [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Hauser MD. The liver and the moral organ. Soc Cogn Affect Neurosci. 2006;1(3):214–220. doi: 10.1093/scan/nsl026. [DOI] [PMC free article] [PubMed] [Google Scholar]
  58. Hume D (1777/2006) An inquiry concerning the principles of morals. Cosimo Classics, New York
  59. Izard CE. The face of emotion. New York: Appleton-Century-Crofts; 1971. [Google Scholar]
  60. Jacob F. Evolution and tinkering. Science. 1977;196(4295):1161–1166. doi: 10.1126/science.860134. [DOI] [PubMed] [Google Scholar]
  61. Jenkins AC, Mitchell JP. Mentalizing under uncertainty: dissociated neural responses to ambiguous and unambiguous mental state inferences. Cereb Cortex. 2010;20(2):404–410. doi: 10.1093/cercor/bhp109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Kant I (1785/1993) Grounding for the Metaphysics of Morals. Hackett Publishing Co, Indianapolis
  63. Keysers C, Gazzola V. Integrating simulation and theory of mind: from self to social cognition. Trends Cogn Sci. 2007;11(5):194–196. doi: 10.1016/j.tics.2007.02.002. [DOI] [PubMed] [Google Scholar]
  64. Killen M, Mulvey KL, Richardson C, Jampol N, Woodward A. The accidental transgressor: morally–relevant theory of mind. Cognition. 2011;119:197–215. doi: 10.1016/j.cognition.2011.01.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  65. Knobe J. Theory of mind and moral cognition: exploring the connections. Trends Cogn Sci. 2005;9(8):357–359. doi: 10.1016/j.tics.2005.06.011. [DOI] [PubMed] [Google Scholar]
  66. Knutson KM, Krueger F, Koenigs M, Hawley A, Escobedo JR, Vasudeva V, Adolphs R, Grafman J. Behavioral norms for condensed moral vignettes. Soc Cogn Affect Neurosci. 2010;5(4):378–384. doi: 10.1093/scan/nsq005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Koenigs M, Tranel D. Irrational economic decision-making after ventromedial prefrontal damage: evidence from the Ultimatum Game. J Neurosci. 2007;27(4):951–956. doi: 10.1523/JNEUROSCI.4606-06.2007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Koenigs M, Young L, Adolphs R, Tranel D, Cushman F, Hauser M, Damasio A. Damage to the prefrontal cortex increases utilitarian moral judgements. Nature. 2007;446(7138):908–911. doi: 10.1038/nature05631. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Kohlberg L (1969) Stage and sequence: the cognitive-developmental approach to socialization. In: Goslin DA (ed) Handbook of socialization theory and research. Rand McNally, Chicago, pp 347–480
  70. Kohlberg L, Levine C, Hewer A (1983) Moral stages: A current formulation and a response to critics. Karger, Basel
  71. Krebs DL. Morality: an evolutionary account. Perspect Psychol Sci. 2008;3(3):149–172. doi: 10.1111/j.1745-6924.2008.00072.x. [DOI] [PubMed] [Google Scholar]
  72. Krueger F, Barbey AK, Grafman J. The medial prefrontal cortex mediates social event knowledge. Trends Cogn Sci. 2009;13(3):103–109. doi: 10.1016/j.tics.2008.12.005. [DOI] [PubMed] [Google Scholar]
  73. Kurth F, Zilles K, Fox PT, Laird AR, Eickhoff SB. A link between the systems: functional differentiation and integration within the human insula revealed by meta-analysis. Brain Struct Funct. 2010;214(5–6):519–534. doi: 10.1007/s00429-010-0255-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Laird AR, Eickhoff SB, Kurth F, Fox PM, Uecker AM, Turner JA, Robinson JL, Lancaster JL, Fox PT. ALE meta-analysis workflows via the Brainmap database: progress towards a probabilistic functional brain atlas. Front Neuroinformatics. 2009;3:23. doi: 10.3389/neuro.11.023.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  75. Laird AR, Eickhoff SB, Li K, Robin DA, Glahn DC, Fox PT. Investigating the functional heterogeneity of the default mode network using coordinate-based meta-analytic modeling. J Neurosci. 2009;29(46):14496–14505. doi: 10.1523/JNEUROSCI.4004-09.2009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Laird AR, Eickhoff SB, Fox PM, Uecker AM, Ray KL, Saenz JJ, Jr, McKay DR, Bzdok D, Laird RW, Robinson JL, Turner JA, Turkeltaub PE, Lancaster JL, Fox PT. The BrainMap Strategy for Standardization, Sharing, and Meta-Analysis of Neuroimaging Data. BMC Res Notes. 2011;4(1):349. doi: 10.1186/1756-0500-4-349. [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. Lamm C, Singer T. The role of anterior insular cortex in social emotions. Brain Struct Funct. 2010;214(5):579–591. doi: 10.1007/s00429-010-0251-3. [DOI] [PubMed] [Google Scholar]
  78. Lamm C, Batson CD, Decety J. The neural substrate of human empathy: effects of perspective-taking and cognitive appraisal. J Cogn Neurosci. 2007;19(1):42–58. doi: 10.1162/jocn.2007.19.1.42. [DOI] [PubMed] [Google Scholar]
  79. Lamm C, Decety J, Singer T. Meta-analytic evidence for common and distinct neural networks associated with directly experienced pain and empathy for pain. Neuroimage. 2011;54(3):2492–2502. doi: 10.1016/j.neuroimage.2010.10.014. [DOI] [PubMed] [Google Scholar]
  80. Leiberg S, Anders S. The multiple facets of empathy: a survey of theory and evidence. Prog Brain Res. 2006;156:419–440. doi: 10.1016/S0079-6123(06)56023-6. [DOI] [PubMed] [Google Scholar]
  81. Leslie AM, Knobe J, Cohen A. Acting intentionally and the side-effect effect: theory of mind and moral judgment. Psychol Sci. 2006;17:421–427. doi: 10.1111/j.1467-9280.2006.01722.x. [DOI] [PubMed] [Google Scholar]
  82. Lieberman MD. Social cognitive neuroscience: a review of core processes. Annu Rev Psychol. 2007;58:259–289. doi: 10.1146/annurev.psych.58.110405.085654. [DOI] [PubMed] [Google Scholar]
  83. Maddock RJ. The retrosplenial cortex and emotion: new insights from functional neuroimaging of the human brain. Trends Neurosci. 1999;22(7):310–316. doi: 10.1016/s0166-2236(98)01374-5. [DOI] [PubMed] [Google Scholar]
  84. Maddock RJ, Garrett AS, Buonocore MH. Remembering familiar people: the posterior cingulate cortex and autobiographical memory retrieval. Neuroscience. 2001;104(3):667–676. doi: 10.1016/s0306-4522(01)00108-7. [DOI] [PubMed] [Google Scholar]
  85. Mar RA. The neural bases of social cognition and story comprehension. Annu Rev Psychol. 2011;62:103–134. doi: 10.1146/annurev-psych-120709-145406. [DOI] [PubMed] [Google Scholar]
  86. Markowitsch HJ. Differential contribution of right and left amygdala to affective information processing. Behav Neurol. 1998;11(4):233–244. doi: 10.1155/1999/180434. [DOI] [PubMed] [Google Scholar]
  87. Mendez MF, Anderson E, Shapira JS. An investigation of moral judgement in frontotemporal dementia. Cogn Behav Neurol. 2005;18(4):193–197. doi: 10.1097/01.wnn.0000191292.17964.bb. [DOI] [PubMed] [Google Scholar]
  88. Mitchell JP (2005) The false dichotomy between simulation and theory–theory: the argument’s error. Trends Cogn Sci 9(8):363–364. doi:10.1016/j.tics.2005.06.010 (author reply 364) [DOI] [PubMed]
  89. Mitchell JP. Mentalizing and Marr: an information processing approach to the study of social cognition. Brain Res. 2006;1079(1):66–75. doi: 10.1016/j.brainres.2005.12.113. [DOI] [PubMed] [Google Scholar]
  90. Mitchell JP. Activity in right temporo-parietal junction is not selective for theory-of-mind. Cereb Cortex. 2008;18(2):262–271. doi: 10.1093/cercor/bhm051. [DOI] [PubMed] [Google Scholar]
  91. Mitchell JP, Macrae CN, Banaji MR. Dissociable medial prefrontal contributions to judgments of similar and dissimilar others. Neuron. 2006;50(4):655–663. doi: 10.1016/j.neuron.2006.03.040. [DOI] [PubMed] [Google Scholar]
  92. Moll J, Schulkin J. Social attachment and aversion in human moral cognition. Neurosci Biobehav Rev. 2009;33(3):456–465. doi: 10.1016/j.neubiorev.2008.12.001. [DOI] [PubMed] [Google Scholar]
  93. Moll J, Oliveira-Souza R, Moll FT, Ignacio FA, Bramati IE, Caparelli-Daquer EM, Eslinger PJ. The moral affiliations of disgust: a functional MRI study. Cogn Behav Neurol. 2005;18(1):68–78. doi: 10.1097/01.wnn.0000152236.46475.a7. [DOI] [PubMed] [Google Scholar]
  94. Moll J, Zahn R, Oliveira-Souza R, Krueger F, Grafman J. Opinion: the neural basis of human moral cognition. Nat Rev Neurosci. 2005;6(10):799–809. doi: 10.1038/nrn1768. [DOI] [PubMed] [Google Scholar]
  95. Moll J, Krueger F, Zahn R, Pardini M, Oliveira-Souza R, Grafman J. Human fronto-mesolimbic networks guide decisions about charitable donation. Proc Natl Acad Sci USA. 2006;103(42):15623–15628. doi: 10.1073/pnas.0604475103. [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Moran JM, Young LL, Saxe R, Lee SM, O’Young D, Mavros PL, Gabrieli JD. Impaired theory of mind for moral judgment in high-functioning autism. Proc Natl Acad Sci USA. 2011;108(7):2688–2692. doi: 10.1073/pnas.1011734108. [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Moretto G, Ladavas E, Mattioli F, Pellegrino G. A psychophysiological investigation of moral judgment after ventromedial prefrontal damage. J Cogn Neurosci. 2010;22(8):1888–1899. doi: 10.1162/jocn.2009.21367. [DOI] [PubMed] [Google Scholar]
  98. Müller VI, Habel U, Derntl B, Schneider F, Zilles K, Turetsky BI, Eickhoff SB. Incongruence effects in crossmodal emotional integration. Neuroimage. 2011;54(3):2257–2266. doi: 10.1016/j.neuroimage.2010.10.047. [DOI] [PMC free article] [PubMed] [Google Scholar]
  99. Phelps EA, O’Connor KJ, Gatenby JC, Gore JC, Grillon C, Davis M. Activation of the left amygdala to a cognitive representation of fear. Nat Neurosci. 2001;4(4):437–441. doi: 10.1038/86110. [DOI] [PubMed] [Google Scholar]
  100. Piaget J (1932) The moral judgment of the child. Kegan Paul, London
  101. Pizarro DA, Bloom P (2003) The intelligence of the moral intuitions: comment on Haidt (2001). Psychol Rev 110(1):193–196 [DOI] [PubMed]
  102. Poldrack RA. Can cognitive processes be inferred from neuroimaging data? Trends Cogn Sci. 2006;10(2):59–63. doi: 10.1016/j.tics.2005.12.004. [DOI] [PubMed] [Google Scholar]
  103. Premack D, Woodruff G. Does the chimpanzee have a theory of mind. Behav Brain Sci. 1978;1(4):515–526. [Google Scholar]
  104. Raine A, Yang Y. Neural foundations to moral reasoning and antisocial behavior. Soc Cogn Affect Neurosci. 2006;1(3):203–213. doi: 10.1093/scan/nsl033. [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Rosenthal R. The file drawer problem and tolerance for null results. Psychol Bull. 1979;86(3):638–641. [Google Scholar]
  106. Sander D, Grafman J, Zalla T. The human amygdala: an evolved system for relevance detection. Rev Neurosci. 2003;14(4):303–316. doi: 10.1515/revneuro.2003.14.4.303. [DOI] [PubMed] [Google Scholar]
  107. Saper CB. Pain as a visceral sensation. Prog Brain Res. 2000;122:237–243. doi: 10.1016/s0079-6123(08)62142-1. [DOI] [PubMed] [Google Scholar]
  108. Saxe R, Kanwisher N (2003) People thinking about thinking people. The role of the temporo-parietal junction in “theory of mind”. Neuroimage 19(4):1835–1842. pii:S1053811903002301 [DOI] [PubMed]
  109. Schacter DL, Addis DR, Buckner RL. Remembering the past to imagine the future: the prospective brain. Nat Rev Neurosci. 2007;8(9):657–661. doi: 10.1038/nrn2213. [DOI] [PubMed] [Google Scholar]
  110. Schilbach L. A second-person approach to other minds. Nat Rev Neurosci. 2010;11(6):449. doi: 10.1038/nrn2805-c1. [DOI] [PubMed] [Google Scholar]
  111. Schilbach L, Eickhoff SB, Mojzisch A, Vogeley K. What’s in a smile? Neural correlates of facial embodiment during social interaction. Soc Neurosci. 2008;3(1):37–50. doi: 10.1080/17470910701563228. [DOI] [PubMed] [Google Scholar]
  112. Schilbach L, Eickhoff SB, Rotarska-Jagiela A, Fink GR, Vogeley K. Minds at rest? Social cognition as the default mode of cognizing and its putative relationship to the “default system” of the brain. Conscious Cogn. 2008;17(2):457–467. doi: 10.1016/j.concog.2008.03.013. [DOI] [PubMed] [Google Scholar]
  113. Schilbach L, Wilms M, Eickhoff SB, Romanzetti S, Tepest R, Bente G, Shah NJ, Fink GR, Vogeley K. Minds made for sharing: initiating joint attention recruits reward-related neurocircuitry. J Cogn Neurosci. 2010;22(12):2702–2715. doi: 10.1162/jocn.2009.21401. [DOI] [PubMed] [Google Scholar]
  114. Schilbach L, Eickhoff SB, Cieslik EC, Kuzmanovic B, Vogeley K (2012) Shall we do this together? Social gaze influences action control in a comparison group, but not in individuals with high-functioning autism. Autism. doi:10.1177/1362361311409258 [DOI] [PubMed]
  115. Shamay-Tsoory SG, Tibi-Elhanany Y, Aharon-Peretz J. The ventromedial prefrontal cortex is involved in understanding affective but not cognitive theory of mind stories. Soc Neurosci. 2006;1(3–4):149–166. doi: 10.1080/17470910600985589. [DOI] [PubMed] [Google Scholar]
  116. Shamay-Tsoory SG, Aharon-Peretz J, Perry D. Two systems for empathy: a double dissociation between emotional and cognitive empathy in inferior frontal gyrus versus ventromedial prefrontal lesions. Brain. 2009;132:617–627. doi: 10.1093/brain/awn279. [DOI] [PubMed] [Google Scholar]
  117. Singer T, Lamm C. The social neuroscience of empathy. Ann N Y Acad Sci. 2009;1156:81–96. doi: 10.1111/j.1749-6632.2009.04418.x. [DOI] [PubMed] [Google Scholar]
  118. Singer T, Seymour B, O’Doherty J, Kaube H, Dolan RJ, Frith CD. Empathy for pain involves the affective but not sensory components of pain. Science. 2004;303(5661):1157–1162. doi: 10.1126/science.1093535. [DOI] [PubMed] [Google Scholar]
  119. Soderstrom H. Psychopathy as a disorder of empathy. Eur Child Adolesc Psychiatry. 2003;12:249–252. doi: 10.1007/s00787-003-0338-y. [DOI] [PubMed] [Google Scholar]
  120. Spengler S, Cramon DY, Brass M. Control of shared representations relies on key processes involved in mental state attribution. Hum Brain Mapp. 2009;30(11):3704–3718. doi: 10.1002/hbm.20800. [DOI] [PMC free article] [PubMed] [Google Scholar]
  121. Spreng RN, Mar RA, Kim AS. The common neural basis of autobiographical memory, prospection, navigation, theory of mind, and the default mode: a quantitative meta-analysis. J Cogn Neurosci. 2009;21(3):489–510. doi: 10.1162/jocn.2008.21029. [DOI] [PubMed] [Google Scholar]
  122. Suddendorf T, Corballis MC (2007) The evolution of foresight: What is mental time travel, and is it unique to humans? Behav Brain Sci 30(3):299–313. doi:10.1017/S0140525X07001975 (discussion 313–351) [DOI] [PubMed]
  123. Tangney JP, Stuewig J, Mashek DJ. Moral emotions and moral behavior. Annu Rev Psychol. 2007;58:345–372. doi: 10.1146/annurev.psych.56.091103.070145. [DOI] [PMC free article] [PubMed] [Google Scholar]
  124. Tomasello M. The cultural origins of human cognition. Cambridge: Harvard University Press; 2001. [Google Scholar]
  125. Tulving E. Elements of Episodic Memory. Oxford: Clarendon Press; 1983. [Google Scholar]
  126. Tulving E. Memory and consciousness. Can Psychol. 1985;26:1–12. [Google Scholar]
  127. Turkeltaub PE, Eden GF, Jones KM, Zeffiro TA. Meta-analysis of the functional neuroanatomy of single-word reading: method and validation. Neuroimage. 2002;16(3):765–780. doi: 10.1006/nimg.2002.1131. [DOI] [PubMed] [Google Scholar]
  128. Turkeltaub PE, Eickhoff SB, Laird AR, Fox M, Wiener M, Fox P. Minimizing within-experiment and within-group effects in activation likelihood estimation meta-analyses. Hum Brain Mapp. 2011 doi: 10.1002/hbm.21186. [DOI] [PMC free article] [PubMed] [Google Scholar]
  129. Overwalle F. Social cognition and the brain: a meta-analysis. Hum Brain Mapp. 2009;30(3):829–858. doi: 10.1002/hbm.20547. [DOI] [PMC free article] [PubMed] [Google Scholar]
  130. Overwalle F. A dissociation between social mentalizing and general reasoning. Neuroimage. 2011;54(2):1589–1599. doi: 10.1016/j.neuroimage.2010.09.043. [DOI] [PubMed] [Google Scholar]
  131. Vogt BA. Pain and emotion interactions in subregions of the cingulate gyrus. Nat Rev Neurosci. 2005;6(7):533–544. doi: 10.1038/nrn1704. [DOI] [PMC free article] [PubMed] [Google Scholar]
  132. Wagner U, N’Diaye K, Ethofer T, Vuilleumier P (2012) Guilt-Specific Processing in the Prefrontal Cortex. Cereb Cortex. doi:10.1093/cercor/bhr016 [DOI] [PubMed]
  133. Wicker B, Keysers C, Plailly J, Royet JP, Gallese V, Rizzolatti G. Both of us disgusted in My Insula: The common neural basis of seeing and feeling disgust. Neuron. 2003;40(3):655–664. doi: 10.1016/s0896-6273(03)00679-2. [DOI] [PubMed] [Google Scholar]
  134. Young L, Saxe R. The neural basis of belief encoding and integration in moral judgment. Neuroimage. 2008;40(4):1912–1920. doi: 10.1016/j.neuroimage.2008.01.057. [DOI] [PubMed] [Google Scholar]
  135. Young L, Saxe R. Innocent intentions: a correlation between forgiveness for accidental harm and neural activity. Neuropsychologia. 2009;47(10):2065–2072. doi: 10.1016/j.neuropsychologia.2009.03.020. [DOI] [PubMed] [Google Scholar]
  136. Young L, Cushman F, Hauser M, Saxe R. The neural basis of the interaction between theory of mind and moral judgment. Proc Natl Acad Sci USA. 2007;104(20):8235–8240. doi: 10.1073/pnas.0701408104. [DOI] [PMC free article] [PubMed] [Google Scholar]
  137. Young L, Bechara A, Tranel D, Damasio H, Hauser M, Damasio A. Damage to ventromedial prefrontal cortex impairs judgment of harmful intent. Neuron. 2010;65(6):845–851. doi: 10.1016/j.neuron.2010.03.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  138. Young L, Camprodon JA, Hauser M, Pascual-Leone A, Saxe R. Disruption of the right temporoparietal junction with transcranial magnetic stimulation reduces the role of beliefs in moral judgments. Proc Natl Acad Sci USA. 2010;107(15):6753–6758. doi: 10.1073/pnas.0914826107. [DOI] [PMC free article] [PubMed] [Google Scholar]
  139. Zahn-Waxler C, Radke-Yarrow M, Wagner E, chapman M. Development of concern for others. Dev Psychol. 1992;28:126–136. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials


Articles from Brain Structure & Function are provided here courtesy of Springer

RESOURCES