Skip to main content
BMC Neuroscience logoLink to BMC Neuroscience
. 2014 Apr 1;15:39. doi: 10.1186/1471-2202-15-39

Neural correlates of moral judgments in first- and third-person perspectives: implications for neuroethics and beyond

Mihai Avram 1,2,6, Kristina Hennig-Fast 3, Yan Bao 5,1,2,6, Ernst Pöppel 1,2,5,6, Maximilian Reiser 4, Janusch Blautzik 4, James Giordano 2,7,8,, Evgeny Gutyrchik 1,2,6
PMCID: PMC3991864  PMID: 24742205

Abstract

Background

There appears to be an inconsistency in experimental paradigms used in fMRI research on moral judgments. As stimuli, moral dilemmas or moral statements/ pictures that induce emotional reactions are usually employed; a main difference between these stimuli is the perspective of the participants reflecting first-person (moral dilemmas) or third-person perspective (moral reactions). The present study employed functional magnetic resonance imaging (fMRI) in order to investigate the neural correlates of moral judgments in either first- or third-person perspective.

Results

Our results indicate that different neural mechanisms appear to be involved in these perspectives. Although conjunction analysis revealed common activation in the anterior medial prefrontal cortex, third person-perspective elicited unique activations in hippocampus and visual cortex. The common activation can be explained by the role the anterior medial prefrontal cortex may play in integrating different information types and also by its involvement in theory of mind. Our results also indicate that the so-called "actor-observer bias" affects moral evaluation in the third-person perspective, possibly due to the involvement of the hippocampus. We suggest two possible ways in which the hippocampus may support the process of moral judgment: by the engagement of episodic memory and its role in understanding the behaviors and emotions of others.

Conclusion

We posit that these findings demonstrate that first or third person perspectives in moral cognition involve distinct neural processes, that are important to different aspects of moral judgments. These  results are important to a deepened understanding of neural correlates of moral cognition—the so-called “first tradition” of neuroethics, with the caveat that any results must be interpreted and employed with prudence, so as to heed neuroethics “second tradition” that sustains the pragmatic evaluation of outcomes, capabilities and limitations of neuroscientific techniques and technologies.

Keywords: fMRI, Moral judgment, Perspective, "actor-observer bias", Anterior medial prefrontal cortex, Precuneus, Hippocampus, Theory of mind, Neuroethics

Background

Studies of moral decision-making have been the focus of philosophy, psychology, and more recently, the brain sciences. Examination of the ways that humans (and perhaps other organisms) engage intent, memory, emotion, and reasoning processes relevant to their execution and constraint of conduct toward others, acquisition and use of various resources, survival, and flourishing have become the emphases of sub-disciplines of the cognitive neurosciences, such as neuroeconomics and more specifically, neuroethics. Developing from the older fields of moral philosophy and moral psychology, neuroethics obtains two primary orientations (or so-called “traditions”). The first can be somewhat colloquially described as “..the neuroscience of ethics” [1]. Rather, we offer that a more apt definition of this branch of neuroethics would be: studies of the putative neural substrates and mechanisms involved in proto-moral and moral cognition and behaviors [2-5]. The second “tradition” addresses the ethico-legal and social issues fostered by the use of neuroscience and neurotechnologies in research, medical practice, or public life.

In this latter regard, particular interest has centered upon the use of neuroimaging techniques and technologies to depict, and define neural bases of moral decision-making, if not “morality”, writ-large–as constituent to ongoing criticism of neuroimaging, in general [6]. Still, by recognizing and compensating inherent technical and conceptual limitations [7] iterative progress in neuroimaging technology and method have yielded improvement in outcomes, which sustain this approach as both valid and valuable to elucidating the relative activity of various neural networks in certain types of cognitive tasks and behaviors, including those involved in moral judgments and behaviors - with certain caveats noted and acknowledged [8,9].

Such studies have revealed the complexity of these types of decisions. In the main, focus has shifted from defining moral judgments as purely cognitive processes (i.e. - reason) to revealing more emotion-based processes, and recent results suggest the involvement of both processes in those decisions that are (both subjectively and objectively evaluated as being) morally sensitive and/or responsive [10-15]. What has also become clear is that moral decisions are not uniformly processed by a particular locus, region or network [16,17], but rather are more widely distributed in and across neural fields that are involved in memory, reward, reinforcement, and punishment, rationalization, interoception (e.g.- provocation of and response to various emotions, self-referentiality, etc.), and behavior. For example, Young and Dungan [18] suggest that such brain areas include the medial prefrontal cortex (MPFC) – involved in emotional processing; posterior cingulate cortex (PCC) and precuneus – both involved in self-referential processing, the temporo-parietal junction (TPJ) and/or somewhat larger fields of Brodmann’s area 39 – that are involved in aspects of social processing and/ or theory of mind (ToM).

As well, it is likely that different patterns of neural network activation may be involved in particular types of moral decisions, based upon the nature of the evocative stimuli, situations, and relative involvement of the subject. In this light, a methodological question has recently been raised regarding the viability of the rational and emotional/ intuitionist theories of moral cognition and judgments [19]. These research approaches to moral judgment use different experimental stimuli: “rationalist” protocols use moral dilemmas to study moral judgments, while “emotionalist” protocols employ emotionally-laden statements or pictures to assess what appear to be moral reactions. Is it possible that these approaches elicit distinct processes of moral cognition and lead to different results? Monin and colleagues [19] argue that the focus of reasoning in moral dilemmas is on the decision-making process - a conflict between two moral constructs and/or principles, whereas moral reactions reflect subjects’ emotional responses to particular stimuli and situations that have moral relevance. Of note is that moral dilemma protocols are typically presented in a first person perspective (1PP), while moral reaction protocols are characteristically presented in a third-person perspective (3PP). Thus, we question whether the perspective of the subject(s) toward the moral stimuli is sufficient to evoke differing effects, and elicit distinct patterns of neural network activity.

We opine that using stimuli presented in either 1- or 3PP may elucidate a number of potentially interactive variables that may shed new light on studies of neural mechanisms and processes of moral cognition. To wit, it has been shown that different patterns of neural activity were observed for stimuli presented in either 1- or 3-PP in non-moral visuospatial tasks[20]. During the 1-PP situation, neural activity was increased in the medial prefrontal cortex (MPFC), posterior cingulate cortex (PCC), and temporoparietal junction (TPJ) bilaterally, whereas in the 3-PP situation, neural activity was increased in the medial superior parietal and right premotor cortex.

Furthermore, differences have also been found in social non-moral tasks (which appear to reflect theory of mind, ToM), although these results are somewhat less clear. In a study on the influence of the person's perspective on ToM, 1- and 3-PP-type sentences elicited different patterns of neural activation: 1PP-based stimuli yielded greater activation in the caudate nucleus, while 3PP-based stimuli evoked increased neural activity in the dorsolateral prefrontal cortex (DLPFC). The authors related activity in the caudate nucleus to self-focal cognition, and DLPFC-activity to ToM. Other studies report stronger 3PP activation in the TPJ and dorsal MPFC [21-24] which are regarded as parts of the ToM network.

On the other hand, many of these studies have reported greater activation for the 1PP compared to 3PP in the MPFC and PCC/ precuneus. Ochsner and colleagues compared neural processes involved in inferences about one's own and others emotional states. Concomitant activation was demonstrated in the MPFC, left inferior PFC, PCC/ precuneus and STS/ TPJ [25]. This appeared to reflect recruitment of specific sub-regions in the MPFC, and additional activation in the medial temporal cortex for processing self-emotionality, while the lateral PFC and medial occipital activation appeared to be involved in processing emotional inferences of/about others. We posit that these results suggest that "self-judgments" seem to activate more medial networks, while judgments about others appear to engage more lateral networks. As well, components of both networks have some degree of overlap.

Social psychological studies have repeatedly shown that negative situations elicit a tendency to attribute one's own actions (1PP) to external causes, while attributing other people's (3PP) behaviors to internal causes, a phenomenon referred to as the "actor- observer bias" [26,27]. This may affect results in studies of moral decision-making, given that many such studies have employed negative situations as stimuli [28]. Nadelhoffer and Feltz [27] conducted a behavioral study of the actor-observer bias using a version of Philippa Foot’s [29] iconic "trolley problem" as the moral dilemma stimulus, viz.- a trolley is running out of control toward five people who are on the track and unaware of the looming danger. You have the opportunity to save these five people by throwing a switch and sending the trolley down a different track. However, if you do this, you will then kill one individual who is on the second track (for overview, see also Thomson [30] and for discussion of relevance to neural bases of moral decision-making, see Green [31]). The dilemma was presented either in a 1PP (i.e. - the subject was the actor, actively engaged in throwing the switch to divert the trolley), or in a 3PP (i.e. - the subject was a passive observer who could tell an actor to throw the switch). In the actor condition, 65% of the participants found the action (throwing the switch) to be permissible, whereas 90% of the participants in the observer condition found the action to be morally acceptable. These results imply different psychological processes involved in the two perspectives.

Thus, differential activation of distinct neural networks in response to 1PP- or 3PP-based stimuli is expected. Based on previous studies activation in the medial parts of the default mode network can be anticipated for the 1PP, and more lateral activation (e.g. DLPFC, TPJ) can be expected for the 3PP. However, since common activation for both perspectives has been found in several studies, and the default mode and ToM networks overlap in several regions, shared activation may also be expected. MPFC and PCC/ precuneus seem to be common denominators for the perspectives. Theoretically, the observer condition (3PP) of the "actor- observer bias" would tend to involve attribution of behaviors to internal causes, thus there is an attempt to understand the mind (i.e. - mental processes, in this case, the perceived “morality”) of the "actor". Indeed, ToM has been linked to moral judgments, and may be seen as important to moral evaluations of the actions of others [18].

As well, given that (a) most decisions, inclusive of potentially moral judgments involve some degree of Bayesian processing [32,33]; (b) such processing involves recollection of circumstance, effect and potential consequences in orientation to self, others and situations [2,5,34], and (c) learning and memory have been shown to play significant roles in these processes [35,36], it is likely that neural substrates of memory (e.g.- septo-hippocampal networks) would be involved [37,38]. Studies have fortified this speculation by demonstrating hippocampal activation in tasks involving perception of the emotions and actions of others [39,40]. Accordingly, we posit that hippocampal activation (for the 3PP-, as well as perhaps 1PP-situations) is to be expected. In sum, we hypothesize that the perspective of the subject (i.e.- as either actor (1PP), or observer (3PP)) will evoke differential activity in distinct neural networks that are putatively involved in the particular cognitive aspects of these orientations to moral judgment(s). To test this hypothesis we employed functional magnetic resonance imaging (fMRI) to compare moral judgments posed in 1- and 3PP-based scenarios.

Method

Participants

Sixteen (16) right-handed subjects (9 female, 7 male; mean age 28.25 years) with normal or corrected to normal vision participated in this study. Participants had no reported history of psychiatric or neurological disorder, and were not using psychoactive drugs at the time of the study. The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ethics Committee and Internal Review Board of the Human Science Center of the Ludwig-Maximilians University. Active, written informed consent for participation in the study was obtained from all participants, and subjects received financial compensation for their time.

Stimulus material

Sixty-nine (69) subjects evaluated 72 moral statements for valence and arousal in a pre-study. Half of the statements were presented in the 1PP ("I am a cruel person because I have aggressive thoughts towards my child"), and half were presented in the 3PP "A person who has aggressive thoughts toward his/ her child is cruel"). To assure valid comparisons, a five point Likert scale was used to rate the stimuli for valence, with scores ranging between −2 (unpleasant) and 2 (pleasant), and arousal, with scores ranging between −2 (agitating) and 2 (calming). Extreme values were excluded on an [−1, 1] interval in order to obviate the strongly emotion- laden stimuli, and to compare similar emotional reactions. Only 8 stimuli remained in each category after the pre-study. In order to ensure valid statistical comparisons of valence and arousal, two paired t-tests were used; there were no statistically significant differences between stimuli presented in 1PP narrative (M = −0.82, SD = 0.35) and 3PP narrative (M = −0.82, SD = 0.19), t (7) = 0.05, p > .05 with respect to valence. There were also no statistically significant differences between stimuli presented in 1PP narrative (M = −0.76, SD = 0.30) and 3PP narrative (M = −0.77, SD = 0.22), t (7) = 0.04, p > .05 with respect to arousal. Another paired t-test was used to control for stimulus sentence length. There were no statistically significant differences between stimuli presented in 1PP narrative (M = 8.38, SD = 3.20) and 3PP narrative (M = 10.25, SD = 2.71), t (7) = 1.34, p > .05.

Subjects had to rate the sentences as "right" or "wrong" by relying upon intuition (i.e.- described to them as “a gut-feeling”), and not necessarily their real life experience(s) (e.g. some participants may not have had children), so as to base their answers upon an "as-if” situation (e.g. If I were to have aggressive thoughts towards my child - and, indeed, if I had children - would I be a cruel person?).

Although the stimuli were controlled for length, there may have been differences in sentence construction. For example, in the 1PP narrative, "I am a cruel person because I have aggressive thoughts towards my child", it might seem that the 3PP narrative that would have been the best match would be: "John is a cruel person because he has aggressive thoughts towards his child". However, the actor-observer bias appears to be more prominent in cases where the actor is not known - e.g. a stranger [26]. Therefore, we choose a more abstract expression, namely "a person”. Another condition was also used, in which participants were asked to evaluate a non-moral statement based upon their perception of what they believed to be right or wrong (e.g. "There are people who are friendly"). An additional, "scrambled" condition was also used, in which participants had to push a response button when viewing a sentence composed of random letters. This condition was employed to test whether moral judgments activate a similar pattern when compared to scrambled words as in our previous study [14] and is not directly related to this study.

All stimuli were presented twice during the fMRI experiment.

Procedure

Functional magnetic resonance imaging (fMRI) was used in order to study the 1PP and 3PP types of judgments. A block design was used with 4 conditions (1PP, 3PP, non-moral, and scrambled) and 8 blocks per condition, each block comprising 2 stimuli, presented in white, on a black background. The order of stimuli and blocks was pseudo-randomized. Subjects viewed the stimuli via a mirror attached to the head-coil on a LCD screen behind the scanner. Stimuli were presented for 6000 ms (Presentation, Neurobehavioral Systems, USA), followed by 300 ms displaying a black screen, which in turn was followed by a 1000 ms black screen with a white question mark, in which subjects had to decide whether the statements could be considered right or wrong by pressing a button (Cedrus Lumina response box, Cambridge Research Systems Ltd.). After the two stimuli a black screen was presented for 6000 ms as a break between blocks. This method was used to ensure consistent parameters of cognitive processing in each subject for each presented stimuli. Given these protocols, reaction time analyses were not required.

The study was conducted with a 3T system (Philips ACHIEVA, Germany) at the University Hospital LMU Munich. For anatomical reference, a T1-weighted MPRAGE sequence was performed (TR = 7.4 ms, TE = 3.4 ms, FA = 8°, 301 sagittal slices, FOV = 240 × 256 mm, matrix = 227 × 227, inter-slice gap = 0.6 mm). For BOLD imaging, a T2*-weighted EPI sequence was used (TR = 3000 ms, TE = 35 ms, FA = 90°, 36 axial slices, slice thickness = 3.5 mm, inter-slice gap = 0 mm, ascending acquisition, FOV = 230 × 230 mm, matrix = 76 × 77, in-plane resolution = 3 × 3 mm). In total 229 functional volumes were acquired, 5 being discarded.

Data processing and analysis

The preprocessing and statistical analyses were performed using SPM8 (Wellcome Department of Cognitive Neurology, London, UK). Motion correction, realignment and spatial normalization were performed in the preprocessing analysis. Smoothing was executed using a Gaussian kernel of 8 mm FWHM. The four experimental conditions were modeled by a boxcar function convolved with a hemodynamic response function. In the first level, several single-tailed t-contrasts have been calculated for each subject, condition versus baseline. The individual contrast images were used for a random effect analysis in a second level. A conjunction analysis was performed to identify positive changes in BOLD signal intensity commonly seen in 1PP and 3PP presentations by using contrast images of each condition compared with the non-moral condition. Only activations are reported. Group activation contrasts (uncorrected < .005) were cluster-level corrected by family wise error (FWE) < .05 with a cluster-size threshold of 50 voxels.

Region of interest (ROI) analysis

Parameter estimates of signal intensity were extracted from regions of interest (ROIs) for each subject using MARSeille Boîte À Région d’Intérêt software (MarsBaR v0.42; [43] in the aMPFC, precuneus, TPJ, and hippocampus, with ROIs defined as spheres with 10mm radius centered at the peaks of the parametric activation. Anatomical description was accomplished by referring to the Automatic Anatomic Labeling (AAL) [41] atlas from the Wake Forest University (WFU) Pickatlas (Advanced NeuroScience Imaging Research Laboratory, Winston-Salem, North Carolina, USA). Repeated measures analyses of variance with mean beta values for each subject were done to determine whether neural activity within these regions differed between 1- and 3PP moral judgments and the non-moral condition. Gaussian distribution, homogeneity of variance and covariance and sphericity could be assumed (p > .05). Corrections for multiple comparisons were done by the Bonferroni procedure. Statistical analyses calculated with SPSS Statistics 16.0 (IBM, USA).

Results

Behavioral results

Subjects evaluated the moral statements to be either morally right, or morally wrong.

A chi-square-test revealed a statistically significant difference in yes/ no responses for the two moral conditions, χ2 (1) = 28.96, p < 0.01. The participants found 19% of the 1PP and 51% of the 3PP stimuli to be morally right.

fMRI results

1PP- and 3PP-based judgments were each compared to the non-moral condition (NM). 1PP-based judgments yielded greater activation than NM in the anterior medial prefrontal cortex (aMPFC - BA 10), posterior cingulate cortex (PCC - BA 23) extending in the precuneus (BA 7), and temporoparietal junction (TPJ - BA 39) (Table 1, Figure 1). 3PP-based judgments elicited greater activation in the aMPFC (BA 10), but also in the lingual gyrus (BA 17), middle occipital gyrus (BA 18) and hippocampus (Table 1, Figure 1).

Table 1.

Relative activation table: 1- and non 3PP moral judgments versus non-moral judgments

 
Left
Right
Brain region BA x y z t mm 3 BA x y z t mm 3
1PP > NM
 
 
 
 
 
 
 
 
 
 
 
 
aPFC
10
−6
56
22
5.64
3080
10
12
56
22
3.35
1593
Posterior cingulate cortex
23
−3
−52
31
3.94
378
 
 
 
 
 
 
Precuneus
7
−3
−58
40
4.98
1431
 
 
 
 
 
 
Temporoparietal junction
39
−42
−55
19
5.22
675
 
 
 
 
 
 
3PP > NM
 
 
 
 
 
 
 
 
 
 
 
 
aPFC
10
−15
50
37
4.87
459
10
3
59
28
4.90
1880
Lingual gyrus
18
−33
−85
4
6.69
3726
 
 
 
 
 
 
Middle occipital gyrus
−22
−25
−8
4.28
378
 
 
 
 
 
 
 
Hippocampus 17 6 −82 −5 7.91 4212            

Note. BA – aPFC - anterior prefrontal cortex, Brodmann area, x, y, z – MNI coordinates.

Figure 1.

Figure 1

Neurofunctional correlates of 1- and 3PP moral judgments. (A) 1PP moral judgments versus NM condition, (B) 3PP moral judgments versus NM condition. Anterior Medial Prefrontal Cortex (aMPFC), Temporoparietal Junction (TPJ).

In order to assess overlapping neural activity evoked by the two judgment modalities, a conjunction analysis was used. Common activation for the two judgment modalities (compared to control) was found only in the anterior medial prefrontal cortex x = 3, y = 59, z = 28 (BA 10; cluster size = 3078 mm3, t = 4.93.).Relative activations were generated only by the 3PP > 1PP contrast in: hippocampus bilaterally, and visual cortex - fusiform gyrus (BA 37), middle occipital gyrus (BA 19), and cuneus (BA 18) (Table 2, Figure 2). No activations above threshold were observed in the inversed contrast, 1PP > 3PP.

Table 2.

Relative activation table: 3PP versus 1PP moral judgments

 
Left
Right
Brain region BA x y z t mm 3 BA x y z t mm 3
Hippocampus
 
−36
−22
−14
4.08
1688
 
24
−28
−11
5.24
1836
Fusiform gyrus
37
−33
−52
−17
6.05
2889
19
24
−70
−14
4.43
1832
Middle occipital gyrus
19
−30
−85
16
8.25
2584
19
27
−85
19
4.69
2448
Cuneus 18 12 −88 19 4.90 536            

Note. BA – Brodmann area, x, y, z – MNI coordinates.

Figure 2.

Figure 2

Neurofunctional correlates of 3- vs 1PP moral judgments.

In order to ensure that the effects were related to the 1PP or 3PP moral conditions, and not due to the subtraction of the NM condition, the aMPFC, precuneus, TPJ, and hippocampus were selected for ROI analyses. Overall main effects were observed for all ROIs. For aMPFC (F(2, 30) = 13.17, p < .001, partial η2 = .468), differences were found between 1PP and NM condition (p < .002), and between 3PP and NM conditions (p < .006), but no difference was found between the two moral conditions (p = 1). For precuneus (F(2, 30) = 5.22, p < .011, partial η2 = .258) differences were found between 1PP and NM condition (p < .038), but none between 3PP and the NM condition (p = .057) or between the two moral conditions (p = .544). For TPJ (F(2, 30) = 7.29, p < .003, partial η2 = .327) differences were found between 1PP and NM condition (p < .003), and between 3PP and NM conditions (p < .032). No difference was found between the moral conditions (p = .262). For hippocampus (F(2, 30) = 12.46, p < .0001, partial η2 = .453) differences were observed between 1PP- and 3PP conditions (p < .0001), and between 3PP and NM condition (p < .005). However, no difference was found between NM and 1PP conditions (p = .316) (Figure 3).

Figure 3.

Figure 3

Region of interest analysis: anterior medial prefrontal cortex (aMPFC), precuneus, hippocampus, and temporoparietal junction (TPJ). Error bars denote standard error of the mean.

Discussion

The findings bring to light both common and distinct activations for moral judgments in 1PP and 3PP. A conjunction analysis revealed common activation in the aMPFC for both perspectives. When compared to the non-moral condition, 1PP moral judgments elicited activation in the aMPFC, PCC extending in the precuneus, and TPJ, whereas 3PP moral judgments elicited activation in the aMPFC, hippocampus and visual cortex.

The behavioral results, which revealed that 19% of the stimuli in 1PP- and 51% of the 3PP- stimuli were evaluated as right, seem to concur with Nadelhoffer and Feltz's study [27] showing involvement of the “actor-observer bias”. However, the paucity of imaging research on the “actor-observer bias“ makes it challenging to describe the way in which the neurofunctional correlates of the bias may be contributory to, or form moral judgments.

Even though first and third person perspectives (1PP, 3PP) elicited additional activity (except for aMPFC) in comparison with the non-moral condition (NM), these differences did not withstand the threshold-correction (except for hippocampus and visual cortex) in the direct (3PP- vs.1PP; 1PP vs. 3PP-based comparisons). The findings reveal both common and distinct activations for moral judgments in 1PP and 3PP. A conjunction analysis revealed common activation in the aMPFC for both perspectives. When compared to the non-moral condition, 1PP moral judgments elicited activation in the aMPFC, PCC extending in the precuneus, and TPJ, whereas 3PP moral judgments elicited activation in the aMPFC, hippocampus and visual cortex.

No significant statistical differences in signal activation strength were revealed by the ROI analyses between 1- and 3PP-based presentations in the MPFC, precuneus, and TPJ. The aMPFC has been shown to be involved in the explicit representation of both one’s own mental state, and also the mental states of others [43]. Furthermore, its activity has been consistently demonstrated in social cognition and ToM tasks [42]. Moreover, the aMPFC seems to function in coordination of external and internal stimuli [44].

Theoretically, 1PP presentation should elicit activation in those areas involved in assessing behavior in a given situation. When compared to the non-moral condition, signal activation was elicited in aMPFC, precuneus and right TPJ. Given that in 81% of the cases the subjects evaluated the moral stimuli as wrong; it seems that subjects may have tried to distance themselves from strong emotional stimuli. Koenigsberg et al. [45] found signal activation in the PCC/ precuneus, TPJ, and middle and superior temporal gyrus during emotional-distancing tasks. Since the aMPFC contributes to the integration of emotion in decision-making and planning [46], activation in this area suggests that the stimuli may have elicited emotional processing. An attempt to relate the stimuli to the self also seems probable, due to activation of the precuneus, which has been shown to be involved in types of self-processing (e.g. mental imagery strategies; [47]). However, these strategies also engage precuneus perspective-based cognition. Perspective-based cognition has also been shown to involve the TPJ [48]. That both the precuneus and TPJ are involved in may suggest that subjects attempted to change their perspective when responding to the moral stimuli.

In the 3PP-based condition, subjects appear to evaluate the behavior of others through the inner characteristics of the actor, in accordance with the “actor-observer bias”. Behavioral data suggest that the evaluating standards were less strict, with 51% of the stimuli being rated as morally right. When compared to the non-moral condition neural activation during presentation of moral conditions was found in aMPFC, hippocampus (bilaterally), and visual cortex. That there was almost equal activation in the aMPFC for both 1PP- and 3PP presentations of moral conditions (as based upon ROI analysis) suggests the involvement of similar processes in these decision events. Activation in the visual cortex may be explained by the visual salience of the emotional stimuli presented. [28,49,50]. Due to dense interconnections between the visual cortex and the amygdala, a modulating effect from the amygdala as noted by previous studies seems possible [51].

Recent neuroimaging studies have related hippocampal activity to ToM in understanding the emotions and behaviors of others [39], specifically as related to the facilitative role of the hippocampus, and its implication in inducing and sustaining emotional reactions. Hippocampal activation may also suggest both a possible role of memories and projection of self-knowledge while making emotional judgments regarding others [40] and the viability of declarative memory to integrate relevant information between different inputs about a given event [52]. However, it has been suggested that ToM may be independent of episodic memory [53]. In the present study, the stimuli were not related to typical daily experiences, but rather, represented extreme violence, blasphemy, and questionable sexual behavior.

Therefore, we argue that activation in the 3PP condition may be dependent upon semantic memory, in that factual or general information about the world may contribute to making sense of perceived deviant behavior. Hippocampal activity has also been shown during tasks of semantic memory [54], in retrieval of relevant memories [55] that allow past events to influence present decisions [56]. Taking this into consideration, the presentation of moral situations may trigger the recollection of memories of related situational and/or contextual information that relates to, and could influence present decision-making through a Bayesian mechanism of ecological observation, orientation and action [2,5,34]. While it might be possible that the observed hippocampal activation could, perhaps partially, be explained by different conditions relying more or less on short-term memory, we find it difficult to explain why the 3PP would rely more on short-term memory than the 1PP, since there were no statistical significant differences in assessments of sentence length, valence, or arousal.

Furthermore, an interaction between the ventromedial prefrontal cortex (vmPFC) and hippocampus has been suggested to mediate cognitive evaluations of the moral character of others [57]. Emotional salience is attributed to moral information by the involvement of the vmPFC, while hippocampal networks involved in memory retrieval enable necessary contextual information in order to make an appropriate character judgment. However, given that the vmPFC includes at least the ventral part of Brodmann’s area 10 (BA 10; [58]), and appears to serve a binding function between aMPFC and the amygdala [59], we suggest that BA 10 may have a functional role in integrating emotional information (via enhanced activation of the visual cortex), and recollective aspects of the decision-process; (possibly through hippocampal connections) that are involved in, and/or subserve moral cognition and judgments.

Thus, we posit that the vmPFC plays a role in emotional salience, while the aMPFC contributes to synthesizing the “moral” information, by integrating emotional and recollective information, thereby enabling appropriate strategies in moral decision-making. To summarize, we claim that the involvement of the hippocampus for the 3PP moral judgment can be explained through the results of recent studies that elucidated its role in understanding emotions and behaviors of others, while somewhat more “classical” hippocampal activity (i.e.- memory) plays a role in the recollection of stored related retrograde situational or contextual information. We consider the role of the hippocampus in 3PP moral judgments of crucial importance due to the psychological implications of these functional roles.

There is also a temporal aspect that may be involved, which would support the “actor-observer bias”. If 1PP presentations engage evaluative cognition, then such processing is temporally related to the present [60,61]. The 3PP situation, however, relies on more abstract evaluations, which tend to be more time independent, in which inner characteristics of others may come into play. Moreover, if subjects distance themselves from the stimuli used in 1PP presentations, the time needed to evaluate these stimuli would be shorter than that needed to evaluate the stimuli in the 3PP condition, where memory processing would represent an important function in stimuli assessment.

An important aspect of the present study is the use of novel stimuli. Since moral dilemmas have already been used to study the "actor-observer bias" [27] a different approach, i.e. using moral reactions, may be helpful in extrapolating the findings. For this reason, control of emotional valence and duration of stimuli has been ensured. Such parameters, however, decrease the number of stimuli that were used. This may be problematic; however, due to the novelty of the approach used, a possible limitation in generalization seems suitable in order to gain greater experimental control over the stimuli.

Despite these limitations, the present findings suggest that different neural networks may be involved in, and subserve the perspective one has towards moral situations. A similar case was found for agency in moral judgments, for which different associated emotions were found to rely upon both distinct and overlapping neural substrates [62]. A psychological component, which could explain the neural differences found for moral perspective taking, is the actor-observer bias. Thus, care must be taken when interpreting neuroimaging studies of the neural bases of morality, since the perspective of the participants towards the moral stimuli may indeed elicit distinct neural activation.

In summary, moral stimuli presented in either 1- or 3PP elicit both distinct (e.g. hippocampus, and visual cortex for 3PP) and common patterns of neural activation (e.g. in the self- or ToM networks). These results suggest that differences may be related to the “actor-observer bias”. In the 1PP presentation the stimuli were evaluated with regard to the situation. Since the participants could not control the situation (although it elicited a strong emotional response), we posit that subjects may have attempted to distance themselves from the stimuli by engaging in perspective shifting. The 3PP moral judgments seem to have been evaluated by considering the inner characteristics of the “actors”, through recollection(s) of relevant information and also by engaging in ToM processes.

The overlap in the self- and ToM networks suggests that self-processing may be a basis through which to experience complex emotions about others' mental state [39]. These findings do not imply identical psychological processes for these different perspectives, and do not contradict the suggested involvement of the “actor-observer bias”. We believe that the most important implication of this study is related to distinct mechanisms and processes of moral cognition. To date, research has posed that networks of the so-called “moral brain” are homogenously activated, independent of the eliciting stimuli. This also implies that similar psychological processes subserve moral cognition and/or reasoning, irrespective of perception of, or orientation to the situation [15]. The present results, however, contrast this view, and suggest that different types of stimuli may indeed engage distinct types of neural activity and psychological processing, and that both reflect orientation to the situation, which may be influenced by a host of factors affecting cognitive biasing, inclusive of cultural differences and a variety of social effects.

While it has been offered that moral and ethical judgments and actions are “other-based” (see, for example, MacMurray [63]), it is important to note that any and all decisions - inclusive of moral judgments (affecting others) - emanate from, and in many ways are reciprocal to, and reflective of the self [2,3,5,64-66]. In this light, potentially moral situations are perceived differently depending upon one’s orientation to, and relative involvement in the situation and circumstance, and its effect upon prior experience, past and present reinforcing and rewarding influences, and predicted outcomes and their impact upon self and others [67-69].

The results presented here suggest that while there appears to be something of a core neural network that is involved in the types of moral decisions rendered in this study, the spatial and temporal engagement of elements of this network are peculiar to distinct types and aspects of situation and circumstances. There are several limitations of this study. First, the number of stimuli remaining after the pilot study was rather small. Therefore, we suggest that future studies employ a larger number of stimuli. This would also enable non-repetition of stimuli, thereby avoiding possible learning effects that have been shown to decrease BOLD signal – e.g. in visual cortex, PFC etc. [69,70]. Second, it remains somewhat uncertain to what extent participants attributed external causes to the 1PP, and internal causes to the 3PP, since the subjects were not required to describe the way in which they evaluated the stimuli. Future studies could employ a post-scanning interview during which subjects are asked to explain their decision-making processes.

Conclusion

In conclusion, we opine that the present study suggests differential patterns and mechanisms of 1PP and 3PP moral judgments. Such findings have implications for consideration of how moral decisions are made and morally-relevant acts are tendered (e.g.- “Good Samaritan” acts, “by-stander effects”, etc.), and prompt further inquiry to how patterns of neural activity may affect types and extent of behaviors in morally-relevant situations, and if and how such patterns of activity are stable, modifiable, and/or learned. Yet, we also advocate prudence in interpretation of these and related findings [2-4,7-9], as the limitations of fMRI, like any neurotechnology, must be appreciated (see van Meter [71] for overview).

This encourages engagement of neuroethics’ second tradition, namely, an analysis of the ways that neuroscience and neurotechnology are, can, and should be employed to gain understanding of cognitions, emotions and behaviors, and how such information is used (in medicine, law and the public sphere). Indeed, while findings such as those presented in this study may be provocative, care must be taken in extrapolating such information to real-world circumstances, so as to avoid over- or under-estimating the role of neurobiology in psychological and social activity, and/or the capabilities of neuroscience and neurotechnology to address and answer perdurable and pressing questions about the “nature” of morality, and other dimensions of cognition, emotion and behavior.

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

All authors contributed to study concept. MA was responsible for data collection, data analysis and interpretation, and manuscript preparation. EG and KF were responsible for data analysis and interpretation and critical review of the manuscript, MR and JB were responsible for data collection and preliminary data evaluation. YB, EP and JG made substantial contributions to interpretation of data, have been involved in developing and revising the manuscript for important intellectual content, and have given final approval of the version to be published.

Contributor Information

Mihai Avram, Email: mihai.avram@med.lmu.de.

Kristina Hennig-Fast, Email: Kristina.Fast@med.uni-muenchen.de.

Yan Bao, Email: baoyan@pku.edu.cn.

Ernst Pöppel, Email: ernst.poeppel@med.uni-muenchen.de.

Maximilian Reiser, Email: Maximilian.Reiser@med.uni-muenchen.de.

Janusch Blautzik, Email: Janusch.Blautzik@med.uni-muenchen.de.

James Giordano, Email: giordano@grp.hwz.uni-muenchen.de.

Evgeny Gutyrchik, Email: evgeny.gutyrchik@med.uni-muenchen.de.

Acknowledgments

This study was supported by a research scholarship of the Bayerische Forschungsstiftung for M.A., and the Clark Foundation Award for JG. The authors thank Liana Buniak for assistance in preparation of this manuscript.

References

  1. Roskies A. Neuroethics for the new millenium. Neuron. 2002;15(1):21–23. doi: 10.1016/S0896-6273(02)00763-8. [DOI] [PubMed] [Google Scholar]
  2. Giordano J. Neuroethics: interacting 'traditions' as a viable meta-ethics. AJOB Neurosci. 2011;15(2):17–19. doi: 10.1080/21507740.2011.559922. [DOI] [Google Scholar]
  3. Giordano J. Neuroethics: traditions, tasks and values. Hum Prospect. 2011;15(1):2–8. [Google Scholar]
  4. Giordano J, Benedikter R. An early - and necessary - flight of the Owl of Minerva: Neuroscience, neurotechnology, human socio-cultural boundaries, and the importance of neuroethics. J Evol Technol. 2012;15(1):14–25. [Google Scholar]
  5. Giordano J, Benedikter R. Topics in Neurobiology of Aggression: Implications for Deterrence. Strategic Multilayer Assessment Group. Washington, DC: SMA Publications; 2013. Toward a systems’ continuum: on the use of neuroscience and neurotechnology to assess and affect aggression, cognition and behavior; pp. 69–85. [Google Scholar]
  6. Uttal WR. The New Phrenology: The limits of localizing cognitive processes in the brain. Cambridge, MA: MIT Press; 2001. [Google Scholar]
  7. Bao Y, Pöppel E. Anthropological universals and cultural specifics: Conceptual and methodological challenges in cultural neuroscience. Neurosci Biobehav Rev. 2012;15:2143–2146. doi: 10.1016/j.neubiorev.2012.06.008. [DOI] [PubMed] [Google Scholar]
  8. Giordano J. In: Technological Innovation in Sensing and Detecting Chemical, Biological, Radiological, Nuclear Threats and Ecological Terrorism. (NATO Science for Peace and Security Series) Vaseashta A, Braman E, editor. NY: Springer; 2012. Integrative convergence in neuroscience: trajectories, problems and the need for a progressive neurobioethics. [Google Scholar]
  9. Giordano J. Neurotechnology: Premises, Potential and Problems. Edited Giordano J. Boca Raton: CRC Press; 2012. Neurotechnology as deimurgical force: avoiding Icarus folly; pp. 1–14. [Google Scholar]
  10. Greene JD, Sommerville RB, Nystrom LE, Darley JM, Cohen JD. An fMRI investigation of emotional engagement in moral judgment. Science. 2001;15:2105–2108. doi: 10.1126/science.1062872. [DOI] [PubMed] [Google Scholar]
  11. Greene J, Haidt J. How (and where) does moral judgment work? Trends Cogn Sci. 2002;15:517–523. doi: 10.1016/S1364-6613(02)02011-9. [DOI] [PubMed] [Google Scholar]
  12. Zaidel DW, Nadal M. Brain Intersections of Aesthetics and Morals: Perspectives from Biology, Neuroscience, and Evolution. Perspect Biol Med. 2011;15:367–380. doi: 10.1353/pbm.2011.0032. [DOI] [PubMed] [Google Scholar]
  13. Tsukiura T, Cabeza R. Shared brain activity for aesthetic and moral judgments: implications for the Beauty-is-Good stereotype. Soc Cogn Affect Neurosci. 2010. doi:10.1093/scan/nsy025. [DOI] [PMC free article] [PubMed]
  14. Avram M, Gutyrchik E, Bao Y, Pöppel E, Reiser M, Blautzik J. Neurofunctional correlates of aesthetic and moral judgments. Neurosci Lett. 2013;15:128–132. doi: 10.1016/j.neulet.2012.11.053. doi:10.1016/j.neulet.2012.11.053. [DOI] [PubMed] [Google Scholar]
  15. Moll J, Zahn R, de Oliveira-Souza R, Krueger F, Grafman J. Opinion: the neural basis of human moral cognition. Nature Rev Neurosci. 2005;15:799–809. doi: 10.1038/nrn1768. [DOI] [PubMed] [Google Scholar]
  16. Casebeer WD. The Neural mechanisms of moral cognition: a multiple- aspect approach to moral judgment and decision-making. BiolPhilos. 2003;15:169–194. [Google Scholar]
  17. Verplaetse J, DeSchrijver J, Vanneste S, Braeckman J. (Eds.). The Moral Brain. Heidelberg: Springer Verlag; 2009. [Google Scholar]
  18. Young L, Dungan J. Where in the brain is morality? Everywhere and maybe nowhere. Soc Neurosci. 2012;15(1):1. doi: 10.1080/17470919.2011.569146. [DOI] [PubMed] [Google Scholar]
  19. Monin B, Pizarro D, Beer JS. Deciding versus reacting: conceptions of moral judgment and the reason- affect debate. Rev Gen Psychol. 2007;15:99–111. [Google Scholar]
  20. Vogeley K, Fink G. Neural correlates of the first-person perspective. Trends Cogn Sci. 2003;15:38–42. doi: 10.1016/S1364-6613(02)00003-7. [DOI] [PubMed] [Google Scholar]
  21. Ruby P, Decety J. Effect of subjective perspective taking during simulation of action: a PET investigation of agency. Nat Neurosci. 2001;15:546–550. doi: 10.1038/87510. [DOI] [PubMed] [Google Scholar]
  22. Ruby P, Decety J. What you believe versus what you think they believe: a neuroimaging study of conceptual perspective-taking. Eur J Neurosci. 2003;15:2475–2480. doi: 10.1046/j.1460-9568.2003.02673.x. doi:10.1046/J.1460-9568.2003.02673.X. [DOI] [PubMed] [Google Scholar]
  23. Ruby P, Decety J. How would you feel versus how do you think she would feel? A neuroimaging study of perspective-taking with social emotions. J Cogn Neurosci. 2004;15:988–999. doi: 10.1162/0898929041502661. doi:10.1162/0898929041502661. [DOI] [PubMed] [Google Scholar]
  24. Ramsey R, Hansen P, Apperly I, Samson D. Seeing it my way or your way: frontoparietal brain areas sustain viewpoint-independent perspective selection processes. J Cogn Neurosci. 2013;15(5):670–684. doi: 10.1162/jocn_a_00345. [DOI] [PubMed] [Google Scholar]
  25. Ochsner KN, Knierim K, Ludlow DH, Hanelin J, Ramachandran T, Glover G. Reflecting upon feelings: An fMRI study of neural systems supporting the attribution of emotion to self and other. J Cog Neurosci. 2004;15:1746–1772. doi: 10.1162/0898929042947829. [DOI] [PubMed] [Google Scholar]
  26. Jones EE, Nisbett RE. The actor and the observer: Divergent perceptions of the causes of behavior. New York: General Learning Press; 1971. [Google Scholar]
  27. Nadelhoffer T, Feltz A. The Actor–Observer Bias and moral intuitions: Adding fuel to Sinnott-Armstrong's fire. Neuroethics. 2008;15(2):133–144. doi: 10.1007/s12152-008-9015-7. [DOI] [Google Scholar]
  28. Takahashi H, Yahata N, Koeda M, Matsuda T, Asai K, Okubo Y. Brain activation associated with evaluative processes of guilt and embarrassment: an fMRI study. Neuroimage. 2004;15:967–974. doi: 10.1016/j.neuroimage.2004.07.054. [DOI] [PubMed] [Google Scholar]
  29. Foot P. The Problem of abortion and the Doctrine of the Double Effect in virtues and vices. Oxford Rev. 1967;15:5–15. [Google Scholar]
  30. Thomson JJ. Killing, letting die, and the trolley problem. The Monist. 1976;15:204–217. doi: 10.5840/monist197659224. [DOI] [PubMed] [Google Scholar]
  31. Greene JD. In: Moral Psychology, Vol. 3: The Neuroscience of Morality. Sinnott-Armstrong W, editor. Cambridge, MA: MIT Press; 2008. The secret joke of Kant’s soul. [Google Scholar]
  32. Lewicki MS, Sejnowksi TJ. In: Advances in Neural Information Processing Systems 9. Mozer M, Jordan M, Petsche T, editor. Cambridge, MA: MIT Press; 1997. Bayesian unsupervised learning of higher order structure; pp. 529–535. [Google Scholar]
  33. Rao RN. Bayesian computation in recurrent neural circuits. Neural Comput. 2004;15(1):1–38. doi: 10.1162/08997660460733976. [DOI] [PubMed] [Google Scholar]
  34. Boyd JR. Destruction and creation. Lecture presented to the U.S. Army Command and General Staff College. 3 September, 1976.
  35. Grossberg S. In: Neurobiology of Attention. Itti L, Rees G, Tsotsos JK, editor. San Diego: Elsevier; 2005. Linking attention to learning, expectation, competition, and consciousness; pp. 652–662. [Google Scholar]
  36. Shadlen M, Gold JI. In: The Cognitive Neurosciences. 3. Gazzaniga MS, editor. Cambridge: MIT Press; 2004. The neurophysiology of decision-making as a window on cognition; pp. 1229–1441. [Google Scholar]
  37. Fumagalli M, Prior A. Functional and clinical neuroanatomy of morality. Brain. 2012;15(7):2006–20021. doi: 10.1093/brain/awr334. [DOI] [PubMed] [Google Scholar]
  38. Raine A, Yang Y. Neural foundations to moral reasoning and anti-social behavior. Soc Cogn Affect Neurosci. 2006;15:203–213. doi: 10.1093/scan/nsl033. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Immordino-Yang MH, Singh V. Hippocampal contributions to the processing of social emotions. Hum Brain Mapp. 2011. doi:10.1002/hbm.21485. [DOI] [PMC free article] [PubMed]
  40. Perry D, Hendler T, Shamay-Tsoory SG. Projecting memories: the role of the hippocampus in emotional mentalizing. Neuroimage. 2011;15:1669–1676. doi: 10.1016/j.neuroimage.2010.08.057. [DOI] [PubMed] [Google Scholar]
  41. Tzourio-Mazoyer N, Landeau B, Papathanassiou D, Crivello F, Etard O, Delcroix N. Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain. Neuroimage. 2002;15(1):273–289. doi: 10.1006/nimg.2001.0978. [DOI] [PubMed] [Google Scholar]
  42. Amodio DM, Frith CD. Meeting of the minds: The medial frontal cortex and social cognition. Nature Re Neurosci. 2006;15:268–277. doi: 10.1038/nrn1884. [DOI] [PubMed] [Google Scholar]
  43. Brett M, Johnsrude I, Owen A. The problem of functional localization in the human brain. Nature Rev Neurosci. 2002;15:243–249. doi: 10.1038/nrn756. [DOI] [PubMed] [Google Scholar]
  44. Gallagher HL, Frith CD. Functional imaging of 'theory of mind'. Trends Cog Sci. 2003;15(2):77–83. doi: 10.1016/S1364-6613(02)00025-6. [DOI] [PubMed] [Google Scholar]
  45. Burgess PW, Simons JS, Dumontheil I, Gilbert SJ. In: Measuring the Mind: Speed, Control, and Age. Duncan J, McLeod P, Phillips L, editor. Oxford: Oxford University Press; 2005. The gateway hypothesis of rostral prefrontal cortex (area 10) function; pp. 215–246. [Google Scholar]
  46. Koenigsberg HW, Fan J, Ochsner KN, Liu X, Guise K, Pizzarello S, Dorantes C, Tecuta L, Guerreri S, Goodman M, New A, Flory J, Siever LJ. Neural correlates of using distancing to regulate emotional responses to social situations. Neuropsychologia. 2010;15(6):1813–1822. doi: 10.1016/j.neuropsychologia.2010.03.002. doi: 10.1016/j.neuropsychologia.2010.03.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  47. Ramnani N, Owen AM. Anterior prefrontal cortex: insights into function from anatomy and neuroimaging. Nature Revs Neurosci. 2004;15(3):184–194. doi: 10.1038/nrn1343. [DOI] [PubMed] [Google Scholar]
  48. Cavanna AE, Trimble MR. The precuneus: a review of its functional anatomy and behavioural correlates. Brain. 2006;15:564–583. doi: 10.1093/brain/awl004. [DOI] [PubMed] [Google Scholar]
  49. Decety J, Sommerville JA. Shared representations between self and others: a social cognitive neuroscience view. Trends Cog Sci. 2003;15:527–533. doi: 10.1016/j.tics.2003.10.004. [DOI] [PubMed] [Google Scholar]
  50. Phan K, Wager T, Taylor S, Liberzon I. Functional neuroanatomy of emotion: a meta-analysis of emotion activation studies in PET and fMRI. Neuroimage. 2002;15:331–348. doi: 10.1006/nimg.2002.1087. [DOI] [PubMed] [Google Scholar]
  51. Sparing R, Mottaghy FM, Ganis G, Thompson WL, Töpper R, Kosslyn SM, Pascual-Leone A. Visual cortex excitability increases during visual mental imagery – a TMS study in healthy human subjects. Brain Res. 2002;15:92–97. doi: 10.1016/S0006-8993(02)02478-2. [DOI] [PubMed] [Google Scholar]
  52. Morris JS, Ohman A, Dolan RJ. Conscious and unconscious emotional learning in the human amygdala. Nature. 1998;15:467–470. doi: 10.1038/30976. [DOI] [PubMed] [Google Scholar]
  53. Croft KE, Duff MC, Anderson SW, Adolphs R, Tranel D. Bilateral amygdala damage is associated with reduced updating of character judgments. Chicago, IL: Paper presented at the annual conference of the Society for Neuroscience; 2009. [Google Scholar]
  54. Rosenbaum RS, Stuss DT, Levine B, Tulving E. Theory of mind is independent of episodic memory. Science. 2007;15:1257. doi: 10.1126/science.1148763. [DOI] [PubMed] [Google Scholar]
  55. Manns JR, Hopkins RO, Reed JM, Kitchener EG, Squire LR. Recognition memory and the human hippocampus. Neuron. 2003;15:171–180. doi: 10.1016/S0896-6273(02)01147-9. [DOI] [PubMed] [Google Scholar]
  56. Strange BA, Fletcher PC, Henson RNA, Friston KJ, Dolan RJ. Segregating the functions of human hippocampus. PNAS USA. 1999;15:4034–4039. doi: 10.1073/pnas.96.7.4034. [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Wood ER, Dudchenko PA, Robitsek RJ, Eichenbaum H. Hippocampal neurons encode information about different types of memory episodes occurring in the same location. Neuron. 2000;15:623–633. doi: 10.1016/S0896-6273(00)00071-4. [DOI] [PubMed] [Google Scholar]
  58. Croft KE, Duff MC, Kovach CK, Anderson SW, Adolphs R, Tranel D. Detestable or marvelous? Neuroanatomical correlates of character judgments. Neuropsychologia. 2010;15(6):1789–1801. doi: 10.1016/j.neuropsychologia.2010.03.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Quirk GJ, Russo GK, Barron JL, Lebron K. The role of ventromedial prefrontal cortex in the recovery of extinguished fear. J Neurosci. 2000;15(16):6225–6231. doi: 10.1523/JNEUROSCI.20-16-06225.2000. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Urry HL, van Reekum CM, Johnstone T, Kalin NH, Thurow ME, Schaefer HS, Jackson CA, Frye CJ, Greischar LL, Alexander AL, Davidson RJ. Amygdala and ventromedial prefrontal cortex are inversely coupled during regulation of negative affect and predict the diurnal pattern of cortisol secretion among older adults. J Neurosci. 2006;15:4415–4425. doi: 10.1523/JNEUROSCI.3215-05.2006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Pöppel E. Pre-semantically defined temporal windows for cognitive processing. Phil Trans Royal Soc B. 1887–1896;15:364. doi: 10.1098/rstb.2009.0015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Pöppel E, Bao Y. In: Subjective Time. Lloyd D, Arstila V, editor. Cambridge, MA: MIT Press; Temporal windows as bridge from objective to subjective time. in Press. [Google Scholar]
  63. Moll J, de Oliveira-Souza R, Garrido GJ, Bramati IE, Caparelli-Daquer EM, Paiva ML, Zahn R, Grafman J. The self as a moral agent: linking the neural bases of social agency and moral sensitivity. Social Neurosci. 2007;15(3-4):336–352. doi: 10.1080/17470910701392024. doi:10.1080/17470910701392024. [DOI] [PubMed] [Google Scholar]
  64. MacMurray J. Persons in Relation. London: Faber and Faber; 1961. [Google Scholar]
  65. Pöppel E. In: Personality from Biological, Cognitive, and Social Perspectives. Maruszewski T, Malgorzata Fajkowska M, Eysenck MM, editor. New York: Eliot Werner Publications, Clinton Corners; 2010. Perceptual identity and personal self: neurobiological reflections; pp. 75–82. [Google Scholar]
  66. Shaver R. Rational Egoism: A Selective and Critical History. Cambridge: Cambridge University Press; 1998. [Google Scholar]
  67. Trivers RL. The evolution of reciprocal altruism. Quarterly Review of Biology. 1971;15:35–57. doi: 10.1086/406755. [DOI] [Google Scholar]
  68. Schino G, Aureli F. A few misunderstandings about reciprocal altruism. Commun Integr Biol. 2010;15(6):561–563. doi: 10.4161/cib.3.6.12977. [DOI] [PMC free article] [PubMed] [Google Scholar]
  69. Stephens C. Modeling reciprocal altruism. Br J Phil Sci. 1996;15(4):533–551. doi: 10.1093/bjps/47.4.533. [DOI] [Google Scholar]
  70. Segaert K, Weber K, De Lange F, Petersson KM, Hagoort P. The suppression of repetition enhancement: A review of fMRI studies. Neuropsychologia. 2013;15:59–66. doi: 10.1016/j.neuropsychologia.2012.11.006. [DOI] [PubMed] [Google Scholar]
  71. VanMeter J. In: Scientific and Philosophical Perspectives in Neuroethics. Giordano J, Gordijn B, editor. Cambridge: Cambridge University Press; 2010. Neuroimaging; pp. 230–239. [Google Scholar]

Articles from BMC Neuroscience are provided here courtesy of BMC

RESOURCES