Skip to main content
Frontiers in Human Neuroscience logoLink to Frontiers in Human Neuroscience
. 2018 Feb 13;12:52. doi: 10.3389/fnhum.2018.00052

Perceived Gaze Direction Modulates Neural Processing of Prosocial Decision Making

Delin Sun 1,2,3,4, Robin Shao 1,2, Zhaoxin Wang 5, Tatia M C Lee 1,2,6,7,*
PMCID: PMC5816754  PMID: 29487516

Abstract

Gaze direction is a common social cue implying potential interpersonal interaction. However, little is known about the neural processing of social decision making influenced by perceived gaze direction. Here, we employed functional magnetic resonance imaging (fMRI) method to investigate 27 females when they were engaging in an economic exchange game task during which photos of direct or averted eye gaze were shown. We found that, when averted but not direct gaze was presented, prosocial vs. selfish choices were associated with stronger activations in the right superior temporal gyrus (STG) as well as larger functional couplings between right STG and the posterior cingulate cortex (PCC). Moreover, stronger activations in right STG was associated with quicker actions for making prosocial choice accompanied with averted gaze. The findings suggest that, when the cue implying social contact is absent, the processing of understanding others’ intention and the relationship between self and others is more involved for making prosocial than selfish decisions. These findings could advance our understanding of the roles of subtle cues in influencing prosocial decision making, as well as shedding lights on deficient social cue processing and functioning among individuals with autism spectrum disorder (ASD).

Keywords: social decision making, eye gaze, fMRI, superior temporal gyrus, posterior cingulate cortex

Introduction

Prosocial behaviors are the cornerstone of a harmonic society (Keltner et al., 2014), and are associated with complex considerations of benefits and intentions of both self and others (Rilling and Sanfey, 2011). Recent work showed that prosocial actions can be promoted in the presence of eyes or eye-like stimuli (Haley and Fessler, 2005; Nettle et al., 2013), suggesting the role of the eyes in effectively biasing social decision making. Gaze is one of the most important conduits of information delivered by eyes, and plays significant roles in social interaction (Itier and Batty, 2009; Carlin and Calder, 2013). However, little is known about the neural processing of social decision making modulated by perceived gaze direction.

Compared to selfish actions, prosocial decisions are more associated with attributions of the intentions and desires of counterparts, the so-called Theory of Mind (ToM; Baron-Cohen et al., 1985). According to this theory, understanding others’ needs and thoughts may promote the engagement in prosocial actions (Dunfield, 2014), while benefiting others may in turn contribute to the development of better ToM (Weller and Lagattuta, 2014). Consistent with this, a recent meta-analysis showed that ToM and prosocial behaviors are positively related in children (Imuta et al., 2016). Evidence also suggests the ability of paying attention to or understanding the information delivered by gaze appears in very early stage of development (Farroni et al., 2002). Someone else’s gaze informs us about the object or place he/she is looking at, and in turn how important or interesting such information is to him/her (Baron-Cohen, 1995; Shimojo et al., 2003; Frischen et al., 2007). Gaze direction has thus been proposed as a privileged stimulus for the attribution of mental state of others (Baron-Cohen, 1995). Direct gaze, usually accompanied with eye contact, indicates that someone is paying attention to us, while averted gaze implies the person is interested in people or objects other than us. On the other hand, Adams and Kleck (2003, 2005) found that direct gaze facilitates the processing of facial expressions indicating approach-oriented emotions (e.g., anger and joy), whereas averted gaze facilitates the processing of expressions implying avoidance-oriented emotions (e.g., fear and sadness). They thus proposed the “Shared Signal Hypothesis”, which postulates that the perception of a specific emotion will be enhanced when gaze direction matches the underlying behavioral intent communicated by that emotion expression. Taken together, direct gaze implies potential social contact and enhances the perception of approach-oriented emotions while averted gaze does not, which could in turn influence how readily we process others’ intentions and our subsequent social decision-making processes. However, no research has directly tested the effect of eye gaze direction on social decision making.

Brain regions related with ToM include the superior temporal gyrus (STG), posterior cingulate cortex (PCC), medial prefrontal cortex (mPFC), temporal-parietal junction (TPJ) and amygdala (Northoff and Bermpohl, 2004; Shaw et al., 2004; Schurz et al., 2014). These areas are widely implicated in social cognitive and decision-making processes (Moll et al., 2005, 2007; Rilling and Sanfey, 2011; Bastin et al., 2016). For example, Rilling et al. (2004) detected stronger activations in mPFC, PCC and TPJ when participants inferred the intent of human counterparts through their feedbacks during economic game tasks. Also, Moll and de Oliveira-Souza (2007) employed written statements describing action scenarios and found that prosocial emotions including guilt, embarrassment and compassion activated mPFC and bilateral STG. In a recent study, Morey et al. (2012) found that brief hypothetical scenarios in which the participants’ actions lead to harmful consequences to others vs. to self were associated with more intense feelings of guilt as well as stronger activations in mPFC, right STG and PCC. Further, amygdala was reported to signal the interaction between gaze direction and perceived facial expression (N’Diaye et al., 2009; Cristinzio et al., 2010; Sato et al., 2010; Ziaei et al., 2016, 2017), suggesting its role in the appraisal of self-relevance. These findings disclosed a positive relationship between prosocial actions/emotions and activations in the ToM brain network, and supported the idea that prosocial behaviors are related with more considerations of others’ thoughts and relationships between self and others.

Here, we were particularly interested in the STG. Existing imaging evidence indicates the right STG as being sensitive to gaze direction (Nummenmaa and Calder, 2009), suggesting its role as an eye direction detector (Baron-Cohen, 1995). Consistent with this, a patient with damage to the right STG showed difficulties in gaze discrimination and gaze-cued attention orientation (Akiyama et al., 2006). Stronger brain activations in right STG were also observed for direct than averted gaze (Calder et al., 2002; Pelphrey et al., 2004; but see Hardee et al., 2008). Previous studies have also found increased activations in the audience’s right STG when an actor tried to deceive the audience about the weight of a box he was lifting (Grezes et al., 2004a), when an actor had a false belief about the weight of the box (Grezes et al., 2004b), and when an actor chose an object he did not like or rejected an object he preferred (Wyk et al., 2009), suggesting that the STG is involved in detecting other’s intentions. Moreover, the STG may collaborate with other ToM areas to process gaze and social information, as it has both anatomical (Parvizi et al., 2006) and functional (Uddin et al., 2009) connections with the PCC. Stronger activations in bilateral STG, mPFC and PCC were reported in healthy participants when viewing direct than averted gaze (von dem Hagen et al., 2014), and greater mPFC-right STG functional connectivity was reported during social emotion, such as embarrassment and guilt, than basic emotion (Burnett and Blakemore, 2009).

In this study, we employed the functional magnetic resonance imaging (fMRI) method to investigate brain responses when participants were making either prosocial or selfish choices against anonymous counterparts in a novel economic exchanging game task (Sun et al., 2016). The effect of eye gaze direction was investigated by showing participants photos of counterparts’ eyes with either direct or averted gaze. In line with existing findings, prosocial vs. selfish choices were proposed to be associated with more considerations of others’ mental states as well as the relationship between self and others. In this context, prosocial choice was defined as behaviors that prioritize the benefit of other social, but not nonsocial (e.g., a robot), agents. We tested two types of relations between perceived gaze direction, social behaviors and the associated neural patterns. If prosocial choices were triggered by direct gaze that cues others’ intentions of social contact, we hypothesized to find stronger activations in right STG and larger functional connectivity between right STG and other ToM areas during making prosocial vs. selfish choices when perceiving direct than averted gaze. On the other hand, averted gaze signals the lack of intentions for social contact, and the viewers may need to make more cognitive efforts to infer the counterpart’s thoughts and to consider the relationship between self and others during making prosocial decisions. We then alternatively hypothesized to find stronger activations in right STG and larger functional connectivity between right STG and other ToM areas during making prosocial vs. selfish choices when detecting averted than direct gaze.

Materials and Methods

Participants

Thirty Chinese female university students (age = 24 ± 2.4 years, range = 20–29 years) participated in this study. Only females were recruited to avoid confounding gender influence in social decision making (Lee et al., 2009; Zhang et al., 2012; Sun et al., 2015b, 2016). All participants were right-handed (Oldfield, 1971) and had normal or corrected-to-normal vision. No participant had metal or medical device implants or any history of neurological or mental disorders. All participants in this study provided written informed consent to participate in procedures reviewed and approved by the local ethical committees at the University of Hong Kong and the East China Normal University. Three participants were excluded due to data recording errors, thus 27 participants were included for final analyses.

Eye Stimuli

There were three types of eye photos: human eyes with a direct gaze, human eyes with an averted gaze and robot’s eyes (Figure 1). The photos of human eyes were collected from 24 volunteers (12 males and 12 females) prior to this study. Each volunteer gave two photos of his or her face with a front view and a neutral expression: one with a direct gaze and the other with an averted gaze (i.e., looking to the left). These photos were put into two sets, each containing six pictures of a male direct gaze, six pictures of a male averted gaze, six pictures of a female direct gaze, and six pictures of a female averted gaze. The two photos from the same volunteer never appeared in the same set to avoid the potential conflicts elicited by different gaze directions from the same volunteer. Half of the participants viewed the eyes in one set, while the other half viewed the other set of eyes. The photo of the robot’s eyes was modified from a cartoon robot’s eyes downloaded from Internet resources. All photos were of identical sizes and adapted to contain only the eye region before being changed into black and white through Adobe Photoshop software (San Jose, CA, USA).

Figure 1.

Figure 1

Task paradigm and eye stimuli. The cyan and purple areas in the vertically stacked bars represented the proportions of reward assigned to the counterpart and the participant, respectively. In each trial, after knowing the total amount to be divided (i.e., the number in the screen) and the counterpart’s offer (represented by the bar on the left-hand side), the participant could accept or reject the proposal by pressing one of two buttons corresponding to the bars in the middle of the screen within 3 s. To accept the counterpart’s proposal is beneficial to both players, while to reject the proposal indicates a plan more advantageous to the participant. Immediately after the choice action, a black line was shown beneath the corresponding bar. The final reward distribution of a trial was presented by the outcome bar on the right-hand side at the last 3 s of the trial. A black line appeared above the outcome bar if the real situation was detected. On the contrary, no line was shown if the detection did not occur. When a rejection was detected, the participant gained nothing in that trial and her area in the outcome bar became black. Under the other conditions, the participant kept her share. If there was no response, or should the response exceed the 3-s decision-making phase, the reward of the trial was sent to the counterpart. Both inter-trial-interval (ITI) and inter-stimulus-interval (ISI) were on average 3 s. One of the three cues was presented during each trial. The cues are human eyes with direct or averted gaze and robot eyes.

Task and Procedure

Each participant received the following instructions before experiment: “You are invited to interact with anonymous counterparts in an online game. You should treat each trial as a single-shot interaction since counterparts will vary across trials and players cannot recognize each other. A photo of either human eyes or a robot’s eyes will be shown on a given trial to represent the counterpart type, i.e., human or robot, but not identity. In each trial, you will receive from a counterpart both a monetary investment as well as an offer on how to divide between you two the final amount, which is the appreciated investment through a computer-mimicked stock market. For simplicity, you will be shown just the final amount but not the initial investment. You can accept the offer or reject it by choosing an alternative plan that is more beneficial to you but less advantageous to the counterpart. After that, the counterpart will have 50-50 chance to know whether you have accepted or rejected his/her offer. If you reject an offer, your share in that trial will be transferred to the counterpart when he/she knows your choice. In the other conditions, you will keep your share. If you make no action, all benefits will be delivered to the counterpart in that trial. Your choices will influence the actual incomes of players. That is to say, both you and the counterpart will get the corresponding amount if you accept the offer. However, if the human counterpart does not know that you have chosen an alternative option, you will gain more than offered and he/she less than offered. All human players but not robot will finally receive real monetary bonuses proportional to the amounts earned during the task”. Based on these instructions, the participant did not know that, in fact, computer programs mimicked all of the responses of human/robot counterparts. This approach was successful in our previous study utilizing similar task procedures (Sun et al., 2016). In order to reduce the influence of value calculation on participants’ choices, expected utility (i.e., reward × probability) on a given trial was equal between accepting and rejecting an offer.

In each trial (Figure 1), following a jittered inter-trial-interval (ITI) of 3 s (Poisson-distributed), the amount of the increased investment was shown on the screen for 3 s (decision phase) during which the participant had to make her choice by pressing one of two buttons with the right index or middle finger. During the decision phase, four vertical bars were displayed on the screen, with the leftmost bar reflecting the counterpart’s offer plan, the middle two bars reflecting the two available options for division to choose from, and the rightmost empty bar signaling the forthcoming outcome (yet to be revealed). One of the choice options corresponded to the counterpart’s proposal, whereas the other option gave the participant a greater potential monetary reward, but carried a 50% risk of gaining nothing. In the option bars, the proportions of reward assigned to the counterpart was represented by the cyan-colored area, and the reward assigned to the participant was represented by the purple-colored area (Figure 1). The number above the option bars indicated the total amount of appreciated investment to be divided. The spatial positions of the two choice options were randomized across trials. Once the participant had made a choice, a black line appeared underneath the selected option bar. After a jittered inter-stimulus interval (ISI) of 3 s (Poisson-distributed), the participant was notified whether the real situation was detected and how much she gained in that trial in the following 3 s (outcome phase). A black line above the outcome bar indicated that the counterpart knew the participant’s actual choice, while no such line was shown if the detection did not occur. When an action of rejecting the counterpart’s offer was detected, the participant gained nothing in that trial and her share in the outcome bar became black. Under the other conditions, the participant kept the share for herself. If the participant failed to make a choice in that trial, or if the response exceeded the 3-s interval, all reward would be sent to the counterpart.

There were a total of 144 trials in the formal task. Photos of a direct gaze, an averted gaze, and a robot’s eyes were each shown for 48 trials in randomized orders. The permutation of the offer—that is, the amount to be divided (a number randomly generated among 80, 100 and 150), the proposed portion of repayment to the counterpart (60%, 65%, 70%), and the location (left or right) of the bars representing two options were balanced across the different types of photos. Before getting into the scanner, each participant was given detailed instructions and completed a minimum of eight practice trials to ensure task comprehension. The photos used in the practice trials were different from those used in the formal task. All participants reported after the task that they believed they were playing with real human-being when the human eye stimuli were shown. The participants were debriefed after the experiment. Each participant was awarded 200 Chinese Yuan as compensation and also 0–100 Chinese Yuan (proportional to the task earnings) as a task bonus. The visual stimuli presentations and response collections were performed through the integrated functional imaging system (IFIS).

Image Acquisition and Preprocessing

All images were acquired using a 3-Tesla Siemens Trio Tim MR scanner with a 12-channel head coil. T2*-weighted functional images were obtained using an EPI pulse sequence without inter-slice gap (33 axial slices parallel to the AC-PC line, TR = 2000 ms, TE = 30 ms, flip angle = 90°, Field of View (FOV) = 192 × 192 mm2, voxel size = 3 × 3 × 4 mm3). A high-resolution anatomical 3D T1-weighted MPRAGE image (192 slices, TR = 2530 ms, TE = 2.4 ms, flip angle = 7°, FOV = 224 × 256 mm2, voxel size = 0.5 × 0.5 × 1 mm3) was also acquired.

Images were preprocessed by using the CONN toolbox1, which calls functions from SPM12 software (Wellcome Department of Imaging Neuroscience, UK), through slice-timing and motion correction, normalization to the MNI (Montreal Neurological Institute) space and, finally, smoothing with an 8-mm full-width half-maximum Gaussian kernel.

Statistical Analyses

Participants made actions (either accept or reject) in more than 94% trials. For our research aims, conditions involving interactions with robot served as a control conditions representing non-social-related processes such as general value-based decision making. We thus subtracted frequency of choice and mean reaction time (RT) related with robot counterpart from the corresponding data associated with human counterparts. No difference in choice frequencies was observed for human and robot counterpart trials (ps > 0.287). The four contrasts of interest were prosocial choice (i.e., accepting offer) accompanied with direct gaze, selfish choice (i.e., rejecting offer) accompanied with direct gaze, prosocial choice accompanied with averted gaze, and selfish choice accompanied with averted gaze. The frequency of choice and RTs were then respectively analyzed by a 2 (gaze direction: direct and averted) × 2 (choice: prosocial and selfish) repeated-measures ANOVA model.

Images were analyzed utilizing the SPM12 software. The general line model (GLM) was used to examine the experimental effects across task events within each participant. The onset of the decision phase was modeled by six regressors with 3-s duration which were combinations of eye stimuli (robot’s eyes, direct human eyes and averted human eyes) and action (accept and reject offers). In addition, one regressor modeled the choice response, one modeled the onset of the outcome phase (3-s duration), and six extra regressors modeling residual head motions were included as nuisances. These regressors were convolved with the SPM canonical hemodynamic response function. High-pass temporal filtering with a cut-off of 128 s was employed to remove low-frequency drifts.

To form within-subject contrasts, consistent with the approach of behavioral data analyses, beta-weight images of regressors of robot counterpart were subtracted from the corresponding images associated with human players. This approach gave four contrast images per participant, i.e., prosocial choice accompanied with direct gaze, selfish choice accompanied with direct gaze, prosocial choice accompanied with averted gaze, and selfish choice accompanied with averted gaze. These contrasts were then entered into a group-level 2 (gaze direction: direct and averted) × 2 (choice: prosocial and selfish) flexible factorial model. Results were voxel-level height thresholded at p < 0.001 and survived family-wise error (FWE) cluster-level correction (p < 0.05) within the whole brain. To specifically test our a priori hypotheses, we also reported findings with voxel-level height threshold at p < 0.001 and survived FWE correction (p < 0.05) within regions of interests (ROIs) including the PCC (Brodmann’s areas 23 and 31; Leech and Sharp, 2014), mPFC (Brodmann’s areas 9, 10, 24, 25 and 32; Murray et al., 2016), and bilateral amygdala constructed using the WFU_PickAtlas toolbox2). We also investigated the findings in bilateral anterior TPJ (center coordinates: left, x = −53, y = −30, z = 10; right, x = 47, y = −35, z = 12) and posterior TPJ (center coordinates: left, x = −53, y = −59, z = 20; right, x = 56, y = −56, z = 18) in spheres with a radius of 8 mm (Schurz et al., 2014). All significant clusters contained more than 5 voxels. Mean beta values of fMRI contrasts were extracted for further analyses from the aforementioned ROIs showing significant task activations, using the MarsBaR toolbox3. Bonferroni method was employed to correct for multiple comparisons during post hoc t tests and fMRI-behavior correlation analyses.

We further investigated the functional coupling between the seed region and the rest of the brain, especially in the ROIs of ToM areas. The seed area was the region showing significant interaction between gaze direction and choice in the fMRI analyses. We performed a generalized psychophysiological interaction (gPPI) analysis through the gPPI toolbox4. Following fMRI analyses, four contrast images per participant were made reflecting the differences between playing against human vs. robot. These contrasts were also entered into a group-level 2 (gaze direction: direct and averted) × 2 (choice: prosocial and selfish) flexible factorial model. Results were voxel-level height thresholded at p < 0.001, FWE cluster-level corrected at p < 0.05, and contained more than 5 voxels. Mean beta values of gPPI contrasts were extracted for further analyses from the significant cluster through the MarsBaR toolbox. Bonferroni method was employed to correct for multiple comparisons during post hoc t tests and behavior-fMRI correlation analyses.

Results

Behavioral Findings

Mean and standard deviations of behavioral measures were organized in Table 1. For our research purposes, behavioral data (i.e., frequency of choice and RT) and image contrasts related with robot counterpart were subtracted from the corresponding data associated with human counterparts. No significant gaze direction effects were found for either frequency of choice (Fs < 2.292, ps > 0.142) or RT (Fs < 0.858, ps > 0.363). Comparisons between human (including both direct and averted gaze conditions) and computer counterparts were organized in the supplementary document.

Table 1.

Frequency of action and reaction time (RT).

Stimuli Robot’s eyes Human direct eyes Human averted eyes
Action Accept Reject Accept Reject Accept Reject
Frequency (%)
Mean 46.9 53.1 47.1 52.9 50.1 49.9
Std 13.5 13.5 12.9 12.9 12.1 12.1
Reaction time (ms)
Mean 1189.2 1143.1 1237.8 1217.9 1224.2 1223.4
Std 161.6 181.5 183.3 205.6 204.0 201.9

fMRI Findings

Averted gaze elicited stronger activations in right STG (Brodmann’s area 22, cluster size = 293 voxels, T value = 4.18, peak MNI coordinates = [52, −46, 8], FWE-corrected cluster-level p value = 0.018) than direct gaze, while there was no significantly stronger activation for direct than averted gaze. No main effect of choice was detected significant. Importantly, a significant interaction between gaze direction and choice was detected in the right STG (Brodmann’s area 42/22, cluster size = 227 voxels, T value = 4.24, peak MNI coordinates = [70, −30, 16], FWE-corrected cluster-level p value = 0.047, see Figure 2A), characterized by stronger activations during prosocial choices vs. selfish choices when presented with averted gaze than direct gaze. Mean beta values extracted from this right STG cluster were greater to prosocial choices than to selfish choices when averted (t(26) = 4.583, p < 0.001 corrected) but not direct (t(26) = −0.459, p > 0.6) gaze was presented (Figure 2B). No significant results were found in the other ROIs. Comparisons between human (including both direct and averted gaze conditions) and computer counterparts were organized in the Supplementary Table S1. Moreover, larger mean betas in the right STG cluster were accompanied with quicker actions (i.e., shorter mean RT) across participants for prosocial choice to averted gaze (Pearson’s R = −0.585, p = 0.008 corrected, Figure 2C). No other correlation in right STG was found significant (all ps > 0.07 uncorrected), see Supplementary Table S2.

Figure 2.

Figure 2

Functional magnetic resonance imaging (fMRI) findings of the interaction between Gaze and Choice. (A) Prosocial vs. selfish choices were associated with stronger activations in right superior temporal gyrus (R STG) for averted than direct gaze. Imaging results were height-thresholded at p < 0.001 and survived p < 0.05 family-wise error (FWE) correction. (B) Larger fMRI beta values averaged within the cluster in R STG were found for prosocial choice than selfish choice when averted gaze was presented (t(26) = 4.583, p = 0.006 corrected). Error bar denotes standard error mean. (C) Significant correlation (Pearson’s R = −0.585, p = 0.008 corrected) was found between fMRI betas in R STG and reaction time (RT) for the condition of prosocial choice accompanied with averted gaze. **p < 0.01, ***p < 0.001.

gPPI Findings

We then investigated the functional couplings between the seed region and the rest of the brain, especially the ToM ROIs. The seed is the right STG cluster showing significant interaction between gaze direction and choice in the fMRI analyses. Prosocial choices vs. selfish choices were accompanied with larger gPPI values in PCC (surviving small-volume FWE correction; Brodmann’s area 31, cluster size = 26 voxels, T value = 4.29, Z value = 4.11, peak MNI coordinates = [−4, −34, 48], FWE-corrected peak-level p value = 0.009, Figure 3A) when averted gaze than direct gaze was shown. Mean beta values extracted from this PCC cluster showed that prosocial choices were accompanied with stronger functional couplings than selfish choices when averted (t(26) = 3.703, p = 0.006 corrected) but not direct (t(26) = −0.672, p > 0.5) gaze was presented (Figure 3B). No significant gPPI results were found in the other ROIs.

Figure 3.

Figure 3

Generalized psychophysiological interaction (gPPI) findings of the interaction between Gaze and Choice. (A) Prosocial vs. selfish choices were associated with larger gPPI values in posterior cingulate cortex (PCC) for averted than direct gaze. Imaging results were height-thresholded at p < 0.001 and survived p < 0.05 FWE correction within an anatomical mask of PCC consisting of Brodmann’s area 23 and 31. (B) Larger gPPI beta values averaged within the cluster in PCC were found for prosocial choice than selfish choice when averted gaze was presented (t(26) = 4.583, p = 0.006 corrected). Error bar denotes standard error mean. **p < 0.01.

Discussion

Eye gaze plays vital roles in many social contexts (Itier and Batty, 2009). However, little is known about its influences on social decision making integral to everyday social functioning (Rilling and Sanfey, 2011). To the best of our knowledge, this is the first study investigating the neural processing of the interaction between social decisions and perceived gaze direction. Consistent with the second a priori hypothesis, we found that prosocial vs. selfish choice elicited stronger activations in right STG and larger right STG-PCC functional connectivity when averted rather than direct gaze was presented. Moreover, stronger activations in right STG was associated with quicker actions for making prosocial choice accompanied with averted gaze. Our findings suggest that both right STG and right STG-PCC functional connections are more involved for making prosocial choices than selfish decisions when the perceived subtle social cues signal a lack of intentions for social contact.

The right STG has been widely reported to play roles in responding to gaze direction (Itier and Batty, 2009), and in detecting, predicting and reasoning about actions and intentions of others (Allison et al., 2000). Importantly, our findings further demonstrate the role of right STG in social decision making. We found stronger activation in right STG for prosocial than selfish choice accompanied with averted gaze. This result suggests that, when the perceived subtle cue through others’ gaze signals a lack of intentions for social contact, the right STG is more involved in inferring about the counterparts’ intentions during making a decision to benefit them. This explanation is further supported by the negative correlation between right STG activation and RT during prosocial choice accompanied with averted gaze. By contrast, no significant differences were found between prosocial and selfish choices accompanied with direct gaze. It is possible that being observed by others’ direct gaze is default in social interaction and carries a relatively constant level of processing about others’ intention regardless of the choice made.

Perception of observations by others has been found to efficiently promote prosocial behaviors (Izuma et al., 2010). It is hypothesized that, when being observed by someone, people make prosocial actions in order to gain social approvals/reputations (Rege and Telle, 2004; Izuma, 2012) and/or to avoid the guilt of harming others (Morey et al., 2012). Consistent with this idea, Izuma (2012) detected more donations to charities and stronger activations in striatum when donating in the presence of observers than in their absence. Direct (vs. averted) gaze is proposed to cue the observations by others and may thus elicit stronger brain activations for prosocial than selfish choices. However, this hypothesis is inconsistent with our findings. It is thus difficult to explain the gaze effect on social decision making through social reputation or guilt.

Previous studies have also detected stronger STG activations in response to the mismatch between one’s motion and the context (Grezes et al., 2004a,b; Wyk et al., 2009), and to moral judgments regarding the events that violate social norms (Prehn et al., 2008; Bahnemann et al., 2010). These findings suggest the roles of STG in reflecting the mismatch between observed actions and the context. If this theory also applies to the conflict between one’s own actions and the context, we would hypothesize to find stronger STG activations for both prosocial decisions accompanied with averted gaze and selfish choices accompanied with direct gaze. The reasoning is that prosocial decision is suboptimal when not being observed by others (represented by averted gaze implying higher chance of “getting away with” potential punishment), and selfish decision is suboptimal when being observed by others (represented by direct gaze implying high risk of being caught and punished). However, we only detected stronger brain activations for the former but not the latter condition, suggesting that our results cannot be fully interpreted by the conflict between one’s own actions and the context.

We also found larger functional couplings between right STG and PCC during making prosocial (vs. selfish) choices when perceiving averted gaze. The PCC has been widely reported in studies on self-referential processing (Lombardo et al., 2010; Brewer et al., 2013; Schurz et al., 2014). It is possible that more processing of others’ intentions (represented by stronger right STG activations) is accompanied with more processing of the relationship between self and others. This thought is consistent with the previous findings that the PCC as well as the nearby precuneus are associated with retrieval of episodic memory (Gobbini and Haxby, 2007; Sun et al., 2015c). In other words, during making prosocial decisions, when the perceived averted gaze implied lack of interpersonal interaction, participants need to recall their personal experiences in order to infer the others’ intentions.

We did not find any significant results in the other ToM ROIs including mPFC, TPJ and amygdala. First, mPFC has been proposed to play central roles in ToM (Schurz et al., 2014). However, Krause et al. (2012) utilized repetitive transcranial magnetic stimulation (rTMS) to bilateral mPFC, and did not find significant effect on either cognitive or affective ToM performance. They further reported that deep rTMS disrupted affective ToM performance in participants with high empathy, but increased affective ToM performance in those with low empathy, suggesting that the roles of mPFC in affective ToM are modulated by the baseline empathic abilities. In our study, the interaction between perceived gaze direction and social decision making may also be influenced by the level of empathy in participants which we unfortunately did not measure. Second, studies have shown that TPJ is sensitive to prediction error, which means the degree to which current information is inconsistent with expectation, in various domains including social interaction (Behrens et al., 2008; Simon et al., 2017). People in multiple-rounds social interactions may employ TPJ to guide behaviors based on previous experiences. However, participants in our task paradigm were instructed to treat each trial as a single-shot interaction, and they could not utilize the experiences with a previous counterpart to influence the current interaction. Third, amygdala has been widely reported to play a central role in response to the interaction between perceived gaze direction and facial expression (Cristinzio et al., 2010; Sato et al., 2010; Ziaei et al., 2016). Facial expression is delivered by not only eyes but also other parts of the face such as the mouth (Ekman, 2003; Sun et al., 2015a). In this study, however, the photos of humans were taken merely in neutral expression, and only the eye region was displayed. These might have minimized the roles of mPFC, TPJ and amygdala in our task paradigm. Future studies should further investigate the brain activation and functional connectivity in the ToM ROIs utilizing alternative task paradigms and stimuli, taking into account individual differences in social traits.

We failed to find any significant behavioral result related with either gaze or interaction between gaze and choice in this study. The presence of eye-like stimuli has been repeatedly reported to increases prosocial behaviors such as greater investments (Bente et al., 2014), greater charitable donations (Powell et al., 2012), theft prevention (Bateson et al., 2013) and higher voting rate (Panagopoulos, 2014), the so-called “watching eyes effect” (Nettle et al., 2013). There are two possible explanations for our insignificant findings. First, most of previous laboratory and field studies on the “watching eyes effect” employed procedures in which a participant makes decisions for only one or just a few trials (Haley and Fessler, 2005; Nettle et al., 2013). By contrast, participants in our task made choices in 144 trials. Second, the “watching eyes effect” were observed when eyes or eye-like stimuli were present vs. absent, while eye stimuli were always present in our study. The behavioral and neural mechanisms of gaze direction influence on social decision making may be different from those of the “watching eyes effect”. These need to be tested in future studies.

Our findings also have clinical implications, especially for people with autism spectrum disorders (ASDs) who exhibit impairments in reciprocal social interactions (Baron-Cohen et al., 1997). ASD has been found to relate to ToM impairments (Baron-Cohen et al., 2001). ASD patients show deficits in paying attention to the eye region (Spezio et al., 2007), extracting useful information from the eyes (Nation and Penny, 2008), and understanding others’ mental states (Campbell et al., 2006). A later diagnosis of autism was found to be predicted at 18 months of age by an absence of joint attention (Baron-Cohen et al., 1996), which is a precursor to ToM and reflects the ability of attending to the object cued by another person’s gaze direction (Emery, 2000). Our findings imply that ASD patients have deficits in processing subtle social cues in the environment such as eye gaze direction and/or in utilizing such information for making social decisions. Future work needs to clarify the neural underpinnings of social decision making in ASD patients when perceiving social cues. Further, previous studies suggested that brain stimulation techniques, such as TMS (Oberman et al., 2015) and Transcranial direct current stimulation (tDCS; Amatachaya et al., 2014, 2015), are promising methods of clinical treatment for ASD. These non-invasive brain stimulation techniques influence the activations in a small population of neurons in a targeted brain region. Our findings suggest that the right STG is a potential target for clinical intervention. Furthermore, the brain activation in the right STG and the right STG-PCC functional connections may also be employed to reflect the outputs of clinical treatment.

Our study have several limitations. First, our participants were all Chinese females. Caucasian and East Asian participants were found to fixate on the internal features (especially the eyes) and the center of faces (Blais et al., 2008), respectively, suggesting culture influences on face processing. People in western cultures may thus be more influenced by gaze direction while making social decisions. Future studies are needed for comparing the gaze-orienting effects across cultures and in different gender groups. Second, direct and averted gazes in this study were all shown in photos containing front-view faces, and were restricted within the eye region for simplicity. Previous studies have shown that head orientation (Itier et al., 2007) and dynamic gaze presentation (Putman et al., 2006) influenced the effects of gaze. Future studies should investigate the influence of gaze direction on social decision making in different head orientations and/or using a dynamic gaze (e.g., movies of gaze motion or real human eyes). Third, only left-oriented eyes were employed to represent the averted gaze in this study. Different directions of averted gaze may confer different social meanings and recruited different STG subregions in Calder et al. (2007). Future studies should separately investigate the effects of different averted gaze directions.

In conclusion, our study provides the first evidence that the neural processing of social decision making is influenced by perceived gaze direction. Our findings suggest that, when the perceived gaze direction signals lack of social contact, making prosocial vs. selfish decisions is associated with greater neural processing of inferring others’ intention and understanding the relationship between self and others. These findings also shed light on the deficiencies in processing subtle social cues and social functioning in ASD individuals.

Author Contributions

DS, RS, ZW and TMCL made substantial contributions to the conception or design of the work or the acquisition, analysis, or interpretation of data for the work; drafted the work or revised it critically for important intellectual content; approved the final version to be published; agreed to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Funding. This work was supported by the Research Grant Council Humanities and Social Sciences Prestigious Fellowship (Ref: HKU703-HSS-13) and the National Natural Science Fund of China (Ref: 31070986). The fund agency has no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnhum.2018.00052/full#supplementary-material

References

  1. Adams R. B., Jr., Kleck R. E. (2003). Perceived gaze direction and the processing of facial displays of emotion. Psychol. Sci. 14, 644–647. 10.1046/j.0956-7976.2003.psci_1479.x [DOI] [PubMed] [Google Scholar]
  2. Adams R. B., Jr., Kleck R. E. (2005). Effects of direct and averted gaze on the perception of facially communicated emotion. Emotion 5, 3–11. 10.1037/1528-3542.5.1.3 [DOI] [PubMed] [Google Scholar]
  3. Akiyama T., Kato M., Muramatsu T., Saito F., Umeda S., Kashima H. (2006). Gaze but not arrows: a dissociative impairment after right superior temporal gyrus damage. Neuropsychologia 44, 1804–1810. 10.1016/j.neuropsychologia.2006.03.007 [DOI] [PubMed] [Google Scholar]
  4. Allison T., Puce A., McCarthy G. (2000). Social perception from visual cues: role of the STS region. Trends Cogn. Sci. 4, 267–278. 10.1016/s1364-6613(00)01501-1 [DOI] [PubMed] [Google Scholar]
  5. Amatachaya A., Auvichayapat N., Patjanasoontorn N., Suphakunpinyo C., Ngernyam N., Aree-Uea B., et al. (2014). Effect of anodal transcranial direct current stimulation on autism: a randomized double-blind crossover trial. Behav. Neurol. 2014:173073. 10.1155/2014/173073 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Amatachaya A., Jensen M. P., Patjanasoontorn N., Auvichayapat N., Suphakunpinyo C., Janjarasjitt S., et al. (2015). The short-term effects of transcranial direct current stimulation on electroencephalography in children with autism: a randomized crossover controlled trial. Behav. Neurol. 2015:928631. 10.1155/2015/928631 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Bahnemann M., Dziobek I., Prehn K., Wolf I., Heekeren H. R. (2010). Sociotopy in the temporoparietal cortex: common versus distinct processes. Soc. Cogn. Affect. Neurosci. 5, 48–58. 10.1093/scan/nsp045 [DOI] [PMC free article] [PubMed] [Google Scholar]
  8. Baron-Cohen S. (1995). Mindblindness: An Essay on Autism and Theory of Mind. Cambridge, MA: MIT Press. [Google Scholar]
  9. Baron-Cohen S., Cox A., Baird G., Swettenham J., Nightingale N., Morgan K., et al. (1996). Psychological markers in the detection of autism in infancy in a large population. Br. J. Psychiatry 168, 158–163. 10.1192/bjp.168.2.158 [DOI] [PubMed] [Google Scholar]
  10. Baron-Cohen S., Leslie A. M., Frith U. (1985). Does the autistic-child have a “theory of mind”? Cognition 21, 37–46. 10.1016/0010-0277(85)90022-8 [DOI] [PubMed] [Google Scholar]
  11. Baron-Cohen S., Wheelwright S., Hill J., Raste Y., Plumb I. (2001). The “Reading the Mind in the Eyes” test revised version: a study with normal adults, and adults with Asperger syndrome or high-functioning autism. J. Child Psychol. Psychiatry 42, 241–251. 10.1111/1469-7610.00715 [DOI] [PubMed] [Google Scholar]
  12. Baron-Cohen S., Wheelwright S., Jolliffe T. (1997). Is there a “language of the eyes”? Evidence from normal adults and adults with autism or Asperger Syndrome. Autism Res. 4, 311–331. 10.1080/71375676129360200 [DOI] [Google Scholar]
  13. Bastin C., Harrison B. J., Davey C. G., Moll J., Whittle S. (2016). Feelings of shame, embarrassment and guilt and their neural correlates: a systematic review. Neurosci. Biobehav. Rev. 71, 455–471. 10.1016/j.neubiorev.2016.09.019 [DOI] [PubMed] [Google Scholar]
  14. Bateson M., Callow L., Holmes J. R., Redmond Roche M. L., Nettle D. (2013). Do images of ‘watching eyes’ induce behaviour that is more pro-social or more normative? A field experiment on littering. PLoS One 8:e82055. 10.1371/journal.pone.0082055 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Behrens T. E. J., Hunt L. T., Woolrich M. W., Rushworth M. F. S. (2008). Associative learning of social value. Nature 456, 245–249. 10.1038/nature07538 [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Bente G., Dratsch T., Kaspar K., Häβler T., Bungard O., Al-Issa A. (2014). Cultures of trust: effects of avatar faces and reputation scores on german and arab players in an online trust-game. PLoS One 9:e98297. 10.1371/journal.pone.0098297 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Blais C., Jack R. E., Scheepers C., Fiset D., Caldara R. (2008). Culture shapes how we look at faces. PLoS One 3:e3022. 10.1371/journal.pone.0003022 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Brewer J. A., Garrison K. A., Whitfield-Gabrieli S. (2013). What about the “self” is processed in the posterior cingulate cortex? Front. Hum. Neurosci. 7:647. 10.3389/fnhum.2013.00647 [DOI] [PMC free article] [PubMed] [Google Scholar]
  19. Burnett S., Blakemore S. J. (2009). Functional connectivity during a social emotion task in adolescents and in adults. Eur. J. Neurosci. 29, 1294–1301. 10.1111/j.1460-9568.2009.06674.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Calder A. J., Beaver J. D., Winston J. S., Dolan R. J., Jenkins R., Eger E., et al. (2007). Separate coding of different gaze directions in the superior temporal sulcus and inferior parietal lobule. Curr. Biol. 17, 20–25. 10.1016/j.cub.2006.10.052 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Calder A. J., Lawrence A. D., Keane J., Scott S. K., Owen A. M., Christoffels I., et al. (2002). Reading the mind from eye gaze. Neuropsychologia 40, 1129–1138. 10.1016/s0028-3932(02)00008-8 [DOI] [PubMed] [Google Scholar]
  22. Campbell R., Lawrence K., Mandy W., Mitra C., Jeyakuma L., Skuse D. (2006). Meanings in motion and faces: developmental associations between the processing of intention from geometrical animations and gaze detection accuracy. Dev. Psychopathol. 18, 99–118. 10.1017/s0954579406060068 [DOI] [PubMed] [Google Scholar]
  23. Carlin J. D., Calder A. J. (2013). The neural basis of eye gaze processing. Curr. Opin. Neurobiol. 23, 450–455. 10.1016/j.conb.2012.11.014 [DOI] [PubMed] [Google Scholar]
  24. Cristinzio C., N’Diaye K., Seeck M., Vuilleumier P., Sander D. (2010). Integration of gaze direction and facial expression in patients with unilateral amygdala damage. Brain 133, 248–261. 10.1093/brain/awp255 [DOI] [PubMed] [Google Scholar]
  25. Dunfield K. A. (2014). A construct divided: prosocial behavior as helping, sharing, and comforting subtypes. Front. Psychol. 5:958. 10.3389/fpsyg.2014.00958 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Ekman P. (2003). Emotions Revealed: Recognizing Faces and Feelings to Improve Communication and Emotional Life. New York, NY: Times Books. [Google Scholar]
  27. Emery N. J. (2000). The eyes have it: the neuroethology, function and evolution of social gaze. Neurosci. Biobehav. Rev. 24, 581–604. 10.1016/s0149-7634(00)00025-7 [DOI] [PubMed] [Google Scholar]
  28. Farroni T., Csibra G., Simion G., Johnson M. H. (2002). Eye contact detection in humans from birth. Proc. Natl. Acad. Sci. U S A 99, 9602–9605. 10.1073/pnas.152159999 [DOI] [PMC free article] [PubMed] [Google Scholar]
  29. Frischen A., Bayliss A. P., Tipper S. P. (2007). Gaze cueing of attention: visual attention, social cognition, and individual differences. Psychol. Bull. 133, 694–724. 10.1037/0033-2909.133.4.694 [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Gobbini M. I., Haxby J. V. (2007). Neural systems for recognition of familiar faces. Neuropsychologia 45, 32–41. 10.1016/j.neuropsychologia.2006.04.015 [DOI] [PubMed] [Google Scholar]
  31. Grezes J., Frith C., Passingham R. E. (2004a). Brain mechanisms for inferring deceit in the actions of others. J. Neurosci. 24, 5500–5505. 10.1523/JNEUROSCI.0219-04.2004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Grezes J., Frith C. D., Passingham R. E. (2004b). Inferring false beliefs from the actions of oneself and others: an fMRI study. Neuroimage 21, 744–750. 10.1016/s1053-8119(03)00665-7 [DOI] [PubMed] [Google Scholar]
  33. Haley K. J., Fessler D. M. T. (2005). Nobody’s watching? Subtle cues affect generosity in an anonymous economic game. Evol. Hum. Behav. 26, 245–256. 10.1016/j.evolhumbehav.2005.01.002 [DOI] [Google Scholar]
  34. Hardee J. E., Thompson J. C., Puce A. (2008). The left amygdala knows fear: laterality in the amygdala response to fearful eyes. Soc. Cogn. Affect. Neurosci. 3, 47–54. 10.1093/scan/nsn001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Imuta K., Henry J. D., Slaughter V., Selcuk B., Ruffman T. (2016). Theory of mind and prosocial behavior in childhood: a meta-analytic review. Dev. Psychol. 52, 1192–1205. 10.1037/dev0000140 [DOI] [PubMed] [Google Scholar]
  36. Itier R. J., Batty M. (2009). Neural bases of eye and gaze processing: the core of social cognition. Neurosci. Biobehav. Rev. 33, 843–863. 10.1016/j.neubiorev.2009.02.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Itier R. J., Villate C., Ryan J. D. (2007). Eyes always attract attention but gaze orienting is task-dependent: evidence from eye movement monitoring. Neuropsychologia 45, 1019–1028. 10.1016/j.neuropsychologia.2006.09.004 [DOI] [PubMed] [Google Scholar]
  38. Izuma K. (2012). The social neuroscience of reputation. Neurosci. Res. 72, 283–288. 10.1016/j.neures.2012.01.003 [DOI] [PubMed] [Google Scholar]
  39. Izuma K., Saito D. N., Sadato N. (2010). Processing of the incentive for social approval in the ventral striatum during charitable donation. J. Cogn. Neurosci. 22, 621–631. 10.1162/jocn.2009.21228 [DOI] [PubMed] [Google Scholar]
  40. Keltner D., Kogan A., Piff P. K., Saturn S. R. (2014). The sociocultural appraisals, values, and emotions (SAVE) framework of prosociality: core processes from gene to meme. Annu. Rev. Psychol. 65, 425–460. 10.1146/annurev-psych-010213-115054 [DOI] [PubMed] [Google Scholar]
  41. Krause L., Enticott P. G., Zangen A., Fitzgerald P. B. (2012). The role of medial prefrontal cortex in theory of mind: a deep rTMS study. Behav. Brain Res. 228, 87–90. 10.1016/j.bbr.2011.11.037 [DOI] [PubMed] [Google Scholar]
  42. Lee T. M., Chan C. C., Leung A. W., Fox P. T., Gao J. H. (2009). Sex-related differences in neural activity during risk taking: an fMRI study. Cereb. Cortex 19, 1303–1312. 10.1093/cercor/bhn172 [DOI] [PMC free article] [PubMed] [Google Scholar]
  43. Leech R., Sharp D. J. (2014). The role of the posterior cingulate cortex in cognition and disease. Brain 137, 12–32. 10.1093/brain/awt162 [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Lombardo M. V., Chakrabarti B., Bullmore E. T., Wheelwright S. J., Sadek S. A., Suckling J., et al. (2010). Shared neural circuits for mentalizing about the self and others. J. Cogn. Neurosci. 22, 1623–1635. 10.1162/jocn.2009.21287 [DOI] [PubMed] [Google Scholar]
  45. Moll J., de Oliveira-Souza R. (2007). Moral judgments, emotions and the utilitarian brain. Trends Cogn. Sci. 11, 319–321. 10.1016/j.tics.2007.06.001 [DOI] [PubMed] [Google Scholar]
  46. Moll J., de Oliveira-Souza R., Garrido G. J., Bramati I. E., Caparelli-Daquer E. M. A., Paiva M. L. M. F., et al. (2007). The self as a moral agent: link-ling the neural bases of social agency and moral sensitivity. Soc. Neurosci. 2, 336–352. 10.1080/17470910701392024 [DOI] [PubMed] [Google Scholar]
  47. Moll J., Zahn R., de Oliveira-Souza R., Krueger F., Grafman J. (2005). The neural basis of human moral cognition. Nat. Rev. Neurosci. 6, 799–809. 10.1038/nrn1768 [DOI] [PubMed] [Google Scholar]
  48. Morey R. A., McCarthy G., Selgrade E. S., Seth S., Nasser J. D., LaBar K. S. (2012). Neural systems for guilt from actions affecting self versus others. Neuroimage 60, 683–692. 10.1016/j.neuroimage.2011.12.069 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Murray E. A., Wise S. P., Graham K. S. (2016). “Chapter 1: the history of memory systems,” in The Evolution of Memory Systems: Ancestors, Anatomy, and Adaptations, eds Baum M., Greene C., Butler M. (Oxford: Oxford University Press; ), 22–24. [Google Scholar]
  50. Nation K., Penny S. (2008). Sensitivity to eye gaze in autism: is it normal? Is it automatic? Is it social? Dev. Psychopathol. 20, 79–97. 10.1017/S0954579408000047 [DOI] [PubMed] [Google Scholar]
  51. N’Diaye K., Sander D., Vuilleumier P. (2009). Self-relevance processing in the human amygdala: gaze direction, facial expression, and emotion intensity. Emotion 9, 798–806. 10.1037/a0017845 [DOI] [PubMed] [Google Scholar]
  52. Nettle D., Harper Z., Kidson A., Stone R., Penton-Voak I. S., Bateson M. (2013). The watching eyes effect in the Dictator Game: it’s not how much you give, it’s being seen to give something. Evol. Hum. Behav. 34, 35–40. 10.1016/j.evolhumbehav.2012.08.004 [DOI] [Google Scholar]
  53. Northoff G., Bermpohl F. (2004). Cortical midline structures and the self. Trends Cogn. Sci. 8, 102–107. 10.1016/j.tics.2004.01.004 [DOI] [PubMed] [Google Scholar]
  54. Nummenmaa L., Calder A. J. (2009). Neural mechanisms of social attention. Trends Cogn. Sci. 13, 135–143. 10.1016/j.tics.2008.12.006 [DOI] [PubMed] [Google Scholar]
  55. Oberman L. M., Rotenberg A., Pascual-Leone A. (2015). Use of transcranial magnetic stimulation in autism spectrum disorders. J. Autism Dev. Disord. 45, 524–536. 10.1007/s10803-013-1960-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Oldfield R. C. (1971). The assessment and analysis of handedness: the Edinburgh inventory. Neuropsychologia 9, 97–113. 10.1016/0028-3932(71)90067-4 [DOI] [PubMed] [Google Scholar]
  57. Panagopoulos C. (2014). Watchful eyes: implicit observability cues and voting. Evol. Hum. Behav. 35, 279–284. 10.1016/j.evolhumbehav.2014.02.008 [DOI] [Google Scholar]
  58. Parvizi J., Van Hoesen G. W., Buckwalter J., Damasio A. (2006). Neural connections of the posteromedial cortex in the macaque. Proc. Natl. Acad. Sci. U S A 103, 1563–1568. 10.1073/pnas.0507729103 [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Pelphrey K. A., Viola R. J., Mccarthy G. (2004). When strangers pass—processing of mutual and averted social gaze in the superior temporal sulcus. Psychol. Sci. 15, 598–603. 10.1111/j.0956-7976.2004.00726.x [DOI] [PubMed] [Google Scholar]
  60. Powell K. L., Roberts G., Nettle D. (2012). Eye images increase charitable donations: evidence from an opportunistic field experiment in a supermarket. Ethology 118, 1096–1101. 10.1111/eth.12011 [DOI] [Google Scholar]
  61. Prehn K., Wartenburger I., Mériau K., Scheibe C., Goodenough O. R., Villringer A., et al. (2008). Individual differences in moral judgment competence influence neural correlates of socio-normative judgments. Soc. Cogn. Affect. Neurosci. 3, 33–46. 10.1093/scan/nsm037 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Putman P., Hermans E., Van Honk J. (2006). Anxiety meets fear in perception of dynamic expressive gaze. Emotion 6, 94–102. 10.1037/1528-3542.6.1.94 [DOI] [PubMed] [Google Scholar]
  63. Rege M., Telle K. (2004). The impact of social approval and framing on cooperation in public good situations. J. Public Econ. 88, 1625–1644. 10.1016/s0047-2727(03)00021-5 [DOI] [Google Scholar]
  64. Rilling J. K., Sanfey A. G. (2011). The neuroscience of social decision-making. Annu. Rev. Psychol. 62, 23–48. 10.1146/annurev.psych.121208.131647 [DOI] [PubMed] [Google Scholar]
  65. Rilling J. K., Sanfey A. G., Aronson J. A., Nystrom L. E., Cohen J. D. (2004). The neural correlates of theory of mind within interpersonal interactions. Neuroimage 22, 1694–1703. 10.1016/j.neuroimage.2004.04.015 [DOI] [PubMed] [Google Scholar]
  66. Sato W., Kochiyama T., Uono S., Yoshikawa S. (2010). Amygdala integrates emotional expression and gaze direction in response to dynamic facial expressions. Neuroimage 50, 1658–1665. 10.1016/j.neuroimage.2010.01.049 [DOI] [PubMed] [Google Scholar]
  67. Schurz M., Radua J., Aichhorn M., Richlan F., Perner J. (2014). Fractionating theory of mind: a meta-analysis of functional brain imaging studies. Neurosci. Biobehav. Rev. 42, 9–34. 10.1016/j.neubiorev.2014.01.009 [DOI] [PubMed] [Google Scholar]
  68. Shaw P., Lawrence E. J., Radbourne C., Bramham J., Polkey C. E., David A. S. (2004). The impact of early and late damage to the human amygdala on ‘theory of mind’ reasoning. Brain 127, 1535–1548. 10.1093/brain/awh168 [DOI] [PubMed] [Google Scholar]
  69. Shimojo S., Simion C., Shimojo E., Scheier C. (2003). Gaze bias both reflects and influences preference. Nat. Neurosci. 6, 1317–1322. 10.1038/nn1150 [DOI] [PubMed] [Google Scholar]
  70. Simon K. C. N. S., Gómez R. L., Nadel L., Scalf P. E. (2017). Brain correlates of memory reconsolidation: a role for the TPJ. Neurobiol. Learn. Mem. 142, 154–161. 10.1016/j.nlm.2017.03.003 [DOI] [PubMed] [Google Scholar]
  71. Spezio M. L., Huang P. Y. S., Castelli F., Adolphs R. (2007). Amygdala damage impairs eye contact during conversations with real people. J. Neurosci. 27, 3994–3997. 10.1523/jneurosci.3789-06.2007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Sun D., Chan C. C. H., Fan J. T., Wu Y., Lee T. M. C. (2015a). Are happy faces attractive? The roles of early vs. late processing. Front. Psychol. 6:1812. 10.3389/fpsyg.2015.01812 [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Sun D., Chan C. C. H., Hu Y., Wang Z., Lee T. M. C. (2015b). Neural correlates of outcome processing post dishonest choice: an fMRI and ERP study. Neuropsychologia 68, 148–157. 10.1016/j.neuropsychologia.2015.01.013 [DOI] [PubMed] [Google Scholar]
  74. Sun D., Lee T. M. C., Chan C. C. H. (2015c). Unfolding the spatial and temporal neural processing of lying about face familiarity. Cereb. Cortex 25, 927–936. 10.1093/cercor/bht284 [DOI] [PMC free article] [PubMed] [Google Scholar]
  75. Sun D., Lee T. M. C., Wang Z. X., Chan C. C. H. (2016). Unfolding the spatial and temporal neural processing of making dishonest choices. PLoS One 11:e0153660. 10.1371/journal.pone.0153660 [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Uddin L. Q., Kelly A. M. C., Biswal B. B., Castellanos F. X., Milham M. P. (2009). Functional connectivity of default mode network components: correlation, anticorrelation, and causality. Hum. Brain Mapp. 30, 625–637. 10.1002/hbm.20531 [DOI] [PMC free article] [PubMed] [Google Scholar]
  77. von dem Hagen E. A. H., Stoyanova R. S., Rowe J. B., Baron-Cohen S., Calder A. J. (2014). Direct gaze elicits atypical activation of the theory-of-mind network in autism spectrum conditions. Cereb. Cortex 24, 1485–1492. 10.1093/cercor/bht003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Weller D., Lagattuta K. H. (2014). Children’s judgments about prosocial decisions and emotions: gender of the helper and recipient matters. Child Dev. 85, 2011–2028. 10.1111/cdev.12238 [DOI] [PubMed] [Google Scholar]
  79. Wyk B. C. V., Hudac C. M., Carter E. J., Sobel D. M., Pelphrey K. A. (2009). Action understanding in the superior temporal sulcus region. Psychol. Sci. 20, 771–777. 10.1111/j.1467-9280.2009.02359.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Zhang H. J., Sun D., Lee T. M. (2012). Impaired social decision making in patients with major depressive disorder. Brain Behav. 2, 415–423. 10.1002/brb3.62 [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Ziaei M., Burianová H., Von Hippel W., Ebner N. C., Phillips L. H., Henry J. D. (2016). The impact of aging on the neural networks involved in gaze and emotional processing. Neurobiol. Aging 48, 182–194. 10.1016/j.neurobiolaging.2016.08.026 [DOI] [PubMed] [Google Scholar]
  82. Ziaei M., Ebner N. C., Burianová H. (2017). Functional brain networks involved in gaze and emotional processing. Eur. J. Neurosci. 45, 312–320. 10.1111/ejn.13464 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials


Articles from Frontiers in Human Neuroscience are provided here courtesy of Frontiers Media SA

RESOURCES