Skip to main content
Brain logoLink to Brain
. 2024 Feb 14;147(9):3018–3031. doi: 10.1093/brain/awae050

Anatomo-functional basis of emotional and motor resonance elicited by facial expressions

Maria Del Vecchio 1, Pietro Avanzini 2, Marzio Gerbella 3, Sara Costa 4, Flavia Maria Zauli 5, Piergiorgio d’Orio 6, Elena Focacci 7, Ivana Sartori 8, Fausto Caruana 9,
PMCID: PMC12007602  PMID: 38365267

Abstract

Simulation theories predict that the observation of other’s expressions modulates neural activity in the same centres controlling their production. This hypothesis has been developed by two models, postulating that the visual input is directly projected either to the motor system for action recognition (motor resonance) or to emotional/interoceptive regions for emotional contagion and social synchronization (emotional resonance). Here we investigated the role of frontal/insular regions in the processing of observed emotional expressions by combining intracranial recording, electrical stimulation and effective connectivity.

First, we intracranially recorded from prefrontal, premotor or anterior insular regions of 44 patients during the passive observation of emotional expressions, finding widespread modulations in prefrontal/insular regions (anterior cingulate cortex, anterior insula, orbitofrontal cortex and inferior frontal gyrus) and motor territories (Rolandic operculum and inferior frontal junction). Subsequently, we electrically stimulated the activated sites, finding that (i) in the anterior cingulate cortex and anterior insula, the stimulation elicited emotional/interoceptive responses, as predicted by the ‘emotional resonance model’; (ii) in the Rolandic operculum it evoked face/mouth sensorimotor responses, in line with the ‘motor resonance’ model; and (iii) all other regions were unresponsive or revealed functions unrelated to the processing of facial expressions. Finally, we traced the effective connectivity to sketch a network-level description of these regions, finding that the anterior cingulate cortex and the anterior insula are reciprocally interconnected while the Rolandic operculum is part of the parieto-frontal circuits and poorly connected with the former.

These results support the hypothesis that the pathways hypothesized by the ‘emotional resonance’ and the ‘motor resonance’ models work in parallel, differing in terms of spatio-temporal fingerprints, reactivity to electrical stimulation and connectivity patterns.

Keywords: stereo-electroencephalography, anterior insula, Rolandic operculum, mirror neurons, emotional expressions, facial mimicry


Del Vecchio et al. use stereo-EEG to shed light on the role of the frontal lobes in the processing of observed emotional expressions. They test the two predominant theories of face processing by combining intracranial recording, electrical stimulation, and effective connectivity tracing in surgical patients.

Introduction

The observation of emotional facial expressions triggers a cascade of neural activations beyond the occipito-temporal face network,1,2 involving multiple frontal areas. Over the years, ‘simulation theories’ tried to account for these activations arguing that other’s expressions resonate in the same regions controlling their production, and that such resonance is likely mediated by a mirror-neuron system for facial expressions, similar to the one classically described for hand actions.3-8 This hypothesis is substantiated, at the behavioural level, by the automatic tendency to respond to emotional expressions with unobservable or mild facial changes.9,10

At the system neuroscience level, this perspective has been modelled by two simulationist models, which postulated that the matching between observed and executed emotional expressions occurred either within motor/premotor areas7,11 (motor resonance) or emotional/interoceptive regions12,13 (emotional resonance). Although similar at first glance, the hypothesis that the observation/execution matching takes place in a motor rather than emotional/interoceptive territories has repercussions on its functional contribution to cognitive processes, spanning from action recognition through its covert replication8 (motor resonance) to social synchronization and emotional contagion14 (emotional resonance). Moreover, even assuming that both systems are at work,15 it would remain to be clarified their possible connections or whether the two systems are hierarchically independent.

To shed light on the role of frontal/insular regions in the processing of observed emotional expressions, and to investigate the motor and emotional resonance models of face processing, the present study takes a three-pronged approach that relies on three different tools provided by invasive investigations in drug-resistant epileptic patients, namely intracranial recording, high-frequency electrical stimulation and effective connectivity.

First, we performed direct measurements of the neural activity elicited by the passive observation of static emotional (smiling and fearful) expressions, compared to neutral ones, by recording intracranial event-related potentials (iERPs) from the entire frontal and insular cortices in 44 subjects (5663 recording sites). Such a high sampling allowed us to obtain unique spatio-temporal information on the processing of faces in prefrontal, premotor and anterior insular territories. Despite motor/emotional resonance and facial mimicry may be facilitated by the observation of dynamic (rather the static) expressions,16,17 and frontal activations are more commonly found during explicit emotion recognition (rather than passive observation) tasks,18 the passive observation of static expressions allowed us to implement a highly-controlled and time-locked experiment and, at the same time, to use the same methodological choices adopted in the classical studies of facial mimicry.9

Once the responding sites of interest were identified, we causally explored the functional relevance of each site by reviewing the behavioural and subjective effects elicited by their high-frequency electrical stimulation (HF-ES). Such information is crucial to establish whether the visual information on others’ expressions is projected to a territory involved in face motor functions (as predicted by ‘motor resonance’), to emotional/interoceptive territories (as predicted by ‘emotional resonance’), or to regions unresponsive to HF-ES and likely unrelated to both functions. In the third and last step, we capitalized on the possibility of studying the cortico-cortical evoked potentials (CCEPs) induced by single-pulse electrical stimulation (SPES) to trace the effective connectivity of the responding regions and to provide a network-level description of the involved regions.

Our results make a case for a new hypothesis that the observation of emotional expressions leads to the concomitant modulation of two independent streams of information, differing in terms of spatio-temporal fingerprints, reactivity to electrical stimulation and connectivity patterns.

Materials and methods

Participants

The study involves a cohort of 44 patients affected by drug-resistant epilepsy undergoing stereo-electroencephalographic (SEEG) implantation of intracerebral electrodes as part of the presurgical evaluation19,20 at the ‘Claudio Munari’ Center for Epilepsy Surgery, Niguarda Hospital, Milan, Italy, between May 2015 and April 2021. Investigations included left hemisphere (n = 17), right hemisphere (n = 20), and bilateral (n = 7) implantations. All these patients were recruited for a perceptual task, consisting of passive observation of static emotional expressions (see later).

The selection of patients has been submitted to a series of stringent precautionary measures. The inclusion/exclusion criteria are the same employed in our previous studies.21 The present study received the approval of the Ethical Committee of Niguarda Hospital (ID 939-12.12.2013). Patients were fully informed of the recording procedures and signed informed consent to participate in the study, according to the Declaration of Helsinki.22

Electrode implantation and anatomical reconstruction

SEEG electrodes were implanted only for clinical purposes. The investigated hemispheres, the location and the number of explored sites were based on hypotheses about the seizure onset zone (SOZ) derived clinical history and examination, non-invasive long-term video-EEG monitoring, and neuroimaging.20,23 Each subject underwent brain MRI (Achieva 1.5 T, Philips Healthcare) and CT (O-arm 1000 system, Medtronic) to acquire appropriate sequences for SEEG planning. The duration of the SEEG investigation was based only on clinical needs. Placement of intracerebral electrodes was performed under general anaesthesia by means of a robotized passive tool-holder (Neuromate, Renishaw Mayfield SA). A variable number of platinum–iridium semi-flexible multi-lead intracerebral electrodes, with a diameter of 0.8 mm, a lead length of 2 mm, an inter-leads distance of 1.5 mm and a maximum of 18 leads per electrode (Microdeep intracerebral electrodes, D08, Dixi Medical) were placed and fixed. After implantation, a fine cone-beam CT data set was acquired by using the O-arm and co-registered with the T1-weighted 3D MR image to verify the actual position of the electrodes. The anatomical reconstruction procedure has been described in previous studies from our group.24

Intracranial recordings: experimental paradigm

During the study, the patient sat approximately 60 cm away from the laptop display where the stimuli were presented. Recordings were obtained in a dimly light, quiet room. Stimuli consisted of static grey-scale images of emotionally expressive faces (five women and three men, adults and of the same ethnicity as the participants) depicting two emotional expressions, i.e. smiling and fearful, and a neutral one. Images were taken from Ekman and Friesen’s25 set of pictures of facial affect, i.e. the same stimuli used by Dimberg9 to elicit facial mimicry. While there is a broad consensus that the emergence of facial mimicry is facilitated by the observation of dynamic expressions, static expressions are processed more accurately and quickly than dynamic expressions when the presentation time is brief.17 In addition, static emotional expressions activate motion-sensitive regions, as they imply biological motion.26 The pictures used are referred to as C, GS, JJ, MF, MO, NR, PF and WF in the dataset from Ekman and Freisen.25 An elliptic mask was fitted to reveal the face itself while hiding hair and ears (Fig. 1A). Each stimulus was presented 54 times in a random order. The duration of each stimulus was 500 ms, with a 500 ms interval between stimuli.

Figure 1.

Figure 1

Experimental stimuli and sampling density. (A) The experimental paradigm was composed of static images depicting human faces and followed by a fixation cross. Facial expressions include positive, negative and neutral expressions. (B) The site sampling density is shown on the inflated surface of fs_LR brain template. The colour scale indicates the number of leads within a disk of 2 cm of radius and centred on each node of the mesh. Black boarders indicate are based on the Lausanne2008 (resolution 60) parcellation. The Lausanne2008 (resolution 60) parcellation schema is illustrated on the inflated surface of the FS_LR brain template. AI = anterior insula; Amy = amygdala; cACC = caudal anterior cingulate cortex; cMFC = caudal middle frontal gyrus; IFG = inferior frontal gyrus; IFGop = IFG pars opercularis; IFGorb = IFG pars orbitalis; IFGtri = IFG pars triangularis; IFJ = inferior frontal junction; Lat OFC1 = lateral orbitofrontal cortex 1; med OFC1 = medial orbitofrontal cortex; PI = posterior insula; rMFG = rostral middle frontal gyrus; rACC = rostral anterior cingulate cortex; RO = Rolandic operculum; SF4 = superior frontal 4; vmPFC = ventromedial prefrontal cortex.

Intracranial recordings: quantification and statistical analysis

Data processing

During the experiments, continuous SEEG was recorded with a 1000 Hz sampling rate by means of a 192 channel-EEG device (EEG-1200 Neurofax, Nihon Kohden®). The channels of interest were referred to a lead in a neutral white matter (neutral reference). A band-pass filter (0.015–300 Hz) was applied to avoid any aliasing effect. Each trial was epoched with a (200 ms, 700 ms) time window, with respect to the onset of the stimulus. Recordings were visually inspected by clinicians to ensure the absence of artefacts or epileptic activity. Channels with pathological EEG activity, when present, were discarded. Although this procedure cannot completely exclude the presence of interictal pathological high-frequency oscillations, not always detectable by visual inspection and potentially biomarkers of the epileptic zone,27,28 the fact that each region was sampled on different patients and different electrodes allowed us to make this possibility negligible.

Analyses were focused on iERPs that, compared to other types of SEEG signals (e.g. gamma-band power), are more sensitive and able to better identify modulations of both excitatory and inhibitory neural activity. IERPs were computed for all leads located in the grey matter, using a subtractive baseline correction applied versus the prestimulus interval (−200/0). Data analysis was carried out by using EEGlab toolbox (freely available at https://sccn.ucsd.edu/eeglab/index.php) to implement data segmentation, time-frequency decomposition and averaging across epochs.

Statistical analysis

A preliminary analysis was performed to identify responsive leads among the ones localized in the cortical grey matter. For each lead, we evaluated the significance of the iERPs elicited by each of the two experimental conditions (i.e. smiling and fearful expressions) after dividing each time window in 20 ms long time-bins following the stimulus onset, by applying a t-test to post-event bins versus pre-event baseline (α = 0.05). To decrease the false-positive ratio, only leads with significant responses in at least four consecutive time bins were designated as responsive. Considering that the responsiveness to the presented stimuli can be accounted for by both high-level aspects of face processing and low-level visual features of the stimulus, this result does not tell of the contribution of the responsive regions to the processing of emotional expressions. All leads showing a significant iERP in at least one experimental condition (smiling and fearful) were subsequently considered.

Discrimination between emotional and neutral expressions

To assess whether the response to each of the two emotional expressions (smiling and fearful) was significantly different from that elicited by neutral expressions, we performed two distinct two-way repeated measures ANOVAs using Condition (emotional, neutral) and Time (35 adjacent 20-ms time bins in the 0 ms, 700 ms window) as within-subject factors. For each lead showing a significant interaction and at least one significant main effect, post hoc analysis was conducted by means of a paired t-test according to a planned comparison design. A third two-way repeated measures ANOVA using the same criteria compared smiling and fearful faces, and was applied to all leads located in the frontal/insular regions and showing significant iERP in one of the two experimental conditions (smiling and fearful expressions). Data processing and statistical analysis were performed by using in-build functions implemented in MATLAB 2015a or EEGLAB.29

Continuous maps

To provide a continuous view of the topographic pattern of active leads, we built a circular mask based on the geodesic distance between two cortical points (i.e. the minimum pathway within the grey matter connecting the source and the target nodes). For each cortical node, we defined the nodes within a 1-cm geodesic distance from the original node and weighted the contribution of each node by a sigmoid function. Node weight was defined as a logistic function with unitary amplitude, a steepness of 2 and a midpoint at 7.5 mm. As a result, each node of the cortical mesh was associated with a collection of surrounding nodes, with all nodes within 5 mm of the origin maximally weighted, whereas those between 5 and 10 mm were gradually reduced in weight, to avoid edge effects. By this approach, we computed three different functional variables:24,30,31

  1. Cortical sampling density, i.e. the number of explored leads per cm2, using the fixed number (7) of nodes per lead and the average surface of a disk.

  2. Overall responsiveness, i.e. the number of responsive nodes as a percentage of the number of explored nodes within a disk. Given the fixed number of nodes representing a lead (7), this variable is equivalent to the number of responsive leads as a percentage of the number of explored leads within a disk. This variable provides an overall picture of the cortical responsiveness, ranging from 0% to 100%, directly comparable to neuroimaging studies. Data were thresholded at 10% to exclude the contribution of sparsely responsive regions.

  3. Relative responsiveness, i.e. the proportion of leads significantly selective to emotional expressions, of the overall number of responding leads. These maps provide a topographic picture of the frontal/insular selectivity of the emotional expressions. Data were thresholded at 10% to exclude the contribution of sparsely responsive regions.

Localization and region of interest analysis

The regional responsivity was assessed by (i) parcelling the entire cortical mantle according to the Lausanne2008 (resolution 60) template, which subdivides the entire brain into 129 different cortical and subcortical structures32,33 (Fig. 1); and (ii) estimating, for each region of the template, the percentage of leads showing a significant effect in the ANOVAs over the total number of investigated leads. The selection of this template was a compromise between different needs, as the more narrow-meshed a parcellation is, the more likely it is that effective connectivity data between two specific areas are missing. Frontal/insular regions showing a significant effect in at least 10% of the overall number of recording leads, evaluated on at least four leads, were selected for the study of the effective connectivity (see later). Results for the occipito-temporal network and parietal regions will not be discussed in depth due to our focus on motor/emotional resonances and because of their extensive treatment in previous human intracranial recordings investigations.34-41 Sampling and results relative to these regions are, however, reported in Supplementary Table 1.

High-frequency electrical stimulation

After the recording of spontaneous seizures, HF-ES was performed through the electrodes in many cerebral structures, aimed at both inducing seizures and brain mapping. Bipolar HF-ES of pairs of adjacent leads was carried out by means of biphasic rectangular stimuli of alternating polarity (frequency: 50 Hz; pulse width: 0.5–1 ms; duration: 5 s, current intensity: up to 4 mA). Stimulations were delivered while patients were maintaining the Mingazzini position and speaking aloud, to evaluate upper limbs movements, speech arrest and other behavioural modifications. All the elicited responses were video-recorded and prospectively stored in clinical report documents. Here we collected the available results obtained by HF-ES of the frontal/insular sites showing a significant interaction between emotional (smiling or fearful) and neutral expressions. The aim of this analysis was to better characterize the functional role of the sites modulated by the visual stimulus, and it does not exclude that other responses can be evoked by HF-ES of other frontal/insular regions not activated by the visual task.

Based on our previous work,42,43 we characterized the type of clinical response according to the following categories: sensorimotor manifestation, interoceptive sensation, emotional behaviour, language impairments, unspecific subjective responses and unresponsive sites.

Effective connectivity

The study of the effective connectivity was based on the open-source dataset Functional Brain Tractography Project f-tract (https://f-tract.eu/),44,45 which reports large-scale human brain connectivity maps based on CCEPs recorded in several hundreds of SEEG patients following SPES. The f-tract dataset is adapted to different atlas releases, including the Lausanne2008-60 atlas that we used to compute the regional responsivity in our study.

To study the internal connectivity of the frontal/insular regions involved in the processing of emotional expressions, where most of the responding leads were located, we focused on the data reporting the ‘afferent probability’ and the ‘efferent probability’, that is, the probability of afferent and efferent connectivity between the various regions of interest (ROIs) of the template. As described before, we considered only the regions showing a significant interaction in either positive versus neutral or negative versus neutral comparisons. For each ROI, a map of the afferent and the efferent probability was created. In addition, we displayed the matrix of probability across the regions of interest, and we grouped the distances among patterns of probability using a hierarchical binary cluster tree.

Results

Intracranial recordings

Frontal/insular responsiveness to emotional expressions

Recordings were obtained from 5663 recording leads [Left (L) = 2391; Right (R) = 3272] located in the cortical grey matter, with 2147 (L = 857; R = 1290) exploring the prefrontal, premotor and anterior insular cortices (see Fig. 1B for sampling density maps). Significant iERPs were elicited in ∼20% of these regions, during either smiling or fearful conditions (L = 176, R = 252 and L = 192, R = 215, respectively). Not surprisingly, the percentage of significant leads in these regions was lower than the overall percentage of significant leads recorded over the entire cortical sheet, i.e. including occipital and temporal regions (∼30%, i.e. L = 683, R = 1008 for smiling and L = 633, R = 993 for fearful expressions; Supplementary Fig. 1). Only leads located in prefrontal, premotor and anterior insular cortices showing significant responsiveness were subjected to further analyses.

Selectivity to smiling expressions

The two-way repeated measures ANOVA comparing smiling and neutral faces showed significant interaction, and at least one significant main effect, in 159 of 428 frontal/insular responding leads (L = 75/176; R = 84/252). A chi-squared test suggested the lack of significant lateralization (χ2 = P > 0.05). Interestingly, the percentage of significant leads in these frontal/insular regions (37.1%) was higher than the overall percentage of significant leads recorded over the entire cortical sheet, i.e. including occipital and temporal regions (25.5%; i.e. 427 of 1691 responding leads; see also Supplementary Table 1). Comparing these data with those obtained on the responsiveness (above), it can be deduced that, compared to occipital and temporal regions, frontal/insular regions have a higher selectivity for emotional faces, despite an overall lower responsiveness to the face stimulus.

The regional selectivity was assessed by parcelling the entire cortical mantle according to the Lausanne2008-60 template32 and estimating, for each region of the template, the percentage of leads showing a significant effect in the ANOVAs. Frontal/insular regions in which the significant number of leads is >10% of the recording leads include four territories: (i) the anterior cingulate region, where significant responses were bilaterally observed in the caudal anterior cingulate cortex (cACC), partially invading the territory of the rostral ACC (rACC) and in the adjacent ventromedial prefrontal cortex (vmPFC; labelled as ‘superior frontal 1’ in the Lausanne2008-60 template); (ii) the orbitofrontal region, that includes the lateral orbitofrontal cortex (OFC)1 and lateral OFC2 ROIs of the Lausanne2008-60 template, with a rightward lateralized reactivity; (iii) the dorsolateral prefrontal region, with significant effects in the left pars triangularis of the inferior frontal gyrus (IFG) and the bilateral anterior insula (AI), i.e. IFG-pt/AI henceforth; and finally, (iv) we found a high number of leads having significant reactivity in two premotor regions that, for reasons related to the borders of the parcellation provided by the Lausanne2008-60 template, failed to reach the 10% threshold: (a) in the bilateral Rolandic operculum (RO; located in the ventral part of the postcentral 3 region of the Lausanne2008-60) significant modulation was found only in the ∼6% of the leads, given the unresponsiveness of the leads located in the dorsal part of the same ROI; and (b) bilaterally in between the inferior frontal junction (IFJ), where the leads with significant modulation were distributed over three distinct ROIs of the Lausanne2008-60 parcellation (Fig. 2 and Table 1). Results relative to the occipito-temporal and parietal regions are reported in Supplementary Table 1.

Figure 2.

Figure 2

Prefrontal, premotor and insular regions activated by emotional expressions. Proportion of sites showing significant differences between emotional and neutral responses, of the overall number of the responsive sites, is plotted on the fs_LR brain template. The color scale indicates the percentage of responsive sites within a disk 1 cm in radius and centred on each node of the mesh.

Table 1.

Regions responding to the observation of positive and negative expressions

Smiling versus Neutral Fearful versus Neutral
Right Left Lateral Right Left Lateral
ROI Resp. leads % Resp. leads % χ2 Resp. leads % Resp. leads % χ2
ACC rostral 4 (31) 13% a 4 (18) 22% a P > 0.05 4 (31) 13% a 5 (18) 28% a P > 0.05
ACC caudal 7 (44) 16% a 8 (31) 26% a P > 0.05 2 (44) 5% 0 (31) 0%
Superior frontal_1 6 (61) 10% a 2 (35) 6% P > 0.05 1 (61) 2% 5 (35) 14% a P < 0.05 a
Superior frontal_2 2 (68) 3% 2 (26) 8% 1 (68) 1% 0 (26) 0%
Superior frontal_3 1 (62) 2% 3 (30) 10% 3 (62) 5% 2 (30) 7%
Superior frontal_4 5 (96) 5% 2 (48) 4% 0 (96) 0% 0 (48) 0%
Frontal pole 0 (0) 0% 1 (7) 14% 0 (0) 0% 3 (7) 43%
Lateral orbitofrontal_1 9 (26) 34% a 1 (35) 3% P < 0.05 a 0 (26) 0% 7 (35) 20% a P < 0.05 a
Lateral orbitofrontal_2 5 (43) 12% a 1 (29) 3% P > 0.05 4 (43) 9% a 6 (29) 21% a P > 0.05
Medial orbitofrontal 1 (63) 2% 3 (41) 7% 3 (63) 5% 4 (41) 10% a
Rostral middle frontal_1 5 (63) 8% 4 (36) 11% a P > 0.05 6 (63) 10% a 1 (36) 3%
Rostral middle frontal_2 1 (37) 3% 3 (12) 25% 1 (37) 3% 2 (12) 17%
Rostral middle frontal_3 1 (4) 25% 0 (13) 0% 0 (4) 0% 0 (13) 0%
Caudal middle frontal 2 (88) 2% 0 (62) 0% 5 (88) 6% 0 (62) 0%
IFG pars opercularis 2 (48) 4% 1 (46) 2% 4 (48) 8% 1 (46) 2%
IFG pars orbitalis 3 (36) 8% 3 (28) 11% 2 (36) 6% 0 (28) 0%
IFG pars triangularis 7 (104) 7% 12 (98) 12% a P > 0.05 16 (104) 15% a 15 (99) 15% a P > 0.05
Precentral_1 1 (28) 4% 1 (14) 7% 2 (28) 7% 0 (14) 0%
Precentral_2 3 (40) 8% 1 (23) 4% 0 (40) 0% 1 (23) 4%
Precentral_3 1 (26) 4% 2 (16) 13% 2 (26) 8% 3 (16) 19%
Precentral_4 0 (87) 0% 2 (59) 3% 1 (87) 1% 5 (59) 8%
Rolandic operculum 7 (117) 6% 4 (60) 7% 17 (117) 15% a 6 (60) 10% a P > 0.05
Postcentral_1 0 (6) 0% 0 (2) 0% 0 (6) 0% 0 (2) 0%
Postcentral_2 1 (26) 4% 1 (26) 4% 0 (26) 0% 1 (26) 4%
Paracentral lobule 1 (18) 6% 1 (15) 7% 0 (18) 0% 0 (15) 0%
Insula (anterior) 9 (100) 9% 14 (75) 19% a P > 0.05 18 (100) 18% a 8 (75) 11% a P > 0.05
Insula (posterior) 4 (95) 4% 8 (91) 9% 6 (95) 6% 5 (91) 5%

The table illustrates, for each investigated prefrontal, premotor and anterior insular regions, the results of the comparison between emotional versus neutral expressions. The parcellation is based on the Lausanne2008 (resolution 60) atlas. For both smiling versus neutral and fear versus neutral comparison, we report the number of significant leads over the number of recording leads (indicated within brackets), the percentage and the results of the chi-squared test. ACC = anterior cingulate cortex; IFG = inferior frontal gyrus; Resp = responsive; ROI = region of interest.

aRegions showing a significant effect in at least 10% of the overall number of recording leads, evaluated on at least four leads, are indicated in bold.

Selectivity to fearful expressions

The two-way repeated measures ANOVA comparing fearful and neutral faces showed a significant interaction, and at least one significant main effect, in 167 of 407 frontal/insular responding leads (L = 74 of 192; R = 93 of 215). A chi-squared test assessed the absence of any lateralization (χ2 = P > 0.05). As in the case of smiling expressions, the percentage of significant leads in the frontal/insular regions (41.0%) was higher than the overall percentage of significant leads recorded over the entire cortical sheet (29.8%; i.e. 484 of 1626 responding leads), confirming a higher frontal/insular selectivity for fearful faces, despite an overall lower responsiveness to the face stimulus.

Frontal/insular regions showing a significant effect in at least 10% of the recording leads were similar to those identified by the previous analysis, with few but important differences: (i) in the anterior cingulate region, we found the bilateral reactivity of the rACC, extending to the adjacent left vmPFC (superior frontal 1 in the Lausanne2008-60 template), but we did not find any modulation in the cACC; (ii) in the OFC the modulation was significantly leftward lateralized, rather than rightward lateralized as it was in the smiling versus neutral comparison; (iii) in the AI/IFG region, significant modulation was similar to the case of the smiling versus neutral comparison; and (iv) in the premotor regions we found significant modulation in the RO (in which reactivity reaches the 10% threshold in the fearful versus neutral comparison), and in between the caudal middle frontal gyrus and the IFJ (as in the previous case, this reactivity was distributed over three distinct ROIs of the Lausanne2008-60 parcellation, none of which reach 10% of significant leads; Fig. 2 and Table 1). Results relative to the occipito-temporal and parietal regions are reported in Supplementary Table 1.

Timing of frontal/insular reactivity during the processing of emotional expressions

For each frontal/insular lead showing a significant interaction, we identified the peak of the iERPs elicited by the observation of the smiling/fearful faces. Overall, we found peaks with latencies every 100 ms (±50), that is, at about 100 ms, 200 ms, 300 ms, 400 ms and 500 ms following stimulus onset (Fig. 3), with the majority of iERPs recorded from 300 ms onwards.

Figure 3.

Figure 3

Timing of intracranial event-related potential responses. Left: The figure illustrates, for each site showing significant differences between emotional and neutral responses, the peak latency following stimulus onset. Both left and right contacts are plotted on the left hemisphere of the inflated surface of the FS_LR brain template. Right: Representative intracranial event-related potentials (iERPs) for each of the five peak latencies. Numbers within each frame indicates, for each peak latency, the corresponding percentage of sites.

As shown in Fig. 3, the distribution of the iERP peaks on the cortical sheet does not reveal a sharp topography, although four interesting trends can be noted: (i) earliest peaks (100/200 ms) were more frequently recorded from the OFC and the vmPFC; (ii) peaks at 300 ms were predominant in the IFG-pt; (iii) late potentials (i.e. 400/500 ms) were mainly recorded from the RO and the ACC (rACC and cACC); and (iv) the AI exhibited a mixed pattern, characterized by the presence of both early and late iERPs.

High-frequency electrical stimulation

Electrical stimulations (HF-ES; see the ‘Materials and methods’ section) were performed on 273 (L = 139; R = 134) sites over the 327 sites (L = 160; R = 167) showing a significant effect in the ANOVA comparing emotional (smiling or fearful) and neutral expressions. The remaining sites were not stimulated, coherently with the fact that HF-ES is typically performed by selecting only a few sites for each specific anatomical structure. All the frontal/insular regions of interest have been stimulated by HF-ES.

Behavioural responses or subjective manifestations were elicited in 103 leads (L = 54; R = 49), while the remaining 62% of leads were unresponsive to HF-ES.

Sensorimotor manifestations were found in 50 sites (L = 27; R = 23), almost exclusively located in RO and IFJ. These manifestations affected different body districts, including the face/mouth region (L = 18; R = 11), the eye/neck region (L = 5; R = 3) and hand/arm (L = 4; R = 9). The stimulation of the RO elicited effects limited to the face/mouth region, such as mouth movement inhibition, dysarthric speech, paresthetic sensations to the mouth region and, remarkably, facial movements including smiling expressions. The stimulation of the IFJ elicited eyelid clonicity and contralateral versive eye/neck movements. Finally, hand/arm atony and paraesthesia were more rarely elicited and scattered across different regions (Fig. 4).

Figure 4.

Figure 4

Responses elicited by high-frequency electrical stimulation. Anatomical distribution of the sites whose electrical stimulation elicits behavioural and subjective responses belonging to the seven categories of response, and unresponsive contacts. Both left and right contacts are plotted on the left hemisphere of the inflated surface of the FS_LR brain template.

Interoceptive manifestations, emotional responses and other undescribed subjective phenomena were found in 41 sites (L = 20; R = 21), almost exclusively from AI and cACC. Interoceptive manifestations (L = 7; R = 7) consisted of nausea and bodily sensations and were predominantly elicited from the AI and, to a lesser extent, from the cACC. Emotional responses (L = 7; R = 5) included positive and negative valence emotions. Positive emotions consisted of mirthful laughter and were elicited only from the cACC. Negative emotions included fear and anxiety and were elicited from the AI. Unspecific subjective manifestations (L = 6; R = 9) such as dizziness, sense of confusion and mental blunting were elicited from the AI and the cACC. AI has the lowest stimulation thresholds (Supplementary Fig. 2), although it gives the most negative responses to stimulation. This data excludes the hypothesis that the negative valence of the evoked response is merely attributable to greater stimulation intensities.

Language impairments including tachyphemia and speech arrest were found in 12 sites (L = 7; R = 5), almost exclusively from the ventral part of IFG-pt.

Unresponsive stimulations were found in 171 sites (L = 86; R = 85). They were frequently obtained in all investigated structures but predominant in the vmPFC, the OFC and the dorsal aspect of the IFG-pt. Most unresponsive regions were typically stimulated at 3 mA (Supplementary Fig. 2), while the majority of 1 mA stimulations were clustered in responsive ones (i.e. AI and RO). It follows that the lack of response in the former cannot be traced back to factors related to the stimulation parameters. For this reason, we rated these sites as unrelated, or at least less directly involved, in either motor or emotional/interoceptive functions.

Effective connectivity

We extracted from the open-source dataset Functional Brain Tractography Project f-tract44,45 (https://f-tract.eu/) the ‘afferent and efferent probability values’ of all regions showing a significant interaction in at least 10% of the recording leads in either smiling versus neutral or fearful versus neutral comparisons, plus the precentral subfield where the majority of IFJ leads were located (‘precentral 3’ according to the Lausanne2008-60 template).

The resulting pattern of connectivity, depicted in Fig. 5A, gathers the analysed regions into four predominant clusters that match the above-described cingulate, orbitofrontal, dorsolateral prefrontal and premotor sets of regions. A first (cingulate) cluster encompasses the vmPFC and the rostral and caudate ACC, which represents an independent anatomical unit. A second (orbitofrontal) cluster is constituted by lateral OFC1 and OFC2, which is closely connected with a third (lateral prefrontal) cluster, grouping together the IFG-pt and the AI. A fourth (premotor) cluster is constituted by the RO and IFJ, corresponding to the ventral part of the postcentral_3 ROI and the precentral_3 ROI of the Lausanne2008-60 template, respectively. As illustrated in Supplementary Fig. 3A, the pattern of connectivity relative to the efferent probability values shows very similar segregation, supporting the hypothesis that these connections are predominantly bidirectional.

Figure 5.

Figure 5

Afferent effective connectivity assessed by cortico-cortical evoked potentials (CCEPs). Left: Matrix of the afferent connectivity probability across the regions showing significant differences between emotional and neutral responses. Data were based on the Functional Brain Tractography Project f-tract using the Lausanne2008 (resolutions 60) parcellation schema. Distances are grouped among patterns of probability using a hierarchical binary cluster tree. Right: Afferent connectivity probability (n ≥ 50%) estimated from CCEP data recorded following single-pulse electrical stimulation. The black arrow indicates the region from which CCEPs were recorded. AI = anterior insula; cACC = caudal anterior cingulate cortex; IFGtri = inferior frontal gyrus pars triangularis; IFJ = inferior frontal junction; Lat OFC1 = lateral orbitofrontal cortex 1; Lat OFC2 = lateral orbitofrontal cortex 2; rACC = rostral anterior cingulate cortex; RO = Rolandic operculum; vmPFC = ventromedial prefrontal cortex.

Figure 5B shows, for each region, the afferent connectivity with probability values >0.5. Three results deserve to be mentioned. First, the cingulate, the orbitofrontal and the lateral prefrontal clusters show dense reciprocal interconnections, while the premotor cluster appears to represent an island itself, poorly connected with all the prefrontal ones (values <40% in Fig. 5A). Second, among the regions of interest, the pattern of afferences of the premotor regions RO and IFJ is markedly different and predominantly dependent on the lateral parieto-frontal networks, rather than on prefrontal ones. Third, among the listed afferences, the only one involved in the visual processing of emotional expressions is the amygdala, projecting to both lateral OFC 1 and AI. Notably, all these observations also apply to the analysis of the efferent connectivity (Supplementary Fig. 3B).

Discussion

In the present study, we investigated the contribution of prefrontal, premotor and anterior insular regions to the processing of emotional expressions, by combining intracranial recording, electrical stimulation and effective connectivity. A specific aim of the study was to disentangle between the ‘motor resonance’ model—hypothesizing that emotional expressions activate the observer motor/premotor cortices7—and the ‘emotional resonance’ model—stating that others’ expressions directly activate the emotional/interoceptive system, bypassing the motor system.12

Intracranial recording revealed widespread reactivity in four sectors, three of which are located in prefrontal/insular regions (ACC, OFC and AI/IFG) and one occupying premotor territories (RO and IFJ). While these data show that all these regions take part in the processing of others’ emotional expressions, the electrical stimulation of the same sites indicated that only some of them are directly involved in the production of emotional/interoceptive or motor responses. More specifically, HF-ES revealed that (i) in ACC and AI, the stimulation of the sites responding to emotional expressions elicited emotions and interoceptive sensations, coherently with the prediction of the ‘emotional resonance’ model; (ii) in RO, the stimulation of the sites responding to emotional expressions elicited face/mouth sensorimotor responses, including facial expressions, coherently with the ‘motor resonance’ model; and (iii) other activated regions (i.e. OFC, IFG-pt, IFJ) were largely unresponsive to HF-ES, or revealed functions unrelated to the processing of emotional expressions, suggesting that their recruitment during the passive observation task should be explained by reasons other than those attributable to any (either emotional or motor) resonance mechanism. The study of the effective connectivity showed that the prefrontal/insular sectors, which include ACC and AI where emotional/interoceptive responses were elicited, are reciprocally interconnected. In contrast, the premotor cluster—which includes RO, where face/mouth sensorimotor responses have been elicited—is part of the parieto-frontal circuits and poorly connected with the prefrontal/insular areas.

These results support the hypothesis that the passive observation of emotional expressions recruits (at least) two streams of information, directed toward independent emotional and motor territories, which differs in terms of functional fingerprints and connectivity. All these aspects are discussed in turn below.

The anterior cingulate-insular system as a hub for ‘emotional resonance’

The ACC and the AI are the only regions that were both modulated by the observation of emotional expressions and reacted to HF-ES by eliciting emotional/interoceptive manifestations. These two results indicate that, in these regions, the visual information on other’s emotions is encoded by the same functional territories that contribute to the observer emotional state. The effective connectivity study revealed that the ACC and the AI are tightly interconnected, as also demonstrated by structural and functional connectivity data.46,47 Although these results indicate many functional affinities, the effect of the stimulation also reveals an interesting difference between the two regions. In the AI, the elicited effects were typically negative (e.g. fear, anxiety or nausea) while the stimulation of the cACC often elicited positive emotional behaviours (e.g. mirthful laughter).

The recruitment of the cACC during the observation of positive emotions is in line with previous evidence obtained by neuroimaging studies.16,48-50 Similarly, the evidence that the stimulation of the same territory elicits positive emotions and emotional behaviours was independently demonstrated by electrical stimulation studies.21,43,51-55 The discovery of a shared circuit for the perception and expression of positive emotions has been anticipated by two recent studies. First, in 2020 our group showed that the same ACC sites whose stimulation elicited mirthful laughter show an increase of the high-gamma band (50–150 Hz) during the observation of dynamic, but not static, expressions of happiness.21 The fact that, in the present study, also static expressions were able to elicit a response can be realistically attributed to the different type of signal analysed in the two studies, i.e. iERP versus high-gamma band. The differences between iERPs and high-gamma activity represents a crucial and still unsettled issue, but it is worth mentioning that previous intracranial studies have shown that visual stimuli—including faces—could modulate iERPs but not high-gamma activity in both high-order visual regions and non-visual cortical regions.37,56,57 Moreover, a new study demonstrated that 130 Hz continuous stimulation applied to the cACC biases the processing of static facial stimuli obtained by morphing happy, sad and neutral, leading participants to rate morphed faces as significantly happier during the stimulation of the cACC.53

Similar to the case of the ACC, the reactivity of the AI during the observation of positive and negative expressions is also in line with many neuroimaging6,48,58-60 and intracranial61 investigations. Concerning the electrical stimulation, previous studies demonstrated that, in line with our data, the HF-ES applied to the AI typically elicits unpleasant interoceptive sensations and negative emotions such as disgust and fear,62-65 albeit new data show that positive emotions can also be elicited from this region.66

Taken together, these data make the ACC and the AI the ideal neural substrate for an ‘emotional resonance’ system, that is, a system projecting the visual representation of other’s emotional displays into the corresponding emotional motor pathway, potentially facilitating facial mimicry and emotional contagion and possibly influencing the observer emotional state. The early timing of insular responses and its afferences from the amygdala, demonstrated by our effective connectivity data, makes the AI (along with the OFC) an ideal entry point of emotional visual information, also considering the early reactivity of the amygdala to emotional expressions.67,68 In contrast, the ACC is more likely involved in the behavioural response triggered by the observed emotional expression, that is, facial mimicry and emotional contagion.14 This hypothesis fits with a combined functional MRI-EMG study demonstrating that the spontaneous reaction to happy facial expressions, assessed by increased EMG activity in the zygomaticus major and orbicularis muscles, correlates with blood oxygen level-dependent activity in the ACC.16 Hence, also considering that the ACC direct projection to the facial nucleus,69,70 and its position at the top of the emotional motor pathway for facial expressions,71 it is tempting to speculate that the late timing of the iERPs recorded from the ACC (300–500 ms) could represent a motor potential elicited by the automatic facial mimicry, which is indeed typically recorded at the peripheral level at the same time window.9

The Rolandic operculum and its contribution to ‘motor resonance’

The RO differs from all the other areas activated by our stimuli for being the only region where HF-ES elicited face/mouth sensorimotor responses, including facial expressions. This result is fully in line with previous stimulation data from our and other groups.72,73 Its contribution to the control of volitional facial movements is substantiated by the evidence that its lesion is associated to impaired voluntary (but not spontaneous) movements of the lower face74-76 and that emotional facial expressions elicited by the stimulation of the RO are typically devoid of emotional content.21,73 The quintessential sensorimotor nature of this region is also confirmed by its connectivity with ventral premotor, primary motor and parietal regions, which clearly reveals the belonging of this area to the parieto-frontal circuits involved in the control of goal-related actions,3,77 as well as by its projections to the facial motor nuclei throughout the internal capsule.70

The recruitment of this region during the observation of emotional expressions has been occasionally observed also by neuroimaging investigations.78,79 Interestingly, a recent study investigated the shared circuits for facial emotion production and observation, reporting that this region was activated during the observation and execution of positive (happy) and negative (angry) expressions.80 Additional insights for the role of the RO in the processing of facial expressions come from lesion studies, reporting that RO lesions impair the explicit recognition of both positive and negative emotional expressions.81

Our study shows that the visual input reaches this region very late, never before 300 ms and usually at about 500 ms. As in the case of the ACC, such a late response is compatible with a motor potential evoked by the observed expression. Being facial mimicry recorded at the peripheral level at about 500 ms following the stimulus presentation,9 there must be at least one motor region active in a period compatible with such response—and the data described above make the RO, along with the ACC, the ideal candidate.

All these data pave the way for the hypothesis that emotional expressions trigger two parallel systems. One—controlled by the ACC-AI circuit—associates the observed expression to one’s own emotional/interoceptive representations, as predicted by the ‘emotional resonance’ model. The other—centred on the RO—superimposes the visual information on other’s expression on the observer motor representation, as hypothesized by the ‘motor resonance’ model. Both systems react very late to the observed stimulus and, at least in principle, both of them are in the position to explain the phenomenon of automatic facial mimicry.

Non-mirror regions responding to emotional expressions

The OFC, the pars triangularis of the IFG and the IFJ were activated by the passive observation of emotional expressions, coherently with the face-selective prefrontal patches described in the macaque.82,83 However, all these regions were largely unresponsive to HF-ES, or revealed functions unrelated to the processing of emotional expressions, indicating that the recruitment of these regions during our passive observation task should be explained by reasons other than those attributable to any emotional or motor resonance mechanism.

Orbitofrontal cortex as a potential entry point for emotional information

The OFC was activated by emotional expressions, in line with previous intracranial recordings showing that the primate OFC contains visual neurons reflecting face expression and face identity.84,85 Intracranial ERPs from the OFC were very early (100/200 ms), confirming in humans the same short latencies observed in the macaque OFC using single-neuron recordings.85 The study of its effective connectivity highlighted, besides a tight connection with both ACC and AI, afferences from the amygdala. However, the OFC was unresponsive to the electrical stimulation, showing that this region does not play a predominant role in initiating emotional behaviours or goal-directed actions directly. The lateralization of smiling (right) and fearful (left) expressions is an unexpected finding, possibly due to the rightward and leftward lateralization of rewarding stimuli and fear extinction processes, respectively.86,87 However, SEEG lateralization effects must be taken with a grain of salt as the majority of an inter-hemispheric difference derive from different groups of patients, thus adding on top of the investigated effects the inter-individual variability.88

Taken together, this set of evidence points to the OFC as a potential entry point of the information directed toward the emotional fields of ACC and AI. This interpretation is compatible with the shared view89 that the OFC gathers sensory information about the stimulus identity from the amygdala, evaluates its reward value and, on the basis of its rewarding/non-rewarding nature, guide behavioural choices by modulating other frontal regions including ACC and AI.90-92 Hence we hypothesize that the predominant source of information feeding the ‘emotional resonance’ system is not necessarily mediated by input from the temporal lobe, as in the case of the mirror-neuron system for action recognition, but rather by subcortical routes encompassing the superior colliculus-pulvinar-amygdala pathway,93-98 eventually reaching the anterior cingulate-insular system through the OFC.

Inferior frontal gyrus pars triangularis and inferior frontal junction are not part of the mirror-neuron system

The modulation of the IFG-pt and the IFJ during the processing of emotional expressions is typically interpreted as a sign of mirror-neuron activation.16,80,99-101 The results obtained by electrically stimulating these regions, and their effective connectivity, suggest taking this assumption with a grain of salt.

Concerning the IFG-pt, its effective connectivity is markedly different from what can be expected from a human homologue of the Macaque’s ventral premotor cortex, where mirror neurons have been originally described.102 Indeed, IFG-pt connections are mainly with prefrontal territories, rather than with parieto-frontal circuits as it is in the case of the Macaque mirror-neuron system.103 Most important, the electrical stimulation of this region gave no response, or produced language impairments, showing that the IFG-pt is not involved in the production of facial actions or in the generation of emotional responses.

Concerning the IFJ, this region is close and partially overlapping with the dorsal premotor cortex, commonly conceived as part of the action observation network.104-106 While its effective connectivity confirms its belonging to the parieto-frontal circuits hosting the mirror-neuron system, the stimulation of this territory elicited contralateral versive eye/neck movements and eyelid clonicity, making a strong case for the role of this region in the control of gaze/neck movements driven by attention-demanding stimuli. Remarkably, previous studies showed that the IFJ response to faces appears to be driven primarily by the eyes.107

Starting from the consideration that emotional expressions are a specific class of highly salient stimuli, one may expect that their observation should activate systems involved in reorienting attention when behaviourally relevant targets are detected. Hence, a possible interpretation is that IFG-pt and IFJ may, respectively, take part to the ‘ventral attentional systems’, specialized for the detection of behaviourally relevant stimuli, and the ‘dorsal attentional system’, involved in top-down orienting of attention.108-111 In line with this interpretation, these regions are usually activated during the explicit recognition of emotional facial expressions, rather than their passive views.18

Two mirror systems in the frontal lobe?

Fifteen years ago, proponents of the so-called ‘simulation theory’ argued that others’ emotional expressions trigger in the observer an automatic inner simulation and, at a behavioural level, the automatic facial mimicry of the expression.4,5 The proponents of the simulation theory also provided a limited number of mutually exclusive explanatory models, two of which are still considered by system neuroscientists. The ‘motor resonance’ model assumes that perceiving other’s emotions triggers a multistep serial processing starting with the simulation of the expression in the perceiver's motor system, where facial movements are represented, and continuing with the sorting of this information to brain regions involved in social cognitive functions.7,8,112 The ‘emotional resonance’ model, in contrast, hypothesizes that perceiving emotional displays directly activates the corresponding emotional regions, such as the ACC and the AI,12 bypassing premotor/motor regions.

Although they were often thought of as mutually exclusive models, the present data suggest that the two systems coexist, and that emotional displays simultaneously activate two partially independent networks, served by different streams of information and potentially involved in different functions.

The ‘emotional resonance’ system, anchored to the anterior cingulate-insular network, receives broad and coarse visual information from the OFC, which in turn is the recipient of input from the superior colliculus-pulvinar-amygdala pathway.89,93,98 Once projected to the ACC and AI, the visual input activates specific emotional/interoceptive representations. While the identification of its specific functions is of course beyond the potentialities of our study, our data are compatible with the hypothesis that this system may serve socio-emotional functions, such as the facilitation of emotional contagion, spontaneous mimicry and social synchronization.14,113 This interpretation is compatible with the evidence that lesions of the ACC and OFC impair not only the recognition of others’ emotions,114,115 but also the production of overt socio-emotional displays, deficits in social interactions and changes in social behaviour.116-118

The ‘motor resonance’ system, hosted by the RO, is in contrast part of the parieto-frontal circuits and represents the homologue of the classic mirror-neuron system for hand actions. The RO is not directly connected to any temporal region endowed with visual properties, hence it is likely that—as in the case of the mirror-neuron system for hand actions—the visual input reaches the RO from inferior parietal regions, that are in turn recipient of action-related visual information from superior and middle temporal regions.119 Following the evidence that lesions of the RO impair the explicit recognition of emotional expressions,81 one may speculate that this system contributes to action recognition strategies by exploiting one’s own motor representations as a heuristic strategy to understand others’ communicative intents.

The debate on motor and emotional resonance is inextricably linked to the debate on facial mimicry.9,10 Whether facial mimicry reflects an automatic motor response, or an emotional one, is still unsettled and this issue is in fact the focus of very recent studies.11 From a system neuroscience perspective, it has often been assumed that this automatic facial action could be controlled by the ventral part of the voluntary motor system, where facial movements are represented,7,8 but new data point at a wider network of cortical and subcortical regions.16 Even if we did not record EMG on our patients, our stimuli were the very same as Ekman and Friesen’s25 set of pictures of facial affect used by Dimberg9 in his seminal studies. It is therefore reasonable to assume that the participants of our study were also producing below threshold facial changes. Capitalizing on the late iERPs recorded in the ACC and the RO, on the result of their stimulation, revealing the occasional production of emotional facial expressions, and their well-known descending projections to the facial nuclei,69,120 we hypothesize that these regions are the best candidates for transforming observed expressions into produced expressions. If on the one hand, this consideration allows us to greatly limit the list of areas potentially involved in the genesis of this phenomenon, on the other hand, it does not still clarify whether facial mimicry is the output of motor resonance (via RO) or emotional resonance (via ACC), with all that follows in terms of its contribution to action recognition or social synchronization.

In conclusion, the present study contributes to a longstanding interdisciplinary debate on the simulation mechanisms active during face-to-face social interactions. Capitalizing on the results of intracranial recordings, electrical stimulation and effective connectivity studies, it suggests that the two predominant models in the literature on mirror neuron—namely, the emotional resonance and the motor resonance models—are not mutually exclusive. Rather, they coexist and work in parallel, but they differ in terms of neural bases (ACC-AI versus RO), visual pathways (amygdala-orbitofrontal versus temporo-parietal), content (emotional/interoceptive versus motor representations) and functions (socio-emotional synchronization versus action recognition).

Supplementary Material

awae050_Supplementary_Data

Acknowledgements

The authors are grateful to Luca Bonini and Marco Lanzilotto for comments on an earlier draft of the manuscript.

Contributor Information

Maria Del Vecchio, Institute of Neuroscience, National Research Council of Italy (CNR), 43125 Parma, Italy.

Pietro Avanzini, Institute of Neuroscience, National Research Council of Italy (CNR), 43125 Parma, Italy.

Marzio Gerbella, Department of Medicine and Surgery, University of Parma, 43125 Parma, Italy.

Sara Costa, Department of Medicine and Surgery, University of Parma, 43125 Parma, Italy.

Flavia Maria Zauli, ‘Claudio Munari’ Epilepsy Surgery Center, ASST GOM Niguarda, 20142 Milan, Italy.

Piergiorgio d’Orio, ‘Claudio Munari’ Epilepsy Surgery Center, ASST GOM Niguarda, 20142 Milan, Italy.

Elena Focacci, Department of Medicine and Surgery, University of Parma, 43125 Parma, Italy.

Ivana Sartori, ‘Claudio Munari’ Epilepsy Surgery Center, ASST GOM Niguarda, 20142 Milan, Italy.

Fausto Caruana, Institute of Neuroscience, National Research Council of Italy (CNR), 43125 Parma, Italy.

Data availability

The conditions of our ethics approval do not permit public archiving of individual anonymized raw data. Readers seeking access to the data should contact the corresponding author. Access will be granted to named individuals in accordance with ethical procedures governing the reuse of sensitive data. Specifically, requestors must sign a formal agreement confirming that (i) the user may not use the database for any non-academic purpose; (ii) the document must be signed by a person with a permanent position at an academic institute or publicly funded research institute. Up to five other researchers affiliated with the same institute for whom the signee is responsible may be named at the end of this document which will allow them to work with this dataset; and (iii) the user may not distribute the database or portions thereof in any way.

Funding

This study was supported by Horizon 2020 Framework Programme through grant agreement no. 935539 (Human Brain Project, SGA3) to PA. M.D.V. was supported by the ministero dell'università e della ricerca through a contribution from the Italian National Recovery and Resilience Plan (NRRP), M4C2, funded by the European Union – NextGenerationEU (Project IR0000011, CUP B51E22000150006, “EBRAINS-Italy”).

Competing interests

The authors report no competing interests.

Supplementary material

Supplementary material is available at Brain online.

References

  • 1. Pitcher  D, Ungerleider  LG. Evidence for a third visual pathway specialized for social perception. Trends Cogn Sci. 2021;25:100–110. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. Grill-Spector  K, Weiner  KS, Kay  K, Gomez  J. The functional neuroanatomy of human face perception. Annu Rev Vis Sci. 2017;3:167–196. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Rizzolatti  G, Cattaneo  L, Fabbri-Destro  M, Rozzi  S. Cortical mechanisms underlying the organization of goal-directed actions and mirror neuron-based action understanding. Physiol Rev. 2014;94:655–706. [DOI] [PubMed] [Google Scholar]
  • 4. Goldman  AI. Simulating minds. Oxford University Press; 2006. [Google Scholar]
  • 5. Goldman  AI, Sripada  CS. Simulationist models of face-based emotion recognition. Cognition. 2005;94:193–213. [DOI] [PubMed] [Google Scholar]
  • 6. Wicker  B, Keysers  C, Plailly  J, Royet  JP, Gallese  V, Rizzolatti  G. Both of us disgusted in my insula: The common neural basis of seeing and feeling disgust. Neuron. 2003;40:655–664. [DOI] [PubMed] [Google Scholar]
  • 7. Wood  A, Rychlowska  M, Korb  S, Niedenthal  P. Fashioning the face: Sensorimotor simulation contributes to facial expression recognition. Trends Cogn Sci. 2016;20:227–240. [DOI] [PubMed] [Google Scholar]
  • 8. Niedenthal  PM, Mermillod  M, Maringer  M, Hess  U. The simulation of smiles (SIMS) model: Embodied simulation and the meaning of facial expression. Behav Brain Sci. 2010;33:433–480. [DOI] [PubMed] [Google Scholar]
  • 9. Dimberg  U. Facial reactions to facial expressions. Psychophysiology. 1982;19:643–647. [DOI] [PubMed] [Google Scholar]
  • 10. Hess  U, Fischer  A. Emotional mimicry as social regulation. Personal Soc Psychol Rev. 2013;17:142–157. [DOI] [PubMed] [Google Scholar]
  • 11. Korb  S, Clarke  A, Massaccesi  C, Willeit  M, Silani  G. Facial mimicry is not modulated by dopamine D2/3 and opioid receptor antagonism. Psychopharmacology (Berl).  2023;240:2081–2091. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Gallese  V, Keysers  C, Rizzolatti  G. A unifying view of the basis of social cognition. Trends Cogn Sci. 2004;8:396–403. [DOI] [PubMed] [Google Scholar]
  • 13. Rizzolatti  G, Caruana  F. Some considerations on de Waal and Preston review. Nat Rev Neurosci. 2017;18:769. [DOI] [PubMed] [Google Scholar]
  • 14. Palagi  E, Celeghin  A, Tamietto  M, Winkielman  P, Norscia  I. The neuroethology of spontaneous mimicry and emotional contagion in human and non-human animals. Neurosci Biobehav Rev. 2020;111:149–165. [DOI] [PubMed] [Google Scholar]
  • 15. Caruana  F. Two simulation systems in the human frontal cortex? Disentangling between motor simulation and emotional mirroring using laughter. Cortex. 2022;148:215–217. [DOI] [PubMed] [Google Scholar]
  • 16. Rymarczyk  K, Żurawski  Ł, Jankowiak-Siuda  K, Szatkowska  I. Neural correlates of facial mimicry: Simultaneous measurements of EMG and BOLD responses during perception of dynamic compared to static facial expressions. Front Psychol. 2018;9:52. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Krumhuber  EG, Skora  LI, Hill  HCH, Lander  K. The role of facial movements in emotion recognition. Nat Rev Psychol. 2023;2:283–296. [Google Scholar]
  • 18. Dricu  M, Frühholz  S. Perceiving emotional expressions in others: Activation likelihood estimation meta-analyses of explicit evaluation, passive perception and incidental perception of emotions. Neurosci Biobehav Rev. 2016;71:810–828. [DOI] [PubMed] [Google Scholar]
  • 19. Munari  C, Hoffmann  D, Francione  S, et al.  Stereo-electroencephalography methodology: Advantages and limits. Acta Neurol Scand Suppl. 1994;152:56–67. Discussion 68–9. [DOI] [PubMed] [Google Scholar]
  • 20. Cossu  M, Cardinale  F, Castana  L, et al.  Stereoelectroencephalography in the presurgical evaluation of focal epilepsy: A retrospective analysis of 215 procedures. Neurosurgery. 2005;57:706–718. [PubMed] [Google Scholar]
  • 21. Caruana  F, Avanzini  P, Pelliccia  V, et al.  Mirroring other’s laughter. Cingulate, opercular and temporal contributions to laughter expression and observation. Cortex. 2020;128:35–48. [DOI] [PubMed] [Google Scholar]
  • 22. International Committee of Medical Journal Editors . Statements from the Vancouver Group. BMJ. 1991;302:1194. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Cardinale  F, Rizzi  M, Vignati  E, et al.  Stereoelectroencephalography: Retrospective analysis of 742 procedures in a single centre. Brain. 2019;142:2688–2704. [DOI] [PubMed] [Google Scholar]
  • 24. Avanzini  P, Abdollahi  RO, Sartori  I, et al.  Four-dimensional maps of the human somatosensory system. Proc Natl Acad Sci U S A. 2016;113:E1936–E1943. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Ekman  P, Friesen  W. Pictures of facial affect. Consulting Psychologists Press; 1976. [Google Scholar]
  • 26. Furl  N, Hadj-Bouziane  F, Liu  N, Averbeck  BB, Ungerleider  LG. Dynamic and static facial expressions decoded from motion-sensitive areas in the macaque monkey. J Neurosci. 2012;32:15952–15962. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Frauscher  B, Bartolomei  F, Kobayashi  K, et al.  High-frequency oscillations: The state of clinical research. Epilepsia. 2017;58:1316–1329. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Roehri  N, Bartolomei  F. Are high-frequency oscillations better biomarkers of the epileptogenic zone than spikes?  Curr Opin Neurol. 2019;32:213–219. [DOI] [PubMed] [Google Scholar]
  • 29. Delorme  A, Makeig  S. EEGLAB: An open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J Neurosci Methods. 2004;134:9–21. [DOI] [PubMed] [Google Scholar]
  • 30. Del Vecchio  M, Fossataro  C, Zauli  FM, et al.  Tonic somatosensory responses and deficits of tactile awareness converge in the parietal operculum. Brain. 2021;144:3779–3787. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31. Avanzini  P, Pelliccia  V, Lo Russo  G, Orban  GA, Rizzolatti  G. Multiple time courses of somatosensory responses in human cortex. Neuroimage. 2018;169:212–226. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Daducci  A, Gerhard  S, Griffa  A, et al.  The connectome mapper: An open-source processing pipeline to map connectomes with MRI. PLoS One. 2012;7:e48121. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Hagmann  P, Kurant  M, Gigandet  X, et al.  Mapping human whole-brain structural networks with diffusion MRI. PLoS One. 2007;2:e597. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34. Puce  A, Allison  T, Bentin  S, Gore  JC, McCarthy  G. Temporal cortex activation in humans viewing eye and mouth movements. J Neurosci. 1998;18:2188–2199. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Allison  T, Puce  A, Spencer  DD, McCarthy  G. Electrophysiological studies of human face perception. I: Potentials generated in occipitotemporal cortex by face and non-face stimuli. Cereb Cortex. 1999;9:415–430. [DOI] [PubMed] [Google Scholar]
  • 36. McCarthy  G, Puce  A, Belger  A, Allison  T. Electrophysiological studies of human face perception. II: Response properties of face-specific potentials generated in occipitotemporal cortex. Cereb Cortex. 1999;9:431–444. [DOI] [PubMed] [Google Scholar]
  • 37. Caruana  F, Cantalupo  G, Russo  GL, Mai  R, Sartori  I, Avanzini  P. Human cortical activity evoked by gaze shift observation: An intracranial EEG study. Hum Brain Mapp. 2014;35:1515–1528. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Babo-Rebelo  M, Puce  A, Bullock  D, et al.  Visual information routes in the posterior dorsal and ventral face network studied with intracranial neurophysiology and white matter tract endpoints. Cereb Cortex. 2022;32:342–366. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Hagen  S, Jacques  C, Maillard  L, Colnat-Coulbois  S, Rossion  B, Jonas  J. Spatially dissociated intracerebral maps for face- and house-selective activity in the human ventral occipito-temporal Cortex. Cereb Cortex. 2020;30:4026–4043. [DOI] [PubMed] [Google Scholar]
  • 40. Hagen  S, Lochy  A, Jacques  C, et al.  Dissociated face- and word-selective intracerebral responses in the human ventral occipito-temporal cortex. Brain Struct Funct. 2021;226:3031–3049. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41. Jacques  C, Rossion  B, Volfart  A, et al.  The neural basis of rapid unfamiliar face individuation with human intracerebral recordings. Neuroimage. 2020;221:117174. [DOI] [PubMed] [Google Scholar]
  • 42. Zauli  F, Del Vecchio  M, Russo  S, et al.  The web of laughter: Frontal and limbic projections of the anterior cingulate cortex revealed by cortico-cortical evoked potential from sites eliciting laughter. Philos Trans R Soc Lond B Biol Sci. 2022;377:20210180. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Caruana  F, Gerbella  M, Avanzini  P, et al.  Motor and emotional behaviours elicited by electrical stimulation of the human cingulate cortex. Brain. 2018;141:3035–3051. [DOI] [PubMed] [Google Scholar]
  • 44. Trebaul  L, Deman  P, Tuyisenge  V, et al.  Probabilistic functional tractography of the human cortex revisited. Neuroimage. 2018;181:414–429. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45. David  O, Job  AS, De Palma  L, Hoffmann  D, Minotti  L, Kahane  P. Probabilistic functional tractography of the human cortex. Neuroimage. 2013;80:307–317. [DOI] [PubMed] [Google Scholar]
  • 46. Cormie  MA, Kaya  B, Hadjis  GE, Mouseli  P, Moayedi  M. Insula-cingulate structural and functional connectivity: An ultra-high field MRI study. Cereb Cortex.  2023;33:9787–9801. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Touroutoglou  A, Hollenbeck  M, Dickerson  BC, Feldman Barrett  L. Dissociable large-scale networks anchored in the right anterior insula subserve affective experience and attention. Neuroimage. 2012;60:1947–1958. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48. Fusar-Poli  P, Placentino  A, Carletti  F, et al.  Functional atlas of emotional faces processing: A voxel-based meta-analysis of 105 functional magnetic resonance imaging studies. J Psychiatry Neurosci. 2009;34:418–432. [PMC free article] [PubMed] [Google Scholar]
  • 49. Kesler-West  ML, Andersen  AH, Smith  CD, et al.  Neural substrates of facial emotion processing using fMRI. Cogn Brain Res. 2001;11:213–226. [DOI] [PubMed] [Google Scholar]
  • 50. Killgore  WDS, Yurgelun-Todd  DA. Activation of the amygdala and anterior cingulate during nonconscious processing of sad versus happy faces. Neuroimage. 2004;21:1215–1223. [DOI] [PubMed] [Google Scholar]
  • 51. Bijanki  KR, Manns  JR, Inman  CS, et al.  Cingulum stimulation enhances positive affect and anxiolysis to facilitate awake craniotomy. J Clin Invest. 2019;129:1152–1166. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52. Caruana  F, Avanzini  P, Gozzo  F, Francione  S, Cardinale  F, Rizzolatti  G. Mirth and laughter elicited by electrical stimulation of the human anterior cingulate cortex. Cortex. 2015;71:323–331. [DOI] [PubMed] [Google Scholar]
  • 53. Metzger  BA, Kalva  P, Mocchi  MM, et al.  Intracranial stimulation and EEG feature analysis reveal affective salience network specialization. Brain. 2023;146:4366–4377. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54. Oane  I, Barborica  A, Chetan  F, et al.  Cingulate cortex function and multi-modal connectivity mapped using intracranial stimulation. Neuroimage. 2020;220:117059. [DOI] [PubMed] [Google Scholar]
  • 55. Pelliccia  V, Del Vecchio  M, Avanzini  P, Revay  M, Sartori  I, Caruana  F. 70 years of human cingulate Cortex stimulation. Functions and dysfunctions through the Lens of electrical stimulation. J Clin Neurophysiol. 2023;40:491–500. [DOI] [PubMed] [Google Scholar]
  • 56. Ahveninen  J, Lee  HJ, Yu  HY, et al.  Visual stimuli modulate local field potentials but drive no high-frequency activity in human auditory cortex. bioRxiv. [Preprint] doi: 10.1101/2022.07.18.500417 [DOI] [PMC free article] [PubMed]
  • 57. Vidal  JR, Ossandón  T, Jerbi  K, et al.  Category-specific visual responses: An intracranial study comparing gamma, Beta, alpha, and ERP response selectivity. Front Hum Neurosci. 2010;4:195. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58. Hennenlotter  A, Schroeder  U, Erhard  P, et al.  A common neural basis for receptive and expressive communication of pleasant facial affect. Neuroimage. 2005;26:581–591. [DOI] [PubMed] [Google Scholar]
  • 59. Phillips  ML, Young  AW, Senior  C, et al.  A specific neural substrate for perceiving facial expressions of disgust. Nature. 1997;389:495–498. [DOI] [PubMed] [Google Scholar]
  • 60. Pohl  A, Anders  S, Schulte-Rüther  M, Mathiak  K, Kircher  T. Positive facial affect—An fMRI study on the involvement of Insula and Amygdala. PLoS One. 2013;8:69886. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61. Krolak-Salmon  P, Hénaff  MA, Isnard  J, et al.  An attention modulated response to disgust in human ventral anterior insula. Ann Neurol. 2003;53:446–453. [DOI] [PubMed] [Google Scholar]
  • 62. Mazzola  L, Mauguière  F, Isnard  J. Electrical stimulations of the human Insula: Their contribution to the ictal semiology of insular seizures. J Clin Neurophysiol. 2017;34:307–314. [DOI] [PubMed] [Google Scholar]
  • 63. Jezzini  A, Caruana  F, Stoianov  I, Gallese  V, Rizzolatti  G. Functional organization of the insula and inner perisylvian regions. Proc Natl Acad Sci U S A. 2012;109:10077–10082. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64. Stephani  C, Fernandez-Baca Vaca  G, Maciunas  R, Koubeissi  M, Lüders  HO. Functional neuroanatomy of the insular lobe. Brain Struct Funct. 2011;216:137–149. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65. Isnard  J, Guénot  M, Sindou  M, Mauguière  F. Clinical manifestations of insular lobe seizures: A stereo-electroencephalographic study. Epilepsia. 2004;45:1079–1090. [DOI] [PubMed] [Google Scholar]
  • 66. Villard  C, Dary  Z, Léonard  J, et al.  The origin of pleasant sensations: Insight from direct electrical brain stimulation. Cortex. 2023;164:1–10. [DOI] [PubMed] [Google Scholar]
  • 67. Krolak-Salmon  P, Hénaff  MA, Vighetto  A, Bertrand  O, Mauguière  F. Early amygdala reaction to fear spreading in occipital, temporal, and frontal cortex: A depth electrode ERP study in human. Neuron. 2004;42:665–676. [DOI] [PubMed] [Google Scholar]
  • 68. Méndez-Bértolo  C, Moratti  S, Toledano  R, et al.  A fast pathway for fear in human amygdala. Nat Neurosci. 2016;19:1041–1049. [DOI] [PubMed] [Google Scholar]
  • 69. Morecraft  RJ, Louie  JL, Herrick  JL, Stilwell-Morecraft  KS. Cortical innervation of the facial nucleus in the non-human primate: A new interpretation of the effects of stroke and related subtotal brain trauma on the muscles of facial expression. Brain. 2001;124:176–208. [DOI] [PubMed] [Google Scholar]
  • 70. Gerbella  M, Pinardi  C, Di Cesare  G, Rizzolatti  G, Caruana  F. Two neural networks for laughter: A tractography study. Cereb Cortex. 2021;31:899–916. [DOI] [PubMed] [Google Scholar]
  • 71. Lauterbach  EC, Cummings  JL, Kuppuswamy  PS. Toward a more precise, clinically-informed pathophysiology of pathological laughing and crying. Neurosci Biobehav Rev. 2013;37:1893–1916. [DOI] [PubMed] [Google Scholar]
  • 72. Mălîia  MD, Donos  C, Barborica  A, et al.  Functional mapping and effective connectivity of the human operculum. Cortex. 2018;109:303–321. [DOI] [PubMed] [Google Scholar]
  • 73. Caruana  F, Gozzo  F, Pelliccia  V, Cossu  M, Avanzini  P. Smile and laughter elicited by electrical stimulation of the frontal operculum. Neuropsychologia. 2016;89:364–370. [DOI] [PubMed] [Google Scholar]
  • 74. Hopf  HC, Müller-Forell  W, Hopf  NJ. Localization of emotional and volitional facial paresis. Neurology. 1992;42:1918–1923. [DOI] [PubMed] [Google Scholar]
  • 75. Wild  B, Rodden  FA, Grodd  W, Ruch  W. Neural correlates of laughter and humour. Brain. 2003;126:2121–2138. [DOI] [PubMed] [Google Scholar]
  • 76. Mateos  V, Salas-Puig  J, Campos  DM, Carrero  V, Andermann  F. Acquired bilateral opercular lesions or Foix-Chavany-Marie syndrome and eating epilepsy. J Neurol Neurosurg Psychiatry. 1995;59:559–560. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77. Cipolloni  PB, Pandya  DN. Cortical connections of the frontoparietal opercular areas in the rhesus monkey. J Comp Neurol. 1999;403:431–458. [PubMed] [Google Scholar]
  • 78. Carr  L, Iacoboni  M, Dubeau  MC, Mazziotta  JC, Lenzi  GL. Neural mechanisms of empathy in humans: A relay from neural systems for imitation to limbic areas. Proc Natl Acad Sci U S A. 2003;100:5497–5502. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79. Leslie  KR, Johnson-Frey  SH, Grafton  ST. Functional imaging of face and hand imitation: Towards a motor theory of empathy. Neuroimage. 2004;21:601–607. [DOI] [PubMed] [Google Scholar]
  • 80. Krautheim  JT, Steines  M, Dannlowski  U, et al.  Emotion specific neural activation for the production and perception of facial expressions. Cortex. 2020;127:17–28. [DOI] [PubMed] [Google Scholar]
  • 81. van den Berg  NS, de Haan  EHF, Huitema  RB, et al.  The neural underpinnings of facial emotion recognition in ischemic stroke patients. J Neuropsychol. 2021;15:516–532. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82. Tsao  DY, Schweers  N, Moeller  S, Freiwald  WA. Patches of face-selective cortex in the macaque frontal lobe. Nat Neurosci. 2008;11:877–879. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83. Diehl  MM, Plakke  BA, Albuquerque  ER, Romanski  LM. Representation of expression and identity by ventral prefrontal neurons. Neuroscience. 2022;496:243–260. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84. Kringelbach  ML, Rolls  ET. Neural correlates of rapid reversal learning in a simple model of human social interaction. Neuroimage. 2003;20:1371–1383. [DOI] [PubMed] [Google Scholar]
  • 85. Rolls  ET, Critchley  HD, Browning  AS, Inoue  K. Face-selective and auditory neurons in the primate orbitofrontal cortex. Exp Brain Res. 2006;170:74–87. [DOI] [PubMed] [Google Scholar]
  • 86. O’Doherty  J, Kringelbach  ML, Rolls  ET, Hornak  J, Andrews  C. Abstract reward and punishment representations in the human orbitofrontal cortex. Nat Neurosci. 2001;4:95–102. [DOI] [PubMed] [Google Scholar]
  • 87. Gottfried  JA, Dolan  RJ. Human orbitofrontal cortex mediates extinction learning while accessing conditioned representations of value. Nat Neurosci. 2004;7:1144–1152. [DOI] [PubMed] [Google Scholar]
  • 88. Mercier  MR, Dubarry  AS, Tadel  F, et al.  Advances in human intracranial electroencephalography research, guidelines and good practices. Neuroimage. 2022;260:119438. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89. Rolls  ET. Emotion, motivation, decision-making, the orbitofrontal cortex, anterior cingulate cortex, and the amygdala. Brain Struct Funct. 2023;228:1201–1257. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90. Padoa-Schioppa  C, Conen  KE. Orbitofrontal cortex: A neural circuit for economic decisions. Neuron. 2017;96:736–754. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91. Howard  JD, Kahnt  T. To be specific: The role of orbitofrontal cortex in signaling reward identity. Behav Neurosci. 2021;135:210–217. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 92. Knudsen  EB, Wallis  JD. Taking stock of value in the orbitofrontal cortex. Nat Rev Neurosci. 2022;23:428–438. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 93. Tamietto  M, de Gelder  B. Neural bases of the non-conscious perception of emotional signals. Nat Rev Neurosci. 2010;11:697–709. [DOI] [PubMed] [Google Scholar]
  • 94. Gerbella  M, Caruana  F, Rizzolatti  G. Pathways for smiling, disgust and fear recognition in blindsight patients. Neuropsychologia. 2019;128:6–13. [DOI] [PubMed] [Google Scholar]
  • 95. Ferrari  PF, Gerbella  M, Coudé  G, Rozzi  S. Two different mirror neuron networks: The sensorimotor (hand) and limbic (face) pathways. Neuroscience. 2017;358:300–315. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 96. Tamietto  M, Pullens  P, De Gelder  B, Weiskrantz  L, Goebel  R. Subcortical connections to human amygdala and changes following destruction of the visual cortex. Curr Biol. 2012;22:1449–1455. [DOI] [PubMed] [Google Scholar]
  • 97. Méndez  CA, Celeghin  A, Diano  M, Orsenigo  D, Ocak  B, Tamietto  M. A deep neural network model of the primate superior colliculus for emotion recognition. Philos Trans R Soc B. 2022;377:20210512. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 98. McFadyen  J, Mattingley  JB, Garrido  MI. An afferent white matter pathway from the pulvinar to the amygdala facilitates fear recognition. Elife. 2019;8:e40766. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 99. Hsu  CT, Sato  W, Kochiyama  T, et al.  Enhanced mirror neuron network activity and effective connectivity during live interaction among female subjects. Neuroimage. 2022;263:119655. [DOI] [PubMed] [Google Scholar]
  • 100. Hamzei  F, Vry  MS, Saur  D, et al.  The dual-loop model and the human mirror neuron system: An exploratory combined fMRI and DTI study of the Inferior frontal gyrus. Cereb Cortex. 2016;26:2215–2224. [DOI] [PubMed] [Google Scholar]
  • 101. Li  B, Solanas  MP, Marrazzo  G, et al.  A large-scale brain network of species-specific dynamic human body perception. Prog Neurobiol. 2023;221:102398. [DOI] [PubMed] [Google Scholar]
  • 102. Gallese  V, Fadiga  L, Fogassi  L, Rizzolatti  G. Action recognition in the premotor cortex. Brain. 1996;119:593–609. [DOI] [PubMed] [Google Scholar]
  • 103. Borra  E, Gerbella  M, Rozzi  S, Luppino  G. The macaque lateral grasping network: A neural substrate for generating purposeful hand actions. Neurosci Biobehav Rev. 2017;75:65–90. [DOI] [PubMed] [Google Scholar]
  • 104. Caspers  S, Zilles  K, Laird  AR, Eickhoff  SB. ALE meta-analysis of action observation and imitation in the human brain. Neuroimage. 2010;50:1148–1167. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 105. Molenberghs  P, Cunnington  R, Mattingley  JB. Brain regions with mirror properties: A meta-analysis of 125 human fMRI studies. Neurosci Biobehav Rev. 2012;36:341–349. [DOI] [PubMed] [Google Scholar]
  • 106. Hardwick  RM, Caspers  S, Eickhoff  SB, Swinnen  SP. Neural correlates of action: Comparing meta-analyses of imagery, observation, and execution. Neurosci Biobehav Rev. 2018;94:31–44. [DOI] [PubMed] [Google Scholar]
  • 107. Chan  AWY, Downing  PE. Faces and eyes in human lateral prefrontal cortex. Front Hum Neurosci. 2011;5:51. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 108. Farrant  K, Uddin  LQ. Asymmetric development of dorsal and ventral attention networks in the human brain. Dev Cogn Neurosci. 2015;12:165–174. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109. Corbetta  M, Shulman  GL. Control of goal-directed and stimulus-driven attention in the brain. Nat Rev Neurosci. 2002;3:201–215. [DOI] [PubMed] [Google Scholar]
  • 110. Fox  MD, Corbetta  M, Snyder  AZ, Vincent  JL, Raichle  ME. Spontaneous neuronal activity distinguishes human dorsal and ventral attention systems. Proc Natl Acad Sci U S A. 2006;103:10046. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 111. Corbetta  M, Patel  G, Shulman  GL. The reorienting system of the human brain: From environment to theory of mind. Neuron. 2008;58:306–324. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 112. Chakrabarti  B, Bullmore  E, Baron-Cohen  S. Empathizing with basic emotions: Common and discrete neural substrates. Soc Neurosci. 2006;1:364–384. [DOI] [PubMed] [Google Scholar]
  • 113. Cuccio  V, Caruana  F. Motor simulation of facial expressions, but not emotional mirroring, depends on automatic sensorimotor abduction. In: Magnani  L, ed. Handbook of abductive cognition. Springer; 2022. [Google Scholar]
  • 114. Tsuchida  A, Fellows  LK. Are you upset? Distinct roles for orbitofrontal and lateral prefrontal cortex in detecting and distinguishing facial expressions of emotion. Cereb Cortex. 2012;22:2904–2912. [DOI] [PubMed] [Google Scholar]
  • 115. Heberlein  AS, Padon  AA, Gillihan  SJ, Farah  MJ, Fellows  LK. Ventromedial frontal lobe plays a critical role in facial emotion recognition. J Cogn Neurosci. 2008;20:721–733. [DOI] [PubMed] [Google Scholar]
  • 116. Hadland  K, Rushworth  MF, Gaffan  D, Passingham  R. The effect of cingulate lesions on social behaviour and emotion. Neuropsychologia. 2003;41:919–931. [DOI] [PubMed] [Google Scholar]
  • 117. Hornak  J, Bramham  J, Rolls  ET, et al.  Changes in emotion after circumscribed surgical lesions of the orbitofrontal and cingulate cortices. Brain. 2003;126:1691–1712. [DOI] [PubMed] [Google Scholar]
  • 118. Rudebeck  PH, Buckley  MJ, Walton  ME, Rushworth  MFS. A role for the macaque anterior cingulate gyrus in social valuation. Science. 2006;313:1310–1312. [DOI] [PubMed] [Google Scholar]
  • 119. Nelissen  K, Borra  E, Gerbella  M, et al.  Action observation circuits in the macaque monkey cortex. J Neurosci. 2011;31:3743–3756. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 120. Gothard  KM. The amygdalo-motor pathways and the control of facial expressions. Front Neurosci. 2014;8:43. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

awae050_Supplementary_Data

Data Availability Statement

The conditions of our ethics approval do not permit public archiving of individual anonymized raw data. Readers seeking access to the data should contact the corresponding author. Access will be granted to named individuals in accordance with ethical procedures governing the reuse of sensitive data. Specifically, requestors must sign a formal agreement confirming that (i) the user may not use the database for any non-academic purpose; (ii) the document must be signed by a person with a permanent position at an academic institute or publicly funded research institute. Up to five other researchers affiliated with the same institute for whom the signee is responsible may be named at the end of this document which will allow them to work with this dataset; and (iii) the user may not distribute the database or portions thereof in any way.


Articles from Brain are provided here courtesy of Oxford University Press

RESOURCES