Abstract
Several previous studies have interfered with the observer’s facial mimicry during a variety of facial expression recognition tasks providing evidence in favor of the role of facial mimicry and sensorimotor activity in emotion processing. In this theoretical context, a particularly intriguing facet has been neglected, namely whether blocking facial mimicry modulates conscious perception of facial expressions of emotions. To address this issue, we used a binocular rivalry paradigm, in which two dissimilar stimuli presented to the two eyes alternatingly dominate conscious perception. On each trial, female participants (N = 32) were exposed to a rivalrous pair of a neutral and a happy expression of the same individual through anaglyph glasses in two conditions: in one, they could freely use their facial mimicry, in the other they had to keep a chopstick between their lips, constraining the mobility of the zygomatic muscle and producing ‘noise’ for sensorimotor simulation. We found that blocking facial mimicry affected the perceptual dominance in terms of cumulative time favoring neutral faces, but it did not change the time before the first dominance was established. Taken together, our results open a door to future investigation of the intersection between sensorimotor simulation models and conscious perception of emotional facial expressions.
Subject terms: Psychology, Human behaviour
Introduction
Recognizing others’ emotions and affective states is one of the most extraordinary human abilities. In a purely evolutionary perspective, detecting and understanding—quickly and accurately—others’ emotions, desires and intentions clearly offers adaptive advantages and promotes affiliation, mating and parenting. For a long time, psychological and neuroscientific research has explored the cognitive and neural bases of this ability1–7. Recent neuroimaging research has provided converging evidence that the brain networks for facial expressions processing comprise several cortical and subcortical regions. These include the fusiform face area, the occipital face area, the superior temporal sulcus (regions of the core system of Haxby’s distributed model of face processing8–15), the insula, the amygdala, the inferior frontal gyrus16–21, and several other not strictly face-sensitive regions22.
In this context, motor (or sensorimotor) simulation models propose that the observer’s subthreshold motor simulation of the observed facial expression facilitates recognition and understanding of others’ congruent facial expression23–25. This theoretical view is supported by a substantial body of evidence. On one side it has been highlighted how brain regions supporting motor and somatosensory representations of facial expressions are involved in recognition of emotions in others26, such that, for instance, lesions of these regions are associated with emotion recognition deficits7,27 and, similarly, repetitive transcranial magnetic stimulation disrupting the right somatosensory6, 28 and the right primary motor29 cortices impairs some aspect of emotional face processing; on the other side there is evidence that muscular facial feedback incongruent with the observed expression causes a decrease in emotion recognition accuracy26,30–36, (but also see37–40). The rationale behind this last series of studies is that, if facial mimicry—which can be measured by electromyography41,42—is a manifestation of sensorimotor simulation triggered by the observation of others’ facial expressions, then an experimental manipulation aimed at interfering with it should consequently interfere with the simulation process itself and thus affecting the processing of facial expressions. For example, Ponari et al. (Experiment 135) required participants to identify the emotion (among the basic six categories) expressed by the faces presented one at a time while the production of their facial mimic patterns was manipulated through a Chinese chopstick (which participants had to keep horizontally between the teeth to prevent the movements of the lower portion of the face) or through two stickers near the inner portion of the eyebrows (which participants had to actively try to bring near through an active contraction of the frontal muscles). The results highlighted an impairment of the accuracy in the identification task as a function of the type of facial mimicry manipulation, i.e., that interfering with the muscles of the lower portion of the participants’ faces was associated with an impairment in the processing of happiness, disgust and fear, while that interfering with the muscles of the upper portion of the face was associated with an impairment of the processing of anger and fear.
While the behavioral evidence is accumulating to support the role of sensorimotor activity and facial mimicry in facilitating the recognition of others’ congruent emotional expressions, it remains unclear at what level of the visual processing of emotional faces the sensorimotor activity might exert its influence43,44. In this regard, a recent sensorimotor simulation model has proposed that sensorimotor signals may feedback to visual areas modulating the visual processing of emotional faces from early stages26,45.
Notably, in these theoretical and empirical frameworks an aspect that has been almost totally neglected is that concerning the possible role of facial mimicry in conscious perception of emotional facial expressions. The present study aimed at investigating precisely this aspect of facial mimicry/sensorimotor simulation using binocular rivalry, a paradigm widely used in the studies of visual conscious perception and its neural correlates46–49.
Under the ecological circumstances, slightly discordant visual inputs to the two eyes result in stable stereo experience (but also see50). When the dissimilarity exceeds a certain threshold, periods of perceptual dominance of one stimulus over the other stochastically alternate, such that one monocular image is dominantly consciously experienced while the other is suppressed and invisible47,51. This condition is called binocular rivalry (BR)52.
In the BR paradigm, participants are often required to report their content of consciousness in a continuous manner among several alternatives (see Fig. 1). The time series data of BR report is quite rich and has been dissected into at least three components, each of which can characterize the underlying neural and psychological processes from different perspectives. First, initial percept measures which of the rivaling stimuli dominates first in conscious percept, potentially characterizing any bias or advantage of one stimulus over the other. Second, onset resolution time measures the time of first initial percept button press, characterizing how long it takes the visual system to resolve perceptual ambiguity. In other words, it is an estimate of the time necessary for the brain to select (or inhibit) one percept over the other. These two BR components indicate the “winner” percept (i.e., initial percept) in the initial phase of unresolved competition (i.e., onset resolution time). A third component, which we call cumulative time (CT), measures instead the total time of dominance in awareness of one percept over the other, i.e. the periods of relatively stable resolution following the initial competition.
Previous research has used the BR paradigm to study conscious perception of emotional stimuli, but rarely distinguished the three components of the BR described above. For example, some focused on how emotional faces predominate over neutral faces in BR53–56. Yoon et al.56, expanding on the research by Alpers and Gerdes53, demonstrated that emotional faces, regardless of their valence, are perceived for longer than neutral faces, although the effect was particularly strong (89% increase in dominance) for positive facial expressions (i.e., happy) which predominated also over negative facial expressions (i.e., disgusted). Unfortunately, the authors did not report the other two measures and the data are not openly available for further research.
Crucially, it is unknown if observer’s facial mimicry can influence conscious perception of facial expressions in the BR paradigm. We propose that there can be at least two possible ways, which are not mutually exclusive, to affect the BR. First, facial mimicry can influence the initial ambiguity resolution. Second, facial mimicry may stabilize such representations once they are the current content of consciousness. The present experiment aimed at investigating precisely this aspect of facial mimicry using BR under facial mimicry manipulation with a neutral and an emotional (i.e., happy) expression of the same identity in rivalry. We tested the impact of facial mimicry manipulation on both ambiguity resolution (i.e., initial percept and onset resolution time) and stabilization of conscious contents (i.e., cumulative time). We manipulated participants’ facial mimicry by means of a chopstick to be held firmly between the lips for half the experiment, while in the other half the facial mimicry was free (the order of the two mimicry conditions was counterbalanced between participants; for similar manipulations see34,36,44,45,57,58). This use of the chopstick inhibits the activation of the zygomatic major, associated with mimicry of facial expression of happiness34,59.
With regard to mimicry manipulation, only one previous study60 has investigated the impact of the integration of proprioceptive information from the face on visual awareness of facial expression of emotions using a variant of BR, called breaking continuous flash suppression (b-CFS61–63). In the b-CFS paradigm, a mask with a high contrast dynamic pattern is presented to one eye, thereby effectively suppressing a stimulus of increasing intensity presented to the other eye. After a certain interval of time, the ocular dominance is reversed, and the previous suppressed stimulus becomes visible. Thus, this paradigm measures the time for a stimulus to access consciousness. In a series of three experiments, Korb and colleagues manipulated participants’ facial mimicry using a b-CFS paradigm by requiring them to voluntarily take facial expressions of frown vs. smile vs. relaxed (Experiments 1 and 2) or by measuring spontaneous electromyographic activity (Experiment 3). Overall, the results did not support the hypothesis of a modulation of visual consciousness on the basis of the integration of the proprioceptive activity of the face in conditions of congruence and incongruence with the observed emotion. That is, the time to break suppression did not differ between conditions. This suggests that facial mimicry does not influence the access to consciousness of emotional facial expressions. To note, b-CFS cannot inform on the potential role of facial mimicry in stabilizing conscious representations of facial expressions.
We expected that blocking/altering facial mimicry could interfere with the alternation in BR. In particular, we expected that interfering with the sensorimotor signal (i.e., by means of the active inhibition of facial mimicry) would (1) favor the neutral facial expressions and/or (2) interfere with happy expressions with respect to the condition of free mimicry by either biasing competition and ambiguity resolution at the early stage and/or stabilizing representations following initial resolution at the late stage. This last scenario seems more plausible in the light of the results reported by Korb and colleagues. To test the early account (i.e., ambiguity resolution), we analyzed the frequency and time of the first rivalry report, i.e. initial percept and onset resolution time, respectively64. To test the late account (i.e., stabilization), we analyzed the cumulative time. As an auxiliary although interesting aim, we wanted to replicate the few interesting results previously reported in the literature about an advantage in terms of predominance of emotional stimuli, and in particular of facial expressions of happiness when in rivalry with neutral facial expressions53,55,56.
Since previous studies have suggested that facial mimicry manipulations have a greater impact on female than male participants36,45 (for compatible results see also29), we decided to recruit exclusively female participants in order to maximize power. Therefore, on the basis of the hypotheses introduced in the previous paragraph, from a statistical point of view, the effect that we expected, both for the tests of the early and the late accounts, was an interaction between facial expression and mimicry manipulation, with (1) an advantage for neutral expressions under the condition of blocked mimicry when compared to free mimicry, and/or (2) a disadvantage for happy expressions under the condition of blocked mimicry when compared to free mimicry.
Finally, on an exploratory basis, in the present investigation we also tested the hypothesis of a relationship between alexithymic traits and awareness of emotional expressions. Alexithymia is defined as a difficulty in recognizing one’s own and others’ emotions65. We hypothesized that the different binocular rivalry metrics we analyzed could show a relationship with the levels of alexithymic traits, in the direction of the decreasing in conscious perception for emotional expressions with increasing levels of alexithymia.
Results
Ratings
Evaluation of valence and arousal were performed on individual stimuli at the end of each session. No effect of the session in terms of mimicry manipulation was observed (valence F(1, 27) = 2.28, p = 0.143; arousal F(1, 27) = 0.08, p = 0.775). Valence ratings differed according to a priori expectations, F(1, 27) = 330.02, p < 0.001. Neutral facial expressions were rated close to zero (M = − 0.63; SD = 0.98; range = − 3 to 3), which was more negative than happy (M = 2.15; SD = 1.12; range = − 3 to 3). Happy expressions were rated more positively than neutral expressions, t(27) = 18.16 p < 0.0001. Arousal ratings also differed according to a priori expectations, F(1, 27) = 56.54, p < 0.001. They were lower for neutral expressions (M = 2.74; SD = 1.54; range = 1–7) than for happy expressions (M = 4.97, SD = 1.63; range = 1–7). Neutral facial expressions were rated as significantly less arousing than happy facial expressions, t(27) = 7.52, p < 0.0001.
Testing the early effects of mimicry blocking (i.e., resolution of ambiguity)
For both initial percept (IP) and onset resolution time (ORT), the statistical models included the factors mimicry (free vs. blocked) and the first clear facial expression reported (i.e., IPs; happy vs. neutral), in line with the very definition of resolution of ambiguity between rivalrous percepts.
Initial percept (IP)
In terms of IPs, happy expressions were reported more frequently in both mimicry conditions (blocked: 304 trials; free: 317 trials) than neutral expressions (blocked: 128; free: 113). The odds ratio is not statistically significant, and the estimated value is 0.83 (β = − 0.184, SE = 0.159, 95% CI [− 0.50, 0.13], t − 1.15, p = 0.248).
Onset resolution time (ORT)
Due to the IP distribution, five participants were removed from analysis (i.e., they did not have all possible IPs). Significant differences were found for the mimicry condition independently of the first clear facial expression reported (i.e., IPs), F(1,22) = 6.38, p = 0.019 and for the IPs independently of mimicry condition, F(1,22) = 13.90, p < 0.001. See Fig. 2.
In general, the findings related to the IP and ORT metrics seem not to support the early account, that is, the inhibition of the zygomatic muscle did not influence the resolution of ambiguity in favor of neutral faces and/or at the expense of happy facial expressions. These results seem to be in line with the previous evidence provided by Korb et al.60.
To note, in terms of IP frequency and ORT, emotional faces showed an advantage when compared to neutral faces irrespective of the mimicry conditions.
Testing the late effects of mimicry blocking (i.e., stabilization of conscious contents)
For the cumulative time (CT) metric, the statistical model included the factors mimicry (free vs. blocked) and the reported content (happy vs. neutral vs. mixed).
A significant difference was observed for reported content independently of the mimicry condition, F(1.79, 48.26) = 69.18, p < 0.001. CT for happy facial expressions was longer than CT for neutral faces, t(54) = 9.78, p < 0.0001, and than CT for mixed percepts, t(54) = 10.54, p < 0.0001. This result is in line with previous findings53, 56. The main effect of mimicry did not yield to significant differences, F(1, 27) = 1.2, p = 0.283. In accordance with our hypothesis, CT for reported content showed an opposite trend as a function of the mimicry conditions, that is CT for neutral faces was expected to increase when mimicry was blocked (blocked mimicry condition: M = 3.39 s; SD = 4 s; free mimicry condition: M = 2.69 s; SD = 3.28 s), conversely CT for happy facial expressions was expected to decrease when mimicry was blocked (blocked mimicry condition: M = 7.51 s; SD = 4.96 s; free mimicry condition: M = 7.75 s; SD = 4.87 s). This observation was substantiated by the significant interaction between mimicry manipulation and the reported content, F(1.85,49.86) = 3.53, p = 0.04. However, our hypothesis was only partially supported by the post-hoc comparisons. In particular, the observed interaction was statistically conveyed entirely by a modulation of CTs for neutral expressions as a function of the mimicry manipulation, such that for neutral faces CT in the blocked mimicry condition was longer than CT in free mimicry condition, t(56.6) = 2.78, p = 0.007. The evidence in favor of the alternative hypothesis H1 may be classified, in terms of Bayes factor, as moderate, BF10 = 5.84. In order to substantiate our conclusions we further report the post-hoc comparisons for happy and mixed percepts (t(56.6) = − 0.928, p = 0.357, BF10 = 0.28; t(56.6) = − 1.348, p = 0.18, BF10 = 0.60, respectively). To note, the evidence in favor of the null hypothesis H0 for happy faces, can be classified as moderate, overall suggesting that the pattern of the observed findings is robust. See Fig. 3.
Questionnaires
In exploratory analyses, we tested if ORT and CT were correlated (Pearson, two-sided correlations) with the Toronto Alexithymia Scale (TAS-20) and the Interpersonal Reactivity Index (IRI). No correlations were significant with the IRI scores. TAS-20 is one of the most commonly used measures of alexithymia traits, with a higher TAS-20 score indicating a higher inability to experience their own bodily emotions. Given that these tests were exploratory and since the study was not designed in terms of statistical power to directly answer this question on the relationship between alexithymic traits and BR metrics for facial expressions, we report the results without strict corrections for multiple comparisons, so that future studies can look into the promising correlations in a planned confirmatory testing.
TAS-20 score was negatively correlated with the cumulative time for the happy expressions in the free mimicry condition (i.e., CTfree_happy; r(26) = − 0.42, p = 0.02), and positively correlated with onset resolution times for the happy expressions in the free mimicry condition (i.e., ORTfree_happy; r(26) = 0.49, p = 0.007). The correlation with the cumulative time (i.e., CTfree_happy) indicates that individuals with difficulty in experiencing their own emotion (as indicated by the high traits of alexithymia), tend to maintain emotional (here happy) faces as the content of their consciousness for a shorter time than individuals with lower levels of alexithymia. Furthermore, as indicated by the correlation with ORT (i.e., ORTfree_happy), the initial time for perceptual disambiguation and conscious selection of an emotional face is greater in individuals with higher levels of alexithymia than in individuals with lower levels of alexithymia. This can be interpreted that alexithymic individuals tend to have difficulty and take longer time in resolving ambiguity in the direction of a more emotional face. It is interesting to note that these effects emerged in the condition in which participants could freely use their facial mimicry and these correlations are reduced in the blocked condition (CTblocked_happy: r(26) = − 0.23, p = 0.24 ORTblocked_happy: r(26) = 0.35, p = 0.06), which is along the direction that sensorimotor simulation theory would predict (see “Discussion”).
Discussion
In the present investigation, we wanted to test the hypothesis of a functional role of the observer’s facial mimicry in ambiguity resolution and/or stabilization within awareness of faces with happy and neutral facial expressions during a binocular rivalry task. In particular, in the light of the sensorimotor simulation model by Wood et al.45, we hypothesized that the communication between sensorimotor and visual systems could either modulate the (initial) resolution of ambiguity under conditions of binocular rivalry or that signals from sensorimotor system could stabilize conscious representations of face stimuli in a later stage once the ambiguity is resolved. In order to test our hypothesis, we asked participants to perform a standard binocular rivalry task by presenting a happy and a neutral face (from the same identity) in rivalry. Crucially, we manipulated participants’ facial mimicry such that in one condition of the experiment they performed the task with their facial mimicry blocked by a chopstick to be held between the lips in order to inhibit the contraction of the zygomatic major, that is the muscle mainly involved when smiling; in the other half of the experiment participants could freely use their facial mimicry.
We reasoned that if the signal from the sensorimotor cortices is involved in the early conscious processing of facial expressions, inhibiting the zygomaticus major should have biased the initial competition between the two rivalry stimuli, such that initial resolution of the ambiguity would have been in favor of neutral faces in the blocked mimicry condition when compared to the free mimicry condition. Our primary outcome measures were the frequency of the initial percept (IP) and the onset resolution time (ORT) until the first rivalry as a function of the facial mimicry manipulation. A different but equally interesting scenario foresees that if the sensorimotor signal is integrated with the visual percept only at later stages of processing when the ‘winning’ stimulus is the content of consciousness, then blocking of facial mimicry would have modulated the stabilization of conscious perception of facial expressions. We tested this scenario through the analysis of the cumulative time (CT) as a function of the mimicry manipulation. We reasoned that, under an iterative model scenario (i.e., sensorimotor and visual systems that iteratively share information from very early stages of processing), an impact of mimicry manipulation could have been expected either on the initial percept and/or on the cumulative time.
First, and importantly, we replicated previous findings53,56 (see also55 for similar findings) such that cumulative time for happy facial expressions was longer than cumulative time for neutral faces, regardless of the mimicry manipulation condition. Furthermore, in our initial percept analysis, we found that happy expressions were more frequently perceived as first stimuli compared to neutral faces. These results corroborate the adequacy of the present paradigm and the quality of the present data.
Notably, the interaction between reported content (i.e., happy, neutral and mixed percepts) and mimicry manipulation did yield significant effects for the cumulative time. Further analyses provided evidence that cumulative time for neutral faces increased when mimicry was blocked compared to when participants could freely use their facial mimicry, thus providing only partial support to our hypothesis. In fact, although we observed a modulation of the cumulative time for neutral faces as a function of the mimicry manipulation, we did not observe an effect of the mimicry manipulation on the cumulative time for happy faces. Both of these results are corroborated by the Bayes factor.
How to reconcile these two results in the light of our initial hypothesis and the (sensorimotor) simulation models? A possible explanation could lie in the type of experimental manipulation of facial mimicry we have adopted here. In the condition of active alteration of mimicry (through the chopstick) it is possible that it was the congruent cross-modal match between the sensorimotor feedback and the visual representation of faces with neutral expression that has facilitated their stabilization in awareness. On the other hand, in our experimental design we did not implement a similar condition of facial mimicry congruent with the representation of happy expressions (e.g., by asking participants to actively keep the chopstick between their teeth, activating the zygomatic muscle). Future studies should clarify—possibly also including an electromyographic monitoring—whether it is this cross-modal congruence (sensorimotor-visual) that may facilitate stabilization in awareness for faces with different facial expressions, both neutral and emotional. An alternative explanation could rely on the notion that the simulation mechanism is especially triggered in those conditions in which facial expressions are particularly subtle or ambiguous (see, e.g.,26). Along this line of reasoning, neutral expressions may be conceived as more ambiguous than full expressions of happiness. Indeed, it is known that neutral expressions can be more easily misinterpreted than full emotional expressions, in both healthy66 and clinical populations67–70. In order to test this hypothesis, future studies could implement similar paradigms to that implemented here by varying the intensity of the rivalrous facial expression (for example through the morphing procedure). In light of these considerations, we admit that the present results are only generalizable to neutral facial expressions and to the role that facial mimicry can play in their stabilization in awareness. Nonetheless, we suggest useful future directions in order to investigate the phenomenon of the interaction between sensorimotor simulation and visual awareness that is currently almost totally neglected.
Another aspect that will require future work to provide a complete picture of the generalizability of these results concerns the female sample of the present study. In the light of the novelty of this field of investigation, we deemed appropriate including a sample of only women in order to maximize the probability of observing an effect because of the large body of previous evidence showing that women are more expressive than men71, more accurate in processing emotional expressions72, and selectively impaired in their ability to process expressions of happiness following the disruption of the primary motor cortex by repetitive transcranial magnetic stimulation29. Our results are currently generalizable to women and future experimental designs should include the participants’ gender as a factor.
Although a similar trend to that observed for the cumulative time was observed for the initial percept metric (i.e., a higher frequency for faces with neutral expression in the condition of blocked mimicry compared to the condition of free mimicry and an opposite trend for faces with expression of happiness), the effect was not statistically significant (when corrected for multiple comparisons). These findings indicate that the signal of the sensorimotor system tends not to play a role before the resolution of the ambiguity in favor of one of the two stimuli in rivalry, but rather that the signal from the sensorimotor system mainly acts as a ‘stabilizer’ of the conscious representation of the neutral face once the stimulus is the current content of consciousness (to note, the concept of ‘stabilization’ used here is not to be intended as for Leopold73 and for Pearson and Brascamp74).
At the present, this whole pattern of findings is almost entirely in line with the results previously reported by Korb et al.60 using a breaking continuous flash suppression (b-CFS) task, especially with regard to the early account we tested. However, and importantly, we observed an effect of the mimicry manipulation on the conscious stabilization of the neutral expressions, suggesting indeed a role of the sensorimotor system during a later stage of conscious processing which deserves to be investigated further, exploring the possibility that it can be generalized to emotional expressions.
Taking into account the line of studies on the neural correlates of consciousness and the debate on the temporal onset of consciousness (see, e.g.,75), it is interesting to note that two different temporal loci have been proposed, one “early” (associated in terms of event-related potentials, ERPs, with the visual awareness negativity in the range of the N2 ERP component76–78), and one “late” (associated with P3b/LP ERP component; e.g.79). Instead, as regards the effects of the blocking of facial mimicry on the construction of a visual percept, a previous study by our research group44 has shown an impact of the facial mimicry manipulation on visual working memory representations of neutral and emotional faces in terms of a modulation of the sustained posterior contralateral negativity ERP component detected at occipito-temporo-parietal sites (SPCN; e.g.80–82). To note (a) this ERP component has an onset of about 300 ms post-stimulus, thus supporting the view that visual and sensorimotor information may interact/combine within 300–400 ms following the exposure to a facial expression, and (b) the onset of this ERP component is later than visual awareness negativity (i.e., early temporal locus of the onset of consciousness) and earlier than the P3b/LP (i.e., late temporal locus of the onset of consciousness). Intriguingly, in another study designed to investigate the lower edge of this interaction between sensorimotor and visual systems, it was found that the ERP components P1 and N170 are not modulated as a function of facial mimicry manipulation, except in relation to alexithymic traits, in the direction of a modulation which tends to manifest itself only for individuals with low alexithymic traits43. An attempt to interpret this whole pattern of results could indicate that facial mimicry manipulations may affect high-level visual processing of facial expressions (approximately) after 170 ms and before 300–400 ms. The results of the present study, on the other hand, indicate that blocking of facial mimicry does not have an effect on the early resolution of perceptual ambiguity but rather on later stabilization of representations within awareness, at least for neutral facial expressions. Taken together, these results may suggest that the integration of somatosensory information with visual information occurs after the onset of awareness for facial expressions, thus suggesting an early onset of visual awareness. Considering the different time scales of the investigative phenomena (i.e., binocular rivalry and ERP findings), we propose this interpretation with great caution as a possible starting point for future studies.
Finally, although exploratory, we observed noteworthy correlations between the TAS-20 scores with metrics of binocular rivalry as a function of the mimicry manipulation. Correlation between TAS-20 scores and onset resolution time suggests that highly alexithymic participants tend to be slower in consciously accessing happy facial expressions when their facial mimicry is free and furthermore correlation with cumulative times indicated, instead, that individuals with higher traits of alexithymia under conditions of free mimicry tend to consciously perceive happy facial expressions for a shorter time compared to individuals with lower traits of alexithymia. These findings are perfectly in line with the very definition of alexithymia which is a subclinical condition involving a lack of emotional awareness and a difficulty in identifying (and describing) feelings and facial expressions of emotions65. We do not expand on the subject as the correlations we have investigated are currently exploratory, however we believe that this relationship deserves future investigation.
To conclude, the present results support an effect of facial mimicry on the awareness of faces with a neutral expression, in particular with regard to the stabilization of their conscious representations. However, we also acknowledge the need for future studies to investigate in greater depth the role of sensorimotor simulation on the conscious processing of emotional expressions and to consider both detecting facial electromyographic responses during a binocular rivalry paradigm similar to that employed in the present investigation, and, by means of time-resolved techniques such as electroencephalography, to track the time-course of the interaction between sensorimotor and visual cortices. In particular the latter would provide a more direct test of the early vs. late interaction hypotheses.
Methods
Participants
Thirty-two female healthy participants were recruited among students at the University of Padova (average age in years = 24.7, SD = 4.8, 2 left-handed). The sample size is considered appropriate on the basis of a meta-analysis on facial feedback and emotional experience83. Given the methodological heterogeneity, we selected a subsample of the originally included studies with these characteristics (a) facial mimicry manipulated as a within-subjects factor and (b) happiness as the main emotion. Using Coles et al.83 approach, the estimated effect in terms of Cohen’s d84 is 0.478 (SE = 0.162, 95% CI [0.117 0.839]). Power was estimated using the pwr package85 considering a paired t-test between blocked and free mimicry conditions. For the estimated effect size, a sample size of 30 participants is required to reach an 80% power level (see the online repository for additional details). Since we have hypothesized a specific direction of the effect due to mimicry, in this case we could consider a one-sided paired t-test, which would need a sample size of 28 participants to reach an 80% power level.
All volunteers gave written informed consent in accordance with the Declaration of Helsinki, and all experimental procedures were approved by the local research ethics committee (Comitato Etico della Ricerca Psicologica Area 17, University of Padua) and performed in accordance with its guidelines. Participants had normal or corrected-to-normal vision. Color blindness was assessed using the Ishihara color blindness test86. A total of four participants were excluded from analysis (one did not follow the coding instruction, two failed in holding the chopstick without the use of teeth and one did not complete the experiment). At the end of the experiment, participants completed the Toronto Alexithymia Scale87 (TAS-20) and the Interpersonal Reactivity Index88 (IRI) questionnaires. Scores on both questionnaires were in the normal range (TAS-20: M = 43.6, SD = 11.5 IRI: M = 98.75, SD = 8).
Materials and apparatus
Visual stimuli were displayed using E-Prime 2.0 Software (version 2.0.10.242; Psychology Software Tools, Pittsburgh, PA) on a LG flatron F700B (Brightness: 85; Contrast: 90) 85 Hz monitor. Original stimuli were selected from the Karolinska Directed Emotional Faces89 and included one male (AM10) and one female (AF01) face each displaying a neutral and happy expression. Visual stimuli were presented covering 8 degrees of visual angle in height and width. Images were cropped with the software GIMP (version 2.8.10; see Figs. 1 and 4) 562 × 562 pixels centering the face in the middle of the square. Images of the same individual expressing the two emotions (happy and neutral) were superimposed and shifted by four pixels in order to facilitate the rivalry between two percepts. Happy and neutral faces of the same identity were analyzed and matched (in terms of contrast and luminance histograms) by using Fiji90 and MATLAB (version R2019; see Figs. 1 and 4).
Procedure
Participants seated on a comfortable chair in a silent, temperature-controlled room. They were asked to place their head on a chin rest, with a distance of 70 cm from the screen. Before starting the experiment, the anaglyph filters (red/green) were set and participants were trained. The side of the color filters was counterbalanced across participants. During the experiment, participants were asked to focus on a fixation point located in the middle of the screen. The experiment consisted of one session of four blocks. Each participant performed two blocks where they could freely use their facial mimicry (“free mimicry condition”) and in other two blocks they were asked to hold a chopstick between their lips without using their teeth (“blocked mimicry condition”). The order of the two mimicry conditions was counterbalanced across participants. In each block, four rivalry stimuli were presented in a random order for a total of eight trials (twice per stimulus) in each block (see Fig. 4). Rivalry stimuli were presented for 15 s preceded by a 2-s fixation point and followed by a 3-s black screen. Participants were asked to code what they saw in real time by pressing one of three keys of the keyboard (“b”, “n”, “m”, these keys are adjacent to each other on the standard keyboard in this order from the left to the right). Participants were informed that on each trial they could see one of two faces, and that the appearance might change from one to the other during the trial. Coding instructions were presented before the beginning of the block; the order of the “b” and “m” keys, corresponding to the coding of the “happy” and “neutral” facial expression, was counterbalanced across blocks, while the “n” key always corresponded to the coding of a “mixed” percept. In the middle and at the end of each block a short break was recommended to the participant to reduce any fatigue. At the end of each mimicry conditions (i.e., two blocks) valence and arousal of each stimulus were measured respectively on a − 3/+ 3 and + 1/+ 7 scales.
Data reduction
Firstly, for each trial we extracted the initial percept (IP), namely the first reported percept during the ongoing trial (neutral and happy facial expressions), in order to analyze whether initial percept frequencies changed as a function of the emotionality of the face and/or as a function of the mimicry manipulation. We further computed onset resolution time (ORT), namely the time to code the IP as a mean per emotion per subject. ORT log transformation was used for the analysis. We also computed cumulative times (CTs), as a measure of perceptual predominance, for mixed percept, neutral and happy facial expressions separately. That is, CTs were computed by summing the perceptual duration for each of mixed, neutral and happy percepts segments during a trial. Two participants did not code a total of three trials (Subject 9, one trial in the blocked mimicry condition; Subject 18, two trials in the free mimicry condition).
Data analysis
Differences in stimuli rating for valence and arousal were assessed in separate analyses of variance (ANOVAs) and post hoc comparisons. Differences in mimicry (blocked and free) and IPs (happy, neutral) were assessed for ORT in analyses of variance (ANOVA) and post hoc comparisons. All post hoc comparisons are Bonferroni corrected for multiple comparisons. Differences in mimicry (blocked and free) and facial expression percept (happy, neutral and mixed) were assessed for CT in analyses of variance (ANOVA) and post hoc comparisons. All post hoc comparisons are Bonferroni corrected for multiple comparisons. In order to estimate the probability of the IP as a function of facial mimicry manipulation we applied a mixed-effects logistic regression model with IP (happy or neutral) explained by mimicry condition (blocked vs. free). Subjects were inserted as random effect with a varying intercept. The mimicry effect was evaluated as the odds ratio between free and blocked conditions.
Author contributions
P.S. developed the study concept. P.S., T.Q., and F.G. contributed to the study design. T.Q. and F.G. performed testing and data collection. T.Q. drawn the figures. T.Q. and F.G. performed the data analysis and all the authors interpreted the data. P.S., T.Q. and F.G. drafted the manuscript and N.T. provided critical revision. All authors approved the final version of the manuscript for submission.
Data availability
The dataset and analyses reported in this manuscript are available at Open Science Framework repository: https://osf.io/xk25b/.
Competing interests
The authors declare no competing interests.
Footnotes
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Carr L, Iacoboni M, Dubeau M-C, Mazziotta JC, Lenzi GL. Neural mechanisms of empathy in humans: A relay from neural systems for imitation to limbic areas. Proc. Natl. Acad. Sci. 2003;100:5497–5502. doi: 10.1073/pnas.0935845100. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Caruana F, Borghi AM. Embodied Cognition: una nuova psicologia. Giornale italiano di psicologia. 2013;40:23–48. [Google Scholar]
- 3.Gallese V, Sinigaglia C. What is so special about embodied simulation? Trends Cogn. Sci. 2011;15:512–519. doi: 10.1016/j.tics.2011.09.003. [DOI] [PubMed] [Google Scholar]
- 4.Goldman A, de Vignemont F. Is social cognition embodied? Trends Cogn. Sci. 2009;13:154–159. doi: 10.1016/j.tics.2009.01.007. [DOI] [PubMed] [Google Scholar]
- 5.Niedenthal PM. Embodying emotion. Science. 2007;316:1002–1005. doi: 10.1126/science.1136930. [DOI] [PubMed] [Google Scholar]
- 6.Pitcher D, Garrido L, Walsh V, Duchaine BC. Transcranial magnetic stimulation disrupts the perception and embodiment of facial expressions. J. Neurosci. 2008;28:8929–8933. doi: 10.1523/JNEUROSCI.1450-08.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Wicker B, et al. Both of us disgusted in my insula the common neural basis of seeing and feeling disgust. Neuron. 2003;40:655–664. doi: 10.1016/S0896-6273(03)00679-2. [DOI] [PubMed] [Google Scholar]
- 8.Haxby JV, et al. The distributed human neural system for face perception. Trends Cogn. Sci. 2000;4:223–233. doi: 10.1016/S1364-6613(00)01482-0. [DOI] [PubMed] [Google Scholar]
- 9.Grill-Spector K, Knouf N, Kanwisher N. The fusiform face area subserves face perception, not generic within-category identification. Nat. Neurosci. 2004;7:555–562. doi: 10.1038/nn1224. [DOI] [PubMed] [Google Scholar]
- 10.Winston JS, Henson RNA, Fine-Goulden MR, Dolan RJ. fMRI-adaptation reveals dissociable neural representations of identity and expression in face perception. J. Neurophysiol. 2004;92:1830–1839. doi: 10.1152/jn.00155.2004. [DOI] [PubMed] [Google Scholar]
- 11.Yovel G, Kanwisher N. Face perception domain specific, not process specific. Neuron. 2004;44:889–898. doi: 10.1016/j.neuron.2004.11.018. [DOI] [PubMed] [Google Scholar]
- 12.Ishai A, Schmidt CF, Boesiger P. Face perception is mediated by a distributed cortical network. Brain Res. Bull. 2005;67:87–93. doi: 10.1016/j.brainresbull.2005.05.027. [DOI] [PubMed] [Google Scholar]
- 13.Rotshtein P, Henson RNA, Treves A, Driver J, Dolan RJ. Morphing Marilyn into Maggie dissociates physical and identity face representations in the brain. Nat. Neurosci. 2005;8:107–113. doi: 10.1038/nn1370. [DOI] [PubMed] [Google Scholar]
- 14.Lee LC, et al. Neural responses to rigidly moving faces displaying shifts in social attention investigated with fMRI and MEG. Neuropsychologia. 2010;48:477–490. doi: 10.1016/j.neuropsychologia.2009.10.005. [DOI] [PubMed] [Google Scholar]
- 15.Gobbini MI, et al. Distinct neural systems involved in agency and animacy detection. J. Cognit. Neurosci. 2011;23:1911–1920. doi: 10.1162/jocn.2010.21574. [DOI] [PubMed] [Google Scholar]
- 16.Fox CJ, Iaria G, Barton JJS. Defining the face processing network: Optimization of the functional localizer in fMRI. Hum. Brain Mapp. 2009;30:1637–1651. doi: 10.1002/hbm.20630. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Trautmann SA, Fehr T, Herrmann M. Emotions in motion: Dynamic compared to static facial expressions of disgust and happiness reveal more widespread emotion-specific activations. Brain Res. 2009;1284:100–115. doi: 10.1016/j.brainres.2009.05.075. [DOI] [PubMed] [Google Scholar]
- 18.Furl N, Henson RN, Friston KJ, Calder AJ. Top-down control of visual responses to fear by the amygdala. J. Neurosci. 2013;33:17435–17443. doi: 10.1523/JNEUROSCI.2992-13.2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Furl N, Henson RN, Friston KJ, Calder AJ. Network interactions explain sensitivity to dynamic faces in the superior temporal sulcus. Cereb. Cortex. 2015;25:2876–2882. doi: 10.1093/cercor/bhu083. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Johnston P, Mayes A, Hughes M, Young AW. Brain networks subserving the evaluation of static and dynamic facial expressions. Cortex. 2013;49:2462–2472. doi: 10.1016/j.cortex.2013.01.002. [DOI] [PubMed] [Google Scholar]
- 21.Harris RJ, Young AW, Andrews TJ. Dynamic stimuli demonstrate a categorical representation of facial expression in the amygdala. Neuropsychologia. 2014;56:47–52. doi: 10.1016/j.neuropsychologia.2014.01.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Liang Y, Liu B, Li X, Wang P. Multivariate pattern classification of facial expressions based on large-scale functional connectivity. Front. Hum. Neurosci. 2018;12:94. doi: 10.3389/fnhum.2018.00094. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Bastiaansen JACJ, Thioux M, Keysers C. Evidence for mirror systems in emotions. Philos. Trans. R. Soc. B Biol. Sci. 2009;364:2391–2404. doi: 10.1098/rstb.2009.0058. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Goldman AI, Sripada CS. Simulationist models of face-based emotion recognition. Cognition. 2005;94:193–213. doi: 10.1016/j.cognition.2004.01.005. [DOI] [PubMed] [Google Scholar]
- 25.Likowski KU, et al. Facial mimicry and the mirror neuron system: Simultaneous acquisition of facial electromyography and functional magnetic resonance imaging. Front. Hum. Neurosci. 2012;6:21. doi: 10.3389/fnhum.2012.00214. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Wood A, Rychlowska M, Korb S, Niedenthal P. Fashioning the face: Sensorimotor simulation contributes to facial expression recognition. Trends Cogn. Sci. 2016;20:227–240. doi: 10.1016/j.tics.2015.12.010. [DOI] [PubMed] [Google Scholar]
- 27.Adolphs R, Damasio H, Tranel D, Cooper G, Damasio AR. A role for somatosensory cortices in the visual recognition of emotion as revealed by three-dimensional lesion mapping. J. Neurosci. 2000;20:2683–2690. doi: 10.1523/JNEUROSCI.20-07-02683.2000. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Pourtois G, Grandjean D, Sander D, Vuilleumier P. Electrophysiological correlates of rapid spatial orienting towards fearful faces. Cereb. Cortex. 2004;14:619–633. doi: 10.1093/cercor/bhh023. [DOI] [PubMed] [Google Scholar]
- 29.Korb S, et al. Gender differences in the neural network of facial mimicry of smiles—An rTMS study. Cortex. 2015;70:101–114. doi: 10.1016/j.cortex.2015.06.025. [DOI] [PubMed] [Google Scholar]
- 30.Borgomaneri S, Bolloni C, Sessa P, Avenanti A. Blocking facial mimicry affects recognition of facial and body expressions. PLoS ONE. 2020;15:e0229364. doi: 10.1371/journal.pone.0229364. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Hyniewska S, Sato W. Facial feedback affects valence judgments of dynamic and static emotional expressions. Front. Psychol. 2015;6:291. doi: 10.3389/fpsyg.2015.00291. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Ipser A, Cook R. Blocking facial mimicry reduces perceptual sensitivity for facial expressions. J. Vis. 2015;15:1376. doi: 10.1167/15.12.1376. [DOI] [Google Scholar]
- 33.Neal DT, Chartrand TL. Embodied emotion perception. Soc. Psychol. Pers. Sci. 2011;2:673–678. doi: 10.1177/1948550611406138. [DOI] [Google Scholar]
- 34.Oberman LM, Winkielman P, Ramachandran VS. Face to face: Blocking facial mimicry can selectively impair recognition of emotional expressions. Soc. Neurosci. 2007;2:167–178. doi: 10.1080/17470910701391943. [DOI] [PubMed] [Google Scholar]
- 35.Ponari M, Conson M, D’Amico NP, Grossi D, Trojano L. Mapping correspondence between facial mimicry and emotion recognition in healthy subjects. Emotion. 2012;12:1398–1403. doi: 10.1037/a0028588. [DOI] [PubMed] [Google Scholar]
- 36.Stel M, van Knippenberg A. The role of facial mimicry in the recognition of affect. Psychol. Sci. 2008;19:984–985. doi: 10.1111/j.1467-9280.2008.02188.x. [DOI] [PubMed] [Google Scholar]
- 37.Bourgeois P, Hess U. The impact of social context on mimicry. Biol. Psychol. 2008;77:343–352. doi: 10.1016/j.biopsycho.2007.11.008. [DOI] [PubMed] [Google Scholar]
- 38.Hess U, Fischer A. Emotional mimicry as social regulation. Pers. Soc. Psychol. Rev. 2013;17:142–157. doi: 10.1177/1088868312472607. [DOI] [PubMed] [Google Scholar]
- 39.Hess U, Fischer A. Emotional mimicry: Why and when we mimic emotions. Soc. Person. Psychol. Compass. 2014;8:45–57. doi: 10.1111/spc3.12083. [DOI] [Google Scholar]
- 40.Hess U. Who to whom and why: The social nature of emotional mimicry. Psychophysiology. 2021;58:e13675. doi: 10.1111/psyp.13675. [DOI] [PubMed] [Google Scholar]
- 41.Achaibou A, Pourtois G, Schwartz S, Vuilleumier P. Simultaneous recording of EEG and facial muscle reactions during spontaneous emotional mimicry. Neuropsychologia. 2008;46:1104–1113. doi: 10.1016/j.neuropsychologia.2007.10.019. [DOI] [PubMed] [Google Scholar]
- 42.Dimberg U, Petterson M. Facial reactions to happy and angry facial expressions: Evidence for right hemisphere dominance. Psychophysiology. 2000;37:693–696. doi: 10.1111/1469-8986.3750693. [DOI] [PubMed] [Google Scholar]
- 43.Lomoriello AS, Maffei A, Brigadoi S, Sessa P. Altering sensorimotor simulation impacts early stages of facial expression processing depending on individual differences in alexithymic traits. Brain Cognit. 2021;148:105678. doi: 10.1016/j.bandc.2020.105678. [DOI] [PubMed] [Google Scholar]
- 44.Sessa P, Lomoriello AS, Luria R. Neural measures of the causal role of observers’ facial mimicry on visual working memory for facial expressions. Soc. Cogn. Affect. Neur. 2018;13:1281–1291. doi: 10.1093/scan/nsy095. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Wood A, Lupyan G, Sherrin S, Niedenthal P. Altering sensorimotor feedback disrupts visual discrimination of facial expressions. Psychon. B Rev. 2016;23:1150–1156. doi: 10.3758/s13423-015-0974-5. [DOI] [PubMed] [Google Scholar]
- 46.Blake R, Logothetis NK. Visual competition. Nat. Rev. Neurosci. 2002;3:13–21. doi: 10.1038/nrn701. [DOI] [PubMed] [Google Scholar]
- 47.Alais, D. & Blake, R. (Eds.). Binocular Rivalry. MIT press (2005).
- 48.Tong F, Meng M, Blake R. Neural bases of binocular rivalry. Trends Cogn. Sci. 2006;10:502–511. doi: 10.1016/j.tics.2006.09.003. [DOI] [PubMed] [Google Scholar]
- 49.Maier A, Panagiotaropoulos TI, Tsuchiya N, Keliris GA. Introduction to research topic—Binocular rivalry: A gateway to studying consciousness. Front. Hum. Neurosci. 2012;6:263. doi: 10.3389/fnhum.2012.00263. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Gallagher RM, Tsuchiya N. Third-eye rivalry. I Percep. 2020;11:2041669520927722. doi: 10.1177/2041669520927722. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Brascamp JW, Klink PC, Levelt WJM. The ‘laws’ of binocular rivalry: 50 years of Levelt’s propositions. Vis. Res. 2015;109:20–37. doi: 10.1016/j.visres.2015.02.019. [DOI] [PubMed] [Google Scholar]
- 52.Wheatstone C. Contributions to the physiology of vision—Part the first. On some remarkable and hitherto unobserved phenomena of binocular vision. Abstr. Pap. Print Philos. Trans. R. Soc. Lond. 1843;4:76–77. [Google Scholar]
- 53.Alpers GW, Gerdes ABM. Here is looking at you: Emotional faces predominate in binocular rivalry. Emotion. 2007;7:495–506. doi: 10.1037/1528-3542.7.3.495. [DOI] [PubMed] [Google Scholar]
- 54.Anderson E, Siegel EH, Barrett LF. What you feel influences what you see: The role of affective feelings in resolving binocular rivalry. J. Exp. Soc. Psychol. 2011;47:856–860. doi: 10.1016/j.jesp.2011.02.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Bannerman RL, Milders M, Gelder BD, Sahraie A. Influence of emotional facial expressions on binocular rivalry. Ophthal. Physl. Opt. 2008;28:317–326. doi: 10.1111/j.1475-1313.2008.00568.x. [DOI] [PubMed] [Google Scholar]
- 56.Yoon KL, Hong SW, Joormann J, Kang P. Perception of facial expressions of emotion during binocular rivalry. Emotion. 2009;9:172–182. doi: 10.1037/a0014714. [DOI] [PubMed] [Google Scholar]
- 57.Baumeister J-C, Rumiati RI, Foroni F. When the mask falls: The role of facial motor resonance in memory for emotional language. Acta Psychol. 2015;155:29–36. doi: 10.1016/j.actpsy.2014.11.012. [DOI] [PubMed] [Google Scholar]
- 58.Niedenthal PM, Brauer M, Halberstadt JB, Innes-Ker ÅH. When did her smile drop? Facial mimicry and the influences of emotional state on the detection of change in emotional expression. Cognit. Emot. 2001;15:853–864. doi: 10.1080/02699930143000194. [DOI] [Google Scholar]
- 59.Strack F, Martin LL, Stepper S. Inhibiting and facilitating conditions of the human smile: A nonobtrusive test of the facial feedback hypothesis. J. Pers. Soc. Psychol. 1988;54:768–777. doi: 10.1037/0022-3514.54.5.768. [DOI] [PubMed] [Google Scholar]
- 60.Korb S, Osimo SA, Suran T, Goldstein A, Rumiati RI. Face proprioception does not modulate access to visual awareness of emotional faces in a continuous flash suppression paradigm. Conscious Cogn. 2017;51:166–180. doi: 10.1016/j.concog.2017.03.008. [DOI] [PubMed] [Google Scholar]
- 61.Tsuchiya N, Koch C. Continuous flash suppression reduces negative afterimages. Nat. Neurosci. 2005;8:1096–1101. doi: 10.1038/nn1500. [DOI] [PubMed] [Google Scholar]
- 62.Jiang Y, Costello P, He S. Processing of invisible stimuli: Advantage of upright faces and recognizable words in overcoming interocular suppression. Psychol. Sci. 2006;18:349–355. doi: 10.1111/j.1467-9280.2007.01902.x. [DOI] [PubMed] [Google Scholar]
- 63.Stein T, Hebart MN, Sterzer P. Breaking continuous flash suppression: A new measure of unconscious processing during interocular suppression? Front. Hum. Neurosci. 2011;5:167. doi: 10.3389/fnhum.2011.00167. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Carter O, Cavanagh P. Onset rivalry: Brief presentation isolates an early independent phase of perceptual competition. PLoS ONE. 2007;2:e343. doi: 10.1371/journal.pone.0000343. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Nemiah JC. Alexithymia: A view of the psychosomatic process. Modern Trends Psychosomatic Med. 1976;3:430–439. [Google Scholar]
- 66.Bell CJ, et al. Earthquake brain: Altered recognition and misclassification of facial expressions are related to trauma exposure but not posttraumatic stress disorder. Front. Psychiatry. 2017;8:278. doi: 10.3389/fpsyt.2017.00278. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Eack SM, Mazefsky CA, Minshew NJ. Misinterpretation of facial expressions of emotion in verbal adults with autism spectrum disorder. Autism. 2015;19:308–315. doi: 10.1177/1362361314520755. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Bourke C, Douglas K, Porter R. Processing of facial emotion expression in major depression: A review. Aust. Nz. J. Psychiat. 2010;44:681–696. doi: 10.3109/00048674.2010.496359. [DOI] [PubMed] [Google Scholar]
- 69.Kohler CG, et al. Facial emotion recognition in schizophrenia: Intensity effects and error pattern. Am. J. Psychiat. 2003;160:1768–1774. doi: 10.1176/appi.ajp.160.10.1768. [DOI] [PubMed] [Google Scholar]
- 70.Daros AR, Zakzanis KK, Ruocco AC. Facial emotion recognition in borderline personality disorder. Psychol. Med. 2013;43:1953–1963. doi: 10.1017/S0033291712002607. [DOI] [PubMed] [Google Scholar]
- 71.Kring AM, Gordon AH. Sex differences in emotion: Expression, experience, and physiology. J. Pers. Soc. Psychol. 1998;74:686–703. doi: 10.1037/0022-3514.74.3.686. [DOI] [PubMed] [Google Scholar]
- 72.Hoffmann H, Kessler H, Eppel T, Rukavina S, Traue HC. Expression intensity, gender and facial emotion recognition: Women recognize only subtle facial emotions better than men. Acta Psychol. 2010;135:278–283. doi: 10.1016/j.actpsy.2010.07.012. [DOI] [PubMed] [Google Scholar]
- 73.Leopold DA, Wilke M, Maier A, Logothetis NK. Stable perception of visually ambiguous patterns. Nat. Neurosci. 2002;5:605–609. doi: 10.1038/nn0602-851. [DOI] [PubMed] [Google Scholar]
- 74.Pearson J, Brascamp J. Sensory memory for ambiguous vision. Trends Cogn. Sci. 2008;12:334–341. doi: 10.1016/j.tics.2008.05.006. [DOI] [PubMed] [Google Scholar]
- 75.Förster J, Koivisto M, Revonsuo A. ERP and MEG correlates of visual consciousness: The second decade. Conscious Cogn. 2020;80:102917. doi: 10.1016/j.concog.2020.102917. [DOI] [PubMed] [Google Scholar]
- 76.Koivisto M, Revonsuo A. An ERP study of change detection, change blindness, and visual awareness. Psychophysiology. 2003;40:423–429. doi: 10.1111/1469-8986.00044. [DOI] [PubMed] [Google Scholar]
- 77.Ojanen V, Revonsuo A, Sams M. Visual awareness of low-contrast stimuli is reflected in event-related brain potentials. Psychophysiology. 2003;40:192–197. doi: 10.1111/1469-8986.00021. [DOI] [PubMed] [Google Scholar]
- 78.Wilenius-Emet M, Revonsuo A, Ojanen V. An electrophysiological correlate of human visual awareness. Neurosci. Lett. 2004;354:38–41. doi: 10.1016/j.neulet.2003.09.060. [DOI] [PubMed] [Google Scholar]
- 79.Dehaene S, Changeux J-P. Experimental and theoretical approaches to conscious processing. Neuron. 2011;70:200–227. doi: 10.1016/j.neuron.2011.03.018. [DOI] [PubMed] [Google Scholar]
- 80.Vogel EK, Machizawa MG. Neural activity predicts individual differences in visual working memory capacity. Nature. 2004;428:748–751. doi: 10.1038/nature02447. [DOI] [PubMed] [Google Scholar]
- 81.Sessa P, Luria R, Gotler A, Jolicœur P, Dell’acqua R. Interhemispheric ERP asymmetries over inferior parietal cortex reveal differential visual working memory maintenance for fearful versus neutral facial identities. Psychophysiology. 2011;48:187–197. doi: 10.1111/j.1469-8986.2010.01046.x. [DOI] [PubMed] [Google Scholar]
- 82.Sessa P, et al. Look out for strangers! Sustained neural activity during visual working memory maintenance of other-race faces is modulated by implicit racial prejudice. Soc. Cogn. Affect. Neurosci. 2012;7:314–321. doi: 10.1093/scan/nsr011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83.Coles NA, Larsen JT, Lench HC. A meta-analysis of the facial feedback literature: Effects of facial feedback on emotional experience are small and variable. Psychol. Bull. 2019;145:610–651. doi: 10.1037/bul0000194. [DOI] [PubMed] [Google Scholar]
- 84.Jacob C. The t test for means. Stat. Power Anal. Behav. Sci. 1988;2:20–26. [Google Scholar]
- 85.Champely, S., Ekstrom, C., Dalgaard, P., Gill, J., Weibelzahl, S., Anandkumar, A. & De Rosario, H. pwr: Basic functions for power analysis (2017).
- 86.Shinobu I. Tests for color blindness. Am. J. Ophthalmol. 1918;1:376. [Google Scholar]
- 87.Bagby RM, Parker JDA, Taylor GJ. The twenty-item Toronto Alexithymia scale—I. Item selection and cross-validation of the factor structure. J. Psychosom. Res. 1994;38:23–32. doi: 10.1016/0022-3999(94)90005-1. [DOI] [PubMed] [Google Scholar]
- 88.Davis MH. Measuring individual differences in empathy: Evidence for a multidimensional approach. J. Pers. Soc. Psychol. 1983;44:113–126. doi: 10.1037/0022-3514.44.1.113. [DOI] [Google Scholar]
- 89.Lundqvist D, Flykt A, Öhman A. The Karolinska Directed Emotional Faces (KDEF) Karolinska Institutet; 1998. p. 2. [Google Scholar]
- 90.Schindelin J, et al. Fiji: an open-source platform for biological-image analysis. Nat. Methods. 2012;9:676–682. doi: 10.1038/nmeth.2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The dataset and analyses reported in this manuscript are available at Open Science Framework repository: https://osf.io/xk25b/.