Abstract
Using still pictures of emotional facial expressions as experimental stimuli, reduced amygdala responses or impaired recognition of basic emotions were repeatedly found in people with psychopathic traits. The amygdala also plays an important role in short‐latency facial mimicry responses. Since dynamic emotional facial expressions may have higher ecological validity than still pictures, we compared short‐latency facial mimicry responses to dynamic and static emotional expressions between adolescents with psychopathic traits and normal controls. Facial EMG responses to videos or still pictures of emotional expressions (happiness, anger, sadness, fear) were measured. Responses to 500‐ms dynamic expressions in videos, as well as the subsequent 1500‐ms phase of maximal (i.e., static) expression, were compared between male adolescents with disruptive behavior disorders and high (n = 14) or low (n = 17) callous‐unemotional (CU) traits, and normal control subjects (n = 32). Responses to still pictures were also compared between groups. EMG responses to dynamic expressions were generally significantly smaller in the high‐CU group than in the other two groups, which generally did not differ. These group differences gradually emerged during the 500‐ms stimulus presentation period but in general they were already seen a few hundred milliseconds after stimulus onset. Group differences were absent during the 1500‐ms phase of maximal expression and during exposure to still pictures. Subnormal short‐latency mimicry responses to dynamic emotional facial expressions in the high‐CU group might have negative consequences for understanding emotional facial expressions of others during daily life when human facial interactions are primarily dynamic.
Keywords: callous‐unemotional traits, disruptive behavior disorders, EMG, emotional facial expressions, facial mimicry
Short abstract
During human interactions, short‐latency facial mimicry responses occur to dynamic emotional facial expressions of others. These are preconscious, automatic responses which cannot be voluntarily controlled. They may be important for emotional understanding of others and appear to be subnormal in male adolescents with psychopathic traits.
1. INTRODUCTION
Children and adolescents with disruptive behavior disorders (DBD), including the DSM‐5 categories of oppositional defiant disorder (ODD) and conduct disorder (CD), exhibit various forms of antisocial behavior. The behaviors of ODD (e.g., being angry, blaming others, arguing with adults) are typically of a less severe nature than the behaviors of CD (e.g., physical cruelty, lying, vandalism), but both are associated with significant impairments in social, academic, or occupational functioning (American Psychiatric Association, 2013). Developmental pathways leading to DBD are varied, even within each diagnostic category (Viding & McCrory, 2020). Callous‐unemotional (CU) traits designate a particularly problematic subgroup of DBD individuals with distinct emotional, cognitive, and biological characteristics (Blair, 2013; Blair et al., 2014; Frick & Kemp, 2021; Frick et al., 2014; Salekin, 2017), exhibiting more severe and stable patterns of delinquency (Reidy et al., 2017), and running the risk to develop psychopathy in adulthood (McMahon et al., 2010). CU traits, including lack of empathy, lack of guilt, and low emotional responsiveness, represent the affective dimension of child and adult psychopathy (Frick & Hare, 2001). A recent meta‐analysis of the relationships between CU traits and questionnaire measures of trait‐like empathy demonstrated moderate negative relationships between CU traits and measures of both affective and cognitive empathy (Waller et al., 2020). Yet, studies examining aspects of state‐like empathy have shown more consistent deficiencies in affective than in cognitive empathy in DBD individuals with CU traits (Blair, 2013; Frick & Kemp, 2021; Pijper et al., 2017; Viding & McCrory, 2019). Focusing more closely on emotional responsiveness in youth with conduct problems and CU traits, Northam and Dadds (2020) demonstrated that individuals with high compared to low levels of CU traits are less emotionally responsive, though results were fairly inconsistent. More consistent differences between subtypes were found in studies using older samples (adolescents vs. children), studies using physiological measures (rather than behavioral or self‐report measures), and studies using stimuli evoking other‐oriented emotions (i.e., affective empathy) rather than self‐oriented emotions.
Affective empathy is supposed to be accompanied by primitive or motor empathy in the form of automatic mimicry responses synchronizing emotional facial expressions, vocalizations, postures, and movements of others (Blair, 2005). According to Blair, no definite conclusions can be drawn regarding the role of motor empathy in persons with psychopathic traits since no studies have formally assessed motor empathy in psychopathy. In a preliminary study in a nonclinical student population using automated computerized coding of facial emotional expressions, individuals high on psychopathic traits showed less motor empathy when exposed to negative facial emotional expressions of others than individuals low on psychopathic traits (Khvatskaya & Lenzenweger, 2016). Adolphs (2006) hypothesizes that rapid automatic mimicry responses contribute to recognition of others' emotional facial expressions and may be crucial for the emergence of affective empathy.
Facial mimicry processes may be considered a determinant of emotional contagion (also called empathic arousal; Waller et al., 2020), that is “the cognitive ability to intuit what another person is feeling” (Hatfield et al., 1994, 2009). However, the role of facial mimicry as an essential link in the transmission of basic emotional states from one person to another, and thus its role in emotional contagion and subsequent emotion recognition, is not unequivocal. As apparent from various studies (e.g., Blairy et al., 1999; Hess & Blairy, 2001; Lischetzke et al., 2020; Olszanowski et al., 2020), facial mimicry may be involved in emotional contagion but its role may vary between individuals and may also depend on situational factors such as whether observed emotional expressions are strong or weak, are realistic or posed, can be easily recognized or not, represent a stereotypical emotion or a more complex emotional state, a positive or negative emotion, or a repulsed or non‐repulsed emotional response. In addition, in experimental studies facial mimicry may be dependent on whether still pictures of emotional facial expressions or dynamic expressions with higher ecological validity are used as stimuli.
Whether facial mimicry occurs in daily life will primarily be determined by the social context, implying that mimicry does not automatically occur (Fischer & Hess, 2017; Hess, 2020). As elucidated by these authors, emotional expressions will be mimicked if expressor and mimicker share the perspective that gave rise to the emotion. The emotional or social meaning of facial expressions may be mimicked rather than facial features as such. Mimicry is thus not a priori an automatic process but may be considered a controlled process determined by social interactions. It will not only be determined by positive but also by negative affiliative connections between persons. Affiliative emotions, such as happiness or sadness, may be mimicked stronger than non‐affiliative expressions like anger or disgust. Mimicry is thus primarily dependent on the goal to affiliate with others and not on observing facial motor activities of others. Also factors like competition, conflict, ingroup membership, social power, and personality traits may determine if mimicry strictly occurs. This would imply that in many social contexts mimicry cannot be considered a necessary determinant of emotion recognition.
Nevertheless, besides such social mimicry responses with relatively long response latencies, short‐latency automatic mimicry responses occur which might have diagnostic value for psychopathy. In healthy persons, passively viewing static or dynamic emotional facial expressions induces short‐latency facial mimicry responses. EMG responses with latencies shorter than 500 ms were observed in corrugator supercilii and zygomaticus major muscles following presentations of angry or happy expressions, respectively (Achaibou et al., 2008; de Wied et al., 2006; Dimberg, 1997; Dimberg & Thunberg, 1998). Such short‐latency responses may be considered automatic responses since they are difficult to control voluntarily by advertent suppression or contracting facial muscles with an incongruent action (Dimberg et al., 2002). Such automatic responses also occurred when emotional expressions were not recognized because stimuli were presented only a few tens of milliseconds and backward masked by a neutral stimulus (Bailey & Henry, 2009; Dimberg et al., 2000). This suggests that short‐latency facial mimicry responses are automatic, preconscious responses.
The amygdala might play an important role in short‐latency facial mimicry responses. It is involved in the earliest, automatic responses to emotional facial expressions, in particular dynamic expressions (Adolphs, 2002). Such stimuli may reach the amygdala across different fast‐conducting pathways, being either a direct subcortical pathway from the retina across superior colliculus and pulvinar, or fast‐conducting pathways across early striate or extrastriate visual cortical areas (Garvert et al., 2014; Pessoa & Adolphs, 2010; Tamietto & De Gelder, 2010). Anyway, the amygdala is assumed to play an important role in various short‐latency networks involving subcortical and cortical visual areas (Johnson, 2005; Pessoa & Adolphs, 2010). Within these trajectories, response latencies of the amygdala to emotional facial expressions are shorter than 100 ms (Diano et al., 2017; Méndez‐Bértholo et al., 2016) because it is activated by fast‐conducting magnocellular cells, either through the trajectory across superior colliculus and pulvinar or through fast cortical channels (Tamietto & De Gelder, 2010; Vuilleumier, 2005). Therefore, the amygdala quickly responds to low spatial frequency information in facial expressions and thus responds to more coarse aspects of the expression like actions around eyes and mouth which are prominent in emotional expressions (Adolphs, 2008; Johnson, 2005; Vuilleumier et al., 2003; Whalen et al., 2004).
Rapid activation of this fast amygdalar route majorly has preconscious effects. In healthy persons or patients with cortical blindness it did not lead to conscious experience of emotional stimuli (Tamietto & De Gelder, 2010). This route was also activated when emotional faces were backward masked (Adolphs, 2008; Killgore & Yurgelun‐Todd, 2004; Whalen et al., 1998). Patients with cortical blindness or contralateral neglect were able to recognize emotional facial expressions when stimuli were presented to the blind half of the visual field (Adolphs, 2003a; Pegna et al., 2005). They showed similar mimicry responses to these stimuli and stimuli presented to the intact half of the field (Tamietto et al., 2009). Low spatial frequency information also underlies most visual abilities in infants who can detect coarse facial and emotional cues in the absence of a mature cortical visual system (Vuilleumier et al., 2003).
The amygdala is particularly involved in spontaneous emotional responses of muscles in the upper part of the face like frontalis, corrugator supercilii, and orbicularis oculi. Contrary to lower facial muscles, upper facial muscles are less susceptible to voluntary control by neurons in the contralateral primary motor or premotor cortex and are less well represented in these cortical areas than lower facial muscles (Rinn, 1984). They are largely controlled by direct ipsilateral and contralateral projections from a motor region within the anterior cingulate cortex (Morecraft et al., 2001). This region receives projections from the lateral and accessory basal nuclei of the amygdala (Morecraft et al., 2007). It specifically subserves spontaneous, involuntary emotional expressions and functions independently of voluntary actions being controlled by the primary motor or premotor cortex (Cattaneo & Pavesi, 2014). The voluntary cortical motor system cannot affect a genuine spontaneous emotional motor response. This may explain why spontaneous mimicry responses to emotional facial expressions are difficult to control voluntarily (Dimberg et al., 2002). The amygdalar route across anterior cingulate obviously provides short‐latency responses in facial muscles. In patients with occipital lobe lesions, the photic blink reflex recorded in orbicularis oculi showed an EMG response latency of 50 ms irrespective of whether the light flash was presented to the blind or sighted visual hemifield (Hackley & Johnson, 1996), suggesting the involvement of the fast processing route across the amygdala.
Although there exists a long‐standing discussion whether the amygdala primarily responds to negative facial emotions, particularly fear, more recent studies in healthy persons show that during passive exposure it responds to all negative and positive basic emotions (Adolphs, 2010; Costafreda et al., 2008; Fusar‐Poli et al., 2009; Sergerie et al., 2008). Nevertheless, fearful expressions generally elicited larger responses in central or basolateral amygdalar nuclei than angry or happy faces (Costafreda et al., 2008; Whalen et al., 2001). Compared with healthy persons, patients with amygdala damage were generally impaired in recognition of negative emotions, particularly, but not exclusively, fear (Adolphs, 2003b; Rapcsak et al., 2000; Zald, 2003). This deficit may be related to problems with fast, automatic processing of the eye region without sensory awareness which may be important for rapid detection of threatening situations being characterized by expressions of fear or anger (Adolphs, 2008; Diano et al., 2017; Rotshtein et al., 2010).
In persons with psychopathic traits, impaired recognition of emotional facial expressions occurs. Particularly expressions of fear, sadness, or anger were worser recognized by (generally male) adults (Blair et al., 2004), adolescents (Fairchild et al., 2009; Muñoz, 2008), or children (Blair & Coles, 2000; Blair et al., 2001; Stevens et al., 2001) with such traits. Fear‐recognition deficits in children with these traits could also be demonstrated using a paradigm testing automatic, preconscious detection of emotional facial expressions (Sylvers et al., 2011), suggesting that deficits are not caused by a lack of conscious attention to emotionally salient cues. Although many studies focused on the relationship between psychopathy and poor recognition of fear or sadness, a meta‐analysis of studies in adults, adolescents, or children with psychopathy also showed deficits in recognition of happiness, surprise, anger, and disgust (Dawel et al., 2012). In all studies mentioned here, including those in the meta‐analysis, presentation times of expressions were relatively long (at least 1 s) so that a specific role of early processing stages in impaired recognition could not be evaluated.
The question is whether the amygdala is involved in such recognition deficits. Problems with recognition of distress signals like fearful or sad expressions in youths and adults with psychopathic traits may be related to amygdala dysfunction already existing at a young age (Blair, 2013). Reduced amygdala responses to negative facial expressions were observed in both adults, adolescents, and children with psychopathic tendencies (Blair, 2010; Decety et al., 2014; Jones et al., 2009; Marsh et al., 2008; Passamonti et al., 2010). Although also in these studies presentation times of expressions were relatively long (at least 1 s), in a few studies subnormal amygdala responses obviously occurred at a pre‐attentive level. In children with high CU traits, backward‐masked unrecognized facial expressions of fear being presented for 17 ms induced an abnormally low amygdala response (Viding et al., 2012). Another study performed in youths under conditions of low attentional load found that higher CU traits were related to smaller amygdala responses to fearful expressions being presented for 200 ms (White et al., 2012). In these two studies, high CU traits were thus obviously related to subnormal early amygdala responses to fearful expressions.
As indicated above, the fast route between amygdala and motoneurons of upper facial muscles (cf. Morecraft et al., 2001, 2007) would enable short‐latency facial mimicry responses to emotional facial expressions. Several studies suggest that this route may be differently involved in dynamic and static emotional expressions. In healthy persons, effects of emotional facial expressions on facial mimicry responses, amygdala responses, and emotion recognition differed between dynamic expressions, morphed pictures, and still pictures. Mimicry responses to dynamic expressions of anger, fear, disgust, or happiness were larger than responses to still pictures (Krumhuber et al., 2013; Rymarczyk et al., 2011, 2016; Sato et al., 2008) and showed stronger relationships with experienced negative or positive emotional valence (Sato et al., 2013). Dynamic expressions of fear (Sato et al., 2004), anger, or happiness (Arsalidou et al., 2011) also elicited larger amygdala responses than static expressions. Dynamic expressions like those presented in video clips have several advantages compared with morphed pictures which show linear changes across time within all locations of the face. The natural, nonlinear unfolding of dynamic expressions has a higher ecological validity than morphed pictures and resulted in faster and more accurate emotion recognition as well as higher judgements of experienced emotion intensity (Calvo et al., 2016; Krumhuber et al., 2013).
In the current study we investigated whether adolescents with DBD and high or low CU traits showed abnormalities in short‐latency mimicry responses, particularly to dynamic emotional facial expressions. Dynamic expressions may have large ecological validity since during normal human interactions facial expressions are predominantly dynamic. We used 500‐ms video clips, during which expressions increased from neutral to maximal intensity. Earlier work using morphed pictures showed that the rate of change of pictures influenced rated intensity and naturalness of perceived expressions (Kamachi et al., 2001; Sato & Yoshikawa, 2004; Yoshikawa & Sato, 2008). For expressions of basic emotions, experienced intensity and naturalness became generally higher with a higher rate of change. The optimal duration of expressions was in the range of 500–740 ms for fear, happiness, and anger but appeared to be somewhat longer (900–1000 ms) for sadness (Hoffmann et al., 2010; Sato & Yoshikawa, 2004). Kinematic manipulation of facial expressions of happy, angry, and sad emotional states confirmed that emotion recognition becomes better with increasing speed of angry or happy expressions and with decreasing speed of sad expressions (Sowden et al., 2021). The major question of our study was whether adolescents with high CU traits showed abnormal short‐latency mimicry responses to dynamic emotional expressions. For this purpose we analyzed facial EMG responses during subsequent 100‐ms intervals following onset of 500‐ms video clips.
2. METHOD
The Medical Ethical Committee of the University Medical Center Utrecht approved the study protocol and both parents and adolescents gave written consent prior to participation.
2.1. Participants
The current data were collected during an experimental session in which also responses to empathy‐inducing film clips were investigated which were earlier reported (de Wied et al., 2012). Participants, their recruitment, diagnostic and psychometric characteristics, and inclusion criteria are extensively described in this earlier report. A group of 31 male adolescents (aged 12–15 years) with DBD as set out in the DSM‐IV‐TR (American Psychiatric Association, 2000), including 17 participants with ODD and 14 with CD, participated in this study. They were recruited from special schools for adolescents with severe behavioral problems. The presence of ODD or CD was assessed using the parent version of the Diagnostic Interview Schedule for Children (DISC‐IV, Dutch version) (Ferdinand & van der Ende, 2000). Thirty‐two Male normal control (NC) adolescents were recruited from a regular school.
Parents and teachers of all participants completed the Child Behavior Checklist (CBCL/4‐18; Achenbach, 1991a) and Teacher's Report Form (TRF/4‐18; Achenbach, 1991b), respectively. DBD adolescents obtained significantly higher scores than controls on the CBCL and TRF externalizing and internalizing scales, which confirmed the presence of group differences in conduct problems (see de Wied et al., 2012 for details).
Parents and teachers also completed the Antisocial Process Screening Device (APSD) (Frick & Hare, 2001), in the Dutch translation (de Wied et al., 2014). The APSD is a 20‐item questionnaire designed to measure psychopathic traits in children and adolescents. This scale includes three factors: Callous‐Unemotional (six items), Narcissism (seven items), and Impulsivity (five items). The intercorrelations between parents and teachers for the three dimensions were: r CU = .46, p < .01; r Narcissism = .52, p < .01; r Impulsivity = .51, p < .01. Ratings from parents and teachers were combined by using the highest score for each item (cf. Frick & Hare, 2001). The CU dimension, capturing the callous interpersonal style that is critical to the construct of psychopathy, was used to assign DBD adolescents to groups with high (CU+; n = 14) or low (CU−; n = 17) CU traits. Because Narcissism and Impulsivity were highly correlated in our study (r = .83, p < .001), they were combined to form the I/CP factor, reflecting impulsivity and conduct problems. Internal consistency was acceptable for the CU dimension (α = .71) and good for the I/CP dimension (α = .93).
As we earlier reported (de Wied et al., 2012), the CU+ group obtained significantly higher scores than the CU− group or control group on the APSD total score and CU and I/CP dimensions. The CU− group, in turn, obtained higher scores than controls on the APSD total score and I/CP dimension but not on the CU dimension. There were no group differences in age or intelligence.
2.2. Emotional facial stimuli
Videos of dynamic facial expressions of four different emotions (happiness, anger, sadness, fear) were presented. Two female and two male actors (aged 15–25 years) were trained to produce these expressions according to a fixed time schedule. Videos had a duration of 5.5 s and consisted of five contiguous stages: (1) a 2‐s still picture of the actor's neutral face; (2) a 500‐ms dynamic expression, increasing from neutral to maximal expression; (3) maintaining maximal expression for 1500 ms; (4) a 500‐ms relaxation stage; (5) a 1‐s still picture of the neutral face. During the training, the actors were assisted by a moving time bar at the bottom of the computer screen indicating the duration of the subsequent stages in different colors. The emotional expressions were based on the Facial Action Coding System (FACS) (Ekman & Friesen, 1978). FACS provides a detailed description of the muscular basis and outward manifestation of each expression in terms of so‐called action units (AUs). In the current study, crucial AUs were AU12 (pulling up lip corner; portrayal of happiness), AU4 (lowering eyebrows; anger, sadness, fear), AU1 (raising inner part of eyebrow; fear, sadness), and AU2 (raising outer part of eyebrow; fear) (cf. Deschamps et al., 2012; Ekman, 1979). The actors were trained under supervision of a certified FACS coder (author RZ). Besides the videos of dynamic emotional expressions, a 5‐s video was made of a neutral expression produced by each actor.
Videos of the dynamic facial expressions were evaluated by 15 male adolescents from the normal population (aged 11–16 years; mean age 11.5 years) and 14 female university students (aged 20–46 years; mean age 24.9 years). Emotional expressions were correctly identified in 90.8% of cases by males and 98.2% by females. Figures for correct identification of emotions were 96.7% for happiness, 95.0% for anger, 98.3% for sadness, and 88.1% for fear.
Still pictures of full facial expressions of the four emotions were also presented (Matsumoto & Ekman, 1988), each particular emotion being depicted by two female and two male models. Each picture was presented for 3 s and preceded by a 3‐s neutral expression by the same model with a 0.5‐s interval between pictures.
2.3. Experimental procedure
The experiment started with a 5‐min excerpt of an aquatic video (Coral Sea Dreaming, Small World Music Inc.) to induce a state of relaxation (Piferi et al., 2000). Next, participants were instructed to look at the videos of facial expressions. First, the four video's of neutral expressions were presented in a random order with intervals randomly varying between 1.5 and 2 s. After a pause of 8 s, eight videos (the four emotions each being displayed by a female and a male actor) were presented in a random order, each particular video being presented four times in succession. All 32 videos were presented with intervals randomly varying between 1.5 and 2 s. Thereupon, participants were instructed to look at still pictures of facial expressions. Sixteen pairs of pictures (four emotions depicted by four models), each pair consisting of a neutral and a full emotional expression by the same model, were presented in a random order with intervals between pairs randomly varying between 1.5 and 2 s. The instructions preceding videos and pictures implied just watching the stimuli attentively, avoiding any task demands like identification of gender or emotion.
2.4. Facial EMG recordings and quantification
EMG was bipolarly recorded from the left frontalis (involved in AU1 and AU2), corrugator supercilii (AU4), and zygomaticus major (AU12) muscles using surface Ag/AgCl electrodes (contact area 2 mm diameter, 15 mm distance between electrode centers). EMG signals were antialiasing filtered using a 512‐Hz lowpass filter, digitized at a rate of 1024 Hz, 20‐Hz digitally high‐pass filtered to remove low‐frequency artifacts (van Boxtel, 2001), and bandreject filtered (48–52 Hz) to remove 50‐Hz power line interference. EMG responses were visually inspected for remaining technical artifacts or strong potentials caused by disruptive actions like coughing, sneezing, strong eyeblinks, etcetera. For dynamic expressions, 2.6%, 2.6%, and 1.4% of EMG responses had to be discarded for corrugator, zygomaticus, and frontalis, respectively. For static expressions, these figures were 1.3%, 1.9%, and 1.3%.
EMG responses to videos of dynamic or static emotional expressions produced by the actors were quantified by calculating mean rectified EMG activity during subsequent 100‐ms periods. The duration of this period was based on a study showing that mechanical changes during fast frontalis contractions (as apparent from the so‐called mechanomyogram) rapidly decline at frequencies above 10 Hz (Alves & Chau, 2010). In case of dynamic expressions, raw EMG values were standardized by expressing them as a percentage of mean rectified EMG activity during the 2‐s neutral expression baseline period preceding the emotional expression. Standardized values were first averaged across the four presentations of the same emotion by the same actor and subsequently across the two actors expressing this emotion. If an EMG response to a dynamic expression had to be discarded due to artifacts, responses were averaged across the remaining two or three stimuli presented by the same actor. EMG responses during videos of neutral expressions were expressed as a percentage of mean rectified EMG during the entire 5‐s duration of the video.
Regarding still pictures of full facial expressions, raw EMG values during 100‐ms periods of emotional expressions were expressed as a percentage of mean rectified EMG activity during the preceding neutral expression. Subsequently, percentages were averaged across the four models expressing the same emotion. If a response had to be discarded, responses were averaged across the remaining three models.
2.5. Statistical analysis
Regarding dynamic expressions, mean EMG activity during the 500‐ms dynamic phase of expression was pairwise compared between the three subject groups using t‐tests for independent samples assuming equal or unequal variances. To apply these analyses with a greater temporal resolution, group comparisons were also performed for separate 100‐ms periods within the 500‐ms dynamic phase. Group comparisons were also performed for mean EMG activity level during the subsequent 1500‐ms apex of expression.
Based on studies demonstrating weak recognition of emotional facial expressions (Dawel et al., 2012; Muñoz, 2008), or weak amygdala responses to such expressions (Viding et al., 2012), in high CU groups, we expected smaller EMG responses in CU+ than in CU− or NC groups. We did not expect different EMG responses between CU− and NC because these two groups had comparable scores on the CU dimension (de Wied et al., 2012). Given these expectations, we applied a priori comparisons between pairs of groups given the advantages of such comparisons in terms of statistical power compared with performing analysis of variance with orthogonal follow‐up comparisons between groups (Ruxton & Beauchamp, 2008; Thompson, 1987). Homoscedastic or heteroscedastic one‐tailed t‐tests were thus performed for comparisons between CU+ and CU− or NC, and two‐tailed tests for comparisons between CU− and NC. EMG responses are reported insofar as a muscle is involved in a specific emotional expression, that is, zygomaticus in happy expressions, corrugator in angry, sad, or fearful expressions, and frontalis in sad or fearful expressions. Corrugator responses to happy expressions were also investigated because this muscle shows diminished activity relative to baseline during happy expressions (Deschamps et al., 2012; de Wied et al., 2006; Rymarczyk et al., 2011; Sato et al., 2008).
Similarly to dynamic expressions, EMG activity during the first 500‐ms of the 3‐s still pictures was compared between groups. Such comparisons were also made during the subsequent 2.5‐s period of the stimulus.
3. RESULTS
3.1. Dynamic facial expressions
No group showed clear EMG responses to videos of neutral expressions (Figure 1). Several group differences in facial mimicry responses were observed during the 500‐ms rise time of dynamic expressions (Figure 2a). During happy expressions, the overall increase in zygomaticus activity was significantly weaker in the CU+ group than in CU− (t 19 = −2.19, p = .021, Cohen's d = 1.00) or NC groups (t 44 = −2.48, p = .008, d = 0.75) whereas CU− and NC did not differ (t 22 = 0.88, p = .387, d = 0.38). On a more fine‐grained time scale, significant group differences in increased zygomaticus activity were already observed as early as 200–300 ms following stimulus onset (Table 1). During happy expressions, corrugator showed an overall inhibition of activity during the 500‐ms rise time which was significantly weaker for CU+ than for CU− (t 29 = 2.27, p = .016, d = 0.84) or NC (t 40 = 1.73, p = .046, d = 0.55) whereas CU− and NC did not differ (t 47 = −0.73, p = .471, d = 0.21). On a more fine‐grained time scale, significant group differences were not observed earlier than 400–500 ms following stimulus onset (Table 1).
TABLE 1.
Interval (ms) | 0–100 | 100–200 | 200–300 | 300–400 | 400–500 |
---|---|---|---|---|---|
Zygomaticus response to happy face | |||||
CU+ (% baseline) | 102.9 | 105.0 | 107.0 | 99.6 | 102.6 |
CU− (% baseline) | 108.9 | 110.4 | 122.7 | 123.7 | 127.6 |
NC (% baseline) | 99.9 | 101.8 | 118.7 | 121.9 | 120.5 |
CU+ vs. CU− | t 29 = −1.33 + d = 0.29 | t 27 = −0.72 d = 0.28 | t 27 = −2.10* d = 0.81 | t 18 = −1.92* d = 0.91 | t 41 = −1.75* d = 0.55 |
CU+ vs. NC | t 19 = 1.05 d = 0.48 | t 44 = 0.76 d = 0.48 | t 38 = −1.97* d = 0.64 | t 44 = −3.29*** d = 0.99 | t 41 = −2.74** d = 0.86 |
CU− vs. NC | t 20 = 2.46* d = 1.10 | t 20 = 1.30 d = 0.58 | t 18 = 0.24 d = 0.11 | t 24 = 0.13 d = 0.05 | t 21 = 0.47 d = 0.21 |
Corrugator response to happy face | |||||
CU+ (% baseline) | 105.85 | 97.97 | 93.89 | 89.38 | 92.98 |
CU− (% baseline) | 100.62 | 97.03 | 90.38 | 85.22 | 86.08 |
NC (% baseline) | 102.00 | 97.82 | 92.07 | 86.76 | 87.70 |
CU+ vs. CU− | t 29 = 1.71* d = 0.64 | t 41 = 0.37 d = 0.12 | t 29 = 1.22 d = 0.45 | t 29 = 1.42 + d = 0.53 | t 41 = 2.18* d = 0.68 |
CU+ vs. NC | t 44 = 1.79* d = 0.54 | t 44 = 0.06 d = 0.02 | t 44 = 0.59 d = 0.18 | t 44 = 0.88 d = 0.27 | t 44 = 1.74* d = 0.52 |
CU− vs. NC | t 24 = −0.58 d = 0.24 | t 47 = −0.31 d = 0.09 | t 47 = −0.63 d = 0.12 | t 47 = −0.55 d = 0.16 | t 47 = −0.53 d = 0.15 |
Corrugator response to angry face | |||||
CU+ (% baseline) | 101.69 | 104.40 | 105.44 | 105.66 | 111.20 |
CU− (% baseline) | 105.85 | 115.65 | 129.55 | 123.29 | 121.19 |
NC (% baseline) | 104.94 | 110.03 | 128.35 | 136.25 | 130.30 |
CU+ vs. CU− | t 29 = −1.16 d = 0.43 | t 21 = −1.95* d = 0.85 | t 24 = −2.79** d = 1.14 | t 23 = −2.36* d = 0.98 | t 26 = −1.36 + d = 0.53 |
CU+ vs. NC | t 44 = −1.39 + d = 0 42 | t 41 = −1.67 + d = 0.52 | t 42 = −3.52*** d = 1.09 | t 39 = −3.39*** d = 1.09 | t 43 = −2.38* d = 0.73 |
CU− vs. NC | t 23 = 0.31 d = 0.13 | t 24 = 0.95 d = 0.39 | t 47 = 0.14 d = 0.04 | t 46 = −1.20 d = 0.35 | t 45 = −0.97 d = 0.29 |
Corrugator response to sad face | |||||
CU+ (% baseline) | 99.98 | 99.68 | 102.75 | 107.12 | 102.35 |
CU− (% baseline) | 108.09 | 120.89 | 122.55 | 126.35 | 122.22 |
NC (% baseline) | 101.86 | 104.24 | 106.57 | 117.74 | 119.20 |
CU+ vs. CU− | t 20 = −1.29 d = 0.58 | t 19 = −2.96** d = 1.36 | t 20 = −2.10* d = 0.94 | t 20 = −1.75* d = 0.78 | t 18 = −2.38* d = 1.12 |
CU+ vs. NC | t 44 = −0.80 d = 0.24 | t 38 = −1.59 + d = 0.52 | t 44 = −0.86 d = 0.26 | t 44 = −2.00* d = 0.60 | t 44 = −3.94*** d = 1.19 |
CU− vs. NC | t 17 = 1.04 d = 0.50 | t 19 = 2.31* d = 1.06 | t 19 = 1.74* d = 0.80 | t 19 = 0.79 d = 0.36 | t 23 = 0.34 d = 0.14 |
Frontalis response to sad face | |||||
CU+ (% baseline) | 100.31 | 105.36 | 104.53 | 103.69 | 111.80 |
CU− (% baseline) | 103.99 | 116.78 | 117.62 | 119.65 | 117.10 |
NC (% baseline) | 100.58 | 103.59 | 102.73 | 105.87 | 103.62 |
CU+ vs. CU− | t 27 = −1.14 d = 0.44 | t 29 = −2.54** d = 0.94 | t 29 = −2.67** d = 0.99 | t 29 = −2.80** d = 1.04 | t 29 = −1.00 d = 0.37 |
CU+ vs. NC | t 44 = −0.10 d = 0.03 | t 44 = 0.49 d = 0.15 | t 44 = 0.49 d = 0.15 | t 44 = −0.52 d = 0.16 | t 44 = 1.89 d = 0.57 |
CU− vs. NC | t 47 = 1.12 d = 0.33 | t 47 = 3.52*** d = 1.03 | t 41 = 3.92*** d = 1.22 | t 47 = 3.25** d = 0.95 | t 18 = 1.43 d = 0.67 |
Corrugator response to fearful face | |||||
CU+ (% baseline) | 100.78 | 99.01 | 101.44 | 105.48 | 105.69 |
CU− (% baseline) | 108.33 | 104.25 | 104.53 | 112.37 | 115.24 |
NC (% baseline) | 109.34 | 114.98 | 104.77 | 104.54 | 101.28 |
CU+ vs. CU− | t 21 = −1.87* d = 0.82 | t 29 = −1.26 d = 0.47 | t 29 = −0.77 d = 0.29 | t 23 = −1.30 d = 0.54 | t 27 = −1.38 + d = 0.53 |
CU+ vs. NC | t 39 = −3.81*** d = 1.22 | t 42 = −4.24*** d = 1.31 | t 44 = −0.72 d = 0.22 | t 44 = 0.22 d = 0.07 | t 41 = 0.90 d = 0.28 |
CU− vs. NC | t 22 = −0.25 d = 0.11 | t 47 = −2.31* d = 0.67 | t 47 = −0.05 d = 0.01 | t 47 = 1.27 d = 0.37 | t 24 = 2.17* d = 0.89 |
Frontalis response to fearful face | |||||
CU+ (% baseline) | 105.70 | 105.07 | 101.58 | 104.61 | 108.43 |
CU− (% baseline) | 107.45 | 109.50 | 106.39 | 114.69 | 119.00 |
NC (% baseline) | 107.78 | 111.55 | 108.05 | 107.54 | 108.50 |
CU+ vs. CU− | t 22 = −0.26 d = 0.11 | t 17 = −0.76 d = 0.37 | t 29 = −1.81* d = 0.67 | t 29 = −2.96** d = 1.10 | t 29 = −2.07* d = 0.77 |
CU+ vs. NC | t 44 = −0.44 d = 0.13 | t 36 = −1.67 d = 0.56 | t 44 = −2.49** d = 0.75 | t 44 = −0.97 d = 0.29 | t 44 = −0.02 d = 0.01 |
CU− vs. NC | t 23 = −0.05 d = 0.02 | t 47 = −0.31 d = 0.09 | t 47 = −0.64 d = 0.19 | t 47 = 2.28* d = 0.67 | t 19 = 1.25 d = 0.57 |
p < .10;
p < .05;
p < .01;
p <. 001.
During angry expressions, the overall increase in corrugator activity during the 500‐ms rise time was significantly weaker for CU+ than for CU− (t 19 = −2.52, p = .010, d = 1.16) or NC (t 37 = −3.45, p = .001, d = 1.13) whereas CU− and NC did not differ (t 47 = −0.40, p = .693, d = 0.12). Following stimulus onset, these group differences specifically occurred as early as 100–200 ms for the difference between CU+ and CU−, and 200–300 ms for CU+ versus NC (Table 1).
During sad expressions, the overall increase in corrugator activity during the 500‐ms rise time was significantly weaker for CU+ than for CU− (t 18 = −2.37, p = .015, d = 1.12) or NC (t 42 = −1.87, p = .034, d = 0.58) whereas CU− and NC did not differ (t 47 = 1.40, p = .168, d = 0.41). The difference between CU+ and CU− occurred as early as 100–200 ms following stimulus onset whereas the difference between CU+ and NC occurred not earlier than 300–400 ms following stimulus onset. In addition, activity in CU− was unexpectedly larger than in NC during the 100–300 ms period (Table 1).
During sad expressions, the overall increase in frontalis activity during the 500‐ms rise time did not differ between CU+ and NC (t 44 = 0.59, p = .779, d = 0.18) and there was a tendency toward a significantly smaller overall increase in CU+ than in CU− (t 25 = −1.55; p = .067, d = 0.62). On a time scale with a resolution of 100 ms, however, the increase in CU+ was significantly smaller than in CU− during the 100–400 ms period (Table 1). There was also an overall tendency toward a significantly larger response in CU− relative to NC (t 19 = 2.07; p = .052, d = 0.95). This difference obtained significance during the 100–400 ms period (Table 1).
During fearful expressions, the overall increase in corrugator activity during the 500‐ms rise time did not differ between CU+ and NC (t 42 = −0.89, p = .189, d = 0.27) although in the 0–100 ms and 100–200 ms intervals, activity was significantly larger in NC than in CU+ (Table 1). There was only a tendency toward a significantly weaker increase in CU+ than in CU− (t 29 = −1.62, p = .058, d = 0.60). Responses in CU− and NC did not differ (t 47 = 0.36, p = .723, d = 0.11).
During fearful expressions, the overall increase in frontalis activity during the 500‐ms rise time did not differ between CU+ and CU− (t 19 = −1.24, p = .115, d = 0.57), although on a time scale with a resolution of 100 ms, activity was larger in CU− than in CU+ during the 200–500 ms period (Table 1). The overall increase in activity did not differ between CU+ and NC (t 42 = −1.15, p = .128, d = 0.35), nor between CU− and NC (t 47 = 0.52, p = .606, d = 0.15).
During the 1500‐ms apex of dynamic expressions, the large majority of EMG responses to the various expressions did not show significant group differences (p values ≥ .115) There was only one exception: during happy expressions, CU+ showed a significantly smaller inhibition of corrugator activity than NC (t 44 = 1.78, p = .041, d = 0.54). Moreover, there were several tendencies toward significant group differences in the expected direction. During happy expressions, there was a tendency toward a significantly smaller increase in zygomaticus activity in CU+ compared with NC (t 41 = −1.53, p = .067, d = 0.48), and a smaller inhibition of corrugator activity in CU+ than in CU− (t 29 = 1.45, p = .080, d = 0.54). During angry expressions, CU+ showed a tendency toward a significantly smaller increase in corrugator activity than CU− (t 19 = −1.66, p = .057, d = 0.76) or NC (t 38 = −1.38, p = .088, d = 0.45) groups.
3.2. Static facial expressions
During the initial 500‐ms period of the 3‐s presentations of static expressions, CU+ did not show significantly different EMG responses to the various expressions compared with CU− or NC (Figure 2b). However, CU− and NC showed two response differences. During angry expressions, corrugator activity was significantly smaller in CU− than in NC (t 41 = −2.04, p = .048, d = 0.64). During sad expressions, corrugator activity was smaller in CU− than in NC (t 47 = −2.72, p = .009, d = 0.79).
During the subsequent 2.5‐s period of static expressions, frontalis activity during sad expressions was significantly smaller in CU+ than in NC (t 44 = −1.91, p = .031, d = 0.58). During fearful expressions, frontalis activity was smaller in CU+ than in CU− (t 20 = −2.44, p = .012, d = 1.09) or NC (t 40 = −1.73, p = .046, d = 0.55).
4. DISCUSSION
As indicated in the introduction, the role of facial mimicry in emotional contagion and subsequent emotion recognition is not unequivocal, particularly not within the context of social interactions. Nevertheless, besides such social mimicry responses with relatively long response latencies, our study demonstrated the occurrence of short‐latency responses, which might play a role in early recognition of fast dynamic emotional expressions during daily life. Such responses may have a diagnostic value with regard to psychopathy because of the involvement of the amygdala, particularly in the case of dynamic emotional facial expressions. As outlined in the introduction, the amygdala is involved in early responses to emotional expressions. Therefore, we studied early mimicry responses to posed maximal dynamic expressions with a duration of 500 ms and a large degree of naturalness. We also studied mimicry responses to static emotional expressions, that is, the apex of dynamic expressions and still pictures of emotional expressions. Aside from performing group comparisons between responses across the entire 500‐ms dynamic expression period, we also performed such comparisons for separate 100‐ms periods during the various expressions to find out whether specific short‐latency responses differed between groups.
Adolescents with high CU traits generally showed smaller facial EMG responses to dynamic emotional facial expressions than participants with low CU traits or normal controls, particularly during happy, angry, and sad expressions. Such differences could often be observed already at short latencies (100 or 200 ms) following dynamic stimulus onset. Smaller responses in CU+ did generally not occur during the apex of dynamic expressions. Neither did they occur during the first 500 ms or the entire period of still picture presentation. We therefore conclude that specifically responses to dynamic expressions were subnormal in the CU+ group. Our results are in agreement with the conclusion that children and adolescents with CU traits show an impairment in recognizing emotional cues from other people (Blair et al., 2014). In addition, our results are consistent with the finding that reduced emotional responsiveness in DBD individuals with high relative to low CU traits is mostly demonstrated (a) with physiological measures, (b) in older (i.e., adolescent) samples, and (c) within an other‐oriented context (Northam & Dadds, 2020).
Although our conclusion holds for dynamic expressions of happiness, anger, and sadness, it does not apply to expressions of fear, which did not induce subnormal corrugator or frontalis responses in CU+. This seems paradoxical as the amygdala is sensitive to coarse visual input caused by wide‐open eyes in fearful faces (Vuilleumier et al., 2003; Whalen et al., 2004). However, the lack of differences between CU+ and NC or CU− could hypothetically be explained by a suppression of facial responses to fearful stimuli in the last two groups. The amygdala particularly responds to fearful facial expressions (Calder et al., 1996; Costafreda et al., 2008; Whalen et al., 2001). The central nucleus of the amygdala, being a nodal point of intra‐amygdalar circuits, is involved in fearful responses and projects to the ventrolateral periaqueductal grey area (VLPAG) in the midbrain (Bandler & Shipley, 1994). This area is involved in somatic and autonomic defensive responses to threatening stimuli (Cattaneo & Pavesi, 2014). VLPAG directly or indirectly projects to facial motoneurons (Holstege, 2002). Activity in VLPAG induced by a threatening stimulus mediates a passive coping strategy avoiding both tonic and phasic motor responses and leading to a complete immobility (Walker & Carrive, 2003). This absence of behavioral responses might explain the relatively small facial EMG responses to fearful faces in NC and CU− groups, thereby nullifying the expected difference with the CU+ group.
As explained in the introduction, early mimicry responses to dynamic emotional facial expressions may be considered automatic, preconscious responses that are difficult to suppress. If they would play a role in recognizing emotional facial expressions of other people, they necessarily should have a short latency to follow the natural dynamics of facial expressions occurring during normal human interactions. As revealed in the introduction, impaired recognition of emotional facial expressions by persons with psychopathic traits may be related to smaller amygdala responses, independent of the amount of attention devoted to the stimuli. We found that the CU+ group showed diminished early facial mimicry responses to dynamic happy, angry, or sad expressions whereas responses to static expressions were not abnormal. The latter agrees with a recent population‐based community study in adolescents demonstrating that amygdala responses to still pictures of happy, angry, sad, or fearful faces were not related to CU traits (Dotterer et al., 2020).
Subnormal mimicry responses to dynamic emotional facial expressions in the CU+ group might negatively influence recognition of such expressions during normal social interaction when dynamic expressions are prevalent and therefore crucial for emotion recognition. Subnormal mimicry may be related to smaller amygdala activation, consistent with the observation that amygdala activation is also smaller when facial muscle contractions are blocked by botulinum toxin (Hennenlotter et al., 2009).
It may be questioned whether subnormal mimicry responses to dynamic emotional expressions in the CU+ group—such mimicry responses being earlier designated as “motor empathy” (Blair, 2005)—are related to a lack of emotional empathy since emotional empathy requires that signals of emotional distress, like negative emotional facial expressions, are recognized (Blair, 2005; Blair et al., 2014). Both deficiencies might be based on a common amygdala dysfunction. In healthy persons, individual differences in the amplitude of short‐latency mimicry responses to dynamic or static emotional facial expressions appeared to be positively related to differences in emotional empathy (Dimberg et al., 2011; Dimberg & Thunberg, 2012; Rymarczyk et al., 2016).
A limitation of this study might be that we used videos of acted rather than spontaneous dynamic expressions. However, healthy persons who were able to distinguish acted from spontaneous dynamic smiles showed similar EMG mimicry responses to both types of stimuli (Krumhuber et al., 2014), suggesting that at amygdalar level there is no distinction between both types of expressions.
Another limitation might be that videos of specific emotional expressions were presented four times in succession rather than in a random order. In healthy persons, habituation of amygdala fMRI responses to successive presentations of pictures of emotional faces was repeatedly reported (Breiter et al., 1996; Fischer et al., 2003; Thomas et al., 2001; Whalen et al., 1998, 2001; Wright et al., 2001; Zald, 2003). Amygdala responses to film clips of emotional facial expressions did not show habituation if participants explicitly payed attention to the expression as part of a cognitive or motor task (van der Gaag et al., 2007). Nevertheless, we earlier observed habituation of corrugator EMG responses with repeated presentation of dynamic happy or angry faces in boys with DBD (de Wied et al., 2006).
Although videos and still pictures were presented in separate sections of the experiment and were each preceded by specific instructions, it remains uncertain whether EMG responses to static expressions could have been influenced by prior presentation of dynamic expressions.
5. CONCLUSION
In conclusion, male adolescents with DBD and high CU traits showed significantly smaller short‐latency facial mimicry responses to dynamic facial expressions of positive or negative emotions compared with boys with low CU traits or normal controls, whereas the last two groups did not differ in this respect. Mimicry responses to static facial expressions generally did not differ between groups. The subnormal dynamic mimicry responses in the high CU group may have negative consequences for human interactions during daily life since during such interactions dynamic rather than static expressions are common.
AUTHOR CONTRIBUTIONS
Anton van Boxtel: Conceptualization; Data curation; Formal analysis; Investigation; Methodology; Project administration; Resources; Software; Supervision; Validation; Visualization; Writing‐original draft; Writing‐review & editing. Ruud Zaalberg: Methodology. Minet de Wied: Conceptualization; Data curation; Funding acquisition; Investigation; Methodology; Project administration; Resources; Supervision; Validation; Writing‐review & editing.
van Boxtel, A. , Zaalberg, R. , & de Wied, M. (2021). Subnormal short‐latency facial mimicry responses to dynamic emotional facial expressions in male adolescents with disruptive behavior disorders and callous‐unemotional traits. Psychophysiology, 59, e13945. 10.1111/psyp.13945
REFERENCES
- Achaibou, A. , Pourtois, G. , Schwartz, S. , & Vuilleumier, P. (2008). Simultaneous recording of EEG and facial muscle reactions during spontaneous emotional mimicry. Neuropsychologia, 46, 1104–1113. 10.1016/j.neuropsychologia.2007.10.019 [DOI] [PubMed] [Google Scholar]
- Achenbach, T. M. (1991a). Manual for the child behavior checklist and 1991 profile. University of Vermont, Department of Psychiatry. [Google Scholar]
- Achenbach, T. M. (1991b). Manual for the teacher's report form and 1991 profile. University of Vermont, Department of Psychiatry. [Google Scholar]
- Adolphs, R. (2002). Recognizing emotion from facial expressions: Psychological and neurological mechanisms. Behavioral and Cognitive Neuroscience Reviews, 1, 21–62. 10.1177/1534582302001001003 [DOI] [PubMed] [Google Scholar]
- Adolphs, R. (2003a). Cognitive neuroscience of human social behaviour. Nature Reviews Neuroscience, 4, 165–178. 10.1038/nrn1056 [DOI] [PubMed] [Google Scholar]
- Adolphs, R. (2003b). Is the human amygdala specialized for processing social information? Annals of the New York Academy of Sciences, 985, 326–340. 10.1111/j.1749-6632.2003.tb07091.x [DOI] [PubMed] [Google Scholar]
- Adolphs, R. (2006). Perception and emotion. How we recognize facial expressions. Current Directions in Psychological Science, 15, 222–226. 10.1111/j.1467-8721.2006.00440.x [DOI] [Google Scholar]
- Adolphs, R. (2008). Fear, faces, and the human amygdala. Current Opinion in Neurobiology, 18, 166–172. 10.1016/j.conb.2008.06.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Adolphs, R. (2010). What does the amygdala contribute to social cognition? Annals of the New York Academy of Sciences, 1191, 42–61. 10.1111/j.1749-6632.2010.05445.x [DOI] [PMC free article] [PubMed] [Google Scholar]
- Alves, N. , & Chau, T. (2010). The design and testing of a novel mechanomyogram‐driven switch controlled by small eyebrow movements. Journal of NeuroEngineering and Rehabilitation, 7, 22. 10.1186/1743-0003-7-22 [DOI] [PMC free article] [PubMed] [Google Scholar]
- American Psychiatric Association . (2000). Diagnostic and statistical manual of mental disorders (4th ed., text revision). American Psychiatric Publishing. 10.1002/9780470479216.corpsy0271 [DOI] [Google Scholar]
- American Psychiatric Association . (2013). Diagnostic and statistical manual of mental disorders (5th ed.). American Psychiatric Publishing. 10.1176/appi.books.9780890425596 [DOI] [Google Scholar]
- Arsalidou, M. , Morris, D. , & Taylor, M. J. (2011). Converging evidence for the advantage of dynamic facial expressions. Brain Topography, 24, 149–163. 10.1007/s10548-011-0171-4 [DOI] [PubMed] [Google Scholar]
- Bailey, P. E. , & Henry, J. D. (2009). Subconscious facial mimicry is preserved in older adulthood. Psychology and Aging, 24, 995–1000. 10.1037/a0015789 [DOI] [PubMed] [Google Scholar]
- Bandler, R. , & Shipley, M. T. (1994). Columnar organization of the midbrain periaqueductal gray: Modules for emotional expression? Trends in Neurosciences, 17, 379–389. 10.1016/0166-2236(94)90047-7 [DOI] [PubMed] [Google Scholar]
- Blair, R. J. R. (2005). Responding to the emotions of others: Dissociating forms of empathy through the study of typical and psychiatric populations. Consciousness and Cognition, 14, 698–718. 10.1016/j.concog.2005.06.004 [DOI] [PubMed] [Google Scholar]
- Blair, R. J. R. (2010). Neuroimaging of psychopathy and antisocial behavior: A targeted review. Current Psychiatry Reports, 12, 76–82. 10.1007/s11920-009-0086-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- Blair, R. J. R. (2013). The neurobiology of psychopathic traits in youths. Nature Reviews Neuroscience, 14, 786–799. 10.1038/nrn3577 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Blair, R. J. R. , & Coles, M. (2000). Expressive recognition and behavioural problems in early adolescence. Cognitive Development, 15, 421–434. 10.1016/S0885-2014(01)00039-9 [DOI] [Google Scholar]
- Blair, R. J. R. , Colledge, E. , Murray, L. , & Mitchell, D. G. V. (2001). A selective impairment in the processing of sad and fearful expressions in children with psychopathic tendencies. Journal of Abnormal Child Psychology, 29, 491–498. 10.1023/A:1012225108281 [DOI] [PubMed] [Google Scholar]
- Blair, R. J. R. , Leibenluft, E. , & Pine, D. S. (2014). Conduct disorder and callous‐unemotional traits in youth. New England Journal of Medicine, 371, 2207–2216. 10.1056/NEJMra1315612 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Blair, R. , Mitchell, D. , Peschardt, K. S. , Colledge, E. , Leonard, R. A. , Shine, J. H. , Murray, L. K. , & Perrett, D. I. (2004). Reduced sensitivity to others' fearful expressions in psychopathic individuals. Personality and Individual Differences, 37, 1111–1122. 10.1016/j.paid.2003.10.008 [DOI] [Google Scholar]
- Blairy, S. , Herrera, P. , & Hess, U. (1999). Mimicry and the judgment of emotional facial expressions. Journal of Nonverbal Behavior, 23, 5–44. [Google Scholar]
- Breiter, H. C. , Etcoff, N. L. , Whalen, P. J. , Kennedy, W. A. , Rauch, S. L. , Buckner, R. L. , Strauss, M. M. , Hyman, S. E. , & Rosen, B. R. (1996). Response and habituation of the human amygdala during visual processing of facial expression. Neuron, 17, 875–887. 10.1016/S0896-6273(00)80219-6 [DOI] [PubMed] [Google Scholar]
- Calder, A. J. , Young, A. W. , Rowland, D. , Perrett, D. I. , Hodges, J. R. , & Etcoff, N. L. (1996). Facial emotion recognition after bilateral amygdala damage: Differentially severe impairment of fear. Cognitive Neuropsychology, 13, 699–745. 10.1080/026432996381890 [DOI] [Google Scholar]
- Calvo, M. G. , Avero, P. , Fernández‐Martin, A. , & Recio, G. (2016). Recognition thresholds for static and dynamic emotional faces. Emotion, 16, 1186–1200. 10.1037/emo0000192 [DOI] [PubMed] [Google Scholar]
- Cattaneo, L. , & Pavesi, G. (2014). The facial motor system. Neuroscience and Biobehavioral Reviews, 38, 135–159. 10.1016/j.neubiorev.2013.11.002 [DOI] [PubMed] [Google Scholar]
- Costafreda, S. G. , Brammer, M. J. , David, A. S. , & Fu, C. H. Y. (2008). Predictors of amygdala activation during the processing of emotional stimuli: A meta‐analysis of 385 PET and fMRI studies. Brain Research Reviews, 58, 57–70. 10.1016/j.brainresrev.2007.10.012 [DOI] [PubMed] [Google Scholar]
- Dawel, A. , O'Kearney, R. , McKone, E. , & Palermo, R. (2012). Not just fear and sadness: Meta‐analytic evidence of pervasive emotion recognition deficits for facial and vocal expressions in psychopathy. Neuroscience and Biobehavioral Reviews, 36, 2288–2304. 10.1016/j.neubiorev.2012.08.006 [DOI] [PubMed] [Google Scholar]
- de Wied, M. , van Boxtel, A. , Matthys, W. , & Meeus, W. (2012). Verbal, facial and autonomic responses to empathy‐eliciting film clips by disruptive male adolescents with high versus low callous‐unemotional traits. Journal of Abnormal Child Psychology, 40, 211–223. 10.1007/s10802-011-9557-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- de Wied, M. , van Boxtel, A. , Zaalberg, R. , Goudena, P. P. , & Matthys, W. (2006). Facial EMG responses to dynamic emotional facial expressions in boys with disruptive behavior disorders. Journal of Psychiatric Research, 40, 112–121. 10.1016/j.jpsychires.2005.08.003 [DOI] [PubMed] [Google Scholar]
- de Wied, M. , van der Baan, H. , Raaijmaakers, Q. , de Ruiter, C. , & Meeus, W. (2014). Factor structure and construct validity of the Dutch version of the antisocial process screening device. Journal of Psychopathology and Behavioral Assessment, 36, 84–92. 10.1007/s10862-013-9371-4 [DOI] [Google Scholar]
- Decety, J. , Skelly, L. , Yoder, K. J. , & Kiehl, K. A. (2014). Neural processing of dynamic emotional facial expressions in psychopaths. Social Neuroscience, 9, 36–49. 10.1080/17470919.2013.866905 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Deschamps, P. K. H. , Schutte, I. , Kenemans, J. L. , Matthys, W. , & Schutter, D. J. L. G. (2012). Electromyographic responses to emotional facial expressions in 6–7 year olds: A feasibility study. International Journal of Psychophysiology, 85, 195–199. 10.1016/j.ijpsycho.2012.05.004 [DOI] [PubMed] [Google Scholar]
- Diano, M. , Celeghin, A. , Bagnis, A. , & Tamietto, M. (2017). Amygdala response to emotional stimuli without awareness: Facts and interpretations. Frontiers in Psychology, 7, 2029. 10.3389/fpsyg.2016.02029 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dimberg, U. (1997). Facial reactions: Rapidly evoked emotional responses. Journal of Psychophysiology, 11, 115–123. [Google Scholar]
- Dimberg, U. , Andréasson, P. , & Thunberg, M. (2011). Emotional empathy and facial reactions to facial expressions. Journal of Psychophysiology, 25, 26–31. 10.1027/0269-8803/a000029 [DOI] [Google Scholar]
- Dimberg, U. , & Thunberg, M. (1998). Rapid facial reactions to emotional facial expressions. Scandinavian Journal of Psychology, 39, 39–45. 10.1111/1467-9450.00054 [DOI] [PubMed] [Google Scholar]
- Dimberg, U. , & Thunberg, M. (2012). Empathy, emotional contagion, and rapid facial reactions to angry and facial expressions. PsyCh Journal, 1, 118–127. 10.1002/pchj.4 [DOI] [PubMed] [Google Scholar]
- Dimberg, U. , Thunberg, M. , & Elmehed, K. (2000). Unconscious facial reactions to emotional facial responses. Psychological Science, 11, 86–89. 10.1111/1467-9280.00221 [DOI] [PubMed] [Google Scholar]
- Dimberg, U. , Thunberg, M. , & Grunedal, S. (2002). Facial reactions to emotional stimuli: Automatically controlled emotional responses. Cognition and Emotion, 16, 449–472. 10.1080/02699930143000356 [DOI] [Google Scholar]
- Dotterer, H. L. , Waller, R. , Hein, T. C. , Pardon, A. , Mitchell, C. , Lopez‐Duran, N. , Monk, C. S. , & Hyde, L. W. (2020). Clarifying the link between amygdala functioning during emotion processing and antisocial behaviors versus callous‐unemotional traits within a population‐based community sample. Clinical Psychological Science, 8, 918–935. 10.1177/2167702620922829 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ekman, P. (1979). About brows: Emotional and conversational signals. In Von Cranach M., Foppa K., Lepenies W., & Ploog D. (Eds.), Human ethology (pp. 169–202). Cambridge University Press. [Google Scholar]
- Ekman, P. , & Friesen, W. V. (1978). Facial action coding system: A technique for the measurement of facial action. Consulting Psychologists Press. [Google Scholar]
- Fairchild, G. , van Goozen, S. H. M. , Calder, A. J. , Stollery, S. J. , & Goodyer, I. M. (2009). Deficits in facial expression recognition in male adolescents with early‐onset or adolescence‐onset conduct disorder. Journal of Child Psychology and Psychiatry, 50, 627–636. 10.1111/j.1469-7610.2008.02020.x [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ferdinand, R. F. , & van der Ende, J. (2000). NIMH‐DISC‐IV: Diagnostic interview schedule for children (Authorized Dutch version). Erasmus Medical Center‐Sophia Children's Hospital. [Google Scholar]
- Fischer, A. , & Hess, U. (2017). Mimicking emotions. Current Opinion in Psychology, 17, 151–155. 10.1016/j.copsyc.2017.07.008 [DOI] [PubMed] [Google Scholar]
- Fischer, H. , Wright, C. I. , Whalen, P. J. , McInerney, S. C. , Shin, L. M. , & Rauch, S. L. (2003). Brain habituation during repeated exposure to fearful and neutral faces: A functional MRI study. Brain Research Bulletin, 59, 387–392. 10.1016/S0361-9230(02)00940-1 [DOI] [PubMed] [Google Scholar]
- Frick, P. J. , & Hare, R. D. (2001). Antisocial process screening device. Multi‐Health Systems. [Google Scholar]
- Frick, P. J. , & Kemp, E. E. (2021). Conduct disorders and empathy development. Annual Review of Clinical Psychology, 17, 391–416. 10.1146/annurev-clinpsy-081219-105809 [DOI] [PubMed] [Google Scholar]
- Frick, P. J. , Ray, J. V. , Thornton, L. C. , & Kahn, R. E. (2014). Can callous‐unemotional traits enhance the understanding, diagnosis, and treatment of serious conduct problems in children and adolescents? A comprehensive review. Psychological Bulletin, 140, 1–57. 10.1037/a0033076 [DOI] [PubMed] [Google Scholar]
- Fusar‐Poli, P. , Placentino, A. , Carletti, F. , Landi, P. , Allen, P. , Surguladze, S. , & Politi, P. (2009). Functional atlas of emotional faces processing: A voxel‐based meta‐analysis of 105 functional magnetic resonance imaging studies. Journal of Psychiatry and Neuroscience, 34, 418–432. https://doi.org/2010-01016-002 [PMC free article] [PubMed] [Google Scholar]
- Garvert, M. M. , Friston, K. J. , Dolan, R. J. , & Garrido, M. I. (2014). Subcortical amygdala pathways enable rapid face processing. NeuroImage, 102, 309–316. 10.1016/j.neuroimage.2014.07.047 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Hackley, S. A. , & Johnson, L. N. (1996). Distinct early and late subcomponents of the photic blink reflex: Response characteristics in patients with retrogeniculate lesions. Psychophysiology, 33, 239–251. 10.1111/j.1469-8986.1996.tb00421.x [DOI] [PubMed] [Google Scholar]
- Hatfield, E. , Cacioppo, J. T. , & Rapson, R. L. (1994). Emotional contagion. Cambridge University Press. [Google Scholar]
- Hatfield, E. , Rapson, R. L. , & Yen‐Chi, L. L. (2009). Emotional contagion and empathy. In Decety J. & Ickes W. (Eds.), The social neuroscience of empathy (pp. 19–30). MIT Press. [Google Scholar]
- Hennenlotter, A. , Dresel, C. , Castrop, F. , Ceballos‐Baumann, A. O. , Wohlschläger, A. M. , & Haslinger, B. (2009). The link between facial feedback and neural activity within central circuits of emotion—New insights from botulinum toxin‐induced denervation of frown muscles. Cerebral Cortex, 19, 537–542. 10.1093/cercor/bhn104 [DOI] [PubMed] [Google Scholar]
- Hess, U. (2020). Who to whom and why: The social nature of emotional mimicry. Psychophysiology, 58, e13675. 10.1111/psyp.13675 [DOI] [PubMed] [Google Scholar]
- Hess, U. , & Blairy, S. (2001). Facial mimicry and emotional contagion to dynamic emotional facial expressions and their influence on decoding accuracy. International Journal of Psychophysiology, 40, 129–141. 10.1016/S0167-8760(00)00161-6 [DOI] [PubMed] [Google Scholar]
- Hoffmann, H. , Traue, H. C. , Bachmayr, F. , & Kessler, H. (2010). Perceived realism of dynamic facial expressions of emotion: Optimal durations for the presentation of emotional onsets and offsets. Cognition and Emotion, 24, 1369–1376. 10.1080/02699930903417855 [DOI] [Google Scholar]
- Holstege, G. (2002). Emotional innervation of the facial musculature. Movement Disorders, 17, S12–S16. 10.1002/mds.10050 [DOI] [PubMed] [Google Scholar]
- Johnson, M. H. (2005). Subcortical face processing. Nature Reviews Neuroscience, 6, 766–774. 10.1038/nrnl766 [DOI] [PubMed] [Google Scholar]
- Jones, A. P. , Laurens, K. R. , Herba, C. M. , Barker, G. J. , & Viding, E. (2009). Amygdala hypoactivity to fearful faces in boys with conduct problems and callous‐unemotional traits. American Journal of Psychiatry, 166, 95–102. 10.1176/appi.ajp.2008.07071050 [DOI] [PubMed] [Google Scholar]
- Kamachi, M. , Bruce, V. , Mukaida, S. , Gyoba, J. , Yoshikawa, S. , & Akamatsu, S. (2001). Dynamic properties influence the perception of facial expressions. Perception, 30, 875–887. 10.1068/p3131 [DOI] [PubMed] [Google Scholar]
- Khvatskaya, Y. , & Lenzenweger, M. F. (2016). Motor empathy in individuals with psychopathic traits: A preliminary study. Journal of Personality Disorders, 30, 613–632. 10.1521/pedi_2015_29_219 [DOI] [PubMed] [Google Scholar]
- Killgore, W. D. S. , & Yurgelun‐Todd, D. A. (2004). Activation of the amygdala and anterior cingulate during nonconscious processing of sad versus happy faces. NeuroImage, 21, 1215–1223. 10.1016/j.neuroimage.2003.12.033 [DOI] [PubMed] [Google Scholar]
- Krumhuber, E. G. , Kappas, A. , & Manstead, A. S. R. (2013). Effects of dynamic aspects of facial expressions: A review. Emotion Review, 5, 41–46. 10.1177/1754073912451349 [DOI] [Google Scholar]
- Krumhuber, E. G. , Likowski, K. U. , & Weyers, P. (2014). Facial mimicry of spontaneous and deliberate Duchenne and non‐Duchenne smiles. Journal of Nonverbal Behavior, 38, 1–11. 10.1007/s10919-013-0167-8 [DOI] [Google Scholar]
- Lischetzke, T. , Cugialy, M. , Apt, T. , Eid, M. , & Niedeggen, M. (2020). Are those who tend to mimic facial expressions especially vulnerable to emotional contagion? Journal of Nonverbal Behavior, 44, 133–152. 10.1007/s10919-019-00316-z [DOI] [Google Scholar]
- Marsh, A. A. , Finger, E. C. , Mitchell, D. G. V. , Reid, M. E. , Sims, C. , Kosson, D. S. , Towbin, K. E. , Leibenluft, E. , Pine, D. S. , & Blair, R. J. (2008). Reduced amygdala response to fearful expressions in children and adolescents with callous‐unemotional traits and disruptive behavior disorders. American Journal of Psychiatry, 165, 712–720. 10.1176/appi.ajp.2007.07071145 [DOI] [PubMed] [Google Scholar]
- Matsumoto, D. , & Ekman, P. (1988). Japanese and Caucasian facial expressions of emotion (JACFEE) and neutral faces (JACNEUF). San Francisco State University. [Google Scholar]
- McMahon, R. J. , Witkiewitz, K. , & Kotler, J. S. (2010). Predictive validity of callous‐unemotional traits measured in early adolescence with respect to multiple antisocial outcomes. Journal of Abnormal Psychology, 119, 752–763. 10.1037/a0020796 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Méndez‐Bértolo, C. , Moratti, S. , Toledano, R. , Lopez‐Sosa, F. , Martínez‐Alvarez, R. , Mah, Y. H. , Vuilleumier, P. , Gil‐Nagel, A. , & Strange, B. A. (2016). A fast pathway for fear in human amygdala. Nature Neuroscience, 19, 1041–1049. 10.1038/nn.4324 [DOI] [PubMed] [Google Scholar]
- Morecraft, R. J. , Louie, J. L. , Herrick, J. L. , & Stilwell‐Morecraft, K. S. (2001). Cortical innervation of the facial nucleus in the non‐human primate: A new interpretation of the effects of stroke and related subtotal brain trauma on the muscles of facial expression. Brain, 124, 176–208. 10.1093/brain/124.1.176 [DOI] [PubMed] [Google Scholar]
- Morecraft, R. J. , McNeal, D. W. , Stilwell‐Morecraft, K. S. , Gedney, M. , Ge, J. , Schroeder, C. M. , & van Hoesen, G. W. (2007). Amygdala interconnections with the cingulate motor cortex in the rhesus monkey. Journal of Comparative Neurology, 500, 134–165. 10.1002/cne.21165 [DOI] [PubMed] [Google Scholar]
- Muñoz, L. C. (2008). Callous‐unemotional traits are related to combined deficits in recognizing afraid faces and body poses. Journal of the American Academy of Child and Adolescent Psychiatry, 48, 554–562. 10.1097/CHI.0b013e31819c2419 [DOI] [PubMed] [Google Scholar]
- Northam, J. C. , & Dadds, M. R. (2020). Is callous always cold? A critical review of the literature on emotion and the development of callous‐unemotional traits in children. Clinical Child and Family Psychology Review, 23, 265–283. 10.1007/s10567-019-00309-w [DOI] [PubMed] [Google Scholar]
- Olszanowski, M. , Wróbel, M. , & Hess, U. (2020). Mimicking and sharing emotions: A re‐examination of the link between facial mimicry and emotional contagion. Cognition and Emotion, 34, 367–376. 10.1080/02699931.2019.1611543 [DOI] [PubMed] [Google Scholar]
- Passamonti, L. , Fairchild, G. , Goodyer, I. M. , Hurford, G. , Hagan, C. C. , Rowe, J. B. , & Calder, A. J. (2010). Neural abnormalities in early‐onset and adolescence‐onset conduct disorder. Archives of General Psychiatry, 67, 729–738. 10.1001/archgenpsychiatry.2010.75 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pegna, A. J. , Khateb, A. , Lazeyras, F. , & Seghier, M. L. (2005). Discriminating emotional faces without primary visual cortices involves the right amygdala. Nature Neuroscience, 8, 24–25. 10.1038/nn1364 [DOI] [PubMed] [Google Scholar]
- Pessoa, L. , & Adolphs, R. (2010). Emotion processing and the amygdala: From a “low road” to “many roads” of evaluating biological significance. Nature Reviews Neuroscience, 11, 773–783. 10.1038/nrn2920 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Piferi, R. L. , Kline, K. A. , Younger, J. , & Lawler, K. A. (2000). An alternative approach for achieving cardiovascular baseline: Viewing an aquatic video. International Journal of Psychophysiology, 37, 207–217. 10.1016/S0167-8760(00)00102-1 [DOI] [PubMed] [Google Scholar]
- Pijper, J. , de Wied, M. , van Goozen, S. , & Meeus, W. H. J. (2017). Empathy problems in youth with disruptive behavior disorders, with and without callous unemotional traits. In Centifanti L. C. & Williams D. M. (Eds.), The Wiley handbook of developmental psychopathology (pp. 161–178). Wiley. 10.1002/9781118554470.ch7 [DOI] [Google Scholar]
- Rapcsak, S. Z. , Galper, S. R. , Comer, J. F. , Reminger, S. L. , Nielsen, L. , Kaszniak, A. W. , Verfaellie, M. , Laguna, J. F. , Labiner, D. M. , & Cohen, R. A. (2000). Fear recognition deficits after focal brain damage: A cautionary note. Neurology, 54, 575–581. 10.1212/WNL.54.3.575 [DOI] [PubMed] [Google Scholar]
- Reidy, D. E. , Krusemark, E. , Kosson, D. S. , Kearns, M. C. , Smith‐Darden, J. , & Kiehl, K. A. (2017). The development of severe and chronic violence among youth: The role of psychopathic traits and reward processing. Child Psychiatry and Human Development, 48, 967–982. 10.1007/s10578-017-0720-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rinn, W. E. (1984). The neuropsychology of facial expression: A review of the neurological and psychological mechanisms for producing facial expressions. Psychological Bulletin, 95, 52–77. 10.1037/0033-2909.95.1.52 [DOI] [PubMed] [Google Scholar]
- Rotshtein, P. , Richardson, M. P. , Winston, J. S. , Kiebel, S. J. , Vuilleumier, P. , Eimer, M. , Driver, J. , & Dolan, R. J. (2010). Amygdala damage affects event‐related potentials for fearful faces at specific time windows. Human Brain Mapping, 31, 1089–1105. 10.1002/hbm.20921 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Ruxton, G. D. , & Beauchamp, G. (2008). Time for some thinking about post hoc testing. Behavioral Ecology, 19, 690–693. 10.1093/beheco/arn020 [DOI] [Google Scholar]
- Rymarczyk, K. , Biele, C. , Grabowska, A. , & Majczynski, H. (2011). EMG activity in response to static and dynamic facial expressions. International Journal of Psychophysiology, 79, 330–333. 10.1016/j.ijpsycho.2010.11.001 [DOI] [PubMed] [Google Scholar]
- Rymarczyk, K. , Żurawski, Ł. , Jankowiak‐Siuda, K. , & Szatkowska, I. (2016). Emotional empathy and facial mimicry for static and dynamic expressions of fear and disgust. Frontiers in Psychology, 7, 1853. 10.3389/fpsyg.2016.01853 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Salekin, R. T. (2017). Research Review: What do we know about psychopathic traits in children? Journal of Child Psychology and Psychiatry, 58, 1180–1200. 10.1111/jcpp.12738 [DOI] [PubMed] [Google Scholar]
- Sato, W. , Fujimura, T. , Kochiyama, T. , & Suzuki, N. (2013). Relationships among facial mimicry, emotional experience, and emotion recognition. PLoS ONE, 8, e57889. 10.1371/journal.pone.0057889 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sato, W. , Fujimura, T. , & Suzuki, N. (2008). Enhanced facial EMG activity in response to dynamic facial expressions. International Journal of Psychophysiology, 70, 70–74. 10.1016/j.ijpsycho.2008.06.001 [DOI] [PubMed] [Google Scholar]
- Sato, W. , Kochiyama, T. , Yoshikawa, S. , Naito, E. , & Matsumura, M. (2004). Enhanced neural activity in response to dynamic facial expressions of emotion: An fMRI study. Cognitive Brain Research, 20, 81–91. 10.1016/j.cogbrainres.2004.01.008 [DOI] [PubMed] [Google Scholar]
- Sato, W. , & Yoshikawa, S. (2004). The dynamic aspects of emotional facial expressions. Cognition and Emotion, 18, 701–710. 10.1080/02699930341000176 [DOI] [Google Scholar]
- Sergerie, K. , Chochol, C. , & Armony, J. L. (2008). The role of the amygdala in emotional processing: A quantitative meta‐analysis of functional neuroimaging studies. Neuroscience and Biobehavioral Reviews, 32, 811–830. 10.1016/j.neubiorev.2007.12.002 [DOI] [PubMed] [Google Scholar]
- Sowden, S. , Schuster, B. A. , Keating, C. T. , Fraser, D. S. , & Cook, J. L. (2021). The role of movement kinematics in facial emotion expression production and recognition. Emotion. Advance online publication. 10.1037/emo0000835 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Stevens, D. , Charman, T. , & Blair, R. J. R. (2001). Recognition of emotion in facial expressions and vocal tones in children with psychopathic tendencies. Journal of Genetic Psychology, 162, 201–211. 10.1080/00221320109597961 [DOI] [PubMed] [Google Scholar]
- Sylvers, P. D. , Brennan, P. A. , & Lilienfeld, S. O. (2011). Psychopathic traits and preattentive threat processing in children: A novel test of the fearlessness hypothesis. Psychological Science, 22, 1280–1287. 10.1177/0956797611420730 [DOI] [PubMed] [Google Scholar]
- Tamietto, M. , Castelli, L. , Vighetti, S. , Perrozo, P. , Geminiani, G. , Weiskrantz, L. , & de Gelder, B. (2009). Unseen facial and bodily expressions trigger fast emotional reactions. Proceedings of the National Academy of Sciences, 106, 17661–17666. 10.1073/pnas.0908994106 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tamietto, M. , & de Gelder, B. (2010). Neural bases of the non‐conscious perception of emotional signals. Nature Reviews Neuroscience, 11, 697–709. 10.1038/nrn2889 [DOI] [PubMed] [Google Scholar]
- Thomas, K. M. , Drevets, W. C. , Whalen, P. J. , Eccard, C. H. , Dahl, R. E. , Ryan, N. D. , & Casey, B. J. (2001). Amygdala responses to facial expressions in children and adults. Biological Psychiatry, 49, 309–316. 10.1016/S0006-3223(00)01066-0 [DOI] [PubMed] [Google Scholar]
- Thompson, B. (1987, November). The importance of a priori contrasts in analysis of variance research. Paper presented at the Annual Meeting of the Mid‐South Educational Research Association, Mobile, AL.
- van Boxtel, A. (2001). Optimal signal bandwidth for the recording of surface EMG of facial, jaw, oral, and neck muscles. Psychophysiology, 38, 22–34. 10.1111/1469-8986.3810022 [DOI] [PubMed] [Google Scholar]
- van der Gaag, C. , Minderaa, R. B. , & Keysers, C. (2007). The bold signal in the amygdala does not differentiate between dynamic facial expressions. Social Cognitive and Affective Neuroscience, 2, 93–103. 10.1093/scan/nsm002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Viding, E. , & McCrory, E. (2019). Towards understanding atypical social affiliation in psychopathy. Lancet Psychiatry, 6, 437–444. 10.1016/S2215-0366(19)30049-5 [DOI] [PubMed] [Google Scholar]
- Viding, E. , & McCrory, E. (2020). Disruptive behaviour disorder: The challenge of delineating mechanisms in the face of heterogeneity. American Journal of Psychiatry, 177, 811–817. 10.1176/appi.ajp.2020.20070998 [DOI] [PubMed] [Google Scholar]
- Viding, E. , Sebastian, C. L. , Dadds, M. R. , Lockwood, P. L. , Cecil, C. A. M. , De Brito, S. A. , & McCrory, E. J. (2012). Amygdala response to preattentive masked fear in children with conduct problems: The role of callous‐unemotional traits. American Journal of Psychiatry, 169, 1109–1116. 10.1176/appi.ajp.2012.12020191 [DOI] [PubMed] [Google Scholar]
- Vuilleumier, P. (2005). How brains beware: Neural mechanisms of emotional attention. Trends in Cognitive Sciences, 9, 585–594. 10.1016/j.tics.2005.10.011 [DOI] [PubMed] [Google Scholar]
- Vuilleumier, P. , Armony, J. L. , Driver, J. , & Dolan, R. J. (2003). Distinct spatial frequency sensitivities for processing faces and emotional expressions. Nature Neuroscience, 6, 624–631. 10.1038/nn1057 [DOI] [PubMed] [Google Scholar]
- Walker, P. , & Carrive, P. (2003). Role of the ventrolateral periaqueductal gray neurons in the behavioral and cardiovascular responses to contextual conditioned fear and poststress recovery. Neuroscience, 116, 897–912. 10.1016/S0306-4522(02)00744-3 [DOI] [PubMed] [Google Scholar]
- Waller, R. , Wagner, N. J. , Barstead, M. G. , Subar, A. , Petersen, J. L. , Hyde, J. S. , & Hyde, L. W. (2020). A meta‐analysis of the associations between callous‐unemotional traits and empathy, prosociality, and guilt. Clinical Psychology Review, 75, 101809. 10.1016/j.cpr.2019.101809 [DOI] [PubMed] [Google Scholar]
- Whalen, P. J. , Kagan, J. , Cook, R. G. , Davis, F. C. , Kim, H. , Polis, S. , & Johnstone, T. (2004). Human amygdala responsivity to masked fearful eye whites. Science, 306, 2061. 10.1126/science.1103617 [DOI] [PubMed] [Google Scholar]
- Whalen, P. J. , Rauch, S. L. , Etcoff, N. L. , McInerney, S. C. , Lee, M. B. , & Jenike, M. A. (1998). Masked presentations of emotional facial expressions modulate amygdala activity without explicit knowledge. Journal of Neuroscience, 18, 411–418. 10.1523/jneurosci.18-01-00411.1998 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Whalen, P. J. , Shin, L. M. , McInerney, S. C. , Fischer, H. , Wright, C. I. , & Rauch, S. L. (2001). A functional MRI study of human amygdala responses to facial expressions of fear versus anger. Emotion, 1, 70–83. 10.1037/1528-3542.1.1.70 [DOI] [PubMed] [Google Scholar]
- White, S. F. , Marsh, A. A. , Fowler, K. A. , Schechter, J. C. , Adalio, C. , Pope, K. , Sinclair, S. , Pine, D. S. , & Blair, R. J. R. (2012). Reduced amygdala response in youths with disruptive behavior disorders and psychopathic traits: Decreased emotional response versus increased top‐down attention to nonemotional features. American Journal of Psychiatry, 169, 750–758. 10.1176/appi.ajp.2012.11081270 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Wright, C. L. , Fisher, H. , Whalen, P. J. , McInerney, S. C. , Shin, L. M. , & Rauch, S. L. (2001). Differential prefrontal cortex and amygdala habituation to repeatedly presented emotional stimuli. NeuroReport, 12, 379–383. 10.1097/00001756-200102120-00039 [DOI] [PubMed] [Google Scholar]
- Yoshikawa, S. , & Sato, W. (2008). Dynamic facial expressions of emotion induce representational momentum. Cognitive Affective and Behavioral Neuroscience, 8, 25–31. 10.3758/CABN.8.1.25 [DOI] [PubMed] [Google Scholar]
- Zald, D. H. (2003). The human amygdala and the emotional evaluation of sensory stimuli. Brain Research Reviews, 41, 88–123. 10.1016/S0165-0173(02)00248-5 [DOI] [PubMed] [Google Scholar]