Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2020 Jul 1.
Published in final edited form as: Psychol Sci Public Interest. 2019 Jul;20(1):1–68. doi: 10.1177/1529100619832930

Emotional Expressions Reconsidered: Challenges to Inferring Emotion From Human Facial Movements

Lisa Feldman Barrett 1,2,3,*, Ralph Adolphs 4, Stacy Marsella 1,5,6, Aleix Martinez 7, Seth D Pollak 8
PMCID: PMC6640856  NIHMSID: NIHMS1021596  PMID: 31313636

Abstract

It is commonly assumed that a person’s emotional state can be readily inferred from the person’s facial movements, typically called “emotional expressions” or “facial expressions.” This assumption influences legal judgments, policy decisions, national security protocols, and educational practices, guides the diagnosis and treatment of psychiatric illness, as well as the development of commercial applications, and pervades everyday social interactions as well as research in other scientific fields such as artificial intelligence, neuroscience, and computer vision. In this paper, we survey examples of this widespread assumption, which we refer to as the “common view”, and then examine the scientific evidence for this view with a focus on the six most popular emotion categories used by consumers of emotion research: anger, disgust, fear, happiness, sadness and surprise. The available scientific evidence suggests that people do sometimes smile when happy, frown when sad, scowl when angry, and so on, more than what would be expected by chance. Yet there is substantial variation in how people communicate anger, disgust, fear, happiness, sadness and surprise, across cultures, situations, and even within a single situation. Furthermore, similar configurations of facial movements variably express instances of more than one emotion category. In fact, a given configuration of facial movements, such as a scowl, often communicates something other than an emotional state. Scientists agree that facial movements convey a range of social information and are important for social communication, emotional or otherwise. But our review suggests there is an urgent need for research that examines how people actually move their faces to express emotions and other social information in the variety of contexts that make up everyday life, as well as careful study of the mechanisms by which people perceive instances of emotion in one another. We make specific research recommendations that will yield a more valid picture of how people move their faces to express emotions, and how they infer emotional meaning from facial movements, as situations of everyday life. This research is crucial to provide consumers of emotion research with the translational information they require.

Executive Summary

It is commonly assumed that a person’s face gives evidence of emotions because there is a reliable mapping between a certain configuration of facial movements, called a “facial expression,” and the specific emotional state that it supposedly signals. This common view of facial expressions remains entrenched in consumers of emotion research, as well as in some scientists, despite an emerging consensus among affective scientists that emotional expressions are considerably more context-dependent and variable. Nonetheless, this common view continues to fuel commercial applications in industry and government (e.g., automated detection of emotions from faces), guide how children are taught (e.g., with posters and books showing stereotyped facial expressions), and impact clinical and legal applications (e.g., diagnoses of psychiatric illnesses and courtroom decisions). In this paper, we evaluate the common view of facial expressions against a review of the evidence and conclude that it rests on a number of flawed assumptions and incorrect interpretations of research findings. Our review is the most comprehensive and systematic to date, encompassing studies of healthy adults across cultures, newborns and young children, as well as people who are congenitally blind, and confirms that specific emotion categories -- anger, disgust, fear, happiness, sadness, and surprise – are each expressed with a particular configuration of facial movements, more reliably than would be expected by mere chance, but contrary to the common view, instances of these emotion categories are NOT expressed with facial movements that are sufficiently reliable and specific across contexts, individuals, and cultures to be considered diagnostic displays of any emotional state. Nor do human perceivers, in fact, infer emotions from particular configurations of muscle movements in a sufficiently reliable and specific way that similarly generalizes. Studies of expression production and perception both demonstrate multiple sources of variability that contradict the common view that smiles, scowls, frowns, and the like, are reliable and specific “expressions of emotion.” We conclude the paper with specific recommendations for both scientists and consumers of science.

Introduction

Faces are a ubiquitous part of everyday life for humans. We greet each other with smiles or nods. We have face-to-face conversations on a daily basis, whether in person or via computers. We capture faces with smartphones and tablets, exchanging photos of ourselves and of each other on Instagram, Snapchat, and other social media platforms. The ability to perceive faces is one of the first capacities to emerge after birth: an infant begins to perceive faces within the first few days of life, equipped with a preference for face-like arrangements that allows the brain to wire itself, with experience, to become expert at perceiving faces (Arcaro et al., 2017; Cassia et al., 2004; Grossmann, 2015; Ghandi et al., 2017; Smith et al., 2018; Turati, 2004; but see Young & Burton (2018) for a more qualified claim).1 Faces offer a rich, salient source of information for navigating the social world: they play a role in deciding who to love, who to trust, who to help, and who is found guilty of a crime (Todorov, 2017; Zebrowitz, 1997, 2017; Zhang, Chen & Yang, 2018). Dating back to the ancient Greeks (Aristotle, in 4th century BCE) and Romans (Cicero), various cultures have viewed the human face as a window on the mind. But to what extent can a raised eyebrow, a curled lip, or a narrowed eye reveal what someone is thinking or feeling, allowing a perceiver’s brain to guess what that someone will do next?2 The answers to these questions have major consequences for human outcomes as they unfold in the living room, the classroom, the courtroom and even on the battlefield. They also powerfully shape the direction of research in a broad array of scientific fields, from basic neuroscience to psychiatry research.

Understanding what facial movements might reveal about a person’s emotions is made more urgent by the fact that many people believe we already know. Specific configurations of facial muscle movements appear as if they summarily broadcast or display a person’s emotions, which is why they are routinely referred to as “emotional expressions” and “facial expressions.”3 A simple Google search using the phrase “emotional facial expressions” [see Box 1, in supplementary on-line materials (SOM)] reveals the ubiquity with which, at least in certain parts of the world, people believe that certain emotion categories are reliably signaled or revealed by certain facial muscle movement configurations – a set of beliefs were refer to as the common view (also called the classical view; Barrett, 2017a). Similarly, many cultural products testify to the common view. Here are several examples:

  • Technology companies are investing tremendous resources to figure out how to objectively “read” emotions in people by detecting their presumed facial expressions, such as scowling faces, frowning faces and smiling faces in an automated fashion. Several companies claim to have already done it (e.g., https://www.affectiva.com/what/products/; https://azure.microsoft.com/en-us/services/cognitive-services/emotion/). For example, Microsoft’s Emotion API promises to take video images of a person’s face to detect what that individual is feeling. The application states: “The emotions detected are anger, contempt, disgust, fear, happiness, neutral, sadness, and surprise. These emotions are understood to be cross-culturally and universally communicated with particular facial expressions”(https://azure.microsoft.com/en-us/services/cognitive-services/emotion/).

  • Countless electronic messages are annotated with emojis or emoticons that are schematized versions of the proposed facial expressions for various emotion categories (https://www.apple.com/newsroom/2018/07/apple-celebrates-world-emoji-day/).

  • Putative emotional expressions are taught to preschool children by displaying scowling faces, frowning faces, smiling faces and so on, in posters (e.g., use “feeling chart for children” in a Google image search), games (https://www.amazon.com/s/ref=nb_sb_noss_2?url=search-alias%3Daps&field-keywords=miniland+emotion) and books (e.g., Cain, 2000; Parr, 2005), and on episodes of Sesame Street (among many examples, see https://www.youtube.com/watch?v=ZxfJicfyCdg, https://vimeo.com/108524970, or https://www.youtube.com/watch?v=y28GH2GoIvc).4

  • Television shows (e.g., Lie to Me), movies (e.g., Inside Out) and documentaries (e.g., The Human Face, produced by the British Broadcasting Company) customarily depict certain facial configurations as universal expressions of emotions.

  • Magazine and newspaper articles routinely feature stories in kind: facial configurations depicting a scowl are referred to as “expressions of anger,” facial configurations depicting a smile are referred to as “expressions of happiness,” facial configurations depicting a frown are referred to as “expressions of sadness,” and so on.

  • Agents of the U.S. Federal Bureau of Investigations (FBI) and the Transportation Security Administration (TSA) were trained to detect emotions and other intentions using these facial configurations, with the goal of identifying and thwarting terrorists (Rhonda Heilig, special agent with the FBI, personal communication, December 15, 2014, 11:20 am; https://how-emotions-are-made.com/notes/Screening_of_Passengers_by_Observation_Techniques).5

  • The facial configurations that supposedly diagnose emotional states also figure prominently in the diagnosis and treatment of psychiatric disorders. One of the most widely used task in autism research, the “Reading the Mind in the Eyes Test”, asks patients to match photos of the upper (eye) region of a posed facial configuration with specific mental state words, including emotion words (Baron-Cohen t al., 2001). Treatment plans for people living with autism and other brain disorders often include learning to recognize these facial configurations as emotional expressions (Baron-Cohen et al., 2004; Kouo & Egel, 2016). This training does not generalize well to real-world skills, however (Bergren et al., 2018; Kouo & Egel, 2016).

  • “Reading” the emotions of a defendant (in the words of Supreme Court Justice Anthony Kennedy -- to “know the heart and mind of the offender”) is one pillar of a fair trial in the U.S. legal system and in many legal systems in the Western world (see Riggins v. Nevada, 1992). Legal actors like jurors and judges routinely rely on facial movements to determine the guilt and remorse of a defendant (e.g., Bandes, 2014; Zebrowitz, 1997). For example, defendants who are perceived as untrustworthy receive harsher sentences than they otherwise would (Wilson & Rule, 2015, 2016), and such perceptions are more likely when a person appears to be angry (i.e., facial structure is similar to the hypothesized facial expression of anger, which is a scowl (Todorov, 2017). An incorrect inference about a defendant’s emotional state can cost someone her children, her freedom, or even her life (for recent examples, see Barrett, 2017, beginning on page 183).

But can a person’s emotional state be reasonably inferred from that person’s facial movements? In this paper, we offer a systematic review of the evidence, testing the common view that instances of emotion are signaled with a distinctive configuration of facial movements with enough consistently that it can serve as a diagnostic marker of those instances. We focus our review on evidence pertaining to six emotion categories that have received the lion’s share of attention in the scientific literature -- anger, disgust, fear, happiness, sadness and surprise – and that, correspondingly, are the focus of common view (as evidenced by our Google search, summarized in Box 1, SOM), but our conclusions apply to all emotion categories that have thus far been scientifically studied. We open the paper with a brief discussion of its scope, approach, and intended audience. We then summarize evidence on how people actually move their faces during episodes of emotion, referred to as studies of expression production studies, following which we examine evidence for which emotions are actually inferred from looking at facial movements, referred to as studies of emotion perception. We identify three key shortcomings in the scientific research that have contributed to a general misunderstanding about how emotions are expressed and perceived in facial movements, and that limit the translation of this scientific evidence for other uses:

  1. limited reliability (instances of the same emotion category are neither reliably expressed with or perceived from a common set of facial movements);

  2. lack of specificity (there is no unique mapping between a single configuration of facial movements and instances of the same emotion category); and,

  3. limited generalizability (the effects of context and culture have not been sufficiently documented and accounted for).

We then discuss our conclusions, followed by proposals for consumers on how they might use the existing scientific literature. We also provide recommendations for future research with consumers of emotion research in mind. We have included additional detail on some topics of import or interest in the supplementary on-line materials (SOM).

Scope, Approach and Intended Audience of Paper

The Common View: Reading an Inner Emotional State of Mind From A Set of Unique Facial Movements

In common English parlance, people refer to “emotions” or “an emotion” as if anger, happiness, or any emotion word refers to an object that is highly similar on every occurrence. But an emotion word refers not to a unitary entity, but to a category of instances that vary from one another in their physical features, such as facial expressions and bodily changes, and mental features. Few scientists who study emotion, if any, take the view that every instance of an emotion category, such as anger, is identical to every other instance, sharing a set of necessary and sufficient features across situations, people and cultures. For example, Keltner and Cordaro (2017) recently wrote, “there is no one-to-one correspondence between a specific set of facial muscle actions or vocal cues and any and every experience of emotion” (p. 62). Yet there is considerable scientific debate about the amount of the within-category variation, the specific features that vary, the causes of the within-category variation, and implications of this variation for the nature of emotion (see Figure 1).

Figure 1. Explanatory frameworks guiding the science of emotion: The nature of emotion categories and their concepts.

Figure 1.

Figure is plotted along two dimensions. Horizontal: represents hypotheses about the surface similarities shared by instances of the same emotion category (e.g., the facial movements that express instances of the same emotion category). Vertical: represents hypotheses about the deep similarities in the mechanisms that cause instances of the same emotion category (e.g., to what extent do instances in the same category share deep, causal features?). Colors represent the type of emotion categories that are proposed in each theoretical framework (green = ad hoc, abstract categories; yellow = prototype or theory-based categories; red = natural kind categories).

One popular scientific framework, referred to as the basic emotion approach, hypothesizes that instances of an emotion category are expressed with facial movements that vary, to some degree, around a typical set of movements (called a prototype) (for example, see Table 1). For example, it is hypothesized that in one instance, anger might be expressed with the expressive prototype (e.g., brows furrowed, eyes wide, lips tightened) plus additional facial movements, such as a widened mouth, whereas on other occasions, a facial movement in the prototype might be missing (e.g., anger might be expressed with narrowed eyes or without movement in the eyebrow region; for a discussion, see Box 2, in SOM). Nonetheless, the basic emotion approach still assumes that the core facial configuration – the prototype -- can be used to diagnose a person’s inner emotional state in much the same way that a fingerprint can be used to uniquely recognize a person. More substantial variation in expressions (e.g., smiling in anger, gasping with widened eyes in anger, and scowling not in anger, but in confusion or concentration) is typically explained as the result of some process that is independent of an emotion itself, such as display rules, emotion regulation strategies such as suppressing the expression, or culture-specific dialects (as proposed by various scientists, including Elfenbein, 2013, 2017; Ekman & Cordaro, 2011; Matsumoto, 1990; Matsumoto, Keltner, Shiota, Frank, & O’Sullivan, 2008; Tracy & Randles, 2011).

Table 1.

A comparison of the facial configurations listed as the expressions of selected emotion categories

Proposed Expressive Configurations Described as Facial Action Units
Emotion Category Matsumoto, Keltner, Shiota, O’Sullivan &
Frank (2008)
Cordaro, Sun, Keltner, Kamble,
Huddar & McNeil (2017)
Keltner et al. (in press) Physical Description
Darnin’s (1872) Description Observed in reseearch Reference Configuration Used International Core Pattern
Amusement Not listed Not listed 6, 12, 26 or 27, 55 or 56, a “head bounce” (Shiota, Campos & Keltner, 2003) 6, 7, 12, 16, 25, 26 or 27, 53 6+7+12+25+26+53 Head back, Duchenne smile (6, 7, 12), lips separated, jaw dropped
Anger 4+ 5+ 24+ 38 4 + 5 or 7 +22+23+24 4 +5 + 7 + 23 (Ekman, Levenson & Friesen, 1983) 4, 7 4+5+17+23+24 Brows furrowed, eyes wide, lips tightened and pressed together
Awe Not listed Not listed 1, 5, 26 or 27, 57 and visible inhalation (Shiota et al., 2003) 1, 2, 5, 12, 25, 26 or 27, 53 Not listed
Contempt 9+ 10+ 22+ 41+ 61 or 62 12 (unilateral) + 14 (unilateral) 12 + 14 (Ekman et al., 1983) 4, 14, 25 Not listed
Disgust 10+ 16+ 22+ 25 or 26 9 or 10, 25 or 26 9+15+16 (Ekman et al., 1983) 4, 6, 7, 9, 10, 25, 26 or 27 7+9+19+25+26 Eyes narrowed, nose wrinkled, lips parted, jaw dropped, tongue show
Embarrassment Not listed Not listed 12, 24, 51, 54, 64 (Keltner & Buswell, 1997) 6, 7, 12, 25, 54, participant dampens smile with 23, 24, frown, etc.) 7+12+15+52+54+64 Eyelids narrowed, controlled smile, head turned and down, (not scored with FACS: hand touches face)
Fear 1+2+5+20 1+2+4+5+20, 25 or 26 1+2+4+5+7+20+26 (Ekman et al., 1983) 1, 2, 5, 7, 25, 26 or 27, participant suddenly shifts entire body backwards in chair 1+2+4+5+7+20+25 Eyebrows raised and pulled together, upper eyelid raised, lower eyelid tense, lips parted and stretched
Happiness 6+12 6 + 12 6+12 (Ekman et al., 1983) 6, 7, 12, 16, 25, 26 or 27 6+7+12+25+26 Duchenne smile (6, 7, 12)
Pride Not listed Not listed 6, 12, 24, 53, a straightening of the back and pulling back of the shoulders to expose the chest (Shiota et al., 2003) 7, 12, 53, participant sits up straight 53+64 Head up, eyes down
Sadness 1 + 15 1+15, 4,17 1+4+5 (Ekman et al., 1983) 4, 43, 54 1+4+6+15+17 Brows knitted, eyes slightly tightened, lip corners depressed, lower lip raised
Shame Not listed Not listed 54, 64 (Keltner & Buswell, 1997) 4, 17, 54 54+64 Head down, eyes down
Surprise 1+ 2 + 5+ 25 or 26 1+2+5+25 or 26 1+2+5+26 (Ekman et al., 1983) 1, 2, 5, 25, 26 or 27 1+2+5+25+26 Eyebrows raised, upper eyelid raised, lips parted, jaw dropped

Note. Darwin’s description taken from Matsumoto et al. (2008), Table 13.1. International core patterns (ICPs) refer to expressions of 22 emotion categories that are thought to be conserved across cultures, taken from Cordaro et al. (2017), Tables 4, 5 and 6. A plus sign means “with”; these action units would appear simultaneously. A comma means “sometimes with”; these action units are statistically the most probable to appear, but do not necessarily need to happen simultaneously (David Cordaro, personal communication, 11/11/2018).

By contrast, other scientific frameworks propose that expressions of the same emotion category, such as anger, substantially vary by design, in a way that is tied to the immediate context, which includes the internal context (e.g., the person’s metabolic condition, the past experiences that come to mind, etc.) and the outward context (e.g., whether a person is at work, at school, or at home, who else is present the broader cultural conditions, etc.), both of which vary in dynamic ways over time (see Box 2, SOM). These debates, while useful to scientists, provide little clear guidance for consumers of emotion research who are focused on the practical issue of whether various emotion categories are expressed with facial configurations of sufficient regularity and distinctiveness so that it is possible to read emotion in a person’s face.

The common view of emotional expressions persist, too, because scientists’ actions often don’t follow their claims in a transparent, straightforward way. Many scientists continue to design experiments, use stimuli and publish review papers that, ironically, leave readers with the impression that certain emotion categories each have a single, unique facial expression, even as those same scientists acknowledge that every emotion category can be expressed with a variable set of facial movements. Published studies typically test the hypothesis that there are unique emotion-expression links (for examples, see the reference lists in Elfenbein & Ambady, 2002; Matsumoto, Keltner, Shiota, O’Sullivan & Frank, 2008; Keltner, Sauter, Tracy & Cowen, in press; also see most of the studies reviewed in this paper, e.g., Cordaro, Sun, Keltner, Kamble, Huddar, & McNeil, 2017). The exact facial configuration tested varies slightly from study to study, but a core facial configuration is still assumed (see Table 1 for examples). This pattern of testing the hypothesis that instances of one emotion category are expressed with a single core facial configuration reinforces (perhaps unintentionally) the common view that each emotion category is consistently and uniquely expressed with its own distinctive configuration of facial movements. Review articles (again, perhaps unintentionally) reinforce the impression of unique face-emotion mappings by including tables and figures that display a single, unique facial configuration for each emotion category, referred to as the expression, signal or display for that emotion (Figure 2 presents two recent examples).6 Consumers of this research then assume that a distinctive configuration can be used to diagnose the presence of the corresponding emotion (e.g., that a scowl indicates the presence of anger).

Figure 2. Example figures from recently published papers that reinforce the common belief in diagnostic facial expressions of emotion.

Figure 2.

Figure 2.

A. Adapted from Cordaro et al. (in press), Table 1, with permission. Face photos © Dr. Lenny Kristal. B. Shariff & Tracy, 2011, Figure 2, with permission.

The common view of emotional expressions has also been imported into other scientific disciplines with an interest in understanding emotions, such as neuroscience and artificial intelligence (AI). For example, from a published paper on AI:

“American psychologist Ekman noticed that some facial expressions corresponding to certain emotions are common for all the people independently of their gender, race, education, ethnicity, etc. He proposed the discrete emotional model using six universal emotions: happiness, surprise, anger, disgust, sadness and fear.” (Brodny, Kolakowska, Landowska, Szwoch, Szwoch, & Wróbel, 2016, p. 1, italics in the original)

Similar examples come from our own papers. One paper series of papers focused on the brain structures involved in perceiving emotions from facial configurations (Adolphs, 2002; Adolphs et al., 1994) and the other focused on early life experiences (Pollak et al., 2000; Pollak & Kistler, 2002). These papers were framed in terms of “recognizing facial expressions of emotion” and exclusively presented participants with specific, posed photographs of scowling faces (the presumed facial expression for anger), wide-eyed gasping faces (the presumed facial expression for fear), and so on. Participants were shown faces of different individuals all posing the same facial configuration for each emotion category, ignoring the importance of context. One reason for this flawed approach to investigating the perception of emotion from faces was that then -- at the time these studies were conducted – as now, published experiments, review articles, and stimulus sets were dominated by the common view that certain emotion categories were signaled with an invariant set of facial configurations, referred to as “facial expressions of basic emotions.”

In this paper, we review the scientific evidence that directly tests two beliefs that form the common view of emotional expressions: that certain emotion categories are each routinely expressed by a unique facial configuration and, correspondingly, that people can reliably infer someone else’s emotional state from a set of facial movements. Our discussion is written for consumers of emotion research, whether they be scientists in other fields or non-scientists, who need not have deep knowledge of the various theories, debates, and broad range of findings in the science of emotion, with sufficient pointers to those discussions if they are of interest (see Box 2, SOM).

In discussing what this paper is about – the common view that a person’s inner emotional state is revealed in facial movements -- it bears mentioning what this paper is not about: This paper is not a referendum on “basic emotion” view we briefly mentioned earlier in this section, proposed by the psychologist Paul Ekman and his colleagues, or any other research program or psychologist’s view. Ekman’s theoretical approach has been highly influential in research on emotion for much of the past 50 years. We often cite studies inspired by the basic emotion approach for this reason. In addition, the common view of emotional expressions is also most readily associated with a simplified version of basic emotion approach, as exemplified by the quotes above. Critiques of Ekman’s basic emotion view (and related views) are numerous (e.g., Barrett, 2006a, 2007, 2011; Ortony & Turner, 1990; Russell, 1991, 1994, 1995), as are rejoinders that defend it (e.g., Ekman, 1992, 1994; Izard, 2007). Our paper steps back from this dialogue. We instead take as our focus the existing research on emotional expression and emotion perception and ask whether it is sufficiently strong to justify the way it is increasingly being used by those who consume it.

A Systematic Approach for Evaluating the Scientific Evidence

When you see someone smile and infer that the person is happy, you are making what is known as a reverse inference: you are assuming that the smile reveals something about the person’s emotional state that you cannot access directly (see Figure 3). Reverse inference requires calculating a conditional probability: the probability that a person is in a particular emotion episode (such as happiness) given the observation of a unique set of facial muscle movements (such as a smile). The conditional probability is written as:

p[emotion categorya unique facial configuration])

for example,

p[happinessa smiling facial configuration])

Figure 3. Evaluation criteria: Reliability and specificity in relation to forward and reverse inference.

Figure 3.

Anger and fear are used as the example categories.

Reverse inferences about emotion are ubiquitous in everyday life – whenever you experience someone as emotional, your brain has performed a reverse inference, guessing at the cause of a facial movement when only having access to the movement itself. Every time an app on a phone or computer measures someone’s facial muscle movements, identifies a facial configuration such as a frowning facial configuration, and proclaims that the target person is sad, that app has engaged in reverse inference, such as:

p[sadnessa frowning facial configuration])

Whenever a security agent infers anger from a scowl, the agent has assumed a strong likelihood for

p[angera scowling facial configuration])

Four criteria must be met to justify a reverse inference that a particular facial configuration expresses and therefore reveals a specific emotional state: reliability, specificity, generalizability and validity (explained in Table 2 and Figure 3). These criteria are commonly encountered in the field of psychological measurement and over the last several decades there has been an ongoing dialogue about thresholds for these criteria as they apply in production and perception studies, with some consensus emerging for the first three criteria (see Haidt & Keltner, 1999). Only when a pattern of facial muscle movements strongly satisfies these four criteria can we justify calling it an “emotional expression.” If any of these criteria are not met, then we should instead refer to a facial configuration with more neutral, descriptive terms without making unwarranted inferences, simply calling it a smile (rather than an expression of happiness), a frown (rather than an expression of sadness), a scowl (rather than an expression of anger), and so on.7

Table 2:

Criteria used to evaluate the empirical evidence

Expression Production Emotion Perception
Reliability When a person is sad, the proposed expression (a frowning facial configuration) should be observed more frequently than would be expected by chance. Likewise, for every other emotion category that is subject to a commonsense belief. Reliability is related to a forward inference: given that someone is happy, what is the likelihood of observing a smile, p[set of facial muscle movementsemotion category]. When a person makes a scowling facial configuration, perceivers should consistently infer that the person is angry. Likewise, for every facial configuration that has been proposed as the expression of a specific emotion category. That is, perceivers must consistently make a reverse inference: given that someone is scowling, what is the likelihood that he is angry, p[emotion categoryset of facial muscle movements].
Chance means that facial configurations occur randomly with no predictable relationship to a given emotional state. This would mean that the facial configuration in question carries no information about the presence or absence of an emotion category. For example, in an experiment that observes the facial configurations associated with instances of happiness and anger, chance levels of scowling or smiling would be 50%. Chance means that emotional states occur randomly with no predictable relationship to a given facial configuration. This would mean that the presence or absence of an emotion category cannot be inferred from the presence or absence of the facial configuration. For example, in an experiment that observes how people perceive 51 different facial configurations, chance levels for correctly labeling a scowling face as anger would be 2%.
Reliability also depends on the base rate: how frequently people make a particularly facial configuration overall. For example, if a person frequently makes a scowling facial configuration during an experiment examining the expressions of anger, sadness and fear, he will seem to be consistently scowling in anger when in fact he is scowling indiscriminately. Reliability also depends on the base rate: how frequently people use a particular emotion label or make a particular emotional inference. For example, if a person frequently labels facial configurations as “angry” during an experiment examining scowling, smiling and frowning faces, she will seem to be consistently perceiving anger when in fact she is labeling indiscriminately.
Reliability rates between 70% and 90% provide strong evidence for the commonsense view, between 40% and 69% provide moderate support for the commonsense view, and between 20% and 39% provide weak support (Ekman, 1994; Haidt & Keltner, 1999; Russell, 1994). Reliability rates between 70% and 90% provide strong evidence for the commonsense view, between 40% and 69% provide moderate support for the commonsense view, and between 20% and 39% provide weak support (Ekman, 1994; Haidt & Keltner, 1999; Russell, 1994).
Specificity If a facial configuration is diagnostic of a specific emotion category, then the facial configuration should express instances of one and only one emotion category better than chance; it should not consistently express instances of any other mental event (emotion or otherwise) at better than chance levels. For example, to be considered the expression of anger, a scowling facial configuration must not express sadness, confusion, indigestion, an attempt to socially influence, etc. at better than chance levels. If a frowning facial configuration is perceived as the diagnostic expression of sadness, then a frowning facial configuration should only be labeled as sadness (or sadness should only be inferred from a frowning facial configuration) at above chance levels. And it should not be consistently perceived as expressions of any mental states other than sadness at better than chance levels.
Estimates of specificity, like reliability, depend on base-rates and on how chance levels are defined. Estimates of specificity, like reliability, depend on base-rates and on how chance levels are defined.
Generalizability Patterns of reliability and specificity should replicate across studies, particularly when different populations are sampled, such as infants, congenitally blind individuals and individuals sampled from diverse cultural contexts, including small-scale, remote cultures. High generalizability across different circumstances ensures that scientific findings are generalizable. Patterns of reliability and specificity should replicate across studies, particularly when different populations are sampled, such as infants, congenitally blind individuals and individuals sampled from diverse cultural contexts, including small-scale, remote cultures. High generalizability across different circumstances ensures that scientific findings are generalizable.
Validity Even if a facial configuration is consistently and uniquely observed in relation to a specific emotion category across many studies (strong generalizability), it is necessary to demonstrate that the person in question is really in the expected emotional state. This is the only way that a given facial configuration leads to accurate inferences about a person’s emotional state. A facial configuration is valid as a display or a signal for emotion if and only if it is strongly associated with other measures of emotion, preferably those that are objective and do not rely on anyone’s subjective report (i.e., a facial configuration should be strongly and consistently related to perceiver-independent evidence about the emotional state of the expresser). Even if a facial configuration is consistently and uniquely labeled with a specific emotion word across many studies (strong generalizability), it is necessary to demonstrate that the person making the facial configuration is really in the expected emotional state. This is the only way that a given perception or inference of emotion is accurate. A perceiver can only be said to be recognizing an emotional expression if and only if the person being perceived is verifiably in the expected emotional state.

Note: Reliability is also related to sensitivity, consistency, informational value, and the true positive rate (for further description, see Figure 3). Specificity is related to uniqueness, discreteness, the true negative rate and referential specificity. In principle, we can also ask more parametrically whether there is a link between the intensity of an emotional instance and the intensity of facial muscle contractions, but scientists rarely do.

The Null Hypothesis and the Role of Context

Tests of reliability, specificity, generalizability and validity are almost always compared to what would be expected by sheer chance, if facial configurations (in studies of expression production) and inferences about facial configurations (in studies of emotion perception) occurred randomly with no relation to particular emotional states. In most studies, chance levels constitute the null hypothesis. An example of the null hypothesis for reliability is that people do not scowl when angry more frequently than would be expected by chance.8 If people are observed to scowl more frequently when angry than they would by chance, then the null hypothesis can be rejected based on the reliability of the findings. We can also test the null hypothesis for specificity: If people scowl more frequently than they would by chance not only when angry but also when fearful, sad, confused, hungry, etc., then the null hypothesis for specificity is retained.9

In addition to testing hypotheses about reliability and specificity, tests of generalizability are becoming more common in the research literature, again using the null hypothesis. Questions about generalizability test whether a finding in one experiment is reproduced in other experiments in different contexts, using different experimental methods or sampling people from different populations. There are two crucial questions about generalizability when it comes to the production and perception of emotional expressions: Do the findings from a laboratory experiment generalize to observations in the real world? And, do the findings from studies that sample participants from Westernized, Educated, Industrialized, Rich and Democratic (WEIRD; Henrich, Heine, & Norenzayan, 2010) populations generalize to people who live in small-scale, remote communities?

Questions of validity are almost never addressed in production and perception studies. Even if reliable and specific facial movements are observed across generalizable circumstances, it is a difficult and unresolved question as to whether these facial movements can justify an inference about a person’s emotion state. We have more to say about this later. In this paper, we evaluate the common view by reviewing evidence pertaining to the reliability, specificity, and generalizability of research findings from production and perception studies.

A focus on rejecting the null hypothesis, defined by what would be expected by chance alone, provides necessary but not sufficient support for the common view of emotional expressions. A slightly above chance co-occurrence of a facial configuration and instances of an emotion category, such as scowling in anger – for example, a correlation coefficient around r = .20 to .39 (adapted from Haidt & Keltner, 1999) -- suggests that a person sometimes scowls in anger, but not most or even much of the time. Weak evidence for reliability suggests that other factors not measured in the experiment are likely causing people to scowl during an instance of anger. It also suggests that people may express anger with facial configurations other than a scowl, possibly in reliable and predictable ways. Following common usage, we refer to these unmeasured factors collectively as context. A similar situation can be described for studies of emotion perception: when participants label a scowling facial configuration as “anger” in a weakly reliable way (between .20 and .39 percent of the time; Haidt & Keltner, 1999), then this suggests the possibility of unmeasured context effects.

In principle, context effects make it possible to test the common view by comparing it directly to an alternative hypothesis that a person’s brain will be influenced by other causal factors (as opposed to comparing the findings to random chance). It is possible, for example, that a state of anger is expressed differently depending on various factors that can be studied, including the situational context (such as whether a person is at work, at school, or at home), social factors (such as who else is present in the situation and the relationship between the expresser and the perceiver), the person’s internal physical context (based on how much sleep they had, how hungry they are, etc.), a person’s internal mental context (such as the past experiences that come to mind or the evaluations they make), the temporal context (what just occurred a moment ago), differences between people (such as whether someone is male or female, warm or distant), and the cultural context, such as whether the expression is occurring in a culture that values the rights of individuals (vs. group cohesion), is open and allows for a variety of behaviors in a situation (vs. closed, having more rigid rules of conduct). Other theoretical approaches offer some of these specific alternative hypotheses (see Box 2 in SOM). In practice, however, experiments almost always test the common view against the null hypothesis for reliability and specificity and rarely test specific alternative hypotheses. When context is acknowledged and studied, it is usually examined as a factor that might moderate a common and universal emotional expression, preserving the core assumptions of the common view (e.g., Cordaro et al., 2017; for more discussion, see Box 3, SOM).

A Focus on Six Emotion Categories: Anger, Disgust, Fear, Happiness, Sadness and Surprise

Our critical examination of the research literature in this paper focuses primarily on testing the common view of facial expressions for six emotion categories -- anger, disgust, fear, happiness, sadness and surprise. We do not include a discussion of every emotion category ever studied in the science of emotion. We do not discuss the many emotion categories that exist in non-English speaking cultures, such as gigil, the irresistible urge to pinch or squeeze something cute, or liget, exuberant, collective aggression (for discussion of non-English emotion categories, see Mesquita & Frijda, 1992; Pavlenko, 2014; Russell, 1991). We do not discuss the various emotion categories that have been documented throughout history (e.g., Smith, 2016). Nor do we discuss every English emotion category for which a prototypical facial expression has been suggested. For example, recent studies motivated primarily by the basic emotion approach have suggested that there are “more than six distinct facial expressions …in fact, upwards of 20 multimodal expressions” (Keltner et al., in press, pg. 4), meaning that scientists have proposed a prototypic facial configuration as the facial expression for each of twenty or so emotion categories, including confusion, embarrassment, pride, sympathy, awe, and so on.

The reasons for our focus on six emotion categories are twofold. First, anger, disgust, fear, happiness, sadness and surprise categories anchor common beliefs about emotions and their expressions (as is evident from Box 4, in SOM) and therefore represent the clearest, strongest test of the common view. Second, these six emotion categories have been the primary focus of systematic research for almost a century and therefore provide the largest corpus of scientific evidence that can be evaluated. Unfortunately, the same cannot be said for any of other emotion categories in question. This is a particularly important point when considering the twenty plus emotion categories that are now the focus of research attention. A PsycInfo search for the term “facial expression” combined with “anger, disgust, fear, happiness, sadness, surprise” produced over 700 entries, but a similar search including “love, shame, contempt, hate, interest, distress, guilt” returned less than 70 entries (Duran & Fernandez-Dols, 2018). Almost all cross-cultural studies of emotion perception have focused on just anger, disgust, fear, happiness, sadness and surprise (plus or minus a few) and experiments that measure how people spontaneously move their faces to express instances of emotion categories other than these six remain rare. In particular, there are too few studies that measure spontaneous facial movements during episodes of other emotion categories (i.e., production studies) to conclude anything about reliability and specificity, and there are too few studies of how these additional emotion categories are perceived in small-scale, remote cultures to conclude anything about generalizability. In an era where the generalizability and robustness of psychological findings are under close scrutiny, it seemed prudent to focus on the emotion categories for which there are, by a factor of ten, the largest number of published experiments. Our discussion, which is based on a sample of six emotion categories, generalizes to emotion categories that have been studied, however.10

The proposed expressive facial configurations for each emotion category are presented in Figure 4, and the origin of these facial configurations is discussed in Box 4 in SOM. They originated with Charles Darwin, who stipulated (rather than discovered) that certain facial configurations are expressions of certain emotion categories, inspired by photographs taken by Duchenne and drawings made by the Scottish anatomist Charles Bell (Darwin, 1872). These stipulations largely form the basis of the common view of emotional expressions.

Figure 4. Facial action ensembles for commonsense facial configurations.

Figure 4.

Facial action coding system (FACS) codes that correspond to the commonsense expressive configuration in adults. A is proposed expression for anger and corresponds to prescribed EMFACS code for anger (AUs 4, 5, 7, and 23). B is proposed expression for disgust and corresponds to prescribed EMFACS code for disgust (AU 10). C is proposed expression for fear and corresponds to prescribed EMFACS code for fear (AUs 1, 2, and 5 or 5 and 20). D is proposed expression for happiness and corresponds to prescribed EMFACS code for the so-called Duchenne smile (AUs 6 and 12). E is proposed expression for sadness and corresponds to prescribed EMFACS code for sadness (AUs 1, 4, 11 and 15 or 1, 4, 15 and 17). F is proposed expression for surprise and corresponds to prescribed EMFACS code for surprise (AUs 1, 2, 5, and 26). It was originally proposed that infants express emotions with the same facial configurations as adults. Later research revealed morphological differences between the proposed expressive configurations for adults and infants. Only three out of a possible nineteen proposed configurations for negative emotions from the infant coding scheme were the same as the configurations proposed for adults (Oster et al., 1992). G. adapted from Cordaro et al. (in press), Table 1, with permission. Face photos © Dr. Lenny Kristal. H. adapted from Shariff & Tracy, 2011, Figure 2, with permission.

Producing Facial Expressions of Emotion: A Review of the Scientific Evidence

In this section, we first review the design of a typical experiment where emotions are induced and facial movements are measured. This review highlights several observations to keep in mind as we review the reliability, specificity and generalizability for expressions of anger, disgust, fear, happiness, sadness and surprise in a variety of populations, including adults in both urban and small-scale remote cultures, infants and children, and congenitally blind individuals. Our review is the most comprehensive to date and allows us to comment on whether the scientific findings generalize across different populations of individuals. The value of doing so becomes apparent when we observe how similar conclusions emerge from these research domains.

The Anatomy of a Typical Experiment Designed to Observe People’s Facial Movements During Episodes of Emotion

In the typical expression production experiment, scientists expose participants to objects, images or events that they (the scientists) believe will evoke an instance of emotion. It’s possible, in principle, to evoke a wide variety of instances for a given emotion category (e.g., Wilson-Mendenhall et al., 2015), but in practice, published studies evoke the most typical instances of each category, often elicited with a stimulus that is presented without context (e.g., a photograph, a short movie clip separated from the rest of the film, etc.). Scientists usually include some measure to verify that participants are in the expected emotional state (such as asking participants to describe how they feel by rating their experience against a set of emotion adjectives). They then observe participants’ facial movements during the emotional episode and then quantify how well the measure of emotion predicts the observed facial movements. When done properly, this yields estimates of reliability and specificity, and in principle provides data to assess generalizability. There are limitations to assessing the validity of a facial configuration as an expression of emotion, as we explain below.

Measuring facial movements.

Healthy humans have a common set of 17 facial muscle groups on each side of the face that contract and relax in patterns.11 To create facial movements that are visible to the naked eye, facial muscles contract, changing the distance between facial features (Neth & Martinez, 2009) and shaping skin into folds and wrinkles on an underlying skeletal structure. Even when facial movements look the same to the naked eye, there may be differences in their execution under the skin. There are individual differences in mechanics of making a facial movement, including variation in the anatomical details (e.g., everyone has a slightly different configuration and relative size of the muscles, some people lack certain muscle components, etc.), in the neural control of those muscles (Cattaneo & Pavesi, 2014; Hutto & Vattoth, 2015; Muri, 2015), and in the underlying skeletal structure of the face (discussed in Box 5, in SOM).

There are three common procedures for measuring facial movements in a scientific experiment. The most sensitive, objective measure of facial movements detects the electrical activity from actual muscular contractions, called facial electromyography (again, see Box 5, in SOM). This is a perceiver-independent way of assessing facial movements that detects muscle contractions that are not necessarily visible to the naked eye (Tassinary & Cacioppo 1992). Facial EMG’s utility is unfortunately offset by its impracticality: facial EMG requires placing electrodes on a participant’s face, which can cause skin abrasions. In addition, a person can typically tolerate only a few electrodes on the face at a time. At the writing of this paper, there were relatively few published papers using facial EMG (we identified 123 studies), the overwhelming majority of which sparsely sampled the face, measuring the electrical signals for only a small number of muscles (between one to six); none of the studies measured naturalistic facial movements as they occur outside the lab, in everyday life. As a consequence, we focus our discussion on two other measurement methods: a perceiver-dependent method that describes visible facial movements, called facial actions, which uses human coders who indicate the presence or absence of a facial movement while viewing video recordings of participants, and automated methods for detecting of facial actions from photographs or videos.

Measuring facial movements with human coders.

The Facial Action Coding System, or FACS (Ekman et al., 2002), is a systematic approach to describe what a face looks like when facial movements have occurred. FACS codes describe the presence and intensity of facial movements. Importantly, FACS is purely descriptive and is therefore agnostic about whether those movements might express emotions or any other mental event.12 Human coders train for many weeks to reliably identify specific movements called “action units” or AUs. Each AU is hypothesized to correspond to the contraction of a distinct facial muscle or a distinct grouping of muscles that is visible as a specific facial movement. For example, the raising of the inner corners of the eyebrows (contracting the frontalis muscle pars medialis) corresponds to AU 1. Lowering of the inner corners of the brows (activation of the corrugator supercilii, depressor glabellae and depressor supercilii) corresponds to AU 4. AUs are scored and analyzed as independent elements, but the underlying anatomy of many facial muscles constrains them so they cannot move independently of one another, generating dependencies between AUs (e.g., see Hao, Wang, Peng, & Ji, 2018). Facial action units (AU) and their corresponding list of facial muscles can be found in Table 3. Expert FACS coders approach inter-rater reliabilities of .80 for individual AUs (Jeni, Cohn, & De la Torre, 2013). The first version of FACS (Ekman & Friesen, 1978) was largely based on the work of Swedish anatomist Carl-Herman Hjortsjö who catalogued the facial configurations described by Duchenne (Hjortsjö, 1969). In addition to the updated versions of FACS (Ekman et al., 2002), other facial coding systems have been devised for human infants (Izard et al., 1995; Oster, 2003), chimpanzees (Vick et al., 2007), and macaque monkeys (Parr et al., 2010).13 Figure 4 displays the common FACS codes for the configurations of facial movements that have been proposed as the expression of anger, disgust, fear, happiness, sadness and surprise.

Table 3:

The Facial Action Coding System (FACS; Ekman & Friesen, 1978) codes for adults

AU Description Facial muscles (type of activation)
1 Inner brow raiser Frontalis (pars medialis) graphic file with name nihms-1021596-t0012.jpg
2 Outer brow raiser Frontalis (pars lateralis) graphic file with name nihms-1021596-t0013.jpg
4 Brow lowerer Corrugator supercilii, depressor supercilii graphic file with name nihms-1021596-t0014.jpg
5 Upper lid raiser Levator palpebrae superioris graphic file with name nihms-1021596-t0015.jpg
6 Cheek raiser Orbicularis oculi (pars orbitalis) graphic file with name nihms-1021596-t0016.jpg
7 Lid tightener Orbicularis oculi (pars palpebralis) graphic file with name nihms-1021596-t0017.jpg
9 Nose wrinkle Levator labii superioris alaquae nasi graphic file with name nihms-1021596-t0018.jpg
10 Upper lip raiser Levator labii superioris graphic file with name nihms-1021596-t0019.jpg
11 Nasolabial deepener Zygomaticus minor graphic file with name nihms-1021596-t0020.jpg
12 Lip corner puller Zygomaticus major graphic file with name nihms-1021596-t0021.jpg
13 Cheeks puffer Levator anguli oris graphic file with name nihms-1021596-t0022.jpg
14 Dimpler Buccinator graphic file with name nihms-1021596-t0023.jpg
15 Lip corner depressor Depressor anguli oris graphic file with name nihms-1021596-t0024.jpg
16 Lower lip depressor Depressor labii inferioris graphic file with name nihms-1021596-t0025.jpg
17 Chin raiser Mentalis graphic file with name nihms-1021596-t0026.jpg
18 Lip puckerer Incisivii labii superioris and incisivii labii inferioris graphic file with name nihms-1021596-t0027.jpg
20 Lip stretcher Risorius w/ platysma graphic file with name nihms-1021596-t0028.jpg
22 Lip funneler Orbicularis oris graphic file with name nihms-1021596-t0029.jpg
23 Lip tightener Orbicularis oris graphic file with name nihms-1021596-t0030.jpg
24 Lip pressor Orbicularis oris graphic file with name nihms-1021596-t0031.jpg
25 Lips part Depressor labii inferioris or relaxation of mentalis, or orbicularis oris graphic file with name nihms-1021596-t0032.jpg
26 Jaw drop Masseter, relaxed temporalis and internal terygoid graphic file with name nihms-1021596-t0033.jpg
27 Mouth stretch Pterygoids, digastric graphic file with name nihms-1021596-t0034.jpg
28 Lip suck Orbicularis oris graphic file with name nihms-1021596-t0035.jpg
41 Lid Droop graphic file with name nihms-1021596-t0036.jpg
42 Slit graphic file with name nihms-1021596-t0037.jpg
43 Eyes Closed graphic file with name nihms-1021596-t0038.jpg
44 Squint graphic file with name nihms-1021596-t0039.jpg
45 Blink graphic file with name nihms-1021596-t0040.jpg
46 Wink graphic file with name nihms-1021596-t0041.jpg

Measuring facial movements with automated algorithms.

Human coders require time-consuming, intensive training and practice before they can reliably assign AU codes. After training, it is a slow process to code photographs or videos frame by frame making human FACS coding impractical to use on facial movements as they occur in everyday life. Large inventories of naturalistic photographs and videos, which have been curated only fairly recently (Benitez-Quiroz et al., 2016), would require decades to manually code. This problem is addressed by automated FACS coding systems using computer vision algorithms (Martinez & Du, 2012; Martinez, 2017; Valstar et al., 2017).14 Recently developed computer vision systems have automated the coding of some (but not all) facial AUs (e.g., Benitez-Quiroz et al., in press; Benitez-Quiroz et al., 2017b; Chu et al., 2017; Corneanu et al., 2016; Essa & Pentland, 1997; Martinez, 2017a; Martinez & Du, 2012; Valstar et al., 2017; see Box 6, SOM) making it more feasible to observe facial movements as they occur in everyday life, at least in principle (see Box 7, SOM). Automated FACS coding is accurate (>90%) when compared to the AU codes from expert human coders, provided that the images were captured under ideal laboratory conditions, where faces are viewed from the front, are well illuminated, are not occluded, and are posed in a controlled way (Benitez-Quiroz et al., 2016). Under ideal conditions, accuracy is highest (~99%) when algorithms are tested and trained on images from the same database (Benitez-Quiroz et al., 2016). The best of these algorithms works quite well when trained and tested on images from different databases (~90%), as long as the images are all taken in ideal conditions (Benitez-Quiroz et al., 2016). Accuracy (compared to human FACS coding) decreases substantially more when coding facial actions in still images or in video frames taken in everyday life where conditions are unconstrained and facial configurations are not stereotypical (e.g.,Yitzhak et al., 2017).15 For example, 38 automated FACS coding algorithms were recently trained on one million images (the 2017 EmotioNet Challenge; Benitez-Quiroz et al., 2017a) and evaluated against separate test images which were FACS coded by experts.16 In these less constrained conditions, accuracy dropped below 83% and a combined measure of precision and recall (a measure called F1, ranging from zero to one) was below .65 (Benitez-Quiroz et al., 2017a).17 These results indicate that current algorithms are not accurate enough in their detection of facial AUs to fully substitute for expert coders when describing facial movements in everyday life. Nonetheless, these algorithms offer a distinct practical advantage because they can be used in conjunction with human coders to speed up the study of facial configurations in millions of images in the wild. It is likely that automated methods will continue to improve as better and more robust algorithms are developed and as more diverse face images become available.

Measuring an emotional state.

Once an approach has been chosen for measuring facial movements, a clear test of the common view of emotional expressions depends on having valid measures that reliably and specifically characterize the instances of each emotion category in a generalizable way, to which the measurements of facial muscle movements can be compared. The methods that scientists use to assess people’s emotional states vary in their dependence on human inference, however, which raises questions about the validity of the measures.

Relatively objective measures of an emotional instance.

The more objective end of the measurement spectrum includes dynamic changes in the autonomic nervous system (ANS), such as cardiovascular, respiratory or perspiration changes (measured as variations in skin conductance), and dynamic changes in the central nervous system, such as changes in blood flow or electrical activity in the brain. These measures are thought to be more objective because the measurements themselves (the numbers) do not require a human judgment (i.e., the measurements are perceiver-independent). Only the interpretation of the measurements (their psychological meaning) requires human inference. For example, a human observer does not judge whether skin conductance or neural activity increases or decreases; human judgment only comes into play when the measurements are interpreted for the emotional meaning.

Currently, there are no objective measures, either singly or as a pattern, that reliability and uniquely identify one emotion category from another in a replicable way. Statistical summaries of hundreds of experiments, called meta-analyses, show for example, that currently there is no relationship between an emotion category, such as anger, and a single, specific set of physical changes in ANS that accompany the instances of that category, even probabilistically (the most comprehensive study published to date is Siegel et al., 2018, but for earlier studies see Cacioppo et al., 2000; Stemmler, 2004; also see Box 8, SOM). In anger, for example, blood pressure can go up, go down, or stay the same (i.e., changes in blood pressure are not consistently associated with anger). And a rise in blood pressure is not unique to instances of anger; it also can occur during a range of other emotional episodes (i.e., changes in blood pressure do not specifically occur in anger and only in anger). 18Individual studies often find patterns of ANS measures that distinguish an instance of one emotion category from another, but those patterns don’t replicate and instead vary across studies, even when studies use the same methods and stimuli, and sample from the same population of participants (e.g., compare findings from Kragel & LaBar, 2013 with Stephens, Christie, & Friedman, 2010). Similar within-category variation is routinely observed for changes in neural activity measured with brain imaging (Lindquist et al., 2012) and single neuron recordings (Guillory & Bujarski, 2014). For example, pattern classification studies discover multivariate patterns of activity across the brain for emotion categories such as anger, sadness, fear, and so on, but these patterns do not replicate from study to study (e.g., Kragel & LaBar, 2015; Saarimäki et al., 2016; Wager et al., 2015; for a discussion, see Clark-Polner et al., 2017). This observed variation does not imply that biological variability during emotional episodes is random, but rather that it may be context-dependent (e.g., yellow and green zones of Figure 1). It may also be the case that current biological measures are simply insufficiently sensitive or comprehensive enough to capture situated variation in a precise way. If this is so, then such variation should be considered unexplained, rather than random.

There is a difficult circularity built into these studies that is worth pointing out, and that we encounter again a few paragraphs down: Scientists must use some criterion for identifying when instances of an emotion category are present in the first place (so as to draw conclusions about whether or not emotion categories can be distinguished by different patterns of physical measurements).19 In most studies that attempt to find bodily or neural “signatures” of emotions, the criterion is a subjective one, either reported by the participants or provided by the scientist, which introduces problems of its own, as we discuss in the next section.

Subjective measures of an emotional instance.

Without objective measures to identify the emotional state of a participant, scientists typically rely on the relatively more subjective measures that anchor the other end of the measurement spectrum. The subjective judgments can come from the participants (who complete self-report measures), from other observers (who infer emotion in the participants), or from the scientists themselves (who use a variety of criteria, including commonsense, to infer the presence of an emotional episode). These are all examples of perceiver-dependent measurements because the measurements themselves, as well as their interpretation, directly rely on human inference.

Scientists often rely on their own judgments and intuitions to stipulate when an emotion is present or absent in participants (as Charles Darwin did). For example, snakes and spiders are said to evoke fear. So are situations that involve escaping from a predator. Sometimes scientists stipulate that certain actions indicate the presence of fear, such as freezing or fleeing or even attacking in defense. The conclusions that scientists draw about emotions depends on the validity of their initial assumptions. It is noteworthy that when it comes to emotions, scientists use exactly the same categories as non-scientists, which may give us cause for concern, as forewarned by William James (James, 1890, 1894)20

Inferences about emotional episodes can also come from other people, for example independent samples of study participants, who categorize the situations in which facial movements are observed. Scientists can ask observers to infer when participants are emotional by having them judge subjects’ behavior or tone of voice; for example, see our discussion of Camras et al. (2007) discussed in the section on infants and children, below.

A third common strategy to identify the emotional state of participants is to simply ask them what they are experiencing. Their self-reports of emotional experience then become the criteria for deciding whether an emotional episode is present or absent. Self-reports are often considered imperfect measures of emotion because they depend on subjective judgements and beliefs and require translation into words. In addition, a person can be experiencing an emotional event yet be unaware of it and therefore unable to report on it (i.e., a person can be conscious but unaware of their experience and unable to report it), or may be unable to express how they feel using emotion words, a condition known as alexithymia. Despite questions about their validity, self-reports are the most common measure of emotion that scientists compare to facial AUs.

Human inference and assessing the presence of an emotional state.

At this point, it should be obvious that any measure of an emotional state, to which measurements of facial muscle movements can be compared, itself requires some degree of human inference; what varies is the amount of inference that is required. Herein lies a problem: To properly test the hypothesis that certain facial movements reliably and specifically express emotion, scientists (ironically) must first make a reverse inference that an emotional event is occurring – that is, they infer the emotional instance by observing changes in the body, brain, and behavior (e.g., only if blood pressure consistently and uniquely rises in anger can a rise in blood pressure be used as a marker of anger). Or they infer (a reverse inference) that an event or object evokes an instance of a specific emotion category (e.g., an electric shock elicits fear but not irritation, curiosity, or uncertainty). These reverse inferences are scientifically sound only if measures of emotion reliably, specifically and validly characterize the instances of the emotion category. So, any clear, scientific test of the common view of emotional expressions rests on a set of more basic inferences about whether an emotional episode is present or absent, and any conclusions that come from such a test are only as sound as those basic inferences.

If all measures of emotion (to which measurements of facial muscle movements are compared) rest on human judgment to some degree, then, in principle, this prevents a scientist from being sure that an emotional state is present, which in turn limits the validity of any experiment designed to test whether a facial configuration validly expresses a specific emotion category. All face-emotion associations that are observed in an experiment reflect human consensus, i.e., the degree of agreement between self-judgments (of the participants), expert-judgments (of the scientist), and/or judgments of other observers (of perceivers who are asked to infer emotion in the participants). These types of agreement are often incorrectly referred to as accuracy. We touch on this point again when we discuss studies that test whether certain facial configurations are routinely perceived as expressions of anger, disgust, fear, and so on.

Testing the common view of emotional expressions: Interpreting the scientific observations.

If a specific facial configuration reliably expresses instances of a certain emotion category in any given experiment, then we would expect measurements of the face (e.g., facial AU codes) to co-occur with measurements that indicate that participants are in the target emotional state. In principle, those measures might be more objective, such as ANS changes during an emotional event, or they might be more subjective, deriving from the scientist, from other perceivers who make judgments about the study participants, or from the participants themselves. In practice, however, most experiments compare facial movements to subjective measures of emotion -- a scientist’s judgment about which emotions are evoked by a particular stimulus, perceivers judgments about participants’ emotional states, or participants’ self-reports of emotional experience -- because ANS and other more objective measurements do not themselves distinguish one emotion category from another in a reliable and specific way. For example, in an experiment, scientists might ask: Do the AUs that create a scowling facial configuration co-occur with self-reports of feeling angry? Do the AUs that create a pouting facial configuration co-occur with perceiver’s judgments that participants are sad? Do the AUs that create a wide-eyed gasping facial configuration co-occur when people are exposed to an electric shock? And so on. If such observations suggest that a configuration of muscle movements is reliably observed during episodes of a given emotion category, then those movements are said to express the emotion in question. As we will see, many studies show that some facial configurations occur more often than random chance, but are not observed with a high degree of reliability (according to the criteria from Haidt & Keltner (1999), outlined in Table 2 and Figure 3).

If a specific facial configuration specifically (i.e., uniquely) expresses instances of a certain emotion category in any given experiment, then we would expect to observe little co-occurrence between measurements of the face and measurements indicating the presence of emotional instances from other categories, except what would be expected by chance (again, see Table 2 and Figure 3). For example, in an experiment, scientists might ask: do the AUs that create a scowling facial configuration co-occur with self-reports of feeling sad, confused, or social motives such as dominance? Do the AUs that create a pouting facial configuration co-occur with perceiver’s judgments that participants are angry or afraid? Do the AUs that create a wide-eyed gasping facial configuration co-occur when people are exposed to a competitor whom they are trying to scare? And so on.

If a configuration of facial movements is observed in instances of a certain emotion category in a reliable, specific way within an experiment, so that we can infer that the movements are expressing an instance of the emotion in that study as hypothesized, then scientists can safely infer that the facial movements in question are an expression of that emotion category’s instances in that situation. One more step is required before we can infer that the facial configuration is the expression of that emotion: we must observe a similar pattern of facial configuration-emotion co-occurrences across different experiments, to some extent generalizing across the specific measures and methods used and the participants and contexts sampled. If the facial configuration-emotion co-occurrences replicate across experiments that sample people from the same culture, then the facial configuration in question can be reasonably be referred to as an emotional expression only in that culture; e.g., if a scowling facial configuration co-occurs with measures of anger (and only anger) across most studies conducted on adult participants in the US who are free from illness, then it is reasonable to refer to a scowl as an expression of anger in the US. If facial configuration-emotion co-occurrences generalize across cultures – that is, replicate across experiments that sample a variety of instances of that emotion category in people from different cultures -- then the facial configuration in question can be said to universally express the emotion category in question.

Studies of Healthy Adults from the U.S. and Other Developed Nations

We now review the scientific evidence from studies that document how people spontaneously move their facial muscles during instances of anger, disgust, fear, happiness, sadness and surprise, and how they pose their faces when asked to indicate how they express each emotion category. We examine evidence gathered in the lab and in naturalistic settings, sampling healthy adults who live in a variety of cultural contexts. To evaluate the reliability, specificity and generalizability of the scientific findings, we adapted criteria set out by Haidt & Keltner (1999), as discussed in Table 2.

Spontaneous facial movements in laboratory studies.

A meta-analysis was recently conducted to test the hypothesis that the facial configuration in Figure 4 co-occur, as hypothesized, with specific emotion categories (Duran et al., 2017). This analysis was published in a book chapter. Thirty-seven published articles reported on how people moved their faces when exposed to objects or events that evoke emotion. Most studies included in the meta-analysis were conducted in the laboratory. The findings from these experiments were statistically summarized to assess the reliability of facial movements as expressions of emotion (see Figure 5). In all emotion categories tested, other than fear, participants moved their facial muscles into the expected configuration more consistently than what we would expect by chance. Consistency levels were weak, however, indicating that the proposed facial configurations in Figure 4 have limited reliability (and to some extent, limited generalizability; i.e., a scowling facial configuration is an expression of anger, but not the expression of anger. More often than not, people moved their faces in ways that were not consistent with the hypotheses of the common view. An expanded version of this meta-analysis (Duran & Fernandez-Dols, 2018) analyzed 89 effect sizes from 47 studies totaling 3599 participants, with similar results: the hypothesized facial configurations were observed, with average effect sizes of r = .32 (for the average correlation between the intensity of a facial configuration and a measure of emotion, with correlations for specific emotion categories ranging from .25 to .38, corresponding to weak evidence of reliability) and proportion = .19 (for the average proportion of the times that a facial configuration was observed during an emotional event, with proportions for specific emotion categories ranging from .15 to .25, interpreted as no evidence to weak evidence of reliability).21

Figure 5. Meta-analysis of facial movements during emotional episodes: A summary of effect sizes across studies (Duran et al., 2017).

Figure 5.

Effect sizes are computed as correlations or proportions (as reported in the original experiments). Results include experiments that reported a correspondence between a facial configuration and its hypothesized emotion category and those that reported a correspondence between individual AUs of that facial configuration and the relevant emotion category; meta-analytic summaries for entire ensembles of AUs only (the facial configurations specified in Figure 2) were even lower than those that appear here.

No overall assessment of specificity was reported in either the original or the expanded meta-analysis because most published studies do not report the false positive rate (i.e., the frequency with which a facial AU is observed when an instance of the hypothesized emotion category was not present; see Figure 3). Nonetheless, some striking examples of specificity failures have been documented in the scientific literature. For example, a certain smile, called a “Duchenne” smile, is defined in terms of facial muscle contractions (i.e., in terms of facial morphology): it involves movement of the orbiculari oculis which raises the cheeks and causes wrinkles at the outer corners of the eyes in addition to movement of the zygomatic major which raises the corners of the lips into a smile. A Duchenne smile is thought to be a spontaneous expression of authentic happiness. Research shows that a Duchenne smile can be intentionally produced when people are not happy, however (Gunnery & Hall, 2014; Gunnery et al., 2013; also see Krumhuber & Manstead, 2009), consistent with evidence that Duchenne smiles often occur when people are signaling submission or affiliation rather than solely reflecting happiness (Rychlowska et al., 2017).

Spontaneous facial movements in naturalistic settings.

Studies of facial configuration-emotion category associations in naturalistic settings tend to yield similar results to studies that were conducted in more controlled laboratory settings (Fernandez-Dols, 2017; Fernandez-Dols & Crivelli, 2013). Some studies observe that people express emotions in real world settings by spontaneously making the facial muscle movements proposed in Figure 4, but such observations do not replicate well across studies (e.g., compare Matsumoto & Willingham, 2006 vs. Crivelli, Carrera and Fernandez-Dols, 2015; Rosenberg & Ekman, 1994 vs. Fernandez-Dols, Sanchez, Carrera, & Ruiz-Belda, 1997). For example, two field studies of winning judo fighters recently demonstrated that so-called “Duchenne” smiles were better predicted by whether an athlete was interacting with an audience than the degree of happiness reported after winning their matches (Crivelli, Carrera, & Fernandez-Dols, 2015). Only eight of the 55 winning fighters produced a “Duchenne” smile in Study 1; all occurred during a social interaction. Only 25 out of 119 winning fighters produced a “Duchenne” smile in Study 2, documenting, at best, weak evidence for reliability.

Posed facial movements.

Another source of evidence comes from asking participants sampled from various cultures to deliberately pose the facial configurations that they believe they use to express emotions. In these studies, participants are given a single emotion word or a single, brief statement to describe each emotion category and then asked to freely pose the expression that they believe they make. In this way, they directly examine common beliefs about emotional expressions. For example, one study provided college students from Canada and Gabon (in Central Africa) with dictionary definitions for ten emotion categories. After practicing in front of a mirror, participants posed the facial configurations so that “their friends would be able to understand easily what they feel” and their poses were FACS coded (Elfenbein et al., 2007, p. 134). Similarly, a recent study asked college students in China, India, Japan, Korea, and the US, to pose the facial movements they believe they make when expressing each of 22 emotion categories (Cordaro, Sun, Keltner, Kamble, Huddar & McNeil, 2017). Participants heard a brief scenario describing an event that might cause anger (“You have been insulted, and you are very angry about it”) and then were instructed to pose a facial (and non-verbal, vocal) expression of emotion, as if the events in the scenario were happening to them. Experimenters were present in the testing room as participants posed their responses. Both studies found moderate to strong evidence for a cross-cultural, common expressive pose for anger, fear, and surprise categories, and weak to moderate evidence for the happiness category, with cultural variation around those common poses; the findings were weaker for disgust and sadness categories (Figure 6).

Figure 6: Comparing posed and spontaneous facial movements.

Figure 6:

Results from Table 6, Cordaro et al. (2017), degree of overlap between the hypothesized configuration of facial movements for each emotion category and the “International Core Patterns” derived from participants’ expressive poses; Gabonese participants in Elfenbein et al. (2007), reliability for the anger category is for AU4 + AU5 only; proportion data only from Duran et al., (2017).

Neither study compared participants’ posed expressions to observations of how they actually moved their faces when expressing emotion. Nonetheless, a quick comparison of the findings from both studies and the proportions of spontaneous facial movements made during emotional events (from the Duran et al. (2017) meta-analysis) makes it clear that posed and spontaneous movements differ, sometimes quite substantially (again, see Figure 6). When people pose a facial configuration that they believe expresses an emotion category, they make facial movements that more reliably agree with the hypothesized facial configurations in Figure 6. The same cannot be said of people’s spontaneous facial movements during actual emotional episodes, however (for convergent evidence, see Motley & Camden, 1988; Namba et al., 2016). One possible interpretation of these findings is that posed and spontaneous facial muscle configurations correspond to distinct communication systems. Indeed, there is some evidence that volitional and involuntary facial movements are controlled by different circuits in the skeletomotor system (Rinn, 1984). Another factor that may contribute to the discrepancy between posed and spontaneous facial movements is that people’s beliefs about their own behavior often reflect their stereotypes or beliefs and do not necessarily correspond to how they actually behave in real life (see Robinson & Clore, 2002).

Summary.

Our review of the available evidence thus far is summarized in the first through third data rows in Table 4. The hypothesized facial configurations presented in Figure 4 spontaneously occur with weak reliability during instances of the predicted emotion category, suggesting that they sometimes serve to express the predicted emotion. Furthermore, the specificity of each facial configuration as an expression of a specific emotion category is largely unknown (because it is typically not reported in many studies). In our view, this pattern of findings is most compatible with the interpretation that hypothesized facial configurations are not made reliably or specifically enough to use them to infer a person’s emotional state. We are not suggesting that facial movements are meaningless and devoid of information. Instead, the data suggest that the meaning of any set of facial movements may be much more variable and context-dependent than hypothesized by the common view.

Table 4:

Reliability and specificity: A summary of the evidence

Reliability Specificity
Expression Production
 Adults, Developed, Spontaneous, Lab weak unknown
 Adults, Developed, Spontaneous, Naturalistic weak unknown
 Adults, Developed, Posed weak to strong unknown
 Adults, Remote, Spontaneous unclear unknown
 Adults, Remote, Posed weak to strong unknown
 Newborns, Infants, Toddlers unsupported unsupported
 Congenitally Blind unsupported to weak unsupported
Emotion Perception
 Adults, Developed, Choice-From-Array moderate to strong unknown
 Adults, Developed, Reverse
Correlation (with Choice-From-Array)
moderate moderate
 Adults, Developed, Free-Labeling weak to moderate weak
 Adults, Developed, Virtual Humans unknown unknown
 Adults, Remote, Choice-From-Array (before 2008) moderate to strong unknown
 Adults, Remote, Choice-From-Array (after 2008) weak to moderate unsupported
 Adults, Remote, Free-Labeling (before 2008) unsupported to strong variable
 Adults, Remote, Free-Labeling (after 2008) unsupported unsupported
 Infants, Young Children unsupported unsupported

Note. Criteria were adopted from Haidt & Keltner (1999), who suggest that reliability rates of 70±90% are considered strong evidence for universal emotion perception (following Ekman, 1994a); presumably, this would also hold for studies of expression production. Weak evidence is in the range of 20±40% (following Russell, 1994). By interpolation, reliability between 41% and 69% would be considered moderate evidence for reliability. Reliability estimates below 20% are interpreted as findings that clearly do not support the reliability hypothesis. We also adopted these criteria for specificity findings. Developed = studies of participants from the U.S. and other more urban countries. Spontaneous = spontaneous facial movements. Posed = posed facial configurations. Remote = studies of participants from small-scale, remote samples.

Studies of Healthy Adults Living in Small-Scale, Remote Cultures

The emotion categories that are at the heart of common view– anger, disgust, fear, happiness, sadness and surprise -- were derived from modern US English (Wierzbicka, 2014) and their proposed expressions (in Figure 4) derive from observations of people who live in urbanized, Western settings. Nonetheless, it is hypothesized that these are facial configurations evolved as emotion-specific expressions to signal socially-relevant emotional information (Shariff & Tracy, 2011) in the challenging situations that originated in our hunting and gathering hominin ancestors who lived on the African savannah during the Pleistocene era (Pinker, 1997; Tooby & Cosmides, 1990). It is further hypothesized that these facial configurations should therefore be observed during instances of the predicted emotion categories with strong reliability and specificity in people around the world, although the facial movements might be slightly modified by culture (Cordaro et al., 2017; Ekman, 1972). The strongest test of these hypotheses would be to sample participants who live in remote parts of the world with relatively little exposure to western cultural norms, practices and values (Norenzayan & Heine, 2005; Henrich et al., 2010) and observe their facial movements during emotional episodes.22 In our evaluation of the evidence, we continued to use the criteria summarized by Haidt & Keltner (1999; see Table 2).

Spontaneous facial movements in naturalistic settings.

Our review of scientific studies that systematically measure the spontaneous facial movements in people of small-scale, remote cultures is brief by necessity: there aren’t any. At the time of publication, we were unable to identify even a single published report or manuscript registered on open-access, pre-print services that measured facial muscle movements in people of remote cultures as they experienced emotional events. Scientists have almost exclusively observed how people label facial configurations as emotional expressions (i.e., they study emotion perception, not production) to test the hypothesis that certain facial configurations evolved to express certain emotion categories in a reliable, specific and generalizable (i.e., universal) manner. Later in the paper we return to this issue and discuss the findings from these emotion perception studies.

There are nonetheless several descriptive reports that provide support for the common view of universal emotional expressions (similar to what Valente et al., 2017 refer to as an “observational approach”). For example, the US psychologist Paul Ekman and colleagues curated an archive of photographs of the Fore hunter-gatherers taken during his visits to Papua New Guinea in the 1960s (Ekman,1980). The photographs were taken as people went about their daily activities in the small hamlets of the eastern highlands of Papua New Guinea. Ekman used his knowledge of the situation in which each photograph was taken to assign each facial configuration to an emotion category, leading him to conclude that the Fore expressed emotions with the proposed facial configurations shown in Figure 4. Yet different scientific methods yielded a contrasting conclusion. When Trobriand Islanders living in Papua New Guinea were asked to infer emotions in facial configurations by labeling these photographs in their native language, both by freely offering words and by choosing the best fitting emotion word from a list of nine choices, they did not label the facial configurations as proposed by Ekman and colleagues at above chance levels (Crivelli et al., 2017).23 In fact, the proposed fear expression -- the wide-eyed gasping face -- is actually interpreted as an expression of threat (intent to harm) and anger by the Maori of New Zealand and in the Trobriand Islanders in remote Papua New Guinea (Crivelli & Fridlund, 2016).

A compendium of spontaneous human behavior published by the Austrian ethologist Irenäus Eibl-Eibesfeldt (Eibl-Eibesfeldt, 1989) is sometimes cited as evidence for the hypothesis that certain facial movements are universal signals for specific emotion categories. No systematic coding procedure was used in his investigations, however. Upon close examination, Eibl-Eibesfeldt’s detailed descriptions appear to be more consistent with the studies of people from more industrialized cultures that we reviewed above: people move their faces in a variety of ways during episodes belonging to the same emotion category. For example, as reported by Eibl-Eibesfeldt, a rapid eyebrow raise (called an eyebrow flash) is thought to express friendly recognition in some, but not all, cultures. This movement would be coded with FACS AU 1 (inner brow raise) and AU 2 (outer brow raise) that are part of the proposed expressions for surprise and fear (Ekman et al., 1983), sympathy (Haidt & Keltner, 1999) and awe (Shiota et al., 2003). Even Eibl-Eibesfeldt acknowledged that eyebrow flashes were not unique expressions of specific emotion categories, writing that they also served as a greeting, to invite social contact, as a sign of thanks, an initiation of flirting, and a general indication of “yes” in Samoans and other Polynesians, in the Eipo and Trobriand islanders in Papua New Guinea, and in the Yanomami of South America. In Japan, eyebrow flashes are considered an impolite way for adults to greet one another. In the US and Europe, an eyebrow flash was observed when friends greet one another, but not strangers.

Posed facial movements.

One study read a brief emotion story to people who live in the remote Fore culture of Papua New Guinea and asked each person to “show how his face would appear” if he was the person described in the emotion stories (Ekman, 1972, p. 273; sample size was not reported). Videotapes of nine participants were shown to 34 US college students who were asked to judge which emotion was being expressed. US participants were asked to infer the emotional meaning of the facial poses by choosing an emotion word from six choices provided by the experimenter (called a choice-from-array task, see Table 5). Participants inferred the intended emotional meaning above chance guessing for smiling (happiness, 73%), frowning (sadness, 68%), scowling (anger, 51%), and nose-wrinkling (disgust, 46%), but not for surprise and fear (27% and 18% respectively).

Table 5:

Common tasks for measuring explicit emotion perception

Concerns Additional Observations
General Considerations
Test-retest reliability is rarely evaluated but is critical. A number of contextual factors are known to influence judgments, including a perceiver’s internal state. Test-retest assessments are rarely done for practical reasons.
Participants are typically asked to infer emotional meaning in exaggerated facial configurations. This reduces the ecological validity of the findings for how people infer emotional meaning in faces in the real world. The facial configurations used in most experiments (see Figure 4) are caricatures – they are exaggerated to maximally distinguish one from the another. Caricatures are easier to label (categorize) than are typical stimuli, particularly when the categories in question are highly interrelated (Goldstone, Steyvers, & Rogosky, 2003). Exaggerated facial configurations have greater “source clarity” (Ekman, Friesen & Ellsworth, 1972)
Participants are typically asked to infer emotional meaning in highly selected facial configurations. In early studies, a smaller set of exaggerated facial configurations were culled from much larger sets of posed faces (involving several thousand faces; for a discussion, see Gendron & Barrett, 2017; Russell, 1994).
Participants are typically asked to infer emotional meaning in static, non-moving facial configurations (i.e., in photographs rather than movies). This reduces the ecological validity of the findings for how people infer emotional meaning in faces in the real world. In the real world, people have to infer when a set of movements begin and end; this is called discrimination or detection). There is information in the dynamics of facial movements (Jack & Schyns, 2017; Krumhuber, Kappas & Manstead, 2013), but dynamic facial movements, particularly when they are spontaneous, do not always produce higher levels of agreement in emotion perception studies. Dynamic movements add realism, intensity and improve levels of agreement primarily when movements are degraded or are artificial
Participants are typically asked to infer emotional meaning in posed, rather than spontaneous, facial configurations. Spontaneous or candid facial configurations typically produce much lower levels of agreement in emotion perception studies (e.g., Kayyal & Russell, 2013; Naab & Russell, 2007).
Only a single task used in most experiments (i.e., participants are asked to infer emotion in facial configurations via one method of responding). Ideally, multiple tasks should be used with the same population of participants to see if convergent results are obtained. This approach is rarely taken, but for an example, see Crivelli et al., 2016; Gendron et al., 2014; Gendron et al., 2018).
Most experiments ask participants to infer emotion in a disembodied face, alone, without context. This reduces the ecological validity of the findings for how people infer emotional meaning in faces in the real world. A growing number of experiments now show that context is an important, and sometimes dominant, source of information when people infer emotional meaning in a facial configuration. See Box 3 in SOM. For example, Situational information tends to dominate perception of emotion in faces both when situations are common, everyday (Carrera-Levillain & Fernández-Dols, 1994) and even when situations are more ambiguous than the exaggerated facial configurations being judged (Carroll & Russell, 1996, Study 3).
Many studies do not report evidence about the specificity of emotion perceptions, or the frequency with which people infer the non-intended emotional meaning to a facial configuration.
Until recently, the large majority of experiments included only one pleasant emotion category (happiness) among several unpleasant emotion categories (anger, fear, sadness, etc.). This may be one reason that agreement rates are so high for smiles. In the last few years, experiments are now including a larger variety of pleasant emotion categories (pride, awe, gratitude, etc.), but there continues to be debate over whether or not they expect these emotion categories are expressed with consistent, specific facial configurations.
Choice-From Array: matching photos of facial configurations and emotion words (with or without brief stories)
Response options are limited to those provided in the task
Words influence how the brain processes visual inputs from faces (e.g., Gendron et al., 2012; Doyle & Lindquist, 2018). Stories can prime action perceptions, as well (Gendron et al., in press). More generally, choice-from-array tasks have been shown to encourage biased perceptual responding using a signal detection analysis (e.g., DeCarlo 2012). Choice-from-array tasks are easy and efficient.
The fact that participants are exposed to the same facial configurations and emotion words over and over allows them to learn the intended pairings even if they don’t know them to begin with (Nelson & Russell, 2016).
An emotion word does not necessarily have a unique correspondence to a single emotion category for all people in a given culture (i.e., they may differ in emotional granularity; Barrett, 2004, 2017; Lindquist & Barrett, 2008) or people from different cultures. Concerns about individual word meaning is why choice-from-array using stories is preferable. Also, choice-from-array tasks are usually straightforward for participants to understand.
A small range of answers are pre-determined by the experimenter, making it easier for participants to provide the answers scientists expect. For example, by constraining which words participants were allowed to choose from, frowns were consensually labeled as fear, wide-eyed gasping faces were labeled as surprise (Russell, 1993). Scowling faces are more likely to be perceived as fearful when paired with the description of danger (Carroll & Russell, 1996, Study 1) and appear determined or puzzled depending on the story they are presented with (Carroll & Russell, 1996, Study 2). Choice-from-array responses are easy for scientists to score. Most studies using continuous judgments (rather than forced choice) find that participants do not infer emotional meaning in facial configurations in a yes/no or on/off sort of way (Russell, 1994).
People are asked to make yes/no decisions about assigning a facial configuration to an emotion category. Multiple emotion words may apply to a single configuration (i.e., people might infer more than one emotional meaning in a face), but the option to infer multiple emotional meanings rarely given to participants. Continuous judgments, such as on a Likert-type scale ranging from one to seven, would solve both of these problems, and also allow analysis of the similarity among facial configurations (which evidence shows is important, e.g., Jack et al., 2016; Kayal & Russell, 2013). Similarity allows scientists to discover the emotional meanings that people implicitly assign to a facial configuration, rather than having people explicitly state them (see further discussion of similarity below).
A participant might decide that no emotion word provided applies to a facial configuration, but the option to respond this way is rarely given to participants (they are usually forced to choose an emotion word; for discussion, see Frank & Stennett, 2001). See Cordaro et al. (2016) for an example of this design feature.
If a participant hears a story and is choose between two faces (e.g., a scowl and smile), she can give the expected answer (e.g., scowl) simply by figuring out that smile is NOT correct. For example, after hearing a story about anger, a participant is shown a scowl and a smile and can choose the scowl merely by realizing the smile is not correct (on the basis of valence). This is similar to getting an answer right on a multiple-choice test by eliminating all the alternatives—you don’t actually know the right answer, but you figured it out because of the structure of the task. A similar point can be made about showing a single face and asking participants to label it with a word by selecting from among a small set of options. Participants use a process of elimination strategy: words that are not chosen on prior trials are selected more frequently, inflating agreement levels (DiGirolamo & Russell, 2017). If a participant hears a story about anger and must choose between a scowl and a smile, she can figure out that the scowl is correct merely because she is distinguishing between negative (scowl) and positive (smile). If a participant hears a story about anger and must choose between a scowl and a frown, he can figure out that the scowl is correct merely because he is distinguishing between high arousal (scowl) and low arousal (frown).
In tasks that involve brief stories or vignettes about emotion, only one typical story is offered for each emotion category, making it more difficult to observe any variation within a category.
Free Sorting: photos of facial configurations are sorted into groupings, such that each grouping represents a category
Face-to-Cue Matching: matching photos of facial configurations to a recording of posed vocalization
Most participants still spontaneously use words to guide their sorting and organize their groupings. Ideal for preverbal participants or those with semantic deficits (e.g., Lindquist et al., 2014).
Similarity Judgments Between Pairs of facial configurations
Perceptual Matching: Indicating whether or not two photos of facial configurations belong to the same emotion category
It is inefficient and time consuming to judge the similarity of all pairs of facial configurations. For a set of 100 faces, this requires (100*100)/2 = 5,000 different similarity judgments. Participants can arrange face stimuli on a computer screen and all pairwise similarity judgments can be computed (the SPAM method proposed by Goldstone, 1994; e.g., see Hout et al.,2013). This procedure also solves the problem that the same pair of stimuli will have a different judged similarity depending on which item is presented first if face pairs are presented sequentially presented faces (the judged similarity of two objects, A and B, can depend on the order in which they are presented; the similarity of A vs B is not always judged to be the same as B vs A; Tversky, 1977). Other advantages are that categories can be discovered, rather than prescribed, and verbal associations are minimized. Analyses of similarity judgments typically yield more continuous similarity relations between emotion categories along affective dimensions (see Russell & Barrett, 1999).
Free-Labeling: photos of facial configurations are labeled with words offered by participants (unconstrained by experimenter)
Forcing people to translate faces into words is not a good match, since much of the information from faces cannot be easily captured in words (Ekman, 1994). This is not a special criticism of free labeling studies -- it applies to all studies that ask people to label a face with words, including the choice-from-array tasks.
Facial expressions did not evolve to represent specific verbal labels (Ekman, 1994, p. 270). “Regardless of the language, of whether the culture is Western or Eastern, industrialized or preliterate, these facial expressions are labeled with the same emotion terms: happiness, sadness, anger, fear, disgust, and surprise" (Ekman, 1972, p. 278).
There is no widely accepted method for categorizing freely provided responses. (Ekman, 1994, p. 274). Most scientists group together similar words (synonyms), so that a variety of words can be used to show evidence of a correct response (e.g., a frowning face, which is the proposed expression for sadness, could be labeled as "sad," "grieving," "disappointed," "blue," "despairing," and so on. Scientists routinely use databases that indicate synonyms, like WORDNET (used in Srinivasan & Martinez (2018). Also, it is possible to do data-driven groupings of emotion words into semantic categories (e.g., Jack et al., 2016; Shaver et al., 1987). The more serious problem is that early studies using free-labeling (e.g., Boucher & Carlson, 1980; Izard, 1971) did not provide enough information in the method sections about how freely provided labels were grouped.
Using freely chosen labels in a study of different cultures is difficult because it may be hard to find adequate translations (Ekman, 1994, p. 274). A given emotion word, like sadness, can correspond to different emotion concepts (with different features) in different languages (e.g., Wierzbicka, 1986, 2014). A single emotion word in one language can refer to more than one concept in another language (e.g., Pavlenko, 2014). Some languages have no one-to-one translation for English emotion words and some emotion concepts in other languages are not directly translatable into English emotion words (see Barrett, 2017; Russell, 1991; Jack et al., 2016). This is not a special criticism of free labeling studies – it holds for any experiment that uses emotion words requiring translation, including choice-from-array tasks. A standard solution to this problem is to use both forward and backward translation (e.g., a word spoken in Hadzane is translated into English and then back translated into Hadzane; if there is no broken telephone, then the translation has fidelity). An even better method is to elicit features for the emotion words in question, including typicality of those features, to determine the fidelity of translation (e.g., de Mendoza et al., 2010) Scientifically, issues with translation are manageable if scientists allow phrases to stand in for specific words.
Using only single words will always fail to capture much of the rich information in faces. Participants often provide multiple words or even longer descriptions of situations, behaviors, or behaviors in situations (e.g., see Gendron et al., 2014; Russell, 1994). Such data are time consuming to code and analyze.
Even when participants are told that photographs are of people trying to express an emotion, they often offer non-emotion labels. For example, Izard (1971) found that people offered labels such as deliberating, clowning, skepticism, pain, and so on (as reported in Russell, 1994). This is not necessarily evidence that participants did not understand the task asked of them. It might be evidence that these facial configurations are not specific for expressing emotions.

Note. Response tasks are arrayed in order from those that constrain participants’ responses most, making it difficult to observe evidence that can disconfirm commonsense beliefs about emotion to those that are least constrained, making it easier to observe variation and disconfirm commonsense beliefs. Choice-from-array = participants are shown a facial configuration and asked to infer its emotional meaning by choosing an emotion word from a small set of words; or, participants presented with an emotion word that labels an emotion category (e.g., sadness) or a brief story about a typical instance of an emotion category (e.g. “the boy’s much loved dog just died and he is sad”) along with two or three photographs of faces (typically posed into one of the configurations presented in Figure 4) and then asked to choose the facial configuration that they judge best matches the emotional episode described in the word or vignette. Typically, each emotion category is represented by a single scenario. Free sorting = Participants are given photographs of facial configurations and asked to sort them into emotion categories by piles on a big table or on a computer screen. Pairwise similarity judgments = participants rate the similarity of all possible pairs of face stimuli (e.g., on a scale of 0-6). For detailed design concerns about choice-from-array tasks, see Russell (1994, 1995).

Summary.

Our review of the available evidence from expression production studies in small-scale, remote cultures is inconclusive because there are no systematic, controlled observations that examine how people who live in these cultural contexts spontaneously move their facial muscles during emotional episodes. The evidence that does exist suggests that common beliefs about emotion may share some similarities across urban and small-scale cultural contexts, but more research is needed before any interpretations are warranted. These findings are summarized in the fourth and fifth data rows of Table 4.

Studies of Healthy Infants and Children

The facial movements of infants and young children provide a valuable way to test common beliefs about emotional expressions because, unlike older children and adults, babies cannot exert voluntary control over their spontaneous expressive behaviors, meaning that they are unable to deliberately mask or portray instances of emotion in accordance with social demands. As a general rule, infants understand far more about the world than what they can easily convey through their physical actions, making it difficult for experiments to distinguish between what infants understand, which often exceeds what they can actually do. Experiments must use human inference to determine when an infant is in an emotional state, as is the case in studies of adults (see Human inference and assessing the presence of an emotional state). The presence (or absence) of an instance of emotion is inferred (i.e., stipulated), either by a scientist (who exposes a child to something that is presumed to evoke an emotion episode) or by adult “raters” who infer the emotional meaning of the evoking situation or the child’s body movements and vocalizations (see Subjective measures of an emotional instance). In the latter cases, inferences are measured by asking research participants to label the situation or the child’s emotional state by choosing an emotion word or image from a small set of options, a task known as choice-from-array. We address the strengths and weaknesses of choice-from-array tasks (see Table 6) and the potential risk of confirmatory bias with the use of such methods (see Some observations on interpreting the data, below).

Table 6:

Culturally common facial configurations discovered using the reverse correlation method

Facial
Configuration
AU
Description
Associated Emotion Words –
U.K.
Associated Emotion Words
- China
graphic file with name nihms-1021596-t0042.jpg 6+12+13+14 delighted, joy, happy, cheerful, contempt, pride joyful, delighted, happy, glad, feel well, pleasantly surprised, embarrassed, pride
graphic file with name nihms-1021596-t0043.jpg 4+20+24+43 fear, scared, anxious, upset, miserable, sad, depressed, shame, embarrassed afraid, anxious, distressed, broken-hearted, sorrow and sadness, having a hard time, grief, dismay, anguish, worry, vexed, unhappy, shame, despise
graphic file with name nihms-1021596-t0044.jpg 2+5+26+27 ecstatic, excited, surprised, frightened, terrified amazed, greatly surprised, alarmed and panicky, scared, fear
graphic file with name nihms-1021596-t0045.jpg 7+9+16+22 hate, disgust, fury, rage, anger disgusted, bristle with anger, furious, wild wrath, storm of fury, storm of anger, indignant, rage

Note. Facial configurations extracted using reverse correlation from 62 models of facial configurations. Red coloring indicates stronger AU presence and blue indicates weakest AU presence. Some words and phrases that refer to emotion categories in Chinese are not considered emotion categories in English. Modified from Jack et al. (2016) and reproduced with permission.

With such a strong reliance on human inference, there is a risk that scientists will implicitly confound the measurements made in an experiment with their interpretation of those measurements, in effect over-interpreting infant behavior as reflecting a specific aspect of an emotional event, in part because these young research participants cannot speak for themselves. Some early and influential studies confound the observation of facial movements with their interpreted emotional meaning, leading to the conclusions that babies as young as 7-months of age were capable of producing an expression of anger when, in fact, it is more scientifically correct to say that the babies were scowling. For example, in one study, infants’ facial movements were coded as they were given a cookie, and then the cookie was taken away and placed out of reach although still clearly visible. The babies appeared to scowl when the cookie was removed and not when it was in their mouths (Stenberg, Campos & Emde, 1983). It is certainly possible that this repeated giving and taking away of the treat angered the infants, but the babies might also have been confused or just generally distressed. Without some independent evidence to indicate that a state of anger was induced, we cannot confidently conclude that certain facial movements in an infant reliably express a specific instance of emotion.

The Stenberg et al. study illustrates some of the design issues that have historically been of concern in many studies with infants. First, emotion-inducing situations are often defined with commonsense intuitions rather than objective evidence (e.g., an infant is assumed to become angry when a cookie is taken away). In fact, it is difficult to know how any individual infant at any point in time will construct and react to such an event. Second, when an infant produces a facial movement, a common assumption is used to infer its emotional meaning without additional measures or controls (e.g., when a scowling facial configuration is observed, it is assumed to necessarily be an expression of infant anger, even if there are no data to confirm that a scowl is specific to instances of anger in an infant). In fact, years later, Campos and his team revised their earlier interpretation of their findings as their research program progressed, later concluding that the facial movements in question (infants lowering and drawing together their brows, staring straight ahead, or pressing their lips together) were more generally associated with unpleasantness and distress, and were not reliable expressions of anger (e.g., Camras, Oster et al., 2007).

The inference problem is particularly poignant when fetuses are studied. For example, a study that used 4-D ultrasonography observed 20-week-old fetuses knitting their brows and described the facial movements as expressions of distress (Dondi et al., 2014). Yet the fetuses were producing these facial movements during situations when fetal distress was unlikely. The brow-knitting was observed during noninvasive ultrasound scanning that did not involve perturbation of the fetus and the pregnant women were at rest. Furthermore, the scans were brief in duration and the facial movements were interspersed with other movements that are typically not thought to express negative emotions, such as smiling and mouthing. This is an example of making a scientific inference about an emotion occurring based solely on the facial movements without converging evidence that the organism in question (a fetus) was in a distressed state. Doing so highlights the common but unsound assumption that certain facial movements reliably index instances of the same emotion category.

The study of expression production in infants and children must deal with other design challenges, in addition to the reliance on human inference, that are shared by experiments employing adult participants. In particular, most experiments observe facial movements in a restricted range of laboratory settings rather than in the wide variety of situations that naturally occur in everyday life. The frequent use of only a single stimulus or event to observe facial movements for each emotion category limits the opportunity to discover whether the expression of an emotion category vary systematically with context.

Even with these design considerations, the scientific findings from studies of infants and children parallel those that we encountered from studies on adults: lack of reliability and specificity in facial muscle movements is the norm, not the exception (again, according to the Haidt & Keltner (1999) criteria in Table 2). Although some older studies concluded that infants produce invariant emotional expressions (e.g., Izard, Hembree, Dougherty, & Spirrizi, 1983; Izard et al., 1987; Izard et al., 1995; Lewis, Ramsay & Sullivan, 2006), these conclusions have been largely overturned by more recent work and in many cases have been reinterpreted and revised by the authors themselves (e.g., Lewis, Ramsay & Sullivan, 2006)..

Facial movement in fetuses, infants and young children.

The most detailed research on facial movements in fetuses and newborns has focused on smiles. Human fetuses lower their brows (AU4), raise their cheeks (AU6), wrinkle their noses (AU9), crease their nasolabia (AU11), pull the corners of their lips (AU12), show their tongues (AU19), part their lips (AU25), and stretch their mouths (AU27) -- all of which have been implicated, to some degree, in adult laughter. Infants sometimes produce facial movements that resemble adult laughter when they are in distress and pain (Dondi et al., 2014; Hata et al., 2013; Reissland et al., 2011; Reissland, Francis, & Mason, 2013; Yan et al., 2006). Within 24 hours of birth, infants raise their cheek muscles in response to being touched (Cecchini et al., 2011). But these movements are not specific to smiling; neonates also raise their cheeks (contract the zygomatic muscle) during rapid eye movement (REM) sleep, when drowsy, and during active sleep (Dondi et al., 2007). A neonatal smile with raised cheeks is caused by brainstem activation (Rinn, 1984), reflecting internally generated arousal rather than expressing or communicating an emotion or even a more general feeling of pleasure (Emde & Koenig, 1969; Sroufe, 1996; Wolff, 1987). So, it remains unclear whether fetal or neonatal facial muscle movements have any relationship to specific emotional episodes, as well as more generally to pleasant feelings or to other social meanings (Messinger, 2002).

In fact, it’s not clear that fetal and neonatal facial movements always have a psychological meaning (consistent with a behavioral ecology view of facial movements; Fridlund, 2017). Newborns appear to produce some combinations of facial movements for muscular reasons. For example, infants produce facial movements associated with the proposed expression for “surprise” (open mouth and raised eyebrows) in situations that are unsurprising, just because opening the mouth necessarily raises their eyebrows; conversely, infants do not consistently show the proposed expressive configuration for surprise in contexts that are likely to be surprising (Camras, 1992; Camras et al., 2017). The facial movement that is part of the proposed expression for sadness (brows oblique and drawn together) occurs when infants attempt to lift their heads to direct their gaze (Michel, Camras, & Sullivan, 1992).

In addition, newborns produce many facial movements that co-occur with fussiness, distress, focused attention, and distaste (Oster, 2005). Newborns react to being given sweet versus sour liquids; for example, newborns make a nose-wrinkle movement, which is part of the proposed expressive configuration for disgust, when given a sour liquid (Granchrow et al., 1983). However, other studies show that newborns also make this facial movement when given sweet, salty, sour and bitter tastes (e.g., Rosenstein & Oster, 1988). Still other studies show that nose-wrinkling does not always occur when infants taste lemon juice (i.e., when that facial movement is expected; Bennett et al., 2002). More generally, infants rarely produce consistent facial movements that cleanly map onto any single emotion category. Instead, infants produce a variety of facial configurations, indicating a lack of emotional specificity (Matias & Cohn, 1993).

There are further examples that illustrate how infant facial movements lack strong reliability and specificity. In a study of 11-month old babies from the US, China and Japan, infants saw a toy gorilla head that growled (to induce fear) or their arms were restrained (to induce anger; Camras et al., 2007). Observers judged the infants to be fearful or angry based on their body movements; yet, the infants produced the same facial movements in the two situations.24 In another study, one-year-old infants were videotaped in situations where they were tickled (to elicit joy), tasted sour flavors (to elicit disgust), watched a jack-in-the box (to elicit surprise), had their arm restrained (to elicit anger), and were approached by a masked stranger (to elicit fear) (Bennett, Bendarsky, and Lewis, 2002). Infants whose arms were restrained (to purportedly induce an instance of anger) produced the facial actions associated with the proposed facial configuration for an anger expression only 24 percent of the time (low reliability), and instead 80 infants (54%) produced the facial actions proposed as the expression of surprise, 37 infants (25%) produced the facial actions proposed as the expression of joy, 29 infants (19%) produced the facial actions proposed as the expression of fear, and 28 (18%) produced the facial actions proposed as the expression of sadness. This dramatic lack of specificity was observed for all emotion categories studied. An equal number of babies produced facial movements that are proposed as the expressions of joy, surprise, anger, disgust, and fear categories when a sour liquid was placed on infants’ tongues to elicit disgust. When infants faced a masked stranger, only 20 (13%) produced facial movements that correspond to the proposed expression for fear, compared to 56 infants (37%) who produced facial actions associated with the proposed expression for instances of joy.25

Taken together, these findings suggest that infant facial movements may be associated with the affective features of experience, such as distress or arousal, as originally described by Bridges (1932), or communicate a desire to approach or avoid something (e.g., Lewis, Sullivan & Kim, 2015). Affective features such as valence (ranging from pleasantness to distress) and arousal (ranging from activated to quiescent) are continuous properties of consciousness, just as approach and avoidance are continuous properties of action. These affective features are shared by many instances of different emotion categories, as well as with mental events that are not considered emotional (as discussed in Box 9, in SOM) but are still effective and important for infants.26 Over time, infants likely learn to differentiate mental events with simple affective features into episodes of emotion with additional psychological features that are specific to their socio-cultural contexts, making them maximally effective at eliciting needed responses from their caregivers (Barrett, 2017a; Holodynski & Friedlmeier 2006; Weiss & Nurcombe, 1992; Witherington et al., 2008).

The affective meaning of an infant’s facial movements may, in fact, be the very properties that make these movements so salient to adult observers. When infants move their lips, open their mouths, or constrict their eyes, adults view infants as feeling more positively or negatively depending upon the context (Bolzani et al., 2005). Infant expressions thus do have a reliable link to instrumental effects in the adults who observe them – playing an important role in parent-infant interaction, attachment and the beginnings of social communication (Atzil et al., 2018; Feldman, 2016). For example, if an infant cries with narrowed eyes, adults rate that infant’s emotion as more negative or having an unwanted experience or needing help, but if the infant makes that same eye movement while smiling, adults interpret the infant as experiencing more positive emotion. These data consistently point to the usefulness of facial movements in the communication of arousal and valence (properties of affect; Box 9, SOM). Even when episodes of more specific emotions start to emerge, we don’t yet have evidence that facial movements map reliably and regularly to a specific emotion category.

Young children begin to produce adult-like facial configurations after the first year of life. Even then, however, children’s facial movements continue to lack strong reliability and specificity (Bennett et al., 2002; Camras & Shutter, 2010; Matias & Cohn, 1993; Oster, 2005). Examples of a wide-eyed gasping facial configuration, proposed as the expression of fear (see Figure 4), have rarely been observed or reported in young infants (Witherington et al., 2010). Nor do infants reliably produce a scowling facial configuration, proposed as the expression of anger (again, see Figure 4). Infants scowl when they cry or are about to cry (Camras, Fatani, Fraumeni & Shuster, 2016). A frown (mouth corner depression, AU15) is not reliably and specifically observed when infants are frustrated (Lewis & Sullivan, 2014; Sullivan et al., 2003). A smile (cheek raising and lip corner pulling, AU6 and AU12) is not reliably observed when infants are in visually engaging or mastery situations, or even when they are in pleasant social interactions (Messinger, 2002).

Experiments that observe young children’s facial movements in naturalistic settings find largely the same results as those conducted in controlled laboratory settings. For example, one study trained ethnographic videographers to record a family’s daily activities over four days (Sears et al., 2014). Coders judged whether or not the child from each participating family made a scowling facial configuration (referred to as an expression of anger), a frowning facial configuration (referred to as an expression of sadness), and so on, for the six (presumed) emotion categories included in the study -- happiness, sadness, surprise, disgust, fear, and anger. During instances that were coded as anger (defined as situations that included verbal disagreements/sibling bickering, requests for compliance and/or reprimands from parents, parent refusal of child requests, during homework, and sibling provocation), a variety of facial movements were observed, including frowns, furrowed brows, and eye-rolls, as well as a variety of vocalizations, including shouts and whining, and both nonaggressive and aggressive physical behaviors. Perhaps the most telling observations for our purposes is that expressions of anger were more often vocal than facial. During anger situations, children raised their voices 42% of the time, followed by whining about 21% of the time. By contrast, children made scowling facial configurations only 16.2% of the time.27 Yet even during anger situations, the facial movements were predominantly frowning, which can be part of many different proposed facial configurations. The authors reasoned that children engage in specific behaviors to obtain specific goals, and that behaviors such as whining are more likely to attract attention and possibly change parental behavior than will a facial movement. Indeed, it is easier for parents to ignore a negative facial expression than a whining child in the room! Similar findings for low reliability and specificity of the facial configurations presented in Figure 4 were recently observed in a naturalistic study that videotaped seven to nine-year old children and their mothers discussing a conflict during their visit to the laboratory related to homework, chores, bedtime or interactions with siblings (Castro et al., 2017).

Summary.

Newborns and infants react to the world around them with facial movements. There is not yet sufficient evidence, however, to conclude that these facial movements reliably and specifically express the instances of any specific emotion category (findings summarized in Table 4). When considered alongside vocalizations and body movements, there is consistent evidence that infant facial movements reliably signal distress, interest and arousal, and perhaps serve as a call for help and comfort. In young children, instances of the same emotion category appear to be expressed by a variety of different muscle movements, and the same muscle movements occur during instances of various emotion categories, and even during non-emotional instances. It may be the case that reliability and specificity emerges through learning and development (see Box 10, in SOM), but this remains an open question that awaits future research.

Studies of Congenitally Blind Individuals

Another source of evidence to test the common view comes from observations of facial movements in people who were born blind. The assumption is that people who are blind cannot learn, by watching others, which facial muscles to move when expressing emotion. Based on this assumption, several studies have claimed to find evidence that congenitally blind individuals express emotions with the hypothesized facial configurations in Figure 4 (e.g., blind athletes show expressions that are reliably interpreted as shame and pride, Tracy & Matsumoto, 2008; see also Matsumoto & Willingham, 2009). People who are born blind learn through other sensory modalities, however (for a review, see Bedny & Saxe, 2012), and therefore can learn whatever regularities exist between emotional states and words for facial movements from hearing descriptions in conversation, in books and movies, and by direct instruction.28 As an example of such learning, Olympic athletes who win medals smile only when they know they are watched by other people, such as when they are on the podium facing the audience; in other situations, such as while waiting behind the podium or while on the podium facing away from people but towards a flag, they did not smile (but presumably were still very happy; Fernandez-Dols et al., 1995). Such findings are consistent with the behavioral ecology view of facial expressions, Fridlund, 1991, 2017) and with more recent sociological evidence that smiles are social cues that can communicate different social messages depending on the cultural context (Martin, Rychlowska, Wood and Niedenthal, 2017).

The limitations that apply to studies of emotional expressions in sighted individuals, reviewed throughout this paper, are even more applicable to scientific studies of emotional expressions in the blind.29 Participants are given pre-determined emotion categories that shape their possible responses, and facial movements are often quantified by human judges who have their own biases when making commonsense judgments (e.g., Galati et al., 1997; Galati et al., 2001; Valente et al., 2017). In addition, people who are blind make additional, often unusual movements of the head and the eyes (Chiesa et al., 2015). For example, people who are blind from birth often move their head in unusual ways to better hear objects or echoes. These unusual movements might interfere with or contaminate expressive facial movements. More importantly, they reveal whether a participant is blind or sighted, and this knowledge can bias human raters who are judging the presence or absence of facial movements in emotional situations.

Helpful insights about the facial expressions of congenitally blind individuals comes from a recent review (Valente et al., 2017) that surveyed 21 studies published between 1932 and 2015. These studies observe how blind participants move their faces during instances of emotion and then compared those movements both to the proposed expressive forms in Figure 4 and to the facial movements of sighted people. Both spontaneous facial movements and posed movements were tested. Eight older studies (published between 1932-1977) reported that congenitally blind individuals spontaneously expressed emotions with the proposed facial configurations in Figure 4, but Valente et al. (correctly) questioned the objectivity of these studies because the data were largely based on subjective impressions offered by researchers or their assistants. The 13 studies published between 1980 and 2015 were better designed: they videotaped participants’ facial movements and described them using a formal facial coding system like FACS for adults or a similar coding system for children. These studies are too few in number and have insufficient sample sizes to conduct a formal meta-analysis, but taken together they suggest that, in general, congenitally blind individuals spontaneously moved their faces in similar ways to sighted individuals during instances of emotion: both groups expressed instances of anger, disgust, fear, happiness, sadness or surprise with the proposed expressive configurations (or their individual AUs) in Figure 4 with either weak reliability or no reliability at all, and neither group produced any of the configurations with any specificity (e.g., Galati et al., 2001; Galati et al., 2003; Galati et al., 1997). The lack of specificity is not surprising given that, upon closer inspection, several of the studies discussed in Valente et al. (2017) compared emotion categories that systematically differ in their prototypical affective properties, contrasting facial movements in pleasant vs. unpleasant circumstances (e.g., Cole et al., 1989), or observed facial movements only in pleasant circumstances without distinguishing the facial AUs for the happiness category vs. other positive emotion categories (e.g., Chiesa et al., 2015), such that their findings cannot be interpreted unambiguously as evidence pertaining to emotional expressions, per se.

While congenitally blind and sighted individuals were similar to one another in the variety of their spontaneous facial movements, they differed in their posed facial configurations. After listening to descriptions of situations that were supposed to elicit an instance of anger, sadness, fear, disgust, surprise, and happiness, sited participants posed their faces with the proposed expressive forms for the negative emotion categories in Figure 4 at higher levels of reliability and specificity than did blind participants (Galati et al., 1997; Roch-Levecq, 2006). These findings suggest that congenitally blind individuals have different beliefs about emotional expressions or that their knowledge of social rules for producing those configurations on command differs from those of sighted individuals.

Summary.

The evidence from studies of blind individuals is consistent with the other scientific evidence reviewed so far (Table 4). Even in the absence of visual experience, blind individuals, like sighted individuals, develop the ability to spontaneously make a variety of facial movements to express emotion, and those movements do not reliably and specifically configure as proposed by the common view of emotion (depicted in Figure 4). Learning to voluntarily pose the proposed expressions in Figure 4 does seem to covary with vision, however, further emphasizing that posed and spontaneous expressions should be treated as different phenomena. Further scientific attention is warranted to examine how congenitally blind individuals learn, via other sensory modalities, to express emotions.

Summary of Scientific Evidence on the Production of Facial Expressions

The scientific findings we have reviewed thus far – dealing with how people actually move their faces during emotional events – does not strongly support the common view that people reliably and specifically express instances of emotion categories with spontaneous facial configurations that resemble those proposed in Figure 4. Adults around the world, infants and children and congenitally blind individuals all show much more variability than commonly hypothesized. Studies of posed expressions further suggest that particular facial movements are linked to particular emotions more by consensus and beliefs, rather than by scientific evidence for “emotion expression.” Consequently, the commonly used phrases such as “emotional facial expression,” “emotional expression” or “emotional display” are misleading. More neutral phrases that assume less, such as “facial configuration” or “pattern of facial movements” or even “facial actions,” should be used instead.

We next turn our attention to the question of whether people reliably and specifically infer certain emotions from certain patterns of facial movements, shifting our focus from studies of production to studies of perception. It has been long assumed that emotion perception provide an indirect way of testing the common view of emotion production, because facial expressions, when they are assumed to be displays of internal emotional states, are thought to have co-evolved with the ability to recognize and read them (Ekman, Friesen, & Ellsworth, 1972). For example, Shariff and Tracy (2011) have suggested that emotional expression (production) and emotion perception likely co-evolved as an integrated signaling system (for additional discussion, see Jack, Sun, Delis, Garrod, & Schyns, 2016).30 In the next section, we review the scientific evidence on emotion perception.

Perceiving Emotions from Facial Movements: A Review of the Scientific Evidence

For over a century, an active line of research has directly examined whether people reliably and specifically infer emotional meaning in the facial configurations presented in Figure 4. Most of these studies are interpreted as evidence for people’s ability to recognize or decode emotion in facial configurations, on the assumption that the configurations broadcast or signal emotional information to be recognized or detected. This is yet another example of confusing what is known and what is being tested. A more correct interpretation is that these studies indicate whether people reliably and specifically infer or judge emotion in those facial configurations. This pervasive confusion in the scientific literature may explain why very few studies have actually investigated the processes by which people detect the onset and offset of facial movements and infer emotions in those movements (i.e., few studies consider the mechanisms by which people infer emotional states from detecting and perceiving facial movements) (for discussion, see Martinez, 2017a, 2017b). In this section, we first review the design of typical emotion perception experiments that are used to test the common view that emotions can be reliably and specifically “read out” from facial movements. We also examine whether people infer emotions from the facial movements in dynamic, computer-generated faces, a class of studies that offer a more data-driven way to study emotion perception, and in virtual humans, which provides the opportunity for a more implicit approach to studying emotion perception.

The Anatomy of a Typical Experiment Designed to Observe Whether People Reliably and Specifically Infer Emotion in Facial Movements

For a person - a perceiver -- to infer that another person is in an emotional state by looking at that person’s facial movements, the perceiver must have many competencies. People move their faces continuously (i.e., real human faces are never still), so a perceiver must notice or detect the relevant facial movements in question and discriminate them from other facial movements (that is, the perceiver must be able to set a perceptual boundary to know when the movements begin and end, and, for example, that a scowl is different from a sneer). The perceiver must be able to identify (or segment) the movements as an ensemble or pattern (i.e., bind them together and distinguish them from other movements that are normally inferred to be irrelevant). And the perceiver must be able to infer similarities and differences between different instances of facial movements, as specified by the task (e.g., categorize a group of facial movements as instances expressing anger, fear, etc.). This categorization might involve merely labeling the facial movements, referred to as action identification (describing how a face is moving, such as smiling) or it might involve inferring that a particular mental state caused the actions, referred to as mental inference or mentalizing (inferring why the action is performed, such as a state of happiness; Vallacher & Wegner, 1987). In principle, the categorization could also involve inferring a situational cause for the actions, but in practice, this question is rarely investigated in studies of emotion perception. The overwhelming majority of studies ask participants to make mental inferences, although as we discuss later in this section, there appears to be important cultural variation in whether emotions are perceived as situated actions vs. as mental states that cause actions.

The use of posed configurations of facial movements in assessments of emotion perception.

The majority of the experiments that study emotion perception ask participants to infer emotion in facial configurations that are posed by actors who are not in an emotional state when the photos were taken or by computer-generated humans who have no actual emotional state. As a consequence, it is not possible to assess the accuracy (i.e., validity) of perceivers’ emotional inferences and, correspondingly, data from emotion perception studies cannot be interpreted as support for the validity of common beliefs about emotional expressions. As is the case in studies of expression production, it is more appropriate to interpret participants’ responses in terms of their agreement (or consensus) with common beliefs. Even more serious is the fact that the proposed expressive facial configurations in Figure 4 do not capture the wider range of muscle movements that are observed when people express instances of these emotion categories. A recent study that mined over seven million images from the internet (for method, see Box 7 in SOM; Srinivasan & Martinez, 2018) identified multiple facial configurations associated with the same emotion category label and their synonyms --17 distinct facial configurations were associated with the word “happiness,” five with “anger,” four with “sadness,” four with “surprised,” two with “fear,” and one with “disgust.” The different facial configurations associated with each emotion word were more than mere variations on a universal core expression – they were distinctive sets of facial movements.31

Measuring emotion perception.

The typical emotion perception experiment takes one of several forms, summarized in Table 5. Choice-from-array tasks, in which participants are asked to match photos of facial configurations and emotion words (with or without brief stories), have dominated the study of emotion perception since the 1970s. For example, a meta-analysis of emotion perception studies published in 2002 summarized 87 studies, 83 (95%) of which exclusively used a choice-from-array response method (Elfenbein & Ambady, 2002). This method has been widely criticized for over two decades, however, because they limit the possibility of observing evidence that could disconfirm the common view. Participants are strongly constrained in how they can infer meaning in a facial configuration, such as a photograph of a scowling facial configuration, since their choices are constrained to the options provided in the experiment (usually a small number of emotion words). In fact, the preponderance of choice-from-array tasks in the scientific study of emotion perception has been identified as one important factor that has helped perpetuate and sustain the common view (Russell, 1994). Other tasks exist for assessing emotion perception (see Table 5), including those that use a free-labeling method, where participants are asked to freely nominate words to label photographs of posed facial configurations, rather than choosing a word from a small set of predefined options. For example, upon viewing a scowling configuration, participants might offer words like “angry,” “sad,” “confused,” “hungry,” or even “wanting to avoid a social interaction.” By allowing participants more freedom in how they infer meaning in a facial configuration, free-labeling makes it equally possible to observe evidence that could either support or disconfirm the common view.

Recent innovations in measuring emotion perception use computer generated faces or heads rather than photographs of posed human faces. One method, called reverse correlation, measures participants’ internal model of emotional expressions (i.e., their mental representations of which facial configurations are likely to express instances of emotion) by observing how participants label an avatar head that displays random combinations of animated facial action units (Yu et al., 2012; for a review, see Jack et al., 2018; Jack & Schyns, 2017). As each pattern appears (on a given test trial), participants infer its emotional meaning by choosing an emotion label from a set of options (a choice-from-array response). After thousands of trials, researchers estimate the statistical relationship between the dynamic patterns of facial movements and each emotion word (e.g., disgust) to reveal participants’ beliefs about which facial configurations are likely to express different emotion categories.

A second approach using computer-generated faces would have participants interact with more fully developed virtual humans (Rickel, Marsella, et al., 2003), also known as Embodied Conversational Agents (Cassell et al., 2000). Virtual humans are software-based artifacts that look like and act like people (for examples, see Figure 7). They are similar to characters in video games in their surface appearance and are designed to interact face-to-face with humans using the same verbal and nonverbal behavior that people use to interact with one another. The underlying technologies used to realize virtual humans vary considerably in approach and capability, but most virtual human models can be programmed to make context-sensitive, dynamic facial actions that, when in a person, would typically communicate emotional information to other people (see Box 11 in SOM for discussion). The majority of the scientific studies with virtual humans were not designed to test whether human participants infer specific emotional meaning in a virtual human’s facial movements, but their design makes them useful for studying when and how facial movements take on meaning as emotional expressions: Unlike all the other ways of assessing emotion perception discussed so far, which ask participants to make explicit inferences about the emotional cause of facial configurations, interactions with virtual humans allow scientists to study how a participant implicitly infers emotional meaning during social interactions.

Figure 7. Examples of virtual humans. Virtual humans are software-based artifacts that look like and act like people.

Figure 7.

(A) Feng et al, 2017; (B) Zoll et al., 2006; (C) Hoyt et al., 2003; (D) Marsella et al. 2000.

Testing the common view from observations of whether certain facial configurations are reliably and specifically perceived as expressions of certain emotion categories.

Traditionally, in most experiments, if participants reliably infer an emotional state from a facial configuration (e.g., inferring anger from a scowling facial configuration) at levels that are greater than what would be expected by chance, then this is taken as evidence that people recognize an emotional state in its facial display. It is more scientifically correct, however to interpret this as evidence that people infer an emotional state (i.e., they consistently make a reverse inference) unless the inference has been verified as valid (i.e., the person in the photograph is, indeed, in the expected emotional state). Only when reverse inferences are observed in a reliable and specific way within an experiment can scientists reasonably infer that participants are perceiving an instance of a certain emotion category in a certain facial configuration; technically, the inference holds only for emotion perception as it occurs in the particular situations contained in the experiment (because situations are never randomly sampled). If the emotion perception evidence replicates across experiments that sample people from the same culture, then the interpretation can be generalized to emotion perceptions in that culture. Only when the findings generalize across cultures – that is, replicate across experiments that sample people from different cultures -- is it reasonable to conclude that people universally infer a specific emotional state when perceiving as specific facial configuration. These findings might also be interpreted as evidence about the reliability and specificity of producing emotional expressions if the co-evolution assumption is valid (i.e., that emotional expressions and their perception co-evolved as an integrated signaling system; Ekman et al., 1972; Jack et al., 2016; Shariff & Tracy, 2011).

Studies of Healthy Adults From the U.S. and Other Developed Nations

Studies that measure emotion perception with choice-from-array tasks.

The most recent meta-analysis of emotion perception studies was published in 2002 (Elfenbein & Ambady, 2002). It statistically summarized 87 experiments in which over 22,000 participants from over 20 cultures around the world inferred emotional meaning in facial configurations and other stimuli (such as posed vocalizations). The majority of participants were sampled from larger-scale or developed countries, including Argentina, Brazil. Canada, Chile, China, England, Estonia, Ethiopia, France, Germany, Greece, Indonesia, Ireland, Israel, Italy, Japan, Malaysia, Mexico, the Netherlands, Scotland, Singapore, Sweden, Switzerland, Turkey, the US, Zambia and various Caribbean countries. The majority of studies (95%) used posed facial configurations; only four studies had participants label spontaneous facial movements, a dramatic example of the challenges facing validity that we discussed earlier. All but four studies used a choice-from-array response method to measure emotion inferences, a good example of the challenges facing hypothesis disconfirmation that we discussed earlier.

The results of the meta-analysis, presented in Figure 8a, reveal that perceivers inferred emotions in the facial configurations in Figure 4 in line with the common view, well above chance levels (using the criteria set out by Haidt & Keltner (1999), presented in Table 2).. Results provided strong evidence that, when participants are viewing posed facial configurations made by people from their own culture, they reliably perceived the expected emotion in those configurations: scowling facial configurations were perceived as anger expressions, wide-eyed facial configurations were perceived as fear expressions, and so on, for all six emotion categories. Moderate levels of reliability were observed when perceivers were labeling facial configurations posed by people from other cultures; this difference in reliability between same- and cross-culture differences is referred to as an ingroup advantage (see Box 12, in SOM). The majority of emotion perception studies do not report whether the hypothesized facial configurations are perceived with any specificity (e.g., how likely was a scowl to be perceived as expressing an instance of emotion categories other than anger, or as an instance of a mental category that is not considered emotional). Without information about specificity, no firm conclusions can be drawn about the emotional meaning of the facial configurations in Figure 4, especially for the translational purpose of inferring someone’s emotional state from their facial comportment in real life.

Figure 8. Emotion perception findings.

Figure 8

(A) Average effect sizes for perceptions of facial configurations from Elfenbein & Ambady (2002), in which 95% of the articles summarized used choice-from-array to measure participants’ emotion inferences. (B) Free-labeling of facial configurations across five language groups from Srinivasan & Martinez (2018). IDs chosen represent the best match to the commonsense facial configurations in Figure 4 based on AUs present. No configuration discovered in this study exactly match the AU configurations proposed by Darwin or documented in prior research. Proportion of times participants offered emotion category labels (or their synonyms) are reported. According to standard scientific criteria, universal expressions of emotion should elicit agreement rates that are considerably higher than those reported here, generally in the 70 ± 90% range, even when methodological constraints are relaxed (Haidt & Keltner, 1999). Specificity data were not available for the Elfenbein & Ambady (2002) meta-analysis.

Nonetheless, most of the studies cited in the meta-analysis interpret their reliability findings alone as evidence for the reverse inference of inferring anger from a scowling face, disgust from a nose-wrinkled face, fear from a wide-eyed gasping face, and so on. Such findings may explain why many scientists who study emotion, when surveyed, indicated that they believe compelling evidence exists for the hypothesis that certain emotion categories are each expressed with a unique, universal facial configuration (see Ekman, 2016) and interpret variation in emotional expressions to be caused by cultural learning that modifies what are presumed to be inborn universal expressive patterns (e.g., Cordaro et al., 2017; Ekman, 1972; Elfenbein, 2013). Cultural learning has also been hypothesized to modify how people “decode” facial configurations during emotion perception (Buck, 1984).

Studies that measure emotion perception with free-labeling tasks.

As we foreshadowed, experimental methods that place fewer constraints on participants’ inferences in experiments that measure emotion perception (Table 5) provide considerably less support for the common view of emotional expressions. In the least constrained experimental task, called free-labeling, perceivers freely volunteer a word (emotion or otherwise) that they believe best captures the meaning in a facial configuration rather than choosing from a small set of experimenter-chosen options. In urban samples, participants who freely-label facial configurations produce the expected emotion labels with weak reliability (when labeling spontaneously produced facial configurations) to moderate reliability (when labeling posed facial configurations), and usually reveal weak specificity when it is assessed at all (for examples and discussion, see Russell, 1994; also see Naab & Russell, 2007). For example, when participants from many countries where English, Spanish, Mandarin Chinese, Farsi, Arabic and Russian is spoken as a first language were then asked to freely provide emotion words to label each of 35 facial configurations that had been cross-culturally identified (Srinivasan & Martinez, 2018), their labels provided evidence of a moderately reliable correspondence between facial configurations and emotion categories, but there was no evidence of specificity (see Figure 8b).32 Multiple facial configurations were associated with the same emotion category label (e.g., 17 different facial configurations were associated with the expression of happiness, five with anger, four with sadness, four with surprise, two with fear, and one with disgust). This many-to-many mapping is inconsistent with the common view that the facial configurations in Figure 4 are universally recognized as expressing the hypothesized emotion category, and they give evidence of variation that is far beyond what is proposed by the basic emotion view. Some of this variability may come from variability across different cultures and languages, but there is variability even within a single culture and language. Evidence of this many-to-many mapping is apparent in free-labeling tasks in small-scale, remote samples as well (Gendron et al., 2018), which we discuss in the next section.

Studies that measure emotion perception with the reverse correlation method.

Using a choice-from-array response method with the reverse correlation method is an inductive way to learn people’s beliefs about which facial configurations express an emotion category (for a review, see Jack et al., 2018; Jack & Schyns, 2017). In such studies, participants view thousands of random combinations of AUs that are computer generated on an avatar head and label each one by choosing an emotion word from a set of pre-defined options. All of the facial configurations labeled with the same emotion word (e.g., anger) are then statistically combined for each participant to estimate a belief about which facial movements express the corresponding emotion category. One recent study using the reverse correlation method with U.K. and Chinese participants found evidence of both variation in the facial movements that were judged to express a single emotion category, as well as similarity in the facial movements that were judged to express different categories (Jack et al., 2016). The study first identified groupings of emotion words that are widely discussed in the scientific literature (which, we should note, is dominated by English), corresponding to 30 English words grouped into eight emotion categories for the U.K. sample (happy/excited/love, pride, surprise, fear, contempt/disgust, anger, sad and shame/embarrassed) and 52 Chinese words grouped into twelve categories in the Chinese sample (joyful/excitement, pleasant surprise, great surprise/amazement, shock/alarm, fear, disgust, anger, sad, embarrassment, shame, pride, and despise). The reverse correlation method revealed 62 separate facial configurations: the same emotion category in a given culture was associated with multiple models of facial movements because synonyms of the same emotion category were associated with distinctive models of facial movements. Amidst this variability, Jack and colleagues also found that these 62 separate facial configurations could be summarized as four prototypes which are presented in Table 6, along with the corresponding emotion words that they were frequently associated with. Each prototype was described with a unique set of affective features (combinations of valence, arousal and dominance). When the four estimated configurations are compared with the common view presented in Figure 4, along with the basic emotion hypotheses listed in Table 1, there are some striking similarities: Configuration 1 most closely resembles the proposed expression for happiness, configuration 2 is similar to a combination of the proposed expressions for fear and anger, configuration 3 most closely resembles the proposed expression for surprise, and configuration 4 is similar to a combination of the proposed expressions for disgust and anger. 33 Taken together, these findings suggests that, at the most general level of description, participants’ beliefs about emotional expressions (i.e., their internal models of which facial movements expressed which emotions) were consistent with the common view (indeed, they could be taken to constitute part of the common view), but when examined in finer detail with more granularity, participants’ also believe that there is substantial within-category variation in the facial movements that express instances of the same emotion category. This finding suggests that the way the common view is often described in reviews, depicted in the media, and used in many applications, does not in fact do justice to people’s more detailed beliefs about variability in facial expressions.

Studies that implicitly assess emotion perception during interactions with virtual humans.

Designers typically study how a virtual human’s expressive movements influence an interaction with a human participant. Much of the early research modeling expressive movements in virtual humans focused on endowing them with the facial expressions proposed in Figure 4. A number of studies have endowed virtual humans with blends of these configurations (Bui et al., 2004; Arya et al. 2009). Designers are also inspired by other people’s beliefs about how emotions are expressed. Actors, for example, have been asked to pose facial configurations that they believe express emotions, which are then processed by graphical and machine learning algorithms to craft the relation between emotional states and expressive movements (Alexander et al, 2009). In another study, human subjects used a specially designed software tool to craft animations of facial movements that they believed express certain mental categories, including emotion categories. Then, other human subjects judged the crafted facial configurations (Ochs et al., 2010). Increasingly, data-driven methods are used that place people in emotion-eliciting conditions, capture the facial and body motion and then synthesize animations from those captured motions (Niewadomski et al., 2015; Ding et al., 2014, Wang, Marsella, & Hawkins, 2008).

In general, studies with virtual humans nicely show how the situational context influences how people infer the meaning of facial movements (de Melo et al., 2014). For example, in a game that allowed competition and cooperation (Prisoner’s Dilemma, Pruitt & Kimmel, 1977), a virtual human who smiled after making a competitive move evoked more competitive, less cooperative responses from human participants compared to a virtual human using an identical strategy in the game (tit-for-tat) but that smiled after cooperating. Virtual humans who make a verbal comment about a film that is inconsistent with their facial movements, such as saying they enjoyed the film but grimacing that was quickly followed by a smile, were perceived as less reliable, trustworthy and credible (Rehm & Andre, 2005).

The dynamics of the facial actions, including the relative timing, speed and duration of the individual facial actions, as well as the sequence of facial muscle movements over time, offer information over and above the mere presence or absence of the movements themselves and have an important influence on how human perceivers interpret facial movements (e.g., Ambadar et al., 2009; Keltner 1995; Jack & Schyns, 2017, Krumhuber et al. 2013) and how much they trust a virtual human during a social interaction (Krumhuber et al., 2009). Research with virtual humans has shown that the dynamics of facial muscle movements are critical for them to be perceived as emotional expressions (Niewiadomski et al, 2015; Ochs et al., 2010). These findings are consistent with research showing that the temporal dynamics carry information about the emotional meaning of facial movements that are made by real humans (e.g., Kamachi et al., 2001; Sato & Yoshikawa, 2004; Krumhuber & Kappas, 2005; for a review, see Krumhuber, Kappas & Manstead, 2013).34

Summary.

Whether or not people can reliably perceive emotions in the expressive configurations of Figure 4, as predicted by the common view, depends on how participants are asked to report or register their inferences (see Table 4). Hundreds of experiments have asked participants to infer the emotional meaning of posed, exaggerated facial configurations like those presented in Figure 4 by choosing a single emotion word from a small number of options offered by scientists, called choice-from-array-tasks. This experimental approach tends to generate moderate to strong evidence that people reliably label scowling facial configurations as angry, frowning facial configurations as sad, and so on for all six emotion categories that anchor the common view. Choice-from-array tasks severely limit the possibility of observing evidence that can disconfirm the common view of emotional expressions, however, because they restrict participants’ options for inferring the psychological meaning of facial configurations by offering them a limited set of emotion labels. (As we discuss below, when people are provided with labels other than angry, sad, afraid, as so on, they routinely choose them; e.g., Carroll & Russell, 1996; also see Crivelli et al., 2017). Additionally, the specificity of those judgments is largely unreported. Scientists often go further and interpret the reliability findings from these studies as evidence that scowls are expressions of anger, frowns are expressions of sadness, and so on. This logic is not sound, however, because most of these studies ask participants to infer emotion in posed, static faces which are likely limited in their validity (i.e., people posing facial configurations like those depicted in Figure 4 are unlikely to be in the hypothesized emotional state). Furthermore, other ways of assessing emotion perception, such as the reverse correlation method and free-labeling tasks, find much weaker evidence for reliability and/or specificity of emotion inferences. Instead, they suggest that what people actually infer and believe about facial movements incorporates considerable variability: In short, the common view depicted in many reviews, summaries, the media, and used in numerous applications is not an accurate reflection of what people in fact believe about facial expressions of emotion, when probed in more detail. In the next section, we discuss scientific evidence from studies of emotion perception in small-scale remote cultures, which further undermines the common view.

Studies of Healthy Adults Living in Small-Scale, Remote Cultures

A growing number of studies examine emotion perception in people from remote, non-industrialized groups. A more in-depth review of these studies can be found in Gendron et al. (2018). Our goal here is to summarize the trends found in this line of research (see Table 7).

Table 7:

Summary of cross-cultural emotion perception in small-scale societies

Unsupported Weak Support Moderate Support Strong Support
Culture N Citation Culture N Citation Culture N Citation Culture N Citation
Free-labeling Fore, PNGa 100 Sorenson (1975), Sample 2 c Sadong, Borneo a1 15 Sorenson (1975), Sample 4 b
Bahinemo, PNG 71 Sorenson (1975), Sample 3
Hadza, Tanzania 43 Gendron et al. (2018), Study 1
Trobrianders, PNG 32f Crivelli et al. (2017), Study 1
Cue-to-cue matching Shuar, Ecuador 23 Bryant & Barrett (2008), Study 2
Choice-from array: Matching face and words Fore, PNGa 32 Ekman et al. (1969)c
Mwani, Mozambique 36ef Crivelli, Jarillo et al. (2016), Study 2 Sadong, Borneo a1 15 Ekman et al. (1969) b
Trobrianders, PNG 24f Crivelli et al. (2017), Study 2 Dioula, Burkina 39 Tracy & Robbins (2008), Study 2
Trobrianders, PNG 68 ef Crivelli, Jarillo et al. (2016), Study 1
Trobrianders, PNG 36 f Crivelli, Russell et al. (2016), Study 1a
Choice-from array: Matching face and scenario Hadza, Tanzania 54 Gendron et al. (2018), Study 2 Dani, New Guineaa 34 Described in Ekman, 1972 g
Fore, PNG a1 189, 130 e Ekman & Friesen (1971) d
Fore, PNG a1 189, 130 e Sorenson (1975), Sample 1 d

Notes. Findings summarized for anger, disgust, fear, sadness and surprise; happiness is the only pleasant category tested in all studies but Tracy & Robins (2008) and therefore perception can be (and likely is) guided by distinguishing valence in those studies. All studies used photographs of posed facial configurations that are similar to those in Figure 4, except Crivelli, Jarillo et al. (2016), Study 2 and Crivelli et al. (2017), Study 1. The Bryant & HC Barrett (2008) study was designed to examine emotion perception from vocalizations but is included because perceivers matched them to faces; in addition, participants were tested in a second language (Spanish) in which they received training. All choice-from-array studies did not carefully control whether foils and target facial configurations could be distinguished by valence and/or arousal except Gendron et al. 2018, Study 2. N = sample size. All participants were adults unless otherwise specified as adol=adolescents, ch=children. PNG = Papua New Guinea. Unsupported = reliability and specificity at chance, or any level of reliability above chance combined with evidence of no specificity. Weak support = reliability between 20% and 40% (weak) for at least a single emotion category other than happiness combined above chance specificity for that category or reliability between 41% and 70% (moderate) for at least a single category other than happiness with unknown specificity. Moderate support = reliability between 41% and 70% (moderate) combined with any evidence of above chance specificity those categories or reliability above 70% (strong) for at least a single category other than happiness with unknown specificity. Strong support = strong evidence of reliability (above 70%) and strong evidence of specificity for at least a single emotion category other than happiness.

Superscript a:Specificity levels were not reported.

Superscript a1:Specificity inferred from reported results.

Superscript b:The sample size, marginal means and exact pattern of errors reported for the Sadong samples is identical in Sorenson (1975), Sample 3 and Ekman et al. (1969); Sorenson described using a free-labeling method and Ekman et al. (1969) described using a choice-from-array method in which participants were shown photographs and asked to choose a label from a small list of emotion words; Ekman (1994) indicated, however, that he did not use a free-labeling method, implying that the samples are distinct.

Superscript c:Sorenson (1975), Sample 2 included three groups of Fore participants (those with little, moderate and most other group contact). The pattern of findings is nearly identical for the subgroup with the most contact and the data reported for the Fore in Ekman et al. (1969); again, Sorenson described using a free-labeling method and Ekman et al. (1969) described using a choice-from-array method. It is questionable whether the Sadong and the Fore subgroup should be considered isolated (see Sorenson, 1975, p. 362 and 363), but we include them here to avoid falsely dichotomizing cultures as “isolated from” versus “exposed to” one another (Fridlund, 1994; Gewald, 2010).

Superscript d:these are likely the same sample because the sample sizes and pattern of data are identical for all emotion categories except for the fear category, which is extremely similar, and for the disgust category which includes responses for contempt in Ekman & Friesen (1971) but was kept separate in Sorenson (1975).

Superscript e:participants were children.

Superscript f:participants were adolescents.

Superscript g:The Dani sample reported in Ekman, 1972 is likely a subset of the data from Ekman, Heider, Friesen, and Heider (unpublished manuscript).

Studies that measure emotion perception with choice-from-array tasks.

During the period from 1969 to 1975, somewhere between five and eight small-scale samples from remote cultures in the South Pacific were studied with choice-from-array tasks to investigate whether participants perceived emotional expression in facial movements in a similar way when compared to people from the US and other industrialized countries of the Western world (see Figure 9a). Our uncertainty in the number of samples stems from reporting inconsistencies in the published record (see note to Table 7). We present the findings here according to how the original authors reported their findings, despite the inconsistencies. Five samples performed choice-from-array tasks, three in which participants chose a photographed facial configuration to match one brief vignette that described each emotion category (Ekman, 1972; Ekman & Friesen, 1971; Sorenson, 1975) and two in which they chose a photograph to match an emotion word (Ekman, Sorenson, & Friesen, 1969). All five samples performing some version of a choice-from-array task provided strong evidence in support of cross-cultural reliability of emotion perception in small-scale societies. Evidence for specificity was not reported. Until 2008, all claims that anger, sadness, fear, disgust, happiness and surprise are universally recognized (and therefore are universally expressed) were largely based on three papers (two of them peer reviewed) reporting on four samples (Ekman, 1972; Ekman & Friesen, 1971; Ekman et al., 1969).35

Figure 9. Map of cross-cultural studies of emotion perception in small-scale societies.

Figure 9.

Figure 9.

People in small scale societies typically live in groupings of several hundred to several thousand that maintain autonomy in social, political and economic spheres. (A). Epoch 1 studies, published between 1969 and 1975, were geographically constrained to societies in the South Pacific. (B). Epoch 2 studies, published between 2008 and 2017, sample from a broader geographic range including Africa and South America, and are more diverse in the ecological and social contexts of the societies tested. This type of diversity is a necessary condition for discovering the extent of cultural variation in psychological phenomena (Medin et al., 2017). Reproduced with permission from Gendron et al. (2018).

Since 2008, 10 verifiably separate experiments observing emotional inferences in small-scale societies have been published or submitted for publication. These studies include a greater diversity of social and ecological contexts, including sampling five small-scale societies across Africa and the South Pacific (see Figure 9b) who were tested with a greater diversity of research methods listed in Table 5, including tasks that allow for the possibility of observing cross-cultural variation in emotion perception and therefore the possibility of disconfirming the common view. Six samples registered their emotion inferences using a choice-from-array task, in which participants were given an emotion word and asked to choose the posed facial configuration that best matched it or vice versa (Crivelli, Jarillo et al., 2016; Crivelli, Russell et al., 2016; Crivelli et al., 2017, Study 2; Gendron et al., 2018, Study 2; Tracy & Robins, 2008). Only one study (Tracy & Robins, 2008) reported that participants selected an emotion word to match the facial configurations similar to those in Figure 4 more reliably than what would be expected by chance, and effects ranged from weak (anger and fear) to strong (happiness) with surprise and disgust falling in the moderate range.36 Information about the specificity of emotion inferences was not reported. A close examination of the evidence from four studies by Crivelli and colleagues suggest weak to moderate levels of reliability for inferring happiness in smiling facial configurations (all four studies), sadness in frowning facial configurations (all four studies), fear in gasping, wide-eyed facial configurations (three studies), anger in scowling facial configurations (two studies) and disgust in nose-wrinkled facial configurations (three studies). A detailed breakdown of findings can be found in Box 13, in SOM. None of the studies found specificity for any facial configuration, however, except that smiling was reported as unique to happiness, but that finding did not replicate across samples.37

The final study using a choice-from-array task with people from a small-scale, remote culture is important because it involves the Hadza hunter-gatherers of Tanzania (Gendron et al., 2018, Study 2). 38 The Hadza are a high-value sample for two reasons. First, universal and innate emotional expressions are hypothesized to have evolved to solve the recurring fitness challenges of hunting and gathering in small groups on the African savanna (Pinker, 1997; Shariff & Tracy, 2011; Tooby & Cosmides, 2008); the Hadza offer a rare opportunity to study foragers who are currently living in an ecosystem that is thought to be similar to that of our Paleolithic ancestors.39 Second, the population is rapidly disappearing (http://www.sciencemag.org/news/2018/05/farmers-tourists-and-cattle-threaten-wipe-out-some-world-s-last-hunter-gatherers). Prior to this study, the Hadza had not participated in any studies of emotion perception, although they have been the subject of social cognition research more broadly (H. C. Barrett et al., 2016; Bryant et al., 2016). After listening to a brief story about a typical instance of anger, disgust, fear, happiness, sadness and surprise, Hadza participants chose the expected facial configuration more often than chance if the target and foil were distinguished by the affective property referred to as valence (i.e., a smiling configuration depicting a pleasant state vs. a scowling configuration depicting an unpleasant state, consistent with anthropological studies of emotion (Russell, 1991), linguistic studies (Osgood, May & Miron, 1975) and findings from other recent studies of participants from small-scale societies, such as the Himba (Gendron et al., 2014a, b) and the Trobriand Islanders (Crivelli, Jarillo et al., 2014). (Also see Srinivasan & Martinez, 2018, described in Box 7, who showed that perceivers can reliably infer valence but not arousal in facial configurations). In addition, Hadza participants who had some contact with people from other cultures -- they had some formal schooling or could speak Swahili which is not their native language – were more consistently able to choose the common facial configuration than were those with no formal schooling who spoke minimal Swahili (for a similar finding with Fore participants in a free labeling study, see Table 2 in Sorenson, 1975). Of the 27 Hadza participants who had minimal contact with other cultures, only 12 reliably chose the wide-eyed gasping facial configuration to match the fear story at above chance levels. (Compare this finding to the observation that the hypothesized universal expression for fear – a wide-eyed gasping facial configuration – is understood as an aggressive, threatening display by Trobriand Islanders; Crivelli, Jarillo & Fridlund, 2016, 2017; Crivelli, Russell et al., 2016).

Studies that measure emotion perception with free-labeling tasks.

During the period from 1969 to 1975, between one and three small-scale samples from remote cultures in the South Pacific were studied with free-labeling to investigate emotion perception (reported in Sorenson, 1975; see Table 7). From 2008 onward, two additional studies were conducted, one using spontaneous facial configurations (Crivelli et al., 2017, Study 1) and the other using posed facial configurations (Gendron et al., 2018, Study 2). Overall, all five studies provide no evidence that the facial configurations in Figure 4 evolved to specifically express certain emotion categories. The three free-labeling studies reported in Sorenson (1975) produced variable results. The only replicable finding appears to be that participants labeled smiling facial configurations uniquely as happiness in all studies (as the only pleasant emotion category tested). The two newer free-labeling studies both indicated that participants only rarely spontaneously labeled facial configurations with the expected emotion labels (or their synonyms) above chance levels. Trobriand Islanders did not label the proposed facial configurations for happiness, sadness, anger, surprise or disgust with the expected emotion labels (or their synonyms) at above chance levels (although they did label the faces consistently with other words; Crivelli et al., 2017, Study 1). Hadza participants labeled smiling and scowling facial configurations at above chance levels as happiness (44%) and anger (65%), respectively (Gendron et al., 2018, Study 2). The word “anger” was not used to uniquely label scowling facial configurations, however, and was frequently applied to frowning, nose-wrinkled and gasping facial configurations.

Facial movements carry meaningful information, even if they do not reliably and specifically display internal emotional states.

The more recent studies of people living in small-scale, remote cultures suggest two interesting observations that are worthy of note. First, even though people may not routinely infer anger from scowls, sadness from frowns, and so on, they do reliably infer other social meanings for those facial configurations, because facial movements often carry important information about a person’s inner state, such as their social motives (Crivelli et al., 2016, 2017; Rychlowska et al., 2015; Wood et al., 2016; Yik & Russell, 1999; for a discussion, see Fridlund, 2017; Martin et al., 2017). For example, as we mentioned earlier, Trobriand Islanders consistently labeled wide-eyed gasping faces (the proposed expressive facial configuration for the fear category) as signaling an intent to attack (i.e., a threat; for additional evidence in carvings and masks in a variety of cultures, including Maori, !Kung Bushmen, Himba, Eipo, see Crivelli, Jarillo, & Fridlund, 2016, 2017).

Second, people do not always infer internal psychological states (emotions or otherwise) from facial movements. People who live in non-western cultural contexts, including Himba and Hadza participants, are more likely to assume that other people’s minds are not accessible to them, a phenomenon called opacity of mind in anthropology (Danziger, 2006; Robbins & Rumsey, 2008). Instead, facial movements are perceived as actions that predict future actions in certain situations (e.g., a wide-eyed gasping face is labeled as “looking” (Crivelli et al., 2017; Gendron et al., 2014a; Gendron et al. 2018). Similar observations were unavailable for the earlier studies conducted by Ekman, Friesen and Sorenson because, according to Sorenson (1975), they directed participants to provide emotion terms. When participants spontaneously offered an action label (e.g. “she is just looking”) or a social evaluation (e.g., “he is ugly”, or “he is stupid”), they were asked to provide an “affect term.” Findings like these suggest that there may be profound cultural variation in the type of inferences human perceivers make when looking at other human faces in general, an observation that has been raised by a number of anthropologists and historians.

A note on interpreting the data.

To properly interpret the scientific evidence, it’s crucial to consider the constraints placed on participants by the experimental tasks they are asked to complete, summarized in Table 5. In most urban and in some remote samples, experiments using choice-from-array tasks produce evidence supporting the common view: Participants reliably label scowling facial configurations as angry, smiling facial configurations as happy, and so on. (We don't yet know whether perceivers are uniquely labeling each facial configuration as a specific emotion because most studies don’t report that information.) It has been known for almost a century that choice-from-array tasks help participants obtain a level of reliability in their emotion perceptions that are not routinely seen in studies using methods that allow participants to respond more freely, and this is one reason they were chosen for use in the first place (for a discussion, see Gendron & Barrett, 2009, 2017; Russell, 1994; Widen & Russell, 2013). When participants are offered words for happiness, fear, surprise, anger, sadness, and disgust to register their inferences for a scowling facial configuration, they are prevented from judging a face as expressing other emotion categories (such as confusion or embarrassment), non-emotional mental states (e.g., a social motive, such as rejection or avoidance), or physical events (such as pain, illness or gas), thus inflating reliability rates within the task. When people are provided with other options, they routinely choose them. For example, participants label scowling faces as “determined” or “puzzled,” wide-eyed faces as “hopeful” and gasping faces as “pained” when they are provided with stories about those emotions rather than with stories of anger, surprise and fear (Carroll & Russell, 1996; also see Crivelli et al., 2017). The problem is not with the choice-from-array task per se – it is more with failing to consider alternative explanations for the observations in an experiment and therefore drawing unwarranted conclusions from the data.

Choice-from-array tasks may do more than just limit response options, making it difficult to disconfirm commonsense beliefs. The emotion words provided during the task may actually encourage people to see anger in scowls, sadness in pouts, and so on, or to learn associations between a word (such as “anger”) and a facial configuration (such as a scowl) during the experiment (e.g., Gendron et al., 2015; Hoemann, Crittenden, Ruark, Gendron, & Barrett, in press). The potency of words is discussed in Box 14, in SOM.

Summary.

The pattern of findings from the studies conducted with remote samples replicates and underscores the pattern observed in samples of participants from larger, more urban cultural contexts: Asking perceivers to infer an emotion by matching a facial configuration to an emotion word selected from a small array of options, or telling participants a brief story about a typical instance of an emotion category and asking them to pick a facial configuration from an array of two or three photos, generally inflates agreement rates, producing evidence that is more likely to support the hypothesis of reliable emotion perception when compared to data coming from less constrained response methods such as free labeling (see Table 4). This is particularly true for studies that include only one pleasant emotion category, i.e., happiness, where all foils differ from the target in valence, and therefore the robust reliability and specificity for inferring happiness from smiling in these studies may be the result of participants engaging in valence perception rather than emotion perception, per se. Studies that use less constrained tasks that are designed to more freely discover how people perceive emotion instead yield evidence that generally fails to find support for the common view. Less constrained studies suggest that perceivers infer more than one emotion category from the same facial configuration, infer the same emotion category in a variety of different configurations and often disagree about the set of emotion categories that they infer. Cultural variation in emotion perception is consistent with the variation we observed in the first section of this paper when we reviewed studies of emotional expression production (again, see Table 4), and is even consistent with the basic of face perception, which itself is determined by experience and cultural factors (Caldara, 2016).

Studies of Healthy Infants and Children

Some scientists concur with the common view that infants can read specific instances of emotion in faces from birth (Haviland & Lelwica, 1987; Izard, Woodburn & Finlon, 2010; Leppänen & Nelson, 2009; Walker-Andrews, 2005). However, it is difficult to ascertain whether infants and young children possess the various capacities required to perceive emotion per se: simply detecting and discriminating facial movements is not the same as categorizing them to infer their emotional meaning. This is because it is challenging to design well-controlled experiments that do a good job of distinguishing these two capacities. Infants are preverbal, so scientists use other measurement techniques, such as the amount of time an infant looks at a stimulus, to infer whether infants can discriminate one facial configuration from another, and ultimately, whether infants categorize those configurations as emotionally meaningful (for a brief explanation, see Box 15, in SOM). This approach introduces several possible confounds because of the stimuli used in the experiments: infants and children are typically shown photographs of the proposed expressive forms that are similar to those presented in Figure 4 (e.g., Leppanen et al, 2009; Peltola et al., 2008). Infants are more familiar with some of these configuration than with others (e.g., most infants are more familiar with smiling faces than with scowls or frowns) and familiarity is known to influence perception (see Box 15, in SOM), making it difficult to know which features of a face are holding an infant’s attention (familiarity or novelty) and which might be the basis of categorization in terms of emotional meaning. The configurations proposed for each emotion category also differ in their perceptual features (e.g., the proposed expressions for fear and surprise contain widened eyes whereas the proposed expression for sadness does not), contributing more ambiguity to the interpretation of findings. For example, when an infant discriminates smiling and scowling facial configurations, it is tempting to infer that that the child is discriminating expressions of anger and happiness when in fact that target of discrimination is the presence or absence of teeth in a photograph (Caron, Caron & Myers, 1985). Moreover, the facial configurations in question are usually made from exaggerated facial movements that are not typical of the expressive variation that children actually observe in their everyday lives (Grossman, 2010). Furthermore, unlike adults, infants may have had little or no experience with viewing photographs of anything, including heads of people with no bodies and no context.

The most important and pervasive confound in developmental studies of emotion perception is that most studies are not designed to test whether infants and children discriminate facial configurations according to their emotional meaning or whether they are discriminating affective features (pleasant vs. unpleasant; high arousal vs. low arousal) (see Box 9, SOM). Often, a facial configuration that is intended to depict a pleasant instance of emotion (smiling in happiness) is compared to one that is intended to depict an unpleasant instance of emotion (e.g., scowling in anger, frowning in sadness or gasping in fear), or these configurations are compared to a neutral face at rest (e.g., Leppänen, Richmond, Vogel-Farley, & Nelson, 2009; Leppänen, Moulson, Nelson & Vogel-Farley, 2007; Montague & Walker-Andrews, 2001). (This problem is similar to the one encountered earlier in our discussion of emotion perception studies in adults from small scale societies, in which perceptions of valence can be confused with perceptions of emotion categories). For example, in one study, 16-18 month olds preferred toys paired with smiling faces and avoided toys paired with scowling and gasping faces (Martin et al. 2014); this type of study cannot distinguish between whether infants are differentiating pleasant from unpleasant, approach vs. avoidance, or something about a specific emotion. Another study (Soken & Pick, 1999) reported that seven-month-olds distinguish sadness and anger when looking at faces, but only when the faces were paired with vocalizations. What is unclear is the extent to which the level of arousal or activation conveyed in the acoustic signals were most salient to infants. A recent study suggested that 10-month-old infants can differentiate between the high arousal, unpleasant scowling and nose-wrinkled facial configurations that are proposed as expressions of anger and disgust, suggesting that they can categorize these two facial configurations separately (Ruba et al., 2017). Yet, the scowling and nose-wrinkled facial configurations also differed in the properties besides their proposed emotional meaning: scowling faces showed no teeth, but nose-wrinkled faces were toothy, and it is well known that infants use perceptual features such as “toothiness” to categorize faces (see Caron et al., 1985). If an infant looks longer at a (pleasant) smiling facial configuration after viewing several (unpleasant) scowling faces, this does not necessarily mean that the infant has discriminated and understands “happiness” from “anger”; the infant might have discriminated positive from negative, affective from neutral, familiar from novel, the presence of teeth from the absence, less eye sclera from more, or even different amounts of contrast in the photographs. In the future, experiments must be designed to rule out the possibility that infants are categorizing facial configurations into different groupings based on factors other than emotion to provide a sound basis to infer that infants are processing specific emotional meaning.

As a consequence of these confounds, there is still much to learn about the developmental course of emotion perception abilities. By three months of age, infants can distinguish the facial features (the morphology) in the proposed expressive configurations for happiness, surprise, and anger, and, by seven months, they can discriminate the features in proposed expressive configurations for fear, sadness, and interest. Left uncertain is whether, beyond just discriminating between the mere appearance of particular facial features, infants also understand the emotional meaning that is typically inferred from those features. By seven months of age, infants can reliably infer whether someone is feeling pleasant or unpleasant when facial configurations are accompanied by sensory information from the voice (Flom & Bahrick, 2007; Walker-Andrews & Dickson, 1997). Only a handful of studies have attempted to test whether infants can infer emotional meaning in facial configurations rather than just discriminating between faces with different physical appearances, but they report conflicting results (Schwartz et al., 1985; Serrano et al., 1992). One promising future direction involves measuring the electrical signals (event related potentials, or ERPs) in infant brains as they view the proposed expressive configurations for anger and fear categories (e.g., Kobiella et al., 2008; Hoehl & Striano, 2008). Both of these studies reported differential brain responses to the proposed facial configurations for anger and fear, but their findings did not replicate one another (and for certain measurements, they observed opposing effects; for a broader review, see Grossmann, 2015).

Studies that measure a child’s ability to use an adult caregiver’s facial movements to resolve ambiguous or threatening situations, referred to as social referencing, have been interpreted as evidence of emotion perception in infants. One-year-olds use social referencing to stay in close physical proximity to a caregiver who is expressing negative affect, while infants are more likely to approach novel objects if the caregiver expresses positive affect (Carver & Vaccaro, 2007; Moses et al., 2001; Saarni et al., 2006). Similar results emerge from the caregiver’s tone of voice (Hertenstein & Campos, 2004; Mumme, Fernald, Herrera, 1996). In fact, by 14 months of age, the positive or negative tone of a caregiver’s voice influence what an infant will touch even more so than will a caregiver’s facial movements or the content of what the adult is actually saying (Valliant-Molina & Bahrick, 2012; Vaish & Striano, 2004). These studies clearly suggest that infants can infer the valenced meaning of facial movements, at least when made by live (as opposed to virtual) people who they are familiar with. But, again, these data do not help resolve what, if anything, infants infer about the emotional meaning of facial movements.

Learning to perceive emotions.

Children grow in emotionally rich social environments, making it difficult to run experiments that are capable of testing the common view of emotion perception while also taking into account the possible roles for learning and social experience. Nonetheless, several themes have emerged in the scientific literature, all of which suggest a clear role for learning and context in children’s developing emotion perception capacities.

One hypothesis that continues to be strongly supported by experiments is that children’s capacity to infer emotional meaning in facial movements depends on context (the conditions surrounding the face that may convey information about a face’s meaning). For example, emotion concept learning, as a potent source of internal context, shapes emotion perception capacity (discussed in Boxes 10 and 16 in SOM). There are also developmental changes in how people use context to shape their emotional inferences about facial movements. Children as young as 19 months old can detect facial movements that are emotionally incongruent with a context (Walle & Campos, 2014). For example, when presented with adult facial configurations that are placed on bodies posing an emotional context (e.g., a scowling facial configuration placed on a body holding a soiled diaper), children (aged four, eight, and twelve) moved their eyes back and forth between faces and bodies when deciding how to label the emotional meaning of the faces, whereas adult participants directed their gaze (and overt visual attention) to the face alone, judging its emotional meaning in a way that was independent of the bodily context (Leitzke & Pollak, 2016). The youngest children were equally likely to label the scene based on face or context. The results of this experiment suggest that younger children devote greater attention to contextual information and actively cross-reference facial and contextual cues, presumably to better learn about and understand the emotional meaning those cues.40

Another important source of context that shapes the development of emotion perception in children involves the broader environment in which children grow. Children who grow up in neglectful or abusive environments, where their emotional interactions with caregivers are highly atypical, have a different developmental trajectory than do those growing in more consistently nurturing environments (Bick & Nelson, 2016; Pollak, 2015). Parents from these high-risk families produce unclear or context-inconsistent expressions of emotion (Shackman et al., 2010). Neglected children (who do not receive sufficient social feedback) show delays in perceiving emotions in the ways that adults do (Camras et al., 2006; Pollak et al., 2000), whereas children who are physically abused learn to preferentially attend to and identify facial movements that are associated with threat, such as a scowling facial configuration (Briggs-Gowan et al., 2015; Cicchetti & Curtis, 2005; da Silva Ferreira, Crippa, & de Lima Osório, 2014; Pollak, Vardi, Putzer Bechner, & Curtin, 2005; Shackman & Pollak, 2014; Shackman, Shackman, & Pollak, 2007). Abused children require less perceptual information to infer anger in a scowling configuration (Pollak & Sinha, 2002) and more reliably track the trajectory of facial muscle activations that signal threat (Pollak, Messner, Kistler & Cohn, 2009). Children raised in physically abusive environments also more readily infer anger and threat in ambiguous facial configurations (Pollak & Kistler, 2002) and then require more effortful control to disengage their attention from signs of threat (Pollak & Tolley-Schell, 2003) when compared to children who have not been maltreated. This close attention to scowling faces with knitted eyebrows shapes how abused children understand what facial movements mean. For example, one study found that five-year-old abused children tended to believe that almost any kind of interpersonal situation could result in an adult becoming angry; by contrast, most non-abused children understand that anger is likely in particular interpersonal circumstances (Perlman et al., 2008).

By three years of age, North American children not only start to show reliability in their emotion perceptions but they also begin to show evidence of specificity. They understand that facial movements do not necessarily map on to emotional states, and how someone really feels can be faked or masked. Moreover, they know what facial movements are expected in a particular context and try to produce them despite their feelings. For example, the “disappointing gift” experiments developed by psychologist Pamela Cole and her colleagues demonstrate this well. In one study, preschool-aged children were told they would be rewarded with a gift after they completed a task. Later, children received a beautifully wrapped package that contained a disappointing item, such as a broken pair of cheap sunglasses. When facing a smiling unfamiliar adult who has presented them with a gift, children forced themselves to smile (lip corner pull, cheek raise, and brow raise) and to thank the experimenter. Yet, while the children were smiling, they often kept their eyes focused, down, slumped their shoulders, and made negative statements about the object, indicating that they did not, in fact, feel positive about the situation (Cole, 1986). Moreover, there was no difference in the behavioral responses of visually impaired children when receiving a disappointing gift (Cole, Jenkins, & Shott, 1989). Studies like this one provide a more implicit way of assessing children’s knowledge about emotion perception (i.e., it illustrates the inferences that children expect others to make from their own facial movements).

Summary.

There is currently no clear evidence to support the hypothesis that infants and young children reliability and specifically infer emotion in the proposed expressive configurations for anger, disgust, fear, happiness, sadness and surprise categories (presented in Figure 4; findings summarized in Table 4). A more plausible interpretation of the existing evidence is that young infants infer affective meaning such as valence and arousal in facial configurations. Data from infants and young children obtained using a variety of methods further suggests that emotion perception abilities emerge and are shaped through learning in a social environment. These findings are consistent with evidence that the human face may be evolutionarily privileged to communicate importance or salience. But it is not clear that the expressive configurations proposed for specific emotion categories are similarly privileged in this way.

Summary of Scientific Evidence on the Perception of Emotion in Faces

The scientific findings on perception studies generally replicate those from production studies in failing to strongly support the common view. The one exception to this overall pattern of findings is seen in studies that ask participants match a posed face to an emotion word or scenario. This method produces evidence to support the common view, even when it is applied to completely novel emotion categories with made up expressive cues, opening up interesting questions about the psychological potency of the elements that make up choice-from-array designs (such as the emotion words embedded in the task or the choice of foils on a given trial). These findings reinforce our earlier conclusion that terms like “facial configuration” or “pattern of facial movements” or even “facial actions” are preferred to more loaded terms like “emotional facial expression,” “emotional expression” or “emotional display,” which can be, at best misleading, and at worst, incorrect.

Summary and Recommendations

Evaluation of the Empirical Evidence

The common view that humans around the world reliably produce and recognize certain emotions in specific configurations of facial movements continues to echo within the science of emotion, even as scientists increasingly acknowledge that anger, sadness, happiness and other emotion categories are more variable in their facial expressions. This entrenched common view does more than guide the practice of science. It influences public understanding of emotion, and hence education, clinical practice, and applications in industry. Indeed, it reaches into almost every facet of modern life, including emoticons and movies. Nonetheless, there is insufficient evidence to support it. People do express instances of anger, disgust, fear, happiness, sadness and surprise with the hypothesized facial configurations presented in Figure 4 at above chance levels, suggesting that those facial configurations sometimes serve as expressions of emotion as proposed. However, the reliability of this finding is weak, and there is evidence that the strength of support for the common view varies systematically with the research methods used. The strongest support for the common view -- found in data from urban, industrialized or developed samples completing choice-from-array tasks -- does not show robust generalizability. Evidence for specificity is lacking in almost all research domains. A summary of the scientific evidence is presented in Table 4.

The research findings do not imply that people move their faces randomly or that the configurations in Figure 4 have no psychological meaning. Instead, they reveal that the facial configurations in question are not “fingerprints” or diagnostic displays that reliability and specifically signal particular emotional states regardless of context, person and culture. It is not possible to confidently infer happiness from just a smile, anger from a scowl, or sadness from a frown, as numerous technology tries to do when applying what they mistakenly believe to be the scientific facts.

Instead, the available evidence from different populations and research domains – infants and children, adults living in industrialized countries and in remote cultures, and even individuals who are congenitally blind -- overwhelmingly points to a different conclusion: when facial movements do express emotional states, they are considerably more variable and dependent on context than the common beliefs allows. There appear to be many-to-many mappings between facial configurations and emotion categories (e.g., anger is expressed with a broader range of facial movements than just a scowl and scowls express more than anger). A scowling facial configuration may be an expression of anger in the sense of being a part of anger in a given instance. But a scowling facial configuration is not the expression of anger in any generalizable or universal way. Scowling facial configurations and the others in Figure 4 belong to a much large repertoire of facial movements that express more than one emotion category, and also non-emotional inner states, in a way that is tailored to specific situations and cultural contexts. The face is a powerful tool for social communication (Jack & Schyns, 2017). Facial movements, like reflexive and voluntary motor movements (Barrett & Finlay, in press), are strongly context-dependent. Recent evidence suggests the people’s categories for emotions are flexible and responsive to the types and frequencies of facial movements they are exposed to in their environments (Plate, Wood, Woodard, & Pollak, in press).

The degree of variation suggested by the published evidence also goes well beyond the hypothesis that the facial configurations in Figure 4 are prototypes or typical expressions, and that any observed variation are merely the result of cultural accents, display rules, suppression or other regulatory strategies, differences in induction methods, measurement error, or stochastic noise (as proposed by various scientists, including Elfenbein, 2013, 2017; Ekman & Cordaro, 2011; Levenson, 2011; Matsumoto, 1990; Roseman, 2011; Tracy & Randles, 2011). Instead, the facial configurations in Figure 4 are best thought of as Western gestures, symbols or stereotypes that fail to capture the rich variety with which people spontaneously move their faces to express emotions in everyday life. A stereotype is not a prototype. The distinction is an important one, because a prototype is the most frequent or typical instance of a category (Murphy, 2002), whereas a stereotype is an oversimplified belief that is taken as generally more applicable than it actually is.

The conclusion that emotional expressions are more variable and context-dependent than commonly assumed is also mirrored by the evidence from physiological changes (such as heart rate and skin conductance measures, Box 8, SOM) and even in evidence on the brain basis of emotion (Clark-Polner et al., 2017). The task of science is to systematically document these context-dependent patterns, as well as understand the mechanisms that cause them, so that we can explain and predict them. Clearly, the face is a rich source of information that plays a crucial role in guiding social interaction. Facial movements, when measured in a high dimensional dynamic context, may serve the diagnostic purpose that many consumers of emotion science are looking for (where context can be a cultural context, a specific situation, a person’s learning history or momentary physiological state, or even the temporal context of what just took place a moment ago; Barrett et al., 2011; Gendron et al., 2013).

A Note on the Scientific Literature

Our review identified several broad problems that lurk within the scientific research on facial expressions and that may cause considerable misunderstanding and confusion for consumers of this research. First, statistical standards are commonly adopted that don’t translate well for applying emotion research to other domains, applied or scientific. Showing that people frown when sad or scowl when angry with greater statistical reliability than would be expected by chance may be a scientific finding that warrants publication in a peer-reviewed journal, but above-chance responding is often low in absolute terms, making broad conclusions impossible, particularly for translation to domains of life where a person’s outcomes can be influenced by what emotional meaning perceivers infer. Making inferences based on statistical reliability without concern for specificity and generalizability is similarly problematic. Second, even studies that surmount these common shortcomings often have a mismatch between what is claimed in their conclusions (or in what others claim in reviews or citations of those primary research papers), and what inferences can, in fact, be supported by the results. Third, and relatedly, this mismatch often results from problems in how studies are designed—the particular stimuli used, the tasks used, and the statistical analyses are critically important and constrain what can be observed and inferred in the first place. Fourth, published research on emotional expressions and emotion perception often confounds the measurements made in an experiment with the interpretation of the data, referring without sufficient justification to facial movements as “emotional displays,” “emotional expressions” or even “facial expressions,” rather than “facial configurations,” “facial movements” or “facial actions”; referring to people “detecting” or “recognizing” emotion rather than “perceiving” or “inferring” an emotional state based on some set of cues (facial movements, vocal acoustics, body posture, etc.); and referring to “accuracy” rather than “agreement” or “consensus.”

A Note on Other Emotion Categories

Our conclusions most directly challenge what we have termed the “common view”: that a scowling facial configuration is the expression of anger, a nose-wrinkled facial configuration the expression of disgust, a gasping facial configuration the expression of fear, a smiling facial configuration the expression of happiness, a frowning facial configuration the expression of sadness, and that a startled facial configuration is the expression of surprise. By necessity, we focused on our review of evidence on these six emotion categories, rather than the more than twenty emotion categories that are currently being studied, because studies on these six are far more numerous than for other emotion categories. Nonetheless, some scientists claim that these other emotion categories each have distinctive, universal expressions, facial or otherwise, that is modified or accented by culture (e.g., Cordaro et al., 2017; Keltner et al., in press). In our view, such claims rest on evidence that is subject to the same critique as we offered for the research that we reviewed in detail here. In short, even though our review focused on the six emotion categories that are sometimes referred to as “basic emotions,” our observations and conclusions generalize to studies of other emotion categories that use similar methods.

Recommendations for Consumers of Emotion Research on Applying the Scientific Findings

Presently, many consumers of emotion research assume that certain questions about emotional expressions have been answered satisfactorily when in fact this is not the case. Technology companies, for example, are spending millions of research dollars to build devices to read emotions from faces, erroneously taking the common view as the one that is scientifically best supported. A more accurate description, however, is that their technology detects facial movements, not emotional expressions.41 Corporations like Amazon are exploring virtual human technology to interface with consumers. Virtual humans are used to educate children, train physicians, train the military as well as infer psychological disorders and perhaps eventually even be used to offer treatments. At the moment, the science of emotion is ill-equipped to support these initiatives. Emotional expressions are more variable and context-dependent than originally assumed, and most of the published research was not designed to probe this variation and characterize this context-dependence. As a consequence, right now, the scientific evidence offers less actionable guidance to consumers than is commonly assumed.

In fact, our review of the scientific evidence indicates that very little about how and why certain facial movements express instances of emotion is actually known at a level of detail that such conclusions could be used in important, real-world applications. To help consumers navigate the science of emotion, we offer some tips for how to read experiments and other scientific papers (Table 8).

Table 8:

Recommendations for reading scientific studies about emotion

1. Take note of whether an experiment is studying expressive stereotypes or more variable facial movements.
2. Take note of data on specificity and generalizability; do not focus solely on reliability at above chance levels.
3. Make a distinction between the data in an experiment (what was measured) and how those data are interpreted.
4. Translate “emotional expressions” or “emotional displays” into “facial movements.”
5. Translate “emotion recognition” into “emotion perception” or “emotion inference.”
6. Translate “accuracy” to “agreement,” “consensus” or “reliability.”
7. Give more weight to studies that measure facial movements or study the perception of facial movements made in more naturalistic settings.
8. Take note of studies that measure or manipulate context.
9. Field studies of people from small-scale, remote cultures are often less well-controlled than studies conducted in the laboratory, but they are invaluable in the information that they provide and should be valued.
10. Remember that emotions are not understood as internal states in all cultures. In some cultures they are understood as situated actions.
11. Do not skip the method and results sections and skip to the discussion to learn the results of an experiment. It is important to know what was measured and observed, not just how scientists interpret their measurements.

More generally, companies may well be fundamentally asking the wrong question. Attempts to simply “read out” people’s internal states from an analysis of their facial movements alone, without considering various aspects of context are at best incomplete, and at worst entirely lack validity, no matter how sophisticated the computational algorithms. These technology developments are powerful tools to investigate the expression and perception of emotions, as we discuss below. Right now, however, it is premature to use this technology to reach conclusions about what people feel based on their facial movements--which brings us to recommendations for future research.

Recommendations for Future Scientific Research

Specific, concrete recommendations for future research to capitalize on the opportunity offered by current challenges can be found in Table 9, but we highlight a few general points here. Foremost, the expressive stereotypes that summarize the common view, like those depicted in Figure 4, are ubiquitous in published research, but it’s time to move beyond a science of stereotypes to develop a science of how people actually move their faces to express emotion and the processes by which those movements carry information about emotion to someone else (a perceiver). (See Box 16 in SOM for a discussion of information theory as applied to emotional communication). The stereotypes of Figure 4 must be replaced by a thriving scientific effort to observe and describe the lexicon of context-sensitive ways in which people move their facial muscles to express emotion, and the discovery of when and how people infer emotions in other people’s facial movements.

Table 9:

Recommendations for future research

General Recommendations
▪ Take chances on studies that attempt to go beyond merely supporting traditional views of emotion.
▪ Support papers that attempt to study facial movements in real life, measuring context, sampling across cultures even though these studies are often less well controlled than studies in the laboratory, or may use facial stimuli that are less familiar to reviewers than canonical stimulus sets.
▪ Prioritize multidisciplinary studies that combine classical psychology methods with cognitive neuroscience, machine learning, etc.
▪ Support larger scale studies that bridge the lab and the world, that study individual people across many contexts, and measure emotional episodes in high dimensional detail, including physical, psychological and social features; encourage multiple investigators with different areas of expertise to work together.
▪ Support the development of computational approaches.
▪ Create R&D teams that pair psychologists and cognitive scientists trained in the psychology of emotion with engineers and computer scientists.
▪ Increase opportunities to test innovative methods and novel hypotheses, with the acknowledgement that such approaches are likely to elicit resistance from established scientists in the field of emotion.
▪ Generate more studies to identify the underlying neural mechanisms of the production and perception of facial movements.
▪ Direct funding to thornier but necessary new questions and be critical of projects that perpetuate past errors in emotion research.
▪ Direct healthy skepticism to tests, measures, and interventions that rely upon assumptions about “reading facial expressions of emotion” that seem to ignore published evidence and/or ignore integration of contextual information along with facial cues.
▪ Develop systematic, precise ways to describe and/or manipulate the dynamics of specific facial actions.
Stimulus Selection Recommendations
Limitations in stimulus selection can bias results. ▪ For perception studies, incorporate images from the wild (e.g., from multiple internet sources) to capture the full range of facial movements that humans produce in their everyday lives.

▪ For both production studies (where stimuli are designated to evoke emotion) and perception studies, build variation into stimulus sets so conclusions about emotion categories are not inferred (or evoked) from limited stimuli. Consider randomly sampling a variety of stimuli for a given category and treating stimuli as a random variable.

▪ For production studies, ensure that multiple stimuli per emotion category are used to evoke an emotion.
Little is known about the dynamics of the production and perception of emotion signaling. ▪ For perception studies, use dynamic images rather than rely on still images. For production studies, code the temporal dynamics of facial movements.

▪ Attempt to determine full dynamics and the apex of an emotion signal, changes to AUs as signals emerge and recede, and whether the kinematics of distinct AUs are similar or different across sequences or phases of emotion signaling.

▪ Ensure sufficient temporal resolution to allow for event segmentation to be assessed in perception studies.
The role of context is hotly debated, but rarely measured. ▪ Manipulate (or at least measure) the context in which target stimuli are perceived to evaluate whether data are truly stimulus-specific or influenced by context features.

▪ Describe in a systematic way the differences in context, whether for production or perception studies. Theories about the effects of context cannot be resolved until we address how to measure and quantify context.
Sample Selection Recommendations
Cross-cultural studies can provide powerful insights, but are limited in number and scope. ▪ Quantify, as best as possible, participants’ degree of exposure to the west, as well as the amount and type of formal schooling made available to participants.

▪ Harness technology to collect larger numbers of images and video sequences of facial movements across cultures. Use unlabeled classification approaches to discover emotion categories and their expressive forms, rather than continuing to ask whether other cultures are similar to the US. Remember that emotions and mental inferences may be understood differently in different cultures.
Task and Method Design Recommendations
Measurement versus interpretation of emotion is often blurred in research studies. ▪ Contrast more than one “emotion” category with a baseline, so that conclusions about a specific emotion category are not drawn from a comparison of an emotion versus a no emotion condition.

▪ Compare multiple emotion categories to non-emotion categories in a given study.
New insights about emotion are constrained by reliance on, and assumptions about, traditional categories. ▪ Measure emotional episodes in a multimodal way and attempt to discover explicit criteria for when an emotion is present or absent. Such discovery may require within-person approaches.

▪ Sample broader categories of possible emotion states than the limited categories used in prior research (move beyond categories such happiness, anger, sadness, fear, etc.). Test for variations in intensity within these categories and similarity across categories.

▪ Unless a study design is completely data-driven, explicitly state the theoretical priors of the research team. The distinction is between whether you are seeking to discover versus verify emotion categories. Both approaches are valid, but should be clearly articulated.
Data Analysis Recommendations
Findings are limited by a failure to consider issues related to forward and reverse inference. ▪ Address issues of reliability and specificity when presenting data on emotion expression and emotion perception.

▪ Use formal signal detection analytics and information theoretic measures rather relying on frequency or levels of agreement. Consider using Bayesian methods so that the null hypothesis can be tested directly.

New research on emotion should consider sampling individuals deeply, with high dimensional measurements, across many different situations, times of day, etc.: a big data approach to learn the expressive repertoires of individual people. In the ideal case, videos of people in natural situations could be quantified by automated algorithms for various physical features such as facial movements, posture, gait, and tone of voice. To this we could add the sampling of other physical features such as ambulatory monitoring of autonomic nervous system changes to sample the internal milieu of people’s bodies as they dynamically change over time, ambulatory eye-tracking to assess gaze and attention, ambulatory brain imaging such as EEG and optical brain imaging (fNIRs). The failure to find reliable “fingerprints” for emotion categories stems, at least in part, from the same reason there are no reliable facial movements to express these categories: approaches have ignored meaningful variability due to context. There is also blue tooth technology to capture the physical spaces people inhabit (which can be quantified for various structural and social descriptive features such as how much light and noise they are exposed to), whether they are with another person, how that person reacts, and so on. Rich, multimodal observations could, in principle, be available from videos, which when time-synched with the other physical measurements, could be extremely useful in understanding the conditions for when certain facial movements are made and what they might mean in a given context. Naturally, big data in the absence of hypotheses is not necessarily helpful.

People could be offered the opportunity to annotate their videos with subjective ratings of the features that describe their experiences (whether or not they are identified as emotions). Candidate features are affective properties such as valence and arousal (see Box 9 in SOM). The features might also be appraisals as descriptions of how a situation is experienced (Barrett, Mesquita et al., 2007) and have the potential to add to the high dimensional characterization of what causes facial movements and what they mean.42 Such an approach introduces various technical and modeling challenges, but this sort of deeply inductive approach is now within reach.

Another opportunity for high dimensional sampling involves interactions with virtual humans. Because virtual humans can realize contingent behavior in rich social interactions under strict and precise experimental control, they can provide a richer, more natural context in which to study emotional expressions and emotion perception than is true for traditional laboratory studies, while not losing the experimental control that limits the causal inferences from ethological studies.

To date, this potential has not been exploited to explore the reliability and specificity in context-sensitive relations between facial movements and mental states. As we noted earlier, most of the systems are now designed to teach people a variety of skills, where the goal is not to assess how well participants perceive emotions in facial movements under realistic, socially ambiguous conditions, but instead to program expressive behaviors into virtual humans that will motivate people to learn the needed skills. In these experiments, the psychological realism of facial movements is often secondary to the primary goals of the experiment. A scientist might even program a virtual human with behavior or appearance that is un-natural or infeasible for a human (i.e., that are supernormal) so that a participant can unambiguously interpret and be influenced by the agent’s actions (Tinbergen, 1953; D. Barrett, 2007).

Nonetheless, the scientific approach of observing people as they interact with artificial humans holds great promise for understanding the dynamics and mechanisms of emotion perception and may get us closer to understanding human emotion perception in everyday life. Virtual humans are vivid. Unlike more passive approaches to evoking emotion such as viewing videos or images of facial configurations, a virtual human engages a human participant in a direct, social interaction to elicit perceptual judgments that are either directly reported or inferred from behaviors measured in the participant. Virtual humans are also highly controllable, allowing for more precise experimentation (Blascovich et al., 2002). A virtual human’s facial movements and other details can be repeated across participants offering the potential for robust and replicable observations. Numerous studies have demonstrated that humans are influenced by them (e.g., Baylor & Kim, 2008; Krumhuber et al, 2007; McCall et al., 2009). For example, human learners are more engaged by virtual agents who move their faces (and modulate their voices), leading them (the real humans) to increased sense of self-efficacy (Kim, Baylor, & Shen, 2007). As a consequence, virtual humans potentially “allow for the study of emotion in a rich virtual ecology, a form of synthetic in vivo experimentation” (Marsella & Gratch, 2016). When combined with the high dimensional sampling we described earlier, there is the potential to revolutionize our understanding of emotional expressions by asking different questions than those encouraged by common views. Automated algorithms using data captured from videos offer substantial improvements with a data-driven, unsupervised approach. The result could be the robust descriptions about the context-sensitive nature of emotional expressions that is currently missing, and that would set the stage for a more mechanistic, causal account of emotions and their expressions.

An ethology of emotions and their expressions can also be pursued in the lab. Experiments can go beyond a study of how people move their faces in a single situation chosen to be most typical of a given emotion category. Most studies to date have been designed to observe facial movements in only the most typical situations. Future studies should examine emotional expression and perception across a range of situations that vary systematically in their physical, psychological, and social features, and aim to understand both the various ways that humans acquire the skills to express and perceive emotion, as well as the conditions that can impair the development of these processes.

The shift towards more context-sensitive scientific studies of emotion has already begun (see Box 3 in SOM), but it currently falls short of what we are recommending. Non-scientists (and some scientists) still anchor on the common view and only slowly shift away from it (Tversky & Kahneman, 1974; Wilson et al., 1996). The pervasiveness of the common view supports strong convictions about what it is that faces signal, and people often continue to hold to those convictions even when they are demonstrably wrong (Barrett, 2017a; Todorov, 2017). Such convictions reflect cultural beliefs and stereotypes, however. This state of affairs is not unique to the science of emotional expression or to the science of emotion more generally (Kuhn, 1962).

In our view, the scientific path forward begins with the explicit acknowledgement that we know much less than we thought we did, providing an opportunity to cultivate the spirit of discovery with renewed vigor and take scientific discovery in a new direction (Firestein, 2016). With this context of discovery comes the sobering realization that those of us who cultivate the science of emotion and the consumers who use this research should seriously question the assumptions of the common view and step back from what we thought we knew about reading emotions in faces. Understanding how best to infer someone’s emotional state or predict someone’s future actions from their facial movements awaits the outcomes of future research.

Supplementary Material

1

Acknowledgements

We thank Jose-Miguel Fernandez-Dols and James Russell for providing us with a copy of their edited volume before its publication. This article benefited from specific discussions with Linda Camras, Pamela Cole, Maria Gendron, Alan Fridlund, and Tuan Le Mau. We are grateful to Jennifer Fugate for her assistance with constructing Figure 4, and in particular for her FACS coding the facial configurations presented in Figure 4. We also send thanks to Sheri Widen who provided assistance on Box Figure 10-1. Many thanks also to Linda Camras, Vanessa Castro, Carlos Crivelli and David Cordaro for providing us with additional details of their published experiments. And thanks to Jeff Cohn for his guidance on acceptable inter-rater reliability levels for FACS coding. We are also deeply grateful to those friends and colleagues who invested their time and efforts in commenting on an earlier draft of this paper (although the summaries of published works and conclusions drawn from them reflect only the views of the authors): Linda Camras, Maria Gendron, Rachael Jack, Judy Hall, Ajay Satpute, and Batja Mesquita. The paper was supported by grants to from the U.S. Army Research Institute for the Behavioral and Social Sciences (W911NF-16-1-019), the National Cancer Institute (U01 CA193632) and the National Institute of Mental Health (R01 MH113234 and R01 MH109464) to LFB; the National Science Foundation (CMMI 1638234) to LFB and SM; NIMH grant 2P50MH094258 to RA; the National Institutes of Health (R01-DC-014498, R01-EY-020834) and the Human Frontier Science Program (RGP0036/2016) to AMM; NEI grant R56 EY020834 to AMM and LFB; the National Institute of Mental Health (R01 MH61285), the National Institute of Child Health and Human Development (U54 HD090256) and a James McKeen Cattell Fund Fellowship to SDP; and the Air Force Office of Scientific Research (FA9550-14-1-0364) to SM. The views, opinions, and/or findings contained in this paper are those of the authors and shall not be construed as an official U.S. Department of the Army position, policy, or decision, unless so designated by other documents.

Glossary

Accuracy

Extent to which a participant’s performance corresponds to the intended performance on an experimental task. Critically, this requires proper experimental task design, so that the intended correct performance is perceiver-independent, and not subject to the whims of the experimenter.

Affect

A general property of experience that has at least two features: pleasantness or unpleasantness (valence) and degree of arousal. Affect is part of every waking moment of life and is not specific to instances of emotion, although all emotional experiences have affect at their core.

Appraisal

Scientists use the word “appraisal” either to describe how a situation is experienced (e.g., a situation is experienced as novel) or to refer to a literal cognitive mechanism that causes those features of experience (e.g., an evaluation or judgment of whether or not a situation is novel).

Approach/avoidance

A fundamental dimension of motivated behavior. It is different from valence, which is a dimension of experience rather than of behavior.

Category/Categorization

The psychological grouping of a collection of objects, people or events that are perceived to be similar in some way. May be done consciously or unconsciously. May be explicit (as when applying a verbal label to instances of the grouping) or implicit (treating instances the same way or behaving towards them in the same way).

Choice-from-array tasks

Any judgment task that asks research participants to pick a correct answer from a small selection of options provided by the experimenter. For example, in the study of emotion perception, participants are often shown a posed facial configuration depicting an emotional expression (e.g., a scowl), along with a small selection of emotion words (e.g., “angry,” “sad,” “happy”) and asked to pick the word that best describes the face.

Common view

In this paper, the most predominant view about how emotions are related to facial movements. While difficult to quantify, we characterize it through examples, e.g., an internet Google search (Box 1, SOM). The common view holds that (a) certain emotions categories reliably cause specific patterns of facial muscle movements, and (b) specific configurations of facial muscle movements are diagnostic of certain emotions categories. See Figure 4.

Conditional probability

The probability that an event “X’ will occur given that another event “Y” has already occurred, or p(X/Y). If “X” is a frown and “Y” is sadness, then p(frown/sadness) is the conditional probability that a person will frown when sad. See also forward inference, reverse inference.

Configural (vs featural) perception of a face

The visual analysis of something, like a face, that is holistic, meaning that the face is visually analyzed as a gestalt (or whole unit) that incorporates features and their relations. Featural processing means that individual features are perceived independently, without reference to one another.

Confirmation bias

The tendency to search for, remember, or believe evidence that is consistent with one's existing beliefs or theories, in favor of evidence inconsistent with one’s beliefs or theories.

Congenitally blind

People who are born without vision. In the literature, there is considerable heterogeneity, with some people being truly blind from the moment they are born, but others having severe visual impairments short of complete blindness or becoming blind in infancy. If the cause is peripheral (in the eyes rather than the brain), such individuals may still be able to think and imagine very similarly to sighted individuals.

Consistency

An outcome that does not vary greatly across time, context, or different individuals (see forward inference). Consistency is not accuracy (a group of people can consistently believe something that is wrong).

Discrimination

In psychophysics, to judge that two stimuli are different from one another; separate from identifying what they are (identification) or what they mean (recognition).

Ecological validity

Refers to the extent to which the findings of a research study are able to be generalized to real-life settings; the extent to which an experimental protocol captures valid aspects of the real world (related to in-the-wild).

Emotional episode

A window of time during which there is an emotional instance. Often, but not always, accompanied by an experience of emotion, and sometimes, but not always, involves an emotional expression.

Emotional expression

a facial configuration, bodily movement, or vocal expression that reliability and specifically communicates an emotional state. Many so-called emotional expressions are in fact errors of reverse inference on the part of perceivers (e.g., an actor crying when not sad).

Emotional granularity

experiencing or perceiving emotions according to many different categories (e.g., low granularity = angry, sad, and afraid are all synonyms of “unpleasant;” high granularity = “frustration,” “irritation,” and “rage” are all distinct from one another and from “anger”.

Emotional instance (or instance of emotion)

An event categorized as an emotion. For example, an instance of anger is the categorization of an emotional episode of anger. In cognitive science, an instance is called a “token” and the category is called a “type”. So, an instance of anger is a token of the category anger. (see Emotional episode).

Face inferiority effect

A phenomenon observed in emotion perception studies of toddlers and young children. They have difficulty inferring the causes for emotions depicted in facial movements alone when compared to inferring the causes of emotions depicted with stories or words.

Facial affect coding system (FACS)

A system to describe and quantify visible human facial movements.

Facial configuration

A pattern of visible contractions of multiple muscles in the face; the production analog to configural perception of faces. Configurations can be described objectively (e.g., with FACS coding). Not synonymous with “facial expression”, which requires an inference about how the facial configurations were caused.

Facial expression

A facial configuration that someone infers is expressing an internal state. Facial expressions of emotion are configurations that perceivers reverse infer to have been caused by an internal emotion state; they are thus perceiver-dependent.

Facial movement

A facial configuration that is objectively described in a perceiver-independent way. This description is agnostic about whether the movement expresses an emotion and does not use reverse inference. FACS coding is an example.

Forced-choice task

An experimental task in which a participant must choose between options provided by the experimenter.

Forward inference

Inferring an effect from knowing its cause. An example would be the conditional probability of observing a frown given we know somebody is angry, p(frown∣anger).

Free labeling

An experimental task that is not forced-choice, but in which the participant generates words of her/his choosing.

Generalization

The replication of research findings across different settings, samples, or methods. Generalizability can be weak, for instance if a finding replicates to a limited extent, or strong, if it replicates across very different methods and cultures.

Habituation task

In a habituation task, infants are repeatedly shown objects or images that belong to the same category. When subsequently shown a novel stimulus (one that is not experienced as similar to the others), infants look longer at it. Used to infer how infants categorize stimuli.

In the wild

In the real world (vs. in the lab). Related to ecological validity.

In-group advantage

In sociology and social psychology, an in-group is the social group of which a person psychologically feels they are a member; typically, people have more visual experience and familiarity with in-group members. In-group advantage refers to the often superior ability to perceive faces or voices from one’s in-group, as compared to from an out-group.

Mental inference/mentalizing

Assigning a mental cause to actions; also sometimes referred to as “theory of mind”. The reverse inference of attributing emotions from seeing facial movements is an example of mentalizing.

Meta-analysis

A method for statistically combining findings from many studies.

Multimodal

Combining information from more than one of the senses (e.g., vision and audition).

Null hypothesis

The hypothesis or default position that there is no relationship between dependent and independent variables. The probability of observing results that support the null hypothesis is chance level, i.e. what would obtain if observations are random, or permuted. Consequently, if the null hypothesis is true, the distribution of p-values is uniform (every possible outcome has an equal chance).

Perceiver-dependent

Interpretation of an observation that depends on human judgment. Perceiver dependency can produce conclusions that are consistent across people but not accurate or valid.

Perceiver-independent

An observation that does not depend on human judgment. Although some philosophers argue that all observations require some human judgment, there are degrees of dependency. Judging that a flower vase is rectangular or oval is relatively perceiver-independent, whereas judging whether it looks nice is perceiver-dependent.

Percent agreement

A measure of agreement between raters; high agreement produces high inter-subject consistency. Percent agreement is not the same as percent accuracy, since the former is more perceiver-dependent than the latter.

Perceptual matching task

An experimental task that requires research participants to judge two stimuli, such as two facial configurations, as similar or different. This only requires discrimination, not categorization, recognition, or naming.

Priors

Background beliefs. In the context of Bayes's Theorem, the belief that a hypothesis is true depends not just on the evidence presented but also on the strength of prior beliefs. If a person has a strong prior, this may result in a confirmation bias.

Prototype

The most frequent or most typical instance of a category. Distinct from stereotype: A group of people may have a perceiver-dependent stereotype that is an inaccurate representation of the prototype.

Recognition

Acknowledging something’s existence (which is confirmed to exist by perceiver-independent means). Contrast with perception (which involves inference and interpretation).

Replication

The extent to which new experiments come to the same conclusions as a previous study. Strong replications generalize well: similar conclusions are obtained even when the new experiments use different subject samples, stimuli, or contexts.

Reverse correlation

A psychophysical, data-driven technique for deriving a representation of something (e.g., an image of a facial configuration) by averaging across a large number of judgments.

Reverse inference

Inferring a cause from having observed its purported effect. For instance, inferring that a scowl means someone is angry (the conditional probability, p(anger∣frown)). In general, reverse inference is poorly constrained, since multiple causes are usually compatible with any observation.

Sensory modalities

The different senses: vision, hearing, etc.

Specificity

Research conclusions that include positive as well negative statements. For instance, concluding that a frown signals anger but that other facial movements do not signal anger, and that a frown does not signal emotions other than anger. High specificity helps make reverse inference valid. Ideally, research conclusions feature both high specificity for some domains, and high generalizability for others (e.g., that a frown signals only anger, but does so across all people and cultures).

Statistical learning

Detecting statistical regularities from an environment; learning to recognize patterns.

Stereotype

a widely held but inaccurate belief about a person or category.

Universal

Something that is common or shared by all humans. The source of this commonality (innate or learned) is a separate issue. If an effect is universal, it generalizes across cultures.

Validity

Whether an observed variable actually measures what is claimed. E.g., whether a facial movement indicates an emotion (construct validity), or is specific for a particular emotion (discriminative validity).

Biography

Lisa Feldman Barrett is University Distinguished Professor of Psychology and Director of the Interdisciplinary Affective Science Laboratory (IASLab) at Northeastern University, with research appointments at Harvard Medical School and Massachusetts General Hospital. Her research focuses on the nature of emotion from both psychological and neuroscience perspectives. Dr. Barrett is the recipient of numerous research awards, including the 2018 APS Lifetime Mentor Award and the NIH Director’s Pioneer Award for transformative research. She is an elected fellow of the American Academy of Arts and Sciences, the Royal Society of Canada, the American Association for the Advancement of Science, the Association for Psychological Science, as well as several other scientific societies. She was recently chosen one of the 50 most influential living psychologists in the world by thebestschools.org. Her research has been continuously funded by NIH and NSF for over 20 years. Dr. Barrett also educates lawyers, judges and other legal actors about emotion, neuroscience and the law as part of her work for the Center for Law, Brain and Behavior. In addition to publishing over 200 peer reviewed papers and 50 book chapters, Dr. Barrett has testified before US Congress in support of basic behavioral research funding and has edited five volumes, including the 4th edition of the Handbook of Emotion, published by Guilford Press. Her book, How emotions are made: The secret life of the brain, is published by Houghton Mifflin Harcourt. More about Dr. Barrett’s work can be found at www.affective-science.org and https://lisafeldmanbarrett.com/

Ralph Adolphs is the Bren Chair Professor of Psychology, Neuroscience and Biology at Caltech in Pasadena, California. His graduate studies were in basic cellular electrophysiology (the auditory system of the barn owl), but his post-doctoral work switched to cognitive neuroscience in humans (cognitive neuropsychology with Antonio Damasio at the University of Iowa). His current lab (emotion.caltech.edu) studies emotion and social behavior using a number of approaches, including electrophysiological recordings from the brain in surgical patients, studies of people with autism, and fMRI. He directs the Caltech Brain Imaging Center and chairs the Institutional Review Board at Caltech (which reviews the ethics of human subjects research). He is the co-author (with David Anderson, a Caltech colleague) of the book, "The Neuroscience of Emotion", published by Princeton University Press.

Stacy C Marsella is a Professor in Computer Science and Psychology at Northeastern University in Boston, Massachusetts. His research is in the computational modeling of emotion, cognition and social behavior, both as a basic research method in the study of human behavior as well as the use of these computational models in a range of applications. His current research studies the interplay of emotion and cognition in decision-making as well as the role that beliefs about others and nonverbal behavior play in social interaction. He pursues a highly interdisciplinary approach that integrates methods and models from computer science, psychology and art. His work is also applied to the design of virtual humans. This technology is being applied in education, health interventions and entertainment as well as being used as stimuli for experiments in social psychology. In addition, his work on modeling and simulating human behavior is used in large scale simulations of social behavior. The research is currently funded by NSF, NIH, AFOSR, DARPA and ONR. Dr. Marsella is a member of the Society of Experimental Social Psychologists, International Society for Research on Emotion and the Association for the Advancement of Artificial Intelligence. His work has received numerous awards including the ACM SIGAI Autonomous Agents Research Award and a UNESCO’s promising and innovative digital initiatives award.

Aleix M. Martinez is Professor in the Department of Electrical and Computer Engineering at The Ohio State University (OSU), where he is the founder and director of the Computational Biology and Cognitive Science Lab. He is also affiliated with the Department of Biomedical Engineering, to the Center for Cognitive and Brain Sciences, and to the Center for Cognitive and Behavioral Brain Imaging, where he is a member of the executive committee. He has served as an associate editor of IEEE Transactions on Pattern Analysis and Machine Intelligence, IEEE Transactions on Affective Computing, Image and Vision Computing, and Computer Vision and Image Understanding. He has been an area chair for many top conferences and was a Program co-Chair for CVPR 2014. He is also a member of NIH’s Cognition and Perception study section. Aleix is most known for being the first to define many problems and solutions in face recognition (e.g., recognition under occlusions, expression, imprecise landmark detection) as well as providing a systematic, in-depth study of the production and perception of facial expressions in the wild. He is also well-known for his work on discriminant analysis (e.g., Bayes optimal solutions, subclass-approaches, optimal kernels), and structure from motion (e.g., using kernel mappings to better model nonrigid changes, noise invariance). Aleix received a Google Faculty Award in 2015.

Seth Pollak is the College of Letters and Science Distinguished Professor of Psychology and Professor of Pediatrics, Anthropology, Neuroscience, and Public Affairs at the University of Wisconsin – Madison. He is also an investigator at the University’s Waisman Center for Human Development. Pollak earned dual Ph.D.s from the University of Rochester in brain & cognitive sciences and in child clinical psychology. Professor Pollak has published widely on topics including children’s emotional development, the influences of social risk factors on children’s brain and behavioral development, and the effects of early experience on children’s learning and health. He is a recipient of the American Psychological Association’s Distinguished Early Career Award and the Boyd-McCandless Award for Distinguished Contributions to Child Development. Dr. Pollak is a Fellow of the American Association for the Advancement of Science and the Association for Psychological Science, as well as a member of the Society for Research in Child Development’s Governing Council.

Footnotes

End Notes

1

English does not contain gender-neutral pronouns. As a consequence, we alternate between male and female pronouns.

2

Decades of research in social psychology shows that humans automatically try to predict other people’s behavior by inferring a mental state – this is called mental state inference or mentalizing, such as when inferring someone’s emotional state (e.g., for a review, see Gilbert, 1998). This research suggests that inference and prediction are not separate steps (Smith & DeCoster, 2000).

3

Bolded words appear in the glossary.

4

To be clear, teaching children how to infer emotions in others is not a problem because this skill is related to efficient communication with others. The question is whether children are being taught information that is scientifically valid and generalizable.

5

As of November 10, 2018, a website for the Detego Group indicated that “The methods developed (sic) by Paul Ekman are based on 40 years of research and are being taught to the FBI, CIA, Scotland Yard and more forensics specialists around the world” (http://www.detegogroup.eu/paul-ekman-introduction/?lang=en).

6

This empirical emphasis is largely consistent with scientists’ explicit reports of what they believe, according to a recent survey from 2014. Two-hundred and forty eight scientists who published peer-reviewed papers on the topic of emotion were asked about their views on what the scientific evidence shows. Of the 149 (60%) who responded, 119 (80%) indicated that they believed compelling evidence exists for the hypothesis that certain emotion categories are expressed with universal facial configurations or vocal signals (Ekman, 2016); no questions about variability were included in the survey.

7

In social psychology, this is the distinction between identifying an action and making an inference about the mental cause of the action (Gilbert, 1998; Vallacher & Wegner, 1987).

8

This corresponds to the null hypothesis for the true positive (in Figure 3).

9

To test the specificity hypothesis, we test something called the false positive: that people frequently scowl when not angry, meaning that they scowl more frequently than chance when fearful, sad, confused, hungry, etc. (see Figure 3). Retaining the null hypothesis for the false positive, that people do not scowl more frequently than they would by chance when fearful, sad, confused, hungry, etc., is equivalent to rejecting the null hypothesis (i.e., finding support for) the specificity hypothesis. Rejecting the null hypothesis for the false positive, because people scowl when fearful, sad, confused, hungry, etc., in addition to when angry, is evidence of no specificity (i.e., retaining the null hypothesis for the test of specificity).

10
Our decision to focus on the anger, disgust, fear, happiness, sadness and surprise categories was reinforced by two observations. First, consider a recent poll that asked scientists about their beliefs (Ekman, 2016). Two-hundred and forty eight scientists who published peer-reviewed papers on the topic of emotion were given a list of 18 emotion labels and were asked to indicate which, according to available empirical evidence, have been established as biological categories with universal expressions. Of the 149 (60%) who responded,
“There was high agreement about five emotions … : anger (91%), fear (90%), disgust (86%), sadness (80%), and happiness (76%). Shame, surprise, and embarrassment were endorsed by 40%–50%. Other emotions, currently under study by various investigators drew substantially less support: guilt (37%), contempt (34%), love (32%), awe (31%), pain (28%), envy (28%), compassion (20%), pride (9%), and gratitude (6%).” (Ekman, 2016, p. 32, italics added).

Second, there is no smoking gun in the published research on these additional emotion categories – that is, there are no scientific findings related to the production or perception of facial expressions for those emotion categories that thus far challenge the general conclusions of this paper. Simply put: regardless of how few or how many emotion categories we evaluated, the findings are the same.

11

Different number of facial muscles are reported in various sources depending on how muscles are grouped or divided.

12

From http://erikarosenberg.com/facs/ :”scientists often refer to a set of actions that occur on the face simultaneously as “facial events,” rather than calling them facial expressions. It is more descriptive. The word “expression” suggests that something from the inside becomes observable on the outside. Yet not every facial behavior expresses an internal state – most probably do not.”

14

Box 6 in SOM presents a summary of computer vision algorithms for automatically detecting facial actions.

15

Changes in illumination and face orientation are currently major hurdles.

16

Thirty-eight groups, each with their own face reading algorithm, announced their intention to participate in the challenge (Benitez-Quiroz et al., 2017a). Groups tuned their algorithms on the set of training images that were provided two weeks before the challenge deadline. Final evaluations were done on the testing set only. Of the original 38 groups, only four submitted results before the challenge ended.

17

These accuracy levels might be considered an upper estimate because of the characteristics of the training and test image databases. The methods for choosing the database are described in Benitez-Quiroz et. al 2016), although we provide a few important details here: Note, however, that a number of images are posed and professional taken.

Some facial configurations are exaggerated. Under these idealized circumstances, manual verification of these faces was estimated at 81% accuracy.

18

It is also possible that an individual person has a variety of probabilistic physical changes that reliably and specifically occur during the instances of a single emotion category, but for a number of reasons this hypothesis has not yet been scientifically tested. Specific studies to address this question would be very helpful.

19

There are ways to get around this circularity by using unsupervised, data-driven methods to discover categories, but to date, studies have used supervised approaches where categories are prescribed by human inference.

20

By relying on their own beliefs, scientists are using human consensus to identify when an emotional episode is occurring and which emotion category it belongs to (i.e., when they agree that fear or some other emotion is present, then it is said to be present). It’s important to realize that every single experiment dealing with emotion to date relies on human inference in this way. Consensus inferences are made in many areas of science. In physics and astronomy consensus emerges from expert scientists whose beliefs and assumptions often challenge the common sense view, such as in the case of quantum mechanics, dark matter, and black holes. In other areas of psychology, consensus is used to define many categories, such as memory and attention, as well as psychiatric categories, such schizophrenia and autism. Even defining depression as a mental vs. a physical illness is a matter of consensus rather than objective ground truth. But it is noteworthy that when it comes to emotions, scientists use exactly the same categories as non-scientists, which may give us cause for concern (as forewarned by William James; James, 1890, 1894). For example, compare the findings in Box 8 with the recent survey of scientists who study emotion (Ekman, 2016): 88 out of 149 scientists responded continue to believe that certain emotion categories have universal physiological markers, despite meta-analyses showing otherwise.

21

These meta-analytic findings are consistent with an earlier summary published by Matsumoto et al. (2008): of the 14 studies using rigorous FACS coding by human experts, only five reported that participants spontaneously displayed some or all of the hypothesized AUs during emotions. This is in contrast to the nine studies using the less reliable EM-FACS coding, all of which reported support. These findings suggest that some type of perceptual bias creeps in when observers make configural judgments of whether an AU is present or not (e.g., indicating whether or not a participant is smiling, or displaying “happiness”) than when AUs are coded independently, one at a time.

22

Remote, small-scale cultures are not untouched by western influences. All cultures have some minimal contact with western cultures (and this was also the case for the seminal papers published by Ekman and his colleagues in the 1970s; Gendron & Crivelli, 2017).

23

The Trobriand Islanders are a different ethnic group than the Fore; Trobrianders are subsistence fisherman and horticulturalists living in a small archipelago of islands located 200km from the mainland (the origin of the original Fore who were photographed). As Crivelli et al. make clear in their paper, these findings are a within-nation rather than a within-culture comparison.

24

The value of this particular study is that the researchers not only coded infants’ facial movements but also measured a range of concurrent movements that could support inferences about the infants’ feelings of pleasantness, unpleasantness and level of arousal, termed affect (see Box 9), including increased respiration, withdrawal/leaning away with the body, stilling/freezing, struggling, turning toward the mother, extreme withdrawal, hiding of their faces, squirming, self-stimulation, looking toward mother, pointing at the object, doing a “double-take,” and banging on the table.

25

Bennett et al. (2002) note that when they observed facial actions were thought to be associated with more than one emotion category (e.g., when an infant produced a facial configuration that was a combination of scowling (anger) and pouting (sadness), they interpreted the expression using the facial actions in only the upper region of the face, which indicates that infants’ facial movements were even more variable than reported in the data tables. A footnote in the paper further indicates that infants produced facial movements that were interpreted to reflect “interest” across all of the eliciting situations, but these facial actions were not included in any data analyses (Bennett et al., 2002, footnote 1). Any facial configuration that included AUs stipulated as interest and AUs for another emotion category was coded as an expression of the other emotion category.

26

Also, it is not clear that children find sour foods disgusting (e.g., Stein, Ottenberg, & Roulet, 1958; Rozin et al., 1986).Young children appear to be attracted to many things that adults find disgusting, whereas by the age of five, children have more adult-like behavioral responses and reject them (Rozin et al., 1986). For a discussion of how disgust is learned, see Widen & Russell (2013).

27

In another naturalistic study, videos of children aged four through seven were downloaded from the internet and FACS coded (Camras et al., 2018). The children were playing “the scary maze game”: a child solves maze after maze of increasing difficulty, only to encounter a screaming, demonic girl from the movie The Exorcist (filmed in 1973). The game is generally thought to evoke an instance of fear (hence the name “scary”), but it may also evoke surprise as the scary stimulus makes a sudden unexpected appearance. Children only produced the wide-eyed, gasping configuration (the proposed facial expression of fear) and/or a startled configuration (the proposed facial expression of surprise) with weak reliability (38% and 10%, respectively).

28

By analogy, people who have been blind since birth learn color concepts and the relation between these concepts, such as “red,” “blue,” and “green” are similar to those of sighted people (e.g., congenitally blind individuals understand the US concept for "blue" is more similar to "green" than to "red”; Shepard & Cooper, 1992). The structure of brain regions in visual cortex that represent visual concepts are also virtually indistinguishable in sighted and congenitally blind individuals (Koster-Hale et al., 2014; Wang et al., 2015).

29

The onset and severity of blindness varies hugely across studies. Even a small amount of visual experience in infancy or early childhood will influence brain development and provide experiences for learning about emotions (see earlier section on emotion concept development in infants). Helen Keller, for example, could see and hear until she was 19 months old, providing some initial scaffolding for her later ability to communicate.

30

For example, recently, Ekman (2017) wrote, “Another challenge to the findings of universality came from the anthropologist, Margaret Mead …. Establishing that posed expressions are universal, she said, does not necessarily mean that spontaneous expressions are universal. I replied (Ekman, 1977) that it seemed illogical to presume that people can readily interpret posed facial expressions if they had not seen those facial expressions and experienced them in actual social life” (Ekman, 2017, p. 46).

31

While these findings are instructive, they likely provide a lower limit of the possible real world variation in the facial configurations that express the varied instances of a given emotion category. After all, the internet is a curated version of reality and some frequent facial configurations are likely missing because they are rarely uploaded to the internet. Similarly, some configurations commonly found on the internet might not be commonly observed in the real world.

32

Compare these findings to those from a study that mined images from the internet using a similar but narrower approach, and who had two raters use a choice-from-array method to label the images (Mollahosseini et al., 2016).

33

Configuration 3 also resembles people’s beliefs about the configurations that express fear and awe (i.e., the “international core patterns” reported by Cordaro et al. 2017).

34

More generally, participants are more likely to perceive the intended emotion in the hypothesized facial configurations of Figure 4 when they are displayed on dynamically moving, synthetic faces (Wehrle, Kaiser, Schmidt, & Scherer, 2000), in video footage of posed facial muscle movements (e.g., Ambadar, Schooler & Cohn, 2005; Cunningham & Wallraven, 2009), and even in point-light displays of motion created by facial muscle movements (Bassili, 1979). This “dynamic advantage” sometimes disappears when participants are viewing real human faces (e.g., Fiorentini & Viviani, 2011; Gold, Barker, et al., 2013; Miles & Johnston, 2007; Nelson & Russell, 2011).

35

Ekman & Friesen (1971) was chosen as one of the forty studies that changed psychology (Hicks, 2012) and, along with Ekman et al. (1969) is routinely discussed in introductory psychology textbooks.

36

Dioula participants from Burkina Faso in West Africa showed strong reliability for labeling smiling facial configurations as happiness, moderate reliability for labeling frowning facial configurations as sadness, startled facial configurations as surprise, and nose-wrinkled facial configurations as disgust, and weak reliability for labeling scowling facial configurations as anger and wide-eyed gasping facial configurations as fear.

37

For example, a sample of Trobriand Islanders, who are subsistence horticulturalists and fishermen living in the Trobriand Islands of Papua New Guinea, labeled a scowling facial configuration as anger with above chance reliability (.29% of the time), but also labeled that facial configuration more frequently with “feels like avoiding a social interaction” (.50% of the time) (Crivelli, Russell et al., 2017, Study 2). In fact, the wide-eyed, gasping facial configuration that is thought to be the expression for fear (Figure 4) is understood as an expression of aggression or threat in the Trobriand culture (Crivelli, Jarillo & Fridlund, 2016, 2017; Crivelli, Russell et al., 2016). Trobrianders uniquely labeled smiling facial configurations as happiness across two studies but this finding did not replicate in a third sample nor in a sample of Mwani participants who are subsistence fisherman living on Matemo Island in Mozambigue, Africa.

38

The ancestors of the Hadza are thought to have been continuously practicing a hunting and gathering lifestyle for at least the past 50,000 years in their current region of East Africa. Furthermore, Hadza social structure, mobility, residential patterns, and language have thus far remained largely buffered from their interactions with other ethnic groups (Apicella & Crittenden, 2016; Crittenden & Marlowe, 2008) which have been sustained for at least the past 100 years (Jones, 2016)

39

The wide-eyed gasping stereotype for fear is thought to have evolved for enhanced sensory sampling that supports efficient threat detection (Susskind et al., 2008). Similarly, the nose-wrinkle stereotype for disgust is thought to have evolved in order to limit exposure to noxious stimuli (Chapman & Anderson, 2012; Chapman, Kim, Susskind, & Anderson, 2009).

40

Interestingly, adult perceivers may have overtly looked at the postures less, but other evidence with the same stimuli suggest that different body contexts influenced how adult participants visually scanned the exact same facial configurations; Aviezer et al., 2008). At the other end of the age spectrum, older adults are also more influenced by context when inferring emotional meaning in facial configurations as compared to young adults (Ngo & Isaacowitz, 2015).

41

Some applications will not be affected by context because they are not aiming to use facial movements to infer an individual’s underlying emotional state. These initiatives have very specific applications in mind. For example, detecting pain in patients (Apple), driver drowsiness (Google), creating virtual facial expression stickers or animojis from one’s own facial poses (Facebook, iPhone X), or Alibaba's “smile to pay.”

42

The word “appraisal” has two meanings in the science of emotion. Here, appraisals simply to refer to the descriptive features of how a situation is experienced, such as novelty, goal relevance, etc., without any inference about how those experiential features are caused (e.g., Clore & Ortony, 2008; Ortony & Clore, 2013). The other meaning of appraisal refers to the mechanisms that cause the experiential features as components of emotion (e.g., the component process model of emotion, in which appraisals are considered evaluative “checks” that the human mind uses in a serial fashion; e.g., Scherer, Mortillaro & Mehu, 2017). There is very little evidence that appraisals are, in fact, causal in nature (for a discussion, Parkinson, 1997). In some studies, for example, participants are presented with a written scenario that is assumed to automatically trigger a specific sequence of appraisal checks (i.e., cognitive evaluations), which in turn is hypothesized to produce a specific pattern of facial muscle movements. Notice that the main causal mechanisms here – appraisal checks – are not measured directly but are inferred to have occurred. In other studies, participants are asked to explicitly report on the appraisals they experience, on the assumption that the corresponding “checks” are active. Emerging scientific evidence links appraisals, as descriptive features, to facial movements, although the evidence to date suggests that these relationships are not as consistent as specific as hypothesized (a summary of this research program can be found in in Scherer, Mortillaro & Mehu, 2017).

References

  1. Adolphs R (2002). Neural mechanisms for recognizing emotions. Current Opinion in Neurobiology 12, 169–178. [DOI] [PubMed] [Google Scholar]
  2. Adolphs R (2017). How should neuroscience study emotions? by distinguishing emotion states, concepts, and experiences. Soc Cogn Affect Neurosci, 12(1), 24–31. doi: 10.1093/scan/nsw153 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Adolphs R, Tranel D, Damasio H, and Damasio A (1994). Impaired recognition of emotion in facial expressions following bilateral damage to the human amygdala. Nature, 372, 669–672. [DOI] [PubMed] [Google Scholar]
  4. Arya Ali, DiPaola Steve, and Parush Avi, “Perceptually Valid facial Expressions for Character-Based Applications,” International Journal of Computer Games Technology, vol. 2009, Article ID 462315, 13 pages, 2009. [Google Scholar]
  5. Ambadar Z, Schooler JW, & Cohn JF (2005). Deciphering the enigmatic face: The importance of facial dynamics in interpreting subtle facial expressions. Psychological Science, 16, 403–410. [DOI] [PubMed] [Google Scholar]
  6. Ambadar Z, Cohn JF, & Reed LI (2009). All Smiles are Not Created Equal: Morphology and Timing of Smiles Perceived as Amused, Polite, and Embarrassed/Nervous. Journal of Nonverbal Behavior, 33(1), 17–34. [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Adolphs R & Anderson D (2018). The Neuroscience of Emotion: A New Synthesis. Princeton University Press. [Google Scholar]
  8. Apicella CL, & Crittenden AN (2016). Hunter-Gatherer families and parenting The Handbook of Evolutionary Psychology: John Wiley & Sons, Inc. [Google Scholar]
  9. Arcaro MJ, Schade PF, Vincent JL, Ponce CR, & Livingstone MS (2017). Seeing faces is necessary for face-domain formation. Nature Neuroscience, 20, 1404–1412. [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Atzil S, Gao W, Fradkin I, & Barrett LF (2018). Growing a social brain. Nature Human Behavior, 2, 624–636. [DOI] [PubMed] [Google Scholar]
  11. Atzil S, Touroutoglou A, Rudy T, Salcedo S, Feldman R, Hooker J, Dickerson BC, Catana C & Barrett LF (2017). Dopamine in the medial amygdala network mediates human bonding. Proceedings of the National Academy of Sciences, 114, 2361–2366. [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Aviezer H, & Hassin RR (2017). Inherently ambiguous: An argument for contextualized emotion perception In Fernandez-Dols J-M & Russell JA (Eds.), The science of facial expression (p. 333–349). NY: Oxford U Press. [Google Scholar]
  13. Aviezer H, Hassin RR, Ryan J, Grady C, Susskind J, Anderson A, … Bentin S. (2008). Angry, disgusted, or afraid? Studies on the malleability of emotion perception. Psychological Science, 19(7), 724–732. [DOI] [PubMed] [Google Scholar]
  14. Aviezer H, Trope Y, & Todorov A (2012). Body cues, not facial expressions, discriminate between intense positive and negative emotions. Science, 338(6111), 1225–1229. doi: 10.1126/science.1224313 [DOI] [PubMed] [Google Scholar]
  15. Bach DR, & Dayan P (2017). Algorithms for survival: a comparative perspective on emotions. Nature Reviews Neuroscience, 18, 311. doi: 10.1038/nrn.2017.35 [DOI] [PubMed] [Google Scholar]
  16. Bandes SA (2014). Remorse, demeanor, and the consequences of misinterpretation. Journal of Law, Religion and State, 3(2), 170–199. doi: 10.1163/22124810-00302004 [DOI] [Google Scholar]
  17. Baron-Cohen S, Golan O, Wheelwright S, Hill JJ Mind Reading: the interactive guide to emotions. London, UK: Jessica Kingsley Limited; 2004. [Google Scholar]
  18. Baron-Cohen S, Wheelwright S, Hill J, Raste Y, and Plumb I (2001). The "reading the mind in the eyes" test revised version: a study with normal adults, and adults with asperger syndrome or high-functioning autism. Journal of Child Psychology and Psychiatry 42, 241–251 [PubMed] [Google Scholar]
  19. Barratt D, Rédei AC, Innes-Ker A, & van de Weijer J (2016). Does the Kuleshov effect really exist? Revisiting a classic film experiment on facial expressions and emotional contexts. Perception, 45, 847–874. [DOI] [PubMed] [Google Scholar]
  20. Barrett Deirdre (2007). "Supernormal Stimuli--Why Birds Are Cuckoo". Waistland: The (R)Evolutionary Science Behind Our Weight and Fitness Crisis. New York, New York: W.W. Norton & Co. [Google Scholar]
  21. Barrett HC, Bolyanatz A, Crittenden AN, Fessler DM, Fitzpatrick S, Gurven M, … Pisor A. (2016). Small-scale societies exhibit fundamental variation in the role of intentions in moral judgment. Proceedings of the National Academy of Sciences, 113(17), 4688–4693. [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Barrett LF (2004). Feelings or words? Understanding the content in self-report ratings of emotional experience. Journal of Personality and Social Psychology, 87, 266–281. PMC1351136 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Barrett LF (2006a). Are Emotions Natural Kinds? Perspectives on Psychological Science, 1, 28–58. [DOI] [PubMed] [Google Scholar]
  24. Barrett LF (2006b). Solving the emotion paradox: Categorization and the experience of emotion. Personality and Social Psychology Review, 10, 20–46. [DOI] [PubMed] [Google Scholar]
  25. Barrett LF, Lindquist K, Bliss-Moreau E, Duncan S, Gendron M, Mize J, & Brennan L (2007). Of mice and men: Natural kinds of emotion in the mammalian brain? Perspectives on Psychological Science, 2, 297–312 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Barrett LF (2012). Emotions are real. Emotion, 12, 413–429 [DOI] [PubMed] [Google Scholar]
  27. Barrett LF (2013). Psychological construction: A Darwinian approach to the science of emotion. Emotion Review, 5, 379–389. [Google Scholar]
  28. Barrett LF (2011). Was Darwin wrong about emotional expressions? Current Directions in Psychological Science, 20, 400–406. [Google Scholar]
  29. Barrett LF (2017a). How emotions are made: The secret life of the brain. New York: Houghton-Mifflin-Harcourt. [Google Scholar]
  30. Barrett LF (2017b). The theory of constructed emotion: An active inference account of interoception and categorization. Social Cognitive and Affective Neuroscience, 12, 1–23. 10.1093/scan/nsw154 [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Barrett LF (2017c). Functionalism cannot save the classical view of emotion. Social Cognitive and Affective Neuroscience, 12, 34–36. 10.1093/scan/nsw156 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Barrett LF, & Bliss-Moreau E (2009). Affect as a psychological primitive. Advances in Experimental Social Psychology, 41, 167–218. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Barrett LF & Finlay BL (in press). Concepts, goals and the control of survival-related behaviors. Current Opinion in the Behavioral Sciences. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Barrett LF & Gendron M (2016). The importance of context: Three corrections to Cordaro, Keltner, Tshering, Wangchuk, & Flynn (2016). Emotion, 16, 803–806. [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Barrett LF, & Satpute AB (2017). Historical pitfalls and new directions in the neuroscience of emotion. Neuroscience Letters. [DOI] [PMC free article] [PubMed] [Google Scholar]
  36. Barrett LF, Mesquita B, Ochsner KN, & Gross JJ (2007). The experience of emotion. Annual Review of Psychology, 58, 373–403. PMC1934613 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Barrett LF, Mesquita B, & Gendron M (2011). Context in emotion perception. Current Directions in Psychological Science, 20, 286–290. [Google Scholar]
  38. Barrett LF, Wilson-Mendenhall CD, & Barsalou LW (2015). The conceptual act theory: A road map Chapter in Barrett LF and Russell JA (Eds.), The psychological construction of emotion (p. 83–110). New York: Guilford. [Google Scholar]
  39. Barsalou LW (1983). Ad hoc categories. Memory and Cognition, 11(3), 211–227. [DOI] [PubMed] [Google Scholar]
  40. Barsalou LW (1985). Ideals, central tendency, and frequency of instantiation as determinants of graded structure in categories. J Exp Psychol Learn Mem Cogn, 11(4), 629–654. [DOI] [PubMed] [Google Scholar]
  41. Barsalou LW (2008). Grounded cognition. Annual Review of Psychology, 59, 617–45. [DOI] [PubMed] [Google Scholar]
  42. Barsalou LW, Simmons WK, Barbey AK, & Wilson CD (2003). Grounding conceptual knowledge in modality-specific systems. Trends in Cognitive Sciences, 7(2), 84–91. [DOI] [PubMed] [Google Scholar]
  43. Bassili JN (1979). Emotion recognition: The role of facial movement and the relative importance of upper and lower areas of the face. Journal of Personality and Social Psychology, 37, 2049–2058. [DOI] [PubMed] [Google Scholar]
  44. Baylor AL, & Kim S (2008). The Effects of Agent Nonverbal Communication on Procedural and Attitudinal Learning Outcomes. Paper presented at the International Conference on Intelligent Virtual Agents, Tokyo. [Google Scholar]
  45. Bedny M, Saxe R (2012). Insights into the origins of knowledge from the cognitive neuroscience of blindness. Cognitive Neuropsychology 29: 56–84. [DOI] [PubMed] [Google Scholar]
  46. Bell C (1806). Essays on the Anatomy of Expression in Painting. London: Longman, Hurst,Rees & Orme. [Google Scholar]
  47. Benitez-Quiroz CF, Srinivasan R, & Martinez AM (2016). Emotionet: An accurate, real-time algorithm for the automatic annotation of a million facial expressions in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 5562–5570). [Google Scholar]
  48. Benitez-Quiroz F, Srinivasan R, & Martinez AM (2019). Discriminant Functional Learning of Color Features for the Recognition of Facial Action Units and their Intensities. IEEE transactions on pattern analysis and machine intelligence. [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Benitez-Quiroz CF, Srinivasan R, Feng Q, Wang Y, & Martinez AM (2017a). EmotioNet Challenge: Recognition of facial expressions of emotion in the wild. arXiv preprint arXiv: 170301210. [Google Scholar]
  50. Benitez-Quiroz CF, Wang Y, & Martinez AM (2017b). Recognition of action units in the wild with deep nets and a new global-Local loss. In Proceedings of the International Conference on Computer Vision. [Google Scholar]
  51. Benitez-Quiroz F, Srinivasan R, & Martinez AM (in press). Discriminant Functional Learning of Color Features for the Recognition of Facial Action Units and their Intensities. IEEE transactions on pattern analysis and machine intelligence. [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Bennett DS, Bendersky M, & Lewis M (2002). Facial expressivity at 4 months: A context by expression analysis. Infancy, 3(1), 97–113. [DOI] [PMC free article] [PubMed] [Google Scholar]
  53. Berggren S, Fletcher-Watson Sue, Milenkovic Nina, Marschik Peter B., Bölte Sven & Jonsson Ulf. (2018). Emotion recognition training in autism spectrum disorder: A systematic review of challenges related to generalizability. Developmental Neurorehabilitation, 21,141–154. [DOI] [PubMed] [Google Scholar]
  54. Bick J, & Nelson CA (2016). Early adverse experiences and the developing brain. Neuropsychopharmacology, 41(1), 177–196. [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Blascovich J, Loomis J, Beall A, Swinth K, Hoyt C, & Bailenson JN (2002). Immersive virtual environment technology as a methodological tool for social psychology. Psychological Inquiry, 13, 103–124. [Google Scholar]
  56. Bolzani Dinehart LH, Messinger DS, Acosta SI, Cassel T, Ambadar Z, & Cohn J (2005). Adult perceptions of positive and negative infant emotional expressions. Infancy, 8(3), 279–303. [Google Scholar]
  57. Boucher JD, & Carlson GE (1980). Recognition of facial Expression in three cultures. Journal of cross-cultural psychology, 11(3), 263–280. [Google Scholar]
  58. Bridges CB (1932). The suppressors of purple. Molecular and General Genetics MGG, 60(1), 207–218. [Google Scholar]
  59. Briggs-Gowan MJ, Pollak SD, Grasso D, Voss J, Mian ND, Zobel E, McCarthy KJ, Wakschlag LS, & Pine DS. (2015). Attention bias and anxiety in young children exposed to family violence. Journal of Child Psychology and Psychiatry, 56: 11, 1194–1201. [DOI] [PMC free article] [PubMed] [Google Scholar]
  60. Brodny G, Kolakowska A, Landowska A, Szwoch M, Szwoch W, & Wróbel MR (2016). Comparison of selected off-the-shelf solutions for emotion recognition based on facial expressions. Paper presented at the 9th International Conference on Human System Interaction, Portsmouth, UK, July 6–8, 2016. [Google Scholar]
  61. Bryant GA, & Barrett HC (2008). Vocal emotion recognition across disparate cultures. Journal of Cognition and Culture, 8, 135–148. [Google Scholar]
  62. Bryant GA, Fessler DM, Fusaroli R, Clint E, Aarøe L, Apicella CL, … Chavez B. (2016). Detecting affiliation in colaughter across 24 societies. Proceedings of the National Academy of Sciences, 113(17), 4682–4687. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Buck R (1984). The communication of emotion. New York, NY: Guilford Press. [Google Scholar]
  64. Bui D, Heylen D, Poel M & Nijholt A (2004) Combination of facial movements on a 3D talking head Proceedings Computer Graphics International, IEEE, pp. 284–291. [Google Scholar]
  65. Cacioppo JT, Berntson GG, Larsen JH, Poehlmann KM, & Ito TA (2000). The psychophysiology of emotion In Lewis R & Haviland-Jones JM (Eds.), The handbook of emotion (2 ed., pp. 173–191). New York: Guilford Press. [Google Scholar]
  66. Cain J (2000). The way I feel. Parenting Press. [Google Scholar]
  67. Calbi M, Heimann K, Barratt D, Siri F, Umilta MA, & Gallese V (2017). How Context Influences Our Perception of Emotional Faces: A Behavioral Study on the Kuleshov Effect. Front Psychol, 8, 1684. doi: 10.3389/fpsyg.2017.01684 [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Caldara R (2017). Culture reveals a flexible system for face processing. Current Directions in Psychological Science, 26, 249–255. [Google Scholar]
  69. Calvo MG, Gutierrez-Garcia A, Fernandez-Martin A, and Nummenmaa L (2014). Recognition of facial expressions of emotion is related to their frequency in everyday life. Journal of Nonverbal Behavior, 38, 549–567. [Google Scholar]
  70. Campos JJ, Campos RG, & Barrett KC (1989). Emergent themes in the study of emotional development and emotion regulation. Developmental Psychology, 25(3), 394. [Google Scholar]
  71. Campos JJ, Mumme D, Kermoian R, & Campos RG (1994). A functionalist perspective on the nature of emotion. Japanese Journal of Research on Emotions, 2(1), 1–20. [PubMed] [Google Scholar]
  72. Camras LA (1991). Conceptualizing early infant affect: View II and Reply In Strongman K (Ed.), International review of studies on emotion (pp. 16–28, 33–36). New York: Wiley [Google Scholar]
  73. Camras LA (1992). Expressive development and basic emotions. Cognition & Emotion, 6(3-4), 269–283. [Google Scholar]
  74. Camras LA, & Allison K (1985). Children's understanding of emotional facial expressions and verbal labels. Journal of Nonverbal Behavior, 9(2), 84–94. [Google Scholar]
  75. Camras LA, & Shutter JM (2010). Emotional facial expressions in infancy. Emotion Review, 2(2), 120–129. [Google Scholar]
  76. Camras LA, Castro VL, Halberstadt AG, & Shuster MM (2017). Spontaneously produced facial expressions in infants and children In Fernandez-Dols J-M & Russell JA (Eds.), The science of facial expression (p. 279–296). NY: Oxford U Press. [Google Scholar]
  77. Camras Fatani, Fraumeni & Shuster, 2016. The development of facial expressions In Barrett LF, Lewis M, & Haviland-Jones JM (Eds.). (2016). Handbook of Emotions, 4th Edition (255–271). New York: Guilford. [Google Scholar]
  78. Camras LA, Oster H, Ujiie T, Campos JJ, Bakeman R, & Meng Z (2007). Do infants show distinct negative facial expressions for fear and anger? Emotional expression in 11-month-old European American, Chinese, and Japanese infants. Infancy, 11(2), 131–155. [Google Scholar]
  79. Camras LA, Perlman SB, Fries ABW, & Pollak SD (2006). Post-institutionalized Chinese and Eastern European children: Heterogeneity in the development of emotion understanding. International Journal of Behavioral Development, 30(3), 193–199. [Google Scholar]
  80. Caron RF, Caron AJ, & Myers RA (1985). Do infants see emotional expressions in static faces? Child Development, 56, 1552–1560. [PubMed] [Google Scholar]
  81. Carrera-Levillain P, Fernandez-Dols JM (1994). Neutral faces in context: Their emotional meaning and their function. Journal of Nonverbal Behavior, 18, 281–299. [Google Scholar]
  82. Carroll JM, & Russell JA (1996). Do facial expressions signal specific emotions? Judging emotion from the face in context. Journal of Personality and Social Psychology, 70(2), 205–218. [DOI] [PubMed] [Google Scholar]
  83. Carver LJ, & Vaccaro BG (2007). 12-month-old infants allocate increased neural resources to stimuli associated with negative adult emotion. Developmental Psychology, 43(1), 54. [DOI] [PMC free article] [PubMed] [Google Scholar]
  84. Cassell J, et al. , Eds. (2000). Embodied Conversational Agents. Cambridge, MA, MIT Press. [Google Scholar]
  85. Cassia VM, Turati C & Simion F (2004). Can a nonspecific bias toward top-heavy patterns explain newborns' face preference? Psychological Science 15, 379–383. [DOI] [PubMed] [Google Scholar]
  86. Castro VL, Camras LA, Halberstadt AG, & Shuster M (2017). Children's Prototypic facial Expressions During Emotion-Eliciting Conversations With Their Mothers. Emotion. doi: 10.1037/emo0000354 [DOI] [PMC free article] [PubMed] [Google Scholar]
  87. Cattaneo L, Pavesi G (2014). "The facial motor system." Neuroscience and Biobehavioral Reviews, 38, 135–159. [DOI] [PubMed] [Google Scholar]
  88. Cecchini M, Baroni E, Di Vito C, & Lai C (2011). Smiling in newborns during communicative wake and active sleep. Infant Behavior and Development, 34(3), 417–423. [DOI] [PubMed] [Google Scholar]
  89. Ceulemans E, Kuppens P, & Mechelen IV (2012). Capturing the Structure of Distinct Types of Individual Differences in the Situation-specific Experience of Emotions: The Case of Anger. European Journal of Personality, 26(5), 484–495. [Google Scholar]
  90. Chapman HA & Anderson AK (2013). Things rank and gross in nature: a review and synthesis of moral disgust. Psychological Bulletin, 139, 300–27 [DOI] [PubMed] [Google Scholar]
  91. Chapman HA, Kim DA, Susskind JM, & Anderson AK (2009). In bad taste: Evidence for the oral origins of moral disgust. Science, 323(5918), 1222–1226. [DOI] [PubMed] [Google Scholar]
  92. Chiesa S, Galati D, & Schmidt S (2015). Communicative interactions between visually impaired mothers and their sighted children: analysis of gaze, facial expressions, voice and physical contacts. Child: Care, Health and Development, 41, 6, 1040–1046. [DOI] [PubMed] [Google Scholar]
  93. Chu WS, De la Torre F, & Cohn JF (2017). Selective transfer machine for personalized facial expression analysis. IEEE transactions on pattern analysis and machine intelligence, 39(3), 529–545. [DOI] [PMC free article] [PubMed] [Google Scholar]
  94. Cicchetti D, & Curtis WJ (2005). An event-related potential study of the processing of affective facial expressions in young children who experienced maltreatment during the first year of life. Development and Psychopathology, 17(3), 641–677. [DOI] [PubMed] [Google Scholar]
  95. Clark-Polner E, Johnson T, & Barrett LF (2017). Multivoxel pattern analysis does not provide evidence to support the existence of basic emotions. Cerebral Cortex, 27, 1944–1948. [DOI] [PMC free article] [PubMed] [Google Scholar]
  96. Clore GL & Ortony A (1991). What more is there to emotion concepts than prototypes? Journal of Personality and Social Psychology, 60, 48–50. [Google Scholar]
  97. Clore GL, Ortony A. (2013). Psychological Construction in the OCC Model of Emotion. Emotion Review, 5(4):335–343. doi: 10.1177/1754073913489751. [DOI] [PMC free article] [PubMed] [Google Scholar]
  98. Clore GL, & Ortony A (2008). Appraisal theories: How cognition shapes affect into emotion In Lewis M, Haviland-Jones JM, & Barrett LF (Eds.), Handbook of Emotions (3rd ed., pp. 628–642). New York, NY: Guilford Press. [Google Scholar]
  99. Cole PM (1986). Children's spontaneous control of facial expression. Child Development, 57, 1309–1321. [Google Scholar]
  100. Cole PM, Jenkins PA, & Shott CT (1989). Spontaneous expressive control in blind and sighted children. Child Development, 60, 683–688. [PubMed] [Google Scholar]
  101. Cordaro DT, Sun R, Keltner D, Kamble S, Hoddar N, & McNeil G (2017). Universals and Cultural Variations in 22 Emotional Expressions Across Five Cultures. Emotion. doi: 10.1037/emo0000302 [DOI] [PubMed] [Google Scholar]
  102. Cordaro DT, Sun R, Kamble S, Hodder N, Monroy M Cowen A, Bai Y, & Keltner D. (in press). The recognition of 18 facial-body expressions across nine cultures. Emotion. [DOI] [PMC free article] [PubMed] [Google Scholar]
  103. Corneanu CA, Simon MO, Cohn JF, & Guerrero SE (2016). Survey on rgb, 3d, thermal, and multimodal approaches for facial expression recognition: History, trends, and affect-related applications. IEEE transactions on pattern analysis and machine intelligence, 38(8), 1548–1568. [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Cowen AS & Keltner D (2017). Self-report captures 27 distinct categories of emotion bridged by continuous gradients. Proceedings of the National Academy of Sciences, www.pnas.org/cgi/doi/10.1073/pnas.1702247114 [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Craig AD (2015). How do you feel?: an interoceptive moment with your neurobiological self. Princeton University Press. [Google Scholar]
  106. Crittenden AN, & Marlowe FW (2008). Allomaternal care among the Hadza of Tanzania. Human Nature, 19(3), 249–262. [DOI] [PubMed] [Google Scholar]
  107. Crivelli C, Carrera P, & Fernández-Dols J-M (2015). Are smiles a sign of happiness? Spontaneous expressions of judo winners. Evolution and Human Behavior, 36(1), 52–58. [Google Scholar]
  108. Crivelli C, Jarillo S, & Fridlund AJ (2016). A multidisciplinary approach to research in small-scale societies: Studying emotions and facial expressions in the field. Frontiers in psychology, 7:1073. doi: 10.3389/fpsyg.2016.01073 [DOI] [PMC free article] [PubMed] [Google Scholar]
  109. Crivelli C, Jarillo S, Russell JA, & Fernández-Dols JM (2016). Reading emotions from faces in two indigenous societies. Journal of Experimental Psychology: General, 145(7), 830–843. doi: 10.1037/xge0000172 [DOI] [PubMed] [Google Scholar]
  110. Crivelli C, Russell JA, Jarillo S, & Fernández-Dols JM (2016). The fear gasping face as a threat display in a Melanesian society. Proceedings of the National Academy of Sciences, U.S.A, 113(44), 12403–12407. doi: 10.1073/PNAS.1611622113 [DOI] [PMC free article] [PubMed] [Google Scholar]
  111. Crivelli C, Jarillo S and Fridlund AJ (2016) A Multidisciplinary Approach to Research in Small-Scale Societies: Studying Emotions and Facial Expressions in the Field. Front. Psychol. 7:1073. doi: 10.3389/fpsyg.2016.01073 [DOI] [PMC free article] [PubMed] [Google Scholar]
  112. Cunningham DW, & Wallraven C (2009). Dynamic information for the recognition of conversational expressions. Journal of Vision, 9, 1–17. [DOI] [PubMed] [Google Scholar]
  113. da Silva Ferreira GC, Crippa JA, & de Lima Osório F (2014). Facial emotion processing and recognition among maltreated children: a systematic literature review. Frontiers in psychology, 5, 1460. [DOI] [PMC free article] [PubMed] [Google Scholar]
  114. Danziger K(2006). Universalism and indigenization in the history of modern psychology In Internationalizing the History of Psychology (Brock AC,ed.),pp.208–225, New York University Press [Google Scholar]
  115. Darwin C (1859/1964). On the Origin of Species (A facsimile of the first edition. ed.). Cambridge, MA: Harvard University Press. [Google Scholar]
  116. Darwin C (1872/1965). The Expression of the Emotions in Man and Animals (Chicago: University of Chicago Press; ). [Google Scholar]
  117. Dashiell JF (1927). A new method of measuring reactions to facial expression of emotion. Psychological Bulletin, 24, 174–175. [Google Scholar]
  118. Davidson RJ (1992). Anterior cerebral asymmetry and the nature of emotion. Brain Cogn, 20(1), 125–151. [DOI] [PubMed] [Google Scholar]
  119. de Gelder B (2016). Emotions and the body. New York: Oxford University Press. [Google Scholar]
  120. de Melo C Peter Carnevale, Stephen Read and Jonathan Gratch. Reading people’s minds from emotion expressions in interdependent decision making. Journal of Personality and Social Psychology, 106(1), 2014, pp. 73–88 [DOI] [PubMed] [Google Scholar]
  121. de Mendoza AH, Fernández-Dols JM, Parrott WG, & Carrera P (2010). Emotion terms, category structure, and the problem of translation: The case of shame and vergüenza. Cognition & Emotion, 24(4), 661–680. [Google Scholar]
  122. DeCarlo LT (2012). On a signal detection approach to -alternative forced choice with bias, with maximum likelihood and Bayesian approaches to estimation. Journal of Mathematical Psychology, 56(3), 196–207. doi: 10.1016/j.jmp.2012.02.004 [DOI] [Google Scholar]
  123. Denham SA (1998). Emotional development in young children. Guilford Press. [Google Scholar]
  124. Denison S, Reed C & Xu F (2013) The emergence of probabilistic reasoning in very young infants: evidence from 4.5- and 6-month-old infants. Developmental Psychology, 49(2), 243–249. [DOI] [PubMed] [Google Scholar]
  125. DiGirolamo MA, & Russell JA (2017). The emotion seen in a face can be a methodological artifact: The process of elimination hypothesis. Emotion, 17(3), 538–546. doi: 10.1037/emo0000247 [DOI] [PubMed] [Google Scholar]
  126. Ding Y, Prepin K, Huang J, Pelachaud C, and Artières T, (2014) “Laughter animation synthesis,” in Proceedings of the 2014 International Conference on Autonomous Agents and Multi-agent Systems, ser. AAMAS ’14, 2014, pp. 773–780. [Google Scholar]
  127. Dondi M, Gervasi MT, Valente A, Vacca T, Bogana G, De Beilis I, … & Oster H. (2012, July). Spontaneous facial expressions of distress in fetuses. In 14th European Conference on facial Expression (pp. 16–18). [Google Scholar]
  128. Dondi M, Messinger D, Colle M, Tabasso A, Simion F, Barba BD, & Fogel A (2007). A new perspective on neonatal smiling: differences between the judgments of expert coders and naive observers. Infancy, 12(3), 235–255. [Google Scholar]
  129. Doyle CM, & Lindquist KA (2018). When a word is worth a thousand pictures: Language shapes perceptual memory for emotion. Journal of Experimental Psychology: General, 147(1), 62–73. doi: 10.1037/xge0000361 [DOI] [PubMed] [Google Scholar]
  130. Duchenne CB (1990/1862). The Mechanism of Human Facial Expression (Renard, Paris); reprinted (1990) (Cambridge Univ Press, London: ). [Google Scholar]
  131. Duran JI, & Fernandez-Dols J-M. (2018). Do emotions result in their predicted facial expressions? A meta-analysis of studies on the link between expression and emotion. PsyArXiv preprint, available at https://psyarxiv.com/65qp7 [DOI] [PubMed] [Google Scholar]
  132. Duran JI, Reisenzein R, & Fernandez-Dols J-M. (2017). Coherence between emotions and facial expressions: A research synthesis In Fernandez-Dols J-M & Russell JA (Eds.), The science of facial expression (p. 107–129). NY: Oxford U Press. [Google Scholar]
  133. Edelman GM, & Tononi G (2000). A universe of consciousness: How matter becomes imagination. New York: Basic books. [Google Scholar]
  134. Eibl-Eibesfeldt I (1989) Human Ethology. Aldine deGruyter. [Google Scholar]
  135. Eibl-Eibesfeldt I (1972). Similarities and differences between cultures in expressive movements In Hinde RA, Non-Verbal Communication. Oxford, England: Cambridge U. Press. [Google Scholar]
  136. Ekman P (1972). Universals And Cultural Differences In facial Expressions Of Emotions In Cole J (Ed.), Nebraska Symposium on Motivation, 1971 (pp. 207–283). Lincoln, Nebraska: University of Nebraska Press. [Google Scholar]
  137. Ekman P (1992). An argument for basic emotions. Cognition & Emotion, 6, 169–200. [Google Scholar]
  138. Ekman P (1994). Strong evidence for universals in facial expressions: A reply to Russell’s mistaken critique. Psychological Bulletin, 115, 268–287. [DOI] [PubMed] [Google Scholar]
  139. Ekman P (2016). What scientists who study emotion agree about. Perspectives on Psychology Science, 11(1), 31–34. doi: 10.1177/1745691615596992 [DOI] [PubMed] [Google Scholar]
  140. Ekman P (2017). Facial expressions In Femandez-Dols J-M & Russell JA (Eds.), The science of facial expression (p. 39–56). NY: Oxford U Press. [Google Scholar]
  141. Ekman P, & Cordaro D (2011). What is meant by calling emotions basic. Emotion Review, 3(4), 364–370. [Google Scholar]
  142. Ekman P, & Friesen WV (1971). Constants across cultures in the face and emotion. Journal of Personality and Social Psychology, 17(2), 124. [DOI] [PubMed] [Google Scholar]
  143. Ekman P, Friesen WV (1976) Pictures of Facial Affect (Consulting Psychologists Press, Palo Alto, CA: ). [Google Scholar]
  144. Ekman P, Friesen WV (1978) Facial Action Coding System: A Technique for the Measurement of Facial Movement (Consulting Psychologists Press, Palo Alto, CA: ). [Google Scholar]
  145. Ekman P, & Oster H (1979). Facial expressions of emotion. Annual Review of Psychology, 30(1), 527–554. [Google Scholar]
  146. Ekman P, Friesen W, & Ellsworth P. (1972). Emotion in the human face: guidelines for research and an integration of findings. Elmsford, N.Y.: Pergamon Press. [Google Scholar]
  147. Ekman P, Friesen WV, and Hager JC (2002). Facial Action Coding System: The Manual On CD ROM (Salt Lake City: The Human Face) [Google Scholar]
  148. Ekman P, Levenson RW, & Friesen WV (1983). Autonomic nervous system activity distinguishes among emotions, Science, 221(4616), 1208–1210. [DOI] [PubMed] [Google Scholar]
  149. Ekman P, Sorenson ER, & Friesen WV (1969). Pan-cultural elements in facial displays of emotion. Science, 164(3875), 86–88. [DOI] [PubMed] [Google Scholar]
  150. Ekman P (1980). The Face of Man. New York: Garland Publishing, Inc. [Google Scholar]
  151. Elfenbein H, Beaupre M, Levesque M, & Hess U (2007). Toward a dialect theory: cultural differences in the expression and recognition of posed facial expressions. Emotion, 7, 131–146. [DOI] [PubMed] [Google Scholar]
  152. Elfenbein HA (2013). Nonverbal dialects and accents in facial expressions of emotion. Emotion Review, 5, 90–96. [Google Scholar]
  153. Elfenbein HA (2017). Emotional dialects in the language of emotion In Fernandez-Dols J-M & Russell JA (Eds.), The science of facial expression (p. 479–496). NY: Oxford U Press. [Google Scholar]
  154. Elfenbein HA, & Ambady N (2002). On the universality and cultural specificity of emotion recognition: a meta-analysis. Psychological Bulletin, 128(2), 203–235. [DOI] [PubMed] [Google Scholar]
  155. Emde RN, & Koenig KL (1969). Neonatal smiling and rapid eye movement states. Journal of the American Academy of Child Psychiatry, 8(1), 57–67. [DOI] [PubMed] [Google Scholar]
  156. Essa IA, & Pentland AP (1997). Coding, analysis, interpretation, and recognition of facial expressions. IEEE transactions on pattern analysis and machine intelligence, 19(7), 757–763. [Google Scholar]
  157. Feldman R (2016). The neurobiology of mammalian parenting and the biosocial context of human caregiving. Hormones and Behavior, 77, 3–17. [DOI] [PubMed] [Google Scholar]
  158. Feng D, Jeong D, Krämer N, Miller L and Marsella S "Is It Just Me?: Evaluating Attribution of Negative Feedback as a Function of Virtual Instructor's Gender and Proxemics", in Proceedings of the 16th Conference on Autonomous Agents and MultiAgent Systems, May 2017, pp. 810–818, International Foundation for Autonomous Agents and Multiagent Systems [Google Scholar]
  159. Fernandez-Dols J-M. (2017). Natural facial expression: A view from psychological construction and pragmatics In Fernandez-Dols J-M & Russell JA(Eds.), The science of facial expression (p. 457–378). NY: Oxford U Press. [Google Scholar]
  160. Fernandez-Dols J-M, Sanchez F, Carrera P, & Ruiz-Belda M-A. (1997). Are spontaneous expressions and emotions linked? An experimental test of coherence. Journal of Nonverbal Behavior, 21, 163–177. [Google Scholar]
  161. Fernández-Dols J-M, & Crivelli C (2013). Emotion and expression: Naturalistic studies. Emotion Review, 5(1), 24–29. doi: 10.1177/1754073912457229 [DOI] [Google Scholar]
  162. Fernández-Dols J-M, & Ruiz-Belda M-A (1995). Are smiles a sign of happiness? Gold medal winners at the Olympic Games. Journal of Personality and Social Psychology, 69(6), 1113–1119. [Google Scholar]
  163. Fiorentini C & Viviani P (2011). Is there a dynamic advantage for facial expressions? Journal of Vision, 11, 1–15. [DOI] [PubMed] [Google Scholar]
  164. Firestein S (2012). Ignorance: How it drives science: Oxford University Press. [Google Scholar]
  165. Flom R, & Bahrick LE (2007). The development of infant discrimination of affect in multimodal and unimodal stimulation: The role of intersensory redundancy. Developmental psychology, 43(1), 238. [DOI] [PMC free article] [PubMed] [Google Scholar]
  166. Frank MG, & Stennett J (2001). The forced-choice paradigm and the perception of facial expressions of emotion. Journal of Personality and Social Psychology, 80, 75–85. [DOI] [PubMed] [Google Scholar]
  167. Fridlund A (2017). The behavioral ecology view of facial displays, 25 years later In Fernandez-Dols J-M & Russell JA (Eds.), The science of facial expression (p. 77–92). NY: Oxford U Press. [Google Scholar]
  168. Fridlund AJ (1991). Sociality of solitary smiling: Potentiation by an implicit audience. Journal of Personality and Social Psychology, 60(2), 229–240. [Google Scholar]
  169. Fridlund AJ (1994). Human facial expression: An evolutionary view. San Diego, CA: Academic Press. [Google Scholar]
  170. Friedman BH (2010). Feelings and the body: the Jamesian perspective on autonomic specificity of emotion. Biological Psychology, 84(3), 383–393. [DOI] [PubMed] [Google Scholar]
  171. Galati D, Sini B, Schmidt S, & Tinti C. (2003). Spontaneous facial expressions in congenitally blind and sighted children aged 8–11. Journal of Visual Impairment & Blindness, 97, 418–428. [Google Scholar]
  172. Galati D, Miceli R, & Sini B (2001). Judging and coding facial expression of emotions in congenitally blind children. International Journal of Behavioral Development, 25(3), 268–278. doi: 10.1080/01650250042000393 [DOI] [Google Scholar]
  173. Galati D, Scherer KR, & Ricci-Bitti PE (1997). Voluntary facial expression of emotion: comparing congenitally blind with normally sighted encoders. Journal of Personality and Social Psychology, 73(6), 1363–1379. [DOI] [PubMed] [Google Scholar]
  174. Gandhi TK, Singh AK, Swami P, Ganesh S, & Sinha P (2017). Emergence of categorical face perception after extended early-onset blindness. Proceedings of the National Academy of Sciences, 114 (23) 6139–6143. [DOI] [PMC free article] [PubMed] [Google Scholar]
  175. Gendron M, & Barrett LF (2009). Reconstructing the past: A century of ideas about emotion in psychology. Emotion Review, 1, 316–339. PMC2835158 [DOI] [PMC free article] [PubMed] [Google Scholar]
  176. Gendron M, & Barrett LA (2017). Facing the past: A history of the face in psychological research on emotion perception In Fernandez-Dols J-M & Russell JA (Eds.), The science of facial expression (p. 15–36). NY: Oxford U Press. [Google Scholar]
  177. Gendron M, Crivelli C, & Barrett LF (2018). Universality reconsidered: Diversity in meaning making of facial expressions. Current Directions in Psychological Science. [DOI] [PMC free article] [PubMed] [Google Scholar]
  178. Gendron M, Hoemann K, Crittenden AN, Msafiri S, Ruark G, & Barrett LF (2018). Emotion perception in the Hadza hunter-gatherers. PsyArXiv preprint, available https://psyarxiv.com/pf2q3/ [DOI] [PMC free article] [PubMed] [Google Scholar]
  179. Gendron M, Lindquist K, Barsalou L, & Barrett LF (2012). Emotion words shape emotion percepts. Emotion, 12, 314–325. [DOI] [PMC free article] [PubMed] [Google Scholar]
  180. Gendron M, Mesquita B, & Barrett LF (2013). Emotion perception: Putting a face in context Chapter in Reisberg D(Ed.), Oxford Handbook of Cognitive Psychology (p. 539–556). New York: Oxford University Press. [Google Scholar]
  181. Gendron M, Roberson D, & Barrett LF (2015). Cultural variation in emotion perception is real: A response to Sauter et al. Psychological Science, 26, 357–359. [DOI] [PMC free article] [PubMed] [Google Scholar]
  182. Gendron M, Roberson D, van der Vyver JM, & Barrett LF (2014a). Perceptions of emotion from facial expressions are not culturally universal: evidence from a remote culture. Emotion, 14(2), 251–262. doi: 10.1037/a0036052 [DOI] [PMC free article] [PubMed] [Google Scholar]
  183. Gendron M, Roberson D, van der Vyver JM, & Barrett LF (2014b). Cultural relativity in perceiving emotion from vocalizations. Psychological Science, 25, 911–920. [DOI] [PMC free article] [PubMed] [Google Scholar]
  184. Gewald J-B (2010). Remote but in contact with history and the world. Proceedings of the National Academy of Sciences, 107(18), E75. doi: 10.1073/pnas.1001284107 [DOI] [PMC free article] [PubMed] [Google Scholar]
  185. Gilbert DT (1998). Ordinary personology In Fiske ST & Gardner L (Eds.), The handbook of social psychology (Vol. 1 and 2, pp. 89–150). New York, NY: McGraw-Hill. [Google Scholar]
  186. Gold JM, Barker JD, Barr S, Bittner JL, Bromfield WD, Chu N, et al. (2013). The efficiency of dynamic and static facial expression recognition. Journal of Vision, 13(5): 23, 1–12 [DOI] [PMC free article] [PubMed] [Google Scholar]
  187. Goldstone R (1994). An efficient method for obtaining similarity data. Behavior Research Methods, Instruments, & Computers, 1994; 26:381–386. [Google Scholar]
  188. Goldstone RL, Steyvers M, & Rogosky BJ (2003). Conceptual interrelatedness and caricatures. Memory & Cognition, 31, 169–180. [DOI] [PubMed] [Google Scholar]
  189. Granchrow JR, Steiner JE, & Daher M (1983). Neonatal facial expressions in response to different qualities and intensities of gustatory stimulation. Infant Behavior & Development, 6, 153–157. [Google Scholar]
  190. Gross CT, & Canteras NS (2012). The many paths to fear. Nature Reviews Neuroscience, 13(9), 651–658. [DOI] [PubMed] [Google Scholar]
  191. Gross D, & Harris PL (1988). False beliefs about emotion: Children's understanding of misleading emotional displays. International Journal of Behavioral Development, 11(4), 475–488. [Google Scholar]
  192. Gross JJ, & Barrett LF (2011). Emotion generation and emotion regulation: One or two depends on your point of view. Emotion Review, 3, 8–16. PMC3072688 [DOI] [PMC free article] [PubMed] [Google Scholar]
  193. Grossmann T (2010). The development of emotion perception in face and voice during infancy. Restor Neurol Neurosci, 28(2):219–36 [DOI] [PubMed] [Google Scholar]
  194. Grossmann T (2015). The development of social brain functions in infancy. Psychological Bulletin, 141(6), 1266–1287. doi: 10.1037/bul0000002 [DOI] [PubMed] [Google Scholar]
  195. Guillory SA & Bujarski KA (2014). Exploring emotion using invasive methods: Review of 60 years of human intracranial electrophysiology. Social Cognitive and Affective Neuroscience, doi: 10.1093/scan/nsu002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  196. Gunnery SD, & Hall JA (2014). The Duchenne Smile and Persuasion. Journal of Nonverbal Behavior, 38(2), 181–194. doi: 10.1007/s10919-014-0177-1 [DOI] [Google Scholar]
  197. Gunnery SD, Hall JA, & Ruben MA (2013). The deliberate Duchenne smile: Individual differences in expressive control. Journal of Nonverbal Behavior, 37(1), 29–41. doi: 10.1007/s10919-012-0139-4 [DOI] [Google Scholar]
  198. Gweon H, & Schulz LE (2011). 16-month-olds rationally infer causes of failed actions. Science, 332(6037), 1524. [DOI] [PubMed] [Google Scholar]
  199. Gweon H, Tenenbaum J, & Schulz LE (2010). Infants consider both the sample and the sampling process in inductive generalization. Proceedings of the National Academy of Sciences, 107(20), 9066–9071, [DOI] [PMC free article] [PubMed] [Google Scholar]
  200. Haidt J, & Keltner D (1999). Culture and facial expression: Open-ended methods find more expressions and a gradient of recognition. Cognition & Emotion, 13(3), 225–266. [Google Scholar]
  201. Hao Wang, Peng, & Ji, 2018. Facial action unit recognition augmented by their dependencies. Presented at 2018 13th IEEE International Conference on Automatic Face and Gesture Recognition. [Google Scholar]
  202. Harris PL (1994). The child's understanding of emotion: Developmental change and the family environment. Journal of Child Psychology and Psychiatry, 35(1), 3–28. [DOI] [PubMed] [Google Scholar]
  203. Hata T, Hanaoka U, Mashima M, Ishimura M, Marumo G, & Kanenishi K (2013). Four-dimensional HDlive rendering image of fetal facial expression: a pictorial essay. Journal of Medical Ultrasonics, 40(4), 437–441. [DOI] [PubMed] [Google Scholar]
  204. Haviland JM, & Lelwica M (1987). The induced affect response: 10-week-old infants' responses to three emotion expressions. Developmental Psychology, 23(1), 97–104.. [Google Scholar]
  205. Henrich I, Heine SJ, & Norenzayan A (2010). The weirdest people in the world? Behav Brain Sci, 33(2–3), 61–83; discussion 83–135. doi: 10.1017/S0140525X0999152X [DOI] [PubMed] [Google Scholar]
  206. Hertenstein MI & Campos JJ (2004). The retention effects of an adult’s emotional displays on infant behavior. Child Development, 75: 595–613. [DOI] [PubMed] [Google Scholar]
  207. Hess U, & Hareli S (2017). The social signal value of emotions: The role of contextual factors in social inferences drawn from emotion displays In Fernandez-Dols J-M & Russell JA (Eds.), The science of facial expression (p. 375–392). NY: Oxford U Press. [Google Scholar]
  208. Hicks RR (2012). Forty studies that changed psychology: Explorations into the history of psychological research, 6th Ed. Pearson, Prentice Hall. [Google Scholar]
  209. Hjortsjö CH (1969). Man's face and mimic language. Studen litteratur. [Google Scholar]
  210. Hoehl S, & Striano T (2008). Neural Processing of Eye Gaze and Threat-Related Emotional facial Expressions in Infancy. Child Development, 79(6), 1752–1760. [DOI] [PubMed] [Google Scholar]
  211. Hoemann K, Crittenden AN, Ruark G, Gendron M, & Barrett LF (in press). Context facilitates the cross-cultural perception of emotion. Emotion. [DOI] [PMC free article] [PubMed] [Google Scholar]
  212. Holodynski M, & Friedlmeier W (2006). Development of emotions and emotion regulation (Vol. 8). Springer Science & Business Media. [Google Scholar]
  213. Hortensius R, Schutter DJ, & Harmon-Jones E (2012). When anger leads to aggression: induction of relative left frontal cortical activity with transcranial direct current stimulation increases the anger–aggression relationship. Soc Cogn Affect Neurosci, 7(3), 342–347. [DOI] [PMC free article] [PubMed] [Google Scholar]
  214. Hout MC, Goldinger SD, & Ferguson RW (2013). The versatility of SpAM: A fast, efficient, spatial method of data collection for multidimensional scaling. Journal of Experimental Psychology: General, 142(1), 256–281. [DOI] [PMC free article] [PubMed] [Google Scholar]
  215. Hoyt C, Blascovich J, & Swinth K (2003). Social inhibition in immersive virtual environments. Presence, 12(2), 183–195. [Google Scholar]
  216. Hutto JR, Vattoth S (2015), "A practical review of the muscles of facial mimicry with special emphasis on the superficial musculoaponeurotic system." American Journal of Radiology 204: W19–W26. [DOI] [PubMed] [Google Scholar]
  217. Iwata J, & LeDoux JE (1988). Dissociation of associative and nonassociative concomitants of classical fear conditioning in the freely behaving rat. Behavioral Neuroscience, 102(1), 66–76. [DOI] [PubMed] [Google Scholar]
  218. Izard CE (2007). Basic emotions, natural kinds, emotion schemas, and a new paradigm. Perspectives on Psychological Science, 2, 260–280. [DOI] [PubMed] [Google Scholar]
  219. Izard CE (1971). The face of emotion. Appleton-Century-Crofts. [Google Scholar]
  220. Izard CE (2010). The many meanings/aspects of emotion: Definitions, functions, activation, and regulation. Emotion Review, 2(4), 363–370. [Google Scholar]
  221. Izard CE, & Malatesta CZ (1987). Perspectives on emotional development: Differential emotions theory of early emotional development In Osofsky JD (Ed.), Handbook of infant development (2nd ed., pp. 494–554). New York: Wiley Interscience. [Google Scholar]
  222. Izard CE, Fantauzzo CA, Castle JM, Haynes OM, Rayias MF, & Putnam PH (1995). The ontogeny and significance of infants' facial expressions in the first 9 months of life. Developmental Psychology, 31(6), 997. [Google Scholar]
  223. Izard CE, Hembree E, & Huebner R (1987). Infants’ emotional expressions to acute pain: Developmental changes and stability of individual differences. Developmental Psychology, 23, 105–113. [Google Scholar]
  224. Izard CE, Hembree E, Dougherty L, & Spirrizi C (1983). Changes in 2-to 19-month-old infants’ facial expressions following acute pain. Developmental Psychology, 19(3), 418–426. [Google Scholar]
  225. Izard CE, Woodburn EM, & Finlon KJ (2010). Extending emotion science to the study of discrete emotions in infants. Emotion Review, 2(2), 134–136. [Google Scholar]
  226. Jack RE, & Schyns PG (2017). Toward a social psychophysics of face communication. Annual Review of Psychology, 68, 269–297. doi: 10.1146/annurev-psych-010416-044242 [DOI] [PubMed] [Google Scholar]
  227. Jack RE, Crivelli C, & Wheatley T (2018). Data-Driven Methods to Diversify Knowledge of Human Psychology. Trends in Cognitive Science, 22(1), 1–5. doi: 10.1016/j.tics.2017.10.002 [DOI] [PubMed] [Google Scholar]
  228. Jack RE, Sun W, Delis I, Garrod OG, & Schyns PG (2016). Four not six: Revealing culturally common facial expressions of emotion. Journal of Experimental Psychology: General, 145(6), 708–730. doi: 10.1037/xge0000162 [DOI] [PubMed] [Google Scholar]
  229. James W (1894). The physical basis of emotion. Psychological Review, 1, 516–529. [DOI] [PubMed] [Google Scholar]
  230. James W (1890/2007). The Principles of Psychology (Vol. 1). New York: Dover. [Google Scholar]
  231. Jeni Laszlo, Cohn Jeffrey F, & De la Torre Fernando. (2013). Facing imbalanced data recommendations for the use of performance metrics Proceedings of the Affective Computing and Intelligent Interaction, Geneva, Switzerland. Los Alamitos, CA: IEEE Computer Society. [DOI] [PMC free article] [PubMed] [Google Scholar]
  232. Jones NB (2016). Demography and Evolutionary Ecology of Hadza Hunter-Gatherers (Vol. 71): Cambridge University Press. [Google Scholar]
  233. Kamachi M, Bruce V, Mukaida S, Gyoba J, Yoshikawa S, & Akamatsu S (2001). Dynamic properties influence the perception of facial expressions. Perception, 30, 875–887. [DOI] [PubMed] [Google Scholar]
  234. Kärtner J, Holodynski M, & Wörmann V (2013). Parental ethnotheories, social practice and the culture-specific development of social smiling in infants. Mind, Culture, and Activity, 20(1), 79–95. [Google Scholar]
  235. Kayyal MH, & Russell JA (2013). Americans and Palestinians Judge Spontaneous facial Expressions of Emotion. Emotion, 13(5), 891–904. doi: 10.1037/a0033244 [DOI] [PubMed] [Google Scholar]
  236. Keltner D, & Cordaro DT (2017). Understanding multimodal emotional expressions: Recent advances in Basic Emotion Theory In Femandez-Dols J-M & Russell JA(Eds.), The science of facial expression (p. 57–75). NY: Oxford U Press. [Google Scholar]
  237. Keltner D (1995). Signs of appeasement: Evidence for the distinct displays of embarrassment, amusement, and shame. Journal of Personality and Social Psychology, 68(3), 441–454. [Google Scholar]
  238. Kim J, Hill R, Durlach P, Lane H, Forbell E, Core M, Marsella S, Pynadath D, and and Hart J (2009). "BiLAT: A Game-Based Environment for Practicing Negotiation in a Cultural Context. Special Issue on Ill-Defined Domains", International Journal of AI in Education, vol. 19, no. 3, pp. 289–308 [Google Scholar]
  239. Kim Y, Baylor AL & Shen E (2007). Pedagogical agents as learning companions: The impact of agent affect and gender. Journal of Computer Assisted Learning (JCAL), 23(3), 220–532. [Google Scholar]
  240. Kobiella A, Grossmann T, Reid VM, & Striano T (2008). The discrimination of angry and fearful facial expressions in 7-month-old infants: An event-related potential study. Cognition and Emotion, 22(1), 134–146. [Google Scholar]
  241. Koster-Hale J, Bedny M, & Saxe R (2014). Thinking about seeing: perceptual sources of knowledge are encoded in the theory of mind brain regions of sighted and blind adults. Cognition, 133(1), 65–78. doi: 10.1016/j.cognition.2014.04.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  242. Kouo JL, & Egel AL (2016). The Effectiveness of Interventions in Teaching Emotion Recognition to Children with Autism Spectrum Disorder. Review Journal of Autism and Developmental Disorders, 3(3), 254–265. doi: 10.1007/s40489-016-0081-1 [DOI] [Google Scholar]
  243. Kragel PA & LaBar KS (2013). Multivariate pattern classification reveals autonomic and experiential representations of discrete emotions. Emotion, 13, 681–90. [DOI] [PMC free article] [PubMed] [Google Scholar]
  244. Kragel PA & LaBar KS (2015). Multivariate neural biomarkers of emotional states are categorically distinct. Social Cognitive and Affective Neuroscience, 10, 1437–1448. [DOI] [PMC free article] [PubMed] [Google Scholar]
  245. Kreibig SD (2010). Autonomic nervous system activity in emotion: a review. Biological Psychology, 84(3), 394–421. doi: 10.1016/j.biopsycho.2010.03.010 [DOI] [PubMed] [Google Scholar]
  246. Krumhuber EG & Kappas, (2005). Moving smiles: The role of dynamic components for the perception of the genuiness of smiles. Journal of Nonverbal Behavior, 29, 3–24. [Google Scholar]
  247. Krumhuber EG, Kappas A, & Manstead ASR (2013). Effects of dynamic aspects of facial expressions: A review. Emotion Review, 5, 41–46. [Google Scholar]
  248. Krumhuber E, Manstead ASR, Cosker D, Marshall D, & Rosin PL (2009). Effects of dynamic attributes of smiles in human and synthetic faces: A simulated job interview setting. Journal of Nonverbal Behavior, 33, 1–15. [Google Scholar]
  249. Krumhuber E, Manstead A, Cosker D, Marshall D, Rosin PL, & Kappas A (2007). Facial dynamics as indicators of trustworthiness and cooperative behavior. Emotion, 7(4), 730–735. [DOI] [PubMed] [Google Scholar]
  250. Kuhn T (1962). The structure of scientific revolutions. Chicago, IL: University of Chicago Press. [Google Scholar]
  251. Kuppens P, Tuerlinckx F, Russell JA and Barrett LF (2013). The relationship between valence and arousal in subjective experience. Psychological Bulletin, 139, 917–940. [DOI] [PubMed] [Google Scholar]
  252. Kuppens P, Van Mechelen I, & Rijmen F (2008). Toward disentangling sources of individual differences in appraisal and anger. Journal of Personality, 76(4), 969–1000. [DOI] [PubMed] [Google Scholar]
  253. Kuppens P, Van Mechelen I, Smits DJM, De Boeck P, & Ceulemans E (2007). Individual differences in patterns of appraisal and anger experience. Cognition & Emotion, 21(4), 689–713. doi:Doi 10.1080/02699930600859219 [DOI] [Google Scholar]
  254. Landau B, & Shipley E (2001). Labelling patterns and object naming. Developmental Science, 4(1), 109–118. [Google Scholar]
  255. Lang PJ, Greenwald MK, Bradley MM, & Hamm AO (1993). Looking at pictures: Affective, facial, visceral, and behavioral reactions Psychophysiology, 30, 261–273. [DOI] [PubMed] [Google Scholar]
  256. Lazarus R (1991). Emotion and Adaptation. NY, Oxford University Press. [Google Scholar]
  257. Lebois LAM, Wilson-Mendenhall CD, Simmons WK, Barrett LF, & Barsalou LW (2018). Learning situated emotions. Neuropsychologia. 10.1016/j.neuropsychologia.2018.01.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  258. Lee DH & Anderson AK (2017). Form and function of facial expressive origins In Fernandez-Dols J-M & Russell JA(Eds.), The science of facial expression (p. 173–194). NY: Oxford U Press. [Google Scholar]
  259. Lee DH, Mirza R, Flanagan JG, & Anderson AK (2014). Optical origins of opposing facial expression actions. Psychological Science, 25(3), 745–752. [DOI] [PubMed] [Google Scholar]
  260. Leitzke BT, & Pollak SD (2016). Developmental changes in the primacy of facial cues for emotion recognition. Developmental Psychology, 52(4), 572. [DOI] [PMC free article] [PubMed] [Google Scholar]
  261. Leppänen JM, & Nelson CA (2009). Tuning the developing brain to social signals of emotions. Nature Reviews Neuroscience, 10, 37–47. [DOI] [PMC free article] [PubMed] [Google Scholar]
  262. Leppänen JM, Moulson MC, Vogel-Farley VK, & Nelson CA (2007). An ERP study of emotional face processing in the adult and infant brain. Child Development, 78(1), 232–245. [DOI] [PMC free article] [PubMed] [Google Scholar]
  263. Leppänen JM, Richmond J, Vogel-Farley VK, Moulson MC, & Nelson CA (2009). Categorical representation of facial expressions in the infant brain. Infancy, 14(3), 346–362. [DOI] [PMC free article] [PubMed] [Google Scholar]
  264. Levenson RW (2011). Basic emotion questions. Emotion Review, 3(4), 379–386. [Google Scholar]
  265. Lewis M, & Sullivan MW (Eds.). (2014). Emotional development in atypical children. Psychology Press. [Google Scholar]
  266. Lewis M, Ramsay DS, & Sullivan MW (2006). The relation of ANS and HPA activation to infant anger and sadness response to goal blockage. Developmental Psychobiology, 48(5), 397–405. [DOI] [PMC free article] [PubMed] [Google Scholar]
  267. Lewis M, Sullivan MW, & Kim HMS (2015). Infant approach and withdrawal in response to a goal blockage: Its antecedent causes and its effect on toddler persistence. Developmental Psychology, 51(11), 1553. [DOI] [PMC free article] [PubMed] [Google Scholar]
  268. Lindquist K, & Barrett LF (2008). Emotional complexity. Chapter in Lewis M, Haviland-Jones JM, and Barrett LF (Eds.), The handbook of emotion, 3rd Edition (p. 513–530). New York: Guilford. [Google Scholar]
  269. Lindquist KA, Gendron M, Barrett LF, & Dickerson BC (2014). Emotion perception but not affect perception is impaired with semantic memory loss. Emotion, 14, 375–87. [DOI] [PMC free article] [PubMed] [Google Scholar]
  270. Lindquist KA, Satpute AB, Wager TD, Weber J, & Barrett LF (2016). The brain basis of positive and negative affect: Evidence from a meta-analysis of the human neuroimaging literature. Cerebral Cortex. 26, 1910–22. [DOI] [PMC free article] [PubMed] [Google Scholar]
  271. Lindquist KA, Wager TD, Kober H, Bliss-Moreau E, & Barrett LF (2012). The brain basis of emotion: a meta-analytic review. Behavioral and Brain Sciences, 35(3), 121–143. [DOI] [PMC free article] [PubMed] [Google Scholar]
  272. Marsella S & Gratch J (2016). Computational models of emotion as psychological tools In Barrett LF, Lewis M and Haviliand-Jones J (Eds.) Handbook of Emotions, 4th Edition. New York: Guilford Press. [Google Scholar]
  273. Marsella S, Johnson WL, and LaBore C (2000). Interactive pedagogical drama. In Proceedings of the fourth international conference on Autonomous agents (AGENTS '00) ACM, New York, NY, USA, 301–308. [Google Scholar]
  274. Martin NG, Maza L, McGrath SJ, & Phelps AE (2014). An examination of referential and affect specificity with five emotions in infancy. Infant Behavior and Development, 37(3), 286–297. [DOI] [PubMed] [Google Scholar]
  275. Martin J, Rychlowska M, Wood A, and Niedenthal P (2017). Smiles as multipurpose social signals. Trends in Cognitive Sciences, 21, 864–877. [DOI] [PubMed] [Google Scholar]
  276. Martinez AM (2017a). Computational models of face perception. Current Directions in Psychological Science, 26(3), 263–269. [DOI] [PMC free article] [PubMed] [Google Scholar]
  277. Martinez AM (2017b). Visual perception of facial expressions of emotion. Current Opinion in Psychology, 17, 27–33. [DOI] [PMC free article] [PubMed] [Google Scholar]
  278. Martinez A, & Du S (2012). A model of the perception of facial expressions of emotion by humans: Research overview and perspectives. Journal of Machine Learning Research, 13, 1589–1608. [PMC free article] [PubMed] [Google Scholar]
  279. Matias R, & Cohn JF (1993). Are max-specified infant facial expressions during face-to-face interaction consistent with differential emotions theory? Developmental Psychology, 29(3), 524. [Google Scholar]
  280. Matsumoto D (1990). Cultural similarities and differences in display rules. Motivation and Emotion, 14, 195–214. [Google Scholar]
  281. Matsumoto D & Willingham B (2006). The thrill of victory and the agony of defeat: Spontaneous expressions of medal winners of the 2004 Athens Olympic Games. Journal of Personality and Social Psychology, 91, 568–581. [DOI] [PubMed] [Google Scholar]
  282. Matsumoto D, & Willingham B (2009). Spontaneous facial expressions of emotion of congenitally and noncongenitally blind individuals. Journal of Personality and Social Psychology, 96(1), 1–10. [DOI] [PubMed] [Google Scholar]
  283. Matsumoto D, Keltner D, Shiota M, Frank M, & O'Sullivan M (2008). Facial expressions of emotions In Lewis M, Haviland-Jones JM, & Barrett LF (Eds.), Handbook of Emotion, 3rd Ed. (pp. 211–234). New York: Macmillan. [Google Scholar]
  284. Mayr E (2004). What makes biology unique? Considerations on the autonomy of a scientific discipline. New York: Cambridge University Press. [Google Scholar]
  285. McCall C, Blascovich J, Young A, & Persky S (2009). Proxemic behaviors as predictors of aggression towards Black (but not White) males in an immersive virtual environment. Social Influence, 4(2), 138–154. [Google Scholar]
  286. Medin DL, Ojalehto B, Marin A, & Bang M (2017). Systems of (non-) diversity. Nature Human Behaviour, 1, 0088. [Google Scholar]
  287. Mesquita B & Frijda NH (1992). Cultural variations in emotions: A review. Psychological Bulletin, 2, 179–204. [DOI] [PubMed] [Google Scholar]
  288. Mesquita B, Boiger M, & De Leersnyder J (2016). The cultural construction of emotions. Current Opinion in Psychology, 8, 31–36. [DOI] [PubMed] [Google Scholar]
  289. Messinger DS (2002). Positive and negative: Infant facial expressions and emotions. Current Directions in Psychological Science, 11(1), 1–6. [Google Scholar]
  290. Michel GF, Camras LA, & Sullivan J (1992). Infant interest expressions as coordinative motor structures. Infant Behavior and Development, 15(3), 347–358. [Google Scholar]
  291. Miles L and Johnston L (2007). Detecting Happiness: Perceiver Sensitivity to Enjoyment and Non-Enjoyment Smiles. Journal of Nonverbal Behavior, 31, 259–275. [Google Scholar]
  292. Miller GA (1995). WordNet: a lexical database for English. Communications of the ACM, 38(11), 39–41. [Google Scholar]
  293. Mobbs D, Weiskopf N, Lau EL, Featherstone E, Dolan R, & Frith C (2006). The Kuleshov Effect: The influence of contextual framing on emotional attributions. Social Cognitive and Affective Neuroscience, 1, 95–106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  294. Mollahosseini A, Hassani B, Salvador MJ, Abdollah H, Chan D, & Mahoor MN (2016). Facial expression recognition from world wild web. arXiv preprint, available at https://arxiv.org/abs/1605.03639 [Google Scholar]
  295. Montagu J (1994). The expression of the passions: The origin and influence of Charles Le Brun's lecture on general and particular expression. London: Yale University Press. [Google Scholar]
  296. Montague DP, & Walker-Andrews AS (2001). Peekaboo: a new look at infants' perception of emotion expressions. Developmental Psychology, 37(6), 826. [PubMed] [Google Scholar]
  297. Moses LJ, Baldwin DA, Rosicky JG, & Tidball G (2001). Evidence for referential understanding in the emotions domain at twelve and eighteen months. Child Development, 72(3), 718–735. [DOI] [PubMed] [Google Scholar]
  298. Motley MT, & Camden CT (1988). Facial expressions of emotion: A comparison of posed expressions versus spontaneous expressions in an interpersonal communication setting. Western Journal of Speech Communication, 52, 1–22. [Google Scholar]
  299. Mumme DL, Fernald A, & Herrera C (1996). Infants’ responses to facial and vocal emotional signals in a social referencing paradigm. Child Development, 67(6), 3219–3237. [PubMed] [Google Scholar]
  300. Müri RM (2015). Cortical control of facial expression. The Journal of Comparative Neurology 524, 1578–1585. [DOI] [PubMed] [Google Scholar]
  301. Murphy GL (2002). The big book of concepts. Cambridge, MA: MIT Press. [Google Scholar]
  302. Naab PJ, & Russell JA (2007). Judgments of emotion from spontaneous facial expressions of New Guineans. Emotion, 7(4), 736–744. [DOI] [PubMed] [Google Scholar]
  303. Namba S, Makihara S, Kabir RS, Miyatani M, & Nakao T (2016). Spontaneous facial expressions are different from posed facial expressions: Morphological properties and dynamic sequences. Current Psychology, 36, 593–605. [Google Scholar]
  304. Neidle Carol, Kegl Judy A., Dawn MacLaughlin Benjamin Bahan & Lee Robert G. 2000. The syntax of American Sign Language. Cambridge: MIT Press. [Google Scholar]
  305. Nelson KE, Welsh JA, Trup EMV, & Greenberg MT (2011). Language delays of impoverished preschool children in relation to early academic and emotion recognition skills. First Language, 31, 164–194. [Google Scholar]
  306. Nelson NL & Russell JA (2011). Preschoolers' use of dynamic facial, bodily, and vocal cues to emotion. Journal of Experimental Child Psychology, 110, 52–61. [DOI] [PubMed] [Google Scholar]
  307. Nelson NL, & Russell JA (2013). Universality revisited. Emotion Review, 5, 8–15. [Google Scholar]
  308. Nelson NL, & Russell JA (2016). Building emotion categories: Children use a process of elimination when they encounter novel expressions. Journal of Experimental Child Psychology, 151, 120–130. [DOI] [PubMed] [Google Scholar]
  309. Nelson NL, Hudspeth K, & Russell JA (2013). A story superiority effect for disgust, fear, embarrassment, and pride. British Journal of Developmental Psychology, 31(3), 334–348. [DOI] [PubMed] [Google Scholar]
  310. Neth D, & Martinez AM (2009). Emotion perception in emotionless face images suggests a norm-based representation. Journal of Vision, 9(1), 5. [DOI] [PubMed] [Google Scholar]
  311. Nezlek JB, Vansteelandt K, Van Mechelen I, & Kuppens P (2008). Appraisal-emotion relationships in daily life. Emotion, 8(1), 145–150. [DOI] [PubMed] [Google Scholar]
  312. Ngo N & Isaacowitz DM (2015). Use of context in emotion perception: The role of top-down control, cue type, and perceiver's age. Emotion, 75, 292–302. [DOI] [PubMed] [Google Scholar]
  313. Niewiadomski R, Ding Y, Mancini M, Pelachaud C, Volpe G, Camurri A (2015). Perception of intensity incongruence in synthesized multimodal expressions of laughter, Affective Computing and Intelligent Interaction (ACII), 2015 International Conference on, Xi'an, pages 684–690. [Google Scholar]
  314. Norenzayan A, & Heine SJ (2005). Psychological universals: What are they and how can we know? Psychological Bulletin, 131(5), 763–784. [DOI] [PubMed] [Google Scholar]
  315. Obrist PA (1981). Cardiovascular psychophysiology: A perspective. New York: Plenum. [Google Scholar]
  316. Obrist PA, Webb RA, Sutterer JR, & Howard JL (1970). The cardiac-somatic relationship: some reformulations. Psychophysiology, 6(5), 569–587. [DOI] [PubMed] [Google Scholar]
  317. Ochs M, Niewiadomski R and Pelachaud C. (2010). How a virtual agent should smile?: morphological and dynamic characteristics of virtual agent's smiles. In Proceedings of the 10th international conference on Intelligent virtual agents (IVA'10), Springer-Verlag, Berlin, Heidelberg, 427–440. [Google Scholar]
  318. Ortony A, Clore GL, Collins A (1990). The cognitive structure of emotions. Cambridge University Press. [Google Scholar]
  319. Osgood CE, May WH, & Miron MS (1975). Cross-cultural universals of affective meaning. Urbana: University of Illinois Press. [Google Scholar]
  320. Oster H, Hegley D, & Nagel L (1992). Adult judgments and fine-grained analysis of infant facial expressions: Testing the validity of a priori coding formulas. Developmental Psychology, 28, 1115. [Google Scholar]
  321. Oster H (2005). The repertoire of infant facial expressions: an ontogenetic perspective In Nadel J & Muir D (Eds.), Emotional development: Recent research advances (pp. 261–292). New York, NY, US: Oxford University Press. [Google Scholar]
  322. Oster H (2003). BabyFACS: Facial Action Coding System for Infants and Young Children. Unpublished manual. New York University. [Google Scholar]
  323. Panksepp J (1998). Affective neuroscience: The foundations of human and animal emotions. New York: Oxford University Press. [Google Scholar]
  324. Parkinson B (1997). Untangling the appraisal- emotion connection. Personality and Social Psychology Review, 1(1), 62–79. [DOI] [PubMed] [Google Scholar]
  325. Parr T (2005). The feelings book. New York: Little, Brown books for young readers. [Google Scholar]
  326. Parr LA, Waller BM, Burrows AM, Gothard KM, and Vick S-J (2010). MaqFACS: A muscle-based facial movement coding system for the rhesus Macaque. American Journal of Physical Anthropology 143, 625–630. [DOI] [PMC free article] [PubMed] [Google Scholar]
  327. Pavlenko A (2014). The bilingual mind: And what it tells us about language and thought. New York: Cambridge University Press. [Google Scholar]
  328. Peltola MJ, Leppänen JM, Palokangas T, Hietanen JK. (2008). Fearful faces modulate looking duration and attention disengagement in 7-month-old infants. Developmental Science, 11, 60–68. [DOI] [PubMed] [Google Scholar]
  329. Perlman SB, Kalish CW, and Pollak SD (2008). The role of maltreatment experience in children's understanding of the antecedents of emotion. Cognition and Emotion, 22 (4), 651–670. [Google Scholar]
  330. Pinker S (1997). How the mind works. New York: Norton. [DOI] [PubMed] [Google Scholar]
  331. Plate RC. Fulvio JM; Shutts K; Green CS; Pollak SD. (2018). Probability learning: Changes in behavior across time and development. Child Development, 89, 205–218. [DOI] [PMC free article] [PubMed] [Google Scholar]
  332. Pollak SD (2009) The emergence of emotion: Experience, development and biology. In Cicchetti D and Gunnar MR (Eds.), Minnesota Symposia on Child Psychology: Meeting the challenge of translational research in child psychology (Vol 35, p. 189–238). Hoboken, NJ: Wiley. [Google Scholar]
  333. Pollak SD. (2015) Multilevel developmental approaches to understanding the effects of child maltreatment: Recent advances and future challenges. Development and Psychopathology, 2November27(4 Pt 2):1387–97. [DOI] [PMC free article] [PubMed] [Google Scholar]
  334. Pollak SD, & Kistler DJ (2002). Early experience is associated with the development of categorical representations for facial expressions of emotion. Proceedings of the National Academy of Sciences, 99(13), 9072–9076. [DOI] [PMC free article] [PubMed] [Google Scholar]
  335. Pollak SD, Messner M, Kistler DJ, & Cohn JF (2009). Development of perceptual expertise in emotion recognition. Cognition, 110(2), 242–247. [DOI] [PMC free article] [PubMed] [Google Scholar]
  336. Pollak SD and Tolley-Schell SA(2003). Selective attention to facial emotion in physically abused children. Journal of Abnormal Psychology, 112, 323–338. [DOI] [PubMed] [Google Scholar]
  337. Pollak SD, & Sinha P (2002). Effects of early experience on children's recognition of facial displays of emotion. Developmental Psychology, 38, 784–791. [DOI] [PubMed] [Google Scholar]
  338. Pollak SD, Cicchetti D, Hornung K, & Reed A (2000). Recognizing emotion in faces: Developmental effects of child abuse and neglect. Developmental Psychology, 36, 679–688. PMID: 9886228 [DOI] [PubMed] [Google Scholar]
  339. Pollak SD, Vardi S, Bechner AMP, and Curtin JJ (2005). Physically abused children's regulation of attention in response to hostility. Child Development, 76(5), 968–977. [DOI] [PubMed] [Google Scholar]
  340. Pruitt DG & Kimmel MJ (1977). Twenty years of experimental gaming: Critique, synthesis, and suggestions for the future. Annual Review Psychology 28, 363–392. [Google Scholar]
  341. Quigley KS, & Barrett LF (2014). Is there reliability and specificity of autonomic changes during emotional episodes? Guidance from the conceptual act theory and psychophysiology. Biological Psychology, 98,82–94. [DOI] [PMC free article] [PubMed] [Google Scholar]
  342. Rehm M & Andre’ E. (2005). Catch me if you can: exploring lying agents in social settings. 4th International Joint Conference on Autonomous Agents and Multiagent Systems (AAMAS), ACM, New York, NY, USA, 937–944. [Google Scholar]
  343. Reissland N, Francis B, & Mason J (2013). Can healthy fetuses show facial expressions of “pain” or “distress”? PLoS One, 8(6), e65530. [DOI] [PMC free article] [PubMed] [Google Scholar]
  344. Reissland N, Francis B, Mason J, & Lincoln K (2011). Do facial expressions develop before birth? PLoS One, 6(8), e24081. [DOI] [PMC free article] [PubMed] [Google Scholar]
  345. Rickel R, Marsella S, Gratch J, Hill R, Traum D and Swartout B (2002). Towards a New Generation of Virtual Humans for Interactive Experiences, in IEEE Intelligent Systems July/August 2002, pp. 32–38. [Google Scholar]
  346. Nevada Riggins v. (1992). 504 U.S. 127 (Kennedy, J., concurring), p. 142. [Google Scholar]
  347. Rinn WE (1984). The neuropsychology of facial expression: a review of the neurological and psychological mechanisms for producing facial expressions. Psychological bulletin, 95(1), 52. [PubMed] [Google Scholar]
  348. Robbins J & Rumsey A (2008). Introduction: Cultural and linguistic anthropology and the opacity of other minds. Anthropological Quarterly, 81, 407–420. [Google Scholar]
  349. Robinson MD, & Clore GL (2002). Belief and feeling: evidence for an accessibility model of emotional self-report. Psychological Bulletin, 128(6), 934. [DOI] [PubMed] [Google Scholar]
  350. Roch-Levecq A-C (2006). Production of basic emotions by children with congenital blindness: Evidence for the embodiment of theory of mind. British Journal of Developmental Psychology, 24(3), 507–528. [Google Scholar]
  351. Roseman IJ (2001). A model of appraisal in the emotion system: Integrating theory, research, and applications In Scherer KR, Schorr A, & Johnstone T (Eds.), Appraisal processes in emotion: Theory, methods, research (pp. 68–91). New York: Oxford University Press. [Google Scholar]
  352. Rosenberg EL, & Ekman P (1994). Coherence between expressive and experiential systems in emotion. Cognition & Emotion, 8, 201–229. [Google Scholar]
  353. Rosenberg EL, & Ekman P (1995). Conceptual and methodological issues in the judgment of facial expressions of emotion. Motivation and Emotion, 19(2), 111–138. [Google Scholar]
  354. Rosenstein D, & Oster H (1988). Differential facial responses to four basic tastes in newborns. Child Development, 59, 1555–1568. [PubMed] [Google Scholar]
  355. Röttger-Rössler B, Scheidecker G, Funk L, & Holodynski M (2015). Learning (by) feeling: A cross-cultural comparison of the socialization and development of emotions. Ethos, 43(2), 187–220. [Google Scholar]
  356. Rozin P, Hammer L, Oster H, Horowitz T, & Marmora V (1986). The child's conception of food: Differentiation of categories of rejected substances in the 16 months to 5 year age range. Appetite, 7, 141–151. [DOI] [PubMed] [Google Scholar]
  357. Ruba AL, Johnson KM, Harris LT, & Wilbourn MP (2017). Developmental changes in infants’ categorization of anger and disgust facial expressions. Developmental Psychology, 53(10), 1826. [DOI] [PubMed] [Google Scholar]
  358. Rudoivic O, Lee J, Dai M, Schuller B, & Picard RW (2018). Personalized machine learning for robot perception of affect and engagement in autism therapy. Science Robotics, 3, eaao6760 June 27, 2018. [DOI] [PubMed] [Google Scholar]
  359. Ruiz-Belda MA, Fernandez-Dols J-M, Carrera P, & Barchard K. (2003). Spontaneous facial expressions of happy bowlers and soccer fans. Cognition and Emotion, 17, 315–326. [DOI] [PubMed] [Google Scholar]
  360. Russell JA (1991). Culture and the categorization of emotions. Psychological Bulletin, 110(3), 426–450. [DOI] [PubMed] [Google Scholar]
  361. Russell JA (1993). Forced-choice response format in the study of facial expressions. Motivation and Emotion, 17, 41–51. [Google Scholar]
  362. Russell JA (1994). Is there universal recognition of emotion from facial expressions? A review of the cross-cultural studies. Psychological Bulletin, 115(1), 102–141. [DOI] [PubMed] [Google Scholar]
  363. Russell JA (1995). Facial expressions of emotion: what lies beyond minimal universality? Psychological Bulletin, 118(3), 379–391. [DOI] [PubMed] [Google Scholar]
  364. Russell JA (2003). Core affect and the psychological construction of emotion. Psychological Review, 110(1), 145–172. [DOI] [PubMed] [Google Scholar]
  365. Russell JA, & Barrett LF (1999). Core affect, prototypical emotional episodes, and other things called emotion: Dissecting the elephant. Journal of Personality and Social Psychology, 76, 805–819. [DOI] [PubMed] [Google Scholar]
  366. Russell JA, & Widen SC (2002a). A label superiority effect in children's categorization of facial expressions. Social Development, 11(1), 30–52. [Google Scholar]
  367. Russell JA, & Widen SC (2002b). Words versus faces in evoking preschool children’s knowledge of the causes of emotions. International Journal of Behavioral Development, 26(2), 97–103. [Google Scholar]
  368. Russell JA, Hosie CD, Gray C, Scott C, Hunter N, Banks JS, & Macaualy MC(1998). The development of theory of mind in deaf children. Journal of Child Psychology and Psychiatry, 39, 903–910. [PubMed] [Google Scholar]
  369. Rychlowska M, Jack RE, Garrod OGB, Schyns PG, Martin JD, & Niedenthal PM (2017). Functional smiles: tools for love, sympathy, and war. Psychological Science, 28, 1259–1270. [DOI] [PMC free article] [PubMed] [Google Scholar]
  370. Rychlowska M, Miyamoto Y, Matsumoto D, Hess U, Gilboa-Schechtman E, Kamble S, Muluk H, Masuda T, Niedenthal PM (2015). Heterogeneity of long-history migration explains cultural differences in reports of emotional expressivity and the functions of smiles. Proceedings of the National Academy of Sciences of the United States of America 112(19), E2429–36. [DOI] [PMC free article] [PubMed] [Google Scholar]
  371. Saarimäki H, Gotsopoulos A, Jaaskelainen IP, Lampinen J, Vuilleumier P, Hari R, Sams M, & Nummenmaa L. (2016). Discrete neural signatures of basic emotions. Cerebral Cortex, 26, 2563–2573. [DOI] [PubMed] [Google Scholar]
  372. Saarni C, Campos JJ, Camras LA, & Witherington D (2006). Emotional development: Action, communication, and understanding In Damon W, Lerner RM, & Eisenberg N (Eds.), Handbook of child psychology, social, emotional, and personality development (Vol. 3, p. 226–299). New York: John Wiley & Sons. [Google Scholar]
  373. Sato W, & Yoshikawa S (2004). The dynamic aspects of emotional facial expressions. Cognition and Emotion, 18, 701–710. [Google Scholar]
  374. Scarantino A (2016). The philosophy of emotions and its impact on affective science In Barrett LF, Lewis M and Haviland-Jones JM (Eds.), Handbook of Emotions, 4th Ed. (p. 3–48). New York: Guilford Press. [Google Scholar]
  375. Scarantino A, & Griffiths P (2011). Don’t give up on basic emotions. Emotion Review, 3(4), 444–454. [Google Scholar]
  376. Scherer KR, Mortillaro M, & Mehu M (2017). Facial expression is driven by appraisal and generates appraisal inferences In Femandez-Dols J-M & Russell JA (Eds.), The Science of Facial Expression (p. 353–374). New York: Oxford U Press. [Google Scholar]
  377. Scherer KR, Mortillaro M, Rotondi I, Sergi I, & Trznadel S (2018). Appraisal-driven facial actions as building blocks for emotion inference. Journal of Personality and Social Psychology, 114(3), 358–379. [DOI] [PubMed] [Google Scholar]
  378. Schwartz GM, Izard CE, & Ansul SE (1985). The 5-month-old's ability to discriminate facial expressions of emotion. Infant Behavior and Development, 8(1), 65–77. [Google Scholar]
  379. Searle JR (1992). The rediscovery of the mind: MIT press. [Google Scholar]
  380. Searle JR (2004). Mind: a brief introduction: Oxford University Press. [Google Scholar]
  381. Sears MS, Repetti RL, Reynolds BM, & Sperling JB (2014). A naturalistic observational study of children's expressions of anger in the family context. Emotion, 14(2), 272. [DOI] [PubMed] [Google Scholar]
  382. Serrano JM, Iglesias J, & Loeches A (1992). Visual discrimination and recognition of facial expressions of anger, fear, and surprise in 4-to 6-month-old infants. Developmental Psychobiology, 25(6), 411–425. [DOI] [PubMed] [Google Scholar]
  383. Shackman JE, & Pollak SD (2014). Impact of physical maltreatment on the regulation of negative affect and aggression. Development and Psychopathology, 26, 1021–1033. [DOI] [PMC free article] [PubMed] [Google Scholar]
  384. Shackman JE, Fatani S, Camras LA, Berkowitz MJ, Bachorowski JA, & Pollak SD (2010). Emotion expression among abusive mothers is associated with their children’s emotion processing and problem behaviours. Cognition and Emotion, 24(8), 1421–1430. [DOI] [PMC free article] [PubMed] [Google Scholar]
  385. Shackman JE, Shackman AJ, and Pollak SD (2007). Physical abuse amplifies attention to threat and increases anxiety in children. Emotion, 7, 838–852. [DOI] [PubMed] [Google Scholar]
  386. Shariff AF, & Tracy J (2011). What are emotion expressions for? Current Directions in Psychological Science, 20(6), 395–399. [Google Scholar]
  387. Shaver P, Schwartz J, Kirson D, & O'Connor C (1987). Emotion knowledge: Further exploration of a prototype approach. Journal of Personality and Social Psychology, 52,1061–1086. [DOI] [PubMed] [Google Scholar]
  388. Shepard RN, & Cooper LA (1992). Representation of colors in the blind, color-blind, and normally sighted. Psychological Science, 3(2), 97–104. [Google Scholar]
  389. Shiota MN, Campos B, & Keltner D (2003). The faces of positive emotion: Prototype displays of awe, amusement, and pride. Annals of the New York Academy of Sciences, 1000, 296–299. [DOI] [PubMed] [Google Scholar]
  390. Siegel EH, Sands MK, Van denNoortgate W, Condon P, Chang Y, Dy J, Quigley KS, & Barrett LF (2018). Emotion fingerprints or emotion populations? A meta-analytic investigation of autonomic features of emotion categories. Psychological Bulletin, 144(4), 343–393. [DOI] [PMC free article] [PubMed] [Google Scholar]
  391. Smith ER, & DeCoster J (2000). Dual-process models in social and cognitive psychology: Conceptual integration and links to underlying memory systems. Personality and Social Psychology Review, 4, 108–131. [Google Scholar]
  392. Smith LB, Jayaraman S, Clerkin E, & Yu C (2018). The developing infant creates a curriculum for statistical learning. Trends in Cognitive Sciences, 10.1016/j.tics.2018.02.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  393. Smith TW (2016). The book of human emotion. Little, Brown & Company: New York, NY. [Google Scholar]
  394. Snyder PJ, Kaufman R, Harrison J, & Maruff P (2010). Charles Darwin's emotional expression “experiment” and his contribution to modern neuropharmacology. Journal of the History of the Neurosciences, 19: 2, 158–170, doi: 10.1080/09647040903506679 [DOI] [PubMed] [Google Scholar]
  395. Soken NH, & Pick AD (1999). Infants' perception of dynamic affective expressions: do infants distinguish specific expressions? Child Development, 70(6), 1275–1282. [DOI] [PubMed] [Google Scholar]
  396. Sorenson ER (1975). Culture and the expression of emotion In Williams TR (Ed.), Psychological Anthropology (pp. 361–372). Chicago: Aldine. [Google Scholar]
  397. Srinivasan R, and Martinez AM (2018). Cross-cultural and cultural-specific production and perception of facial expressions of emotion in the wild. IEEE Transactions on Affective Computing. ArXiv preprint, available at https://arxiv.org/pdf/1808.04399.pdf [Google Scholar]
  398. Sroufe AL (1996). Emotional development: The organization of emotional life in the early years. New York, NY: Cambridge University Press. [Google Scholar]
  399. Stebbins G & Delsarte F (1887/2013). Delsarte system of expression. E. S. Werner. [Google Scholar]
  400. Stein M, Ottenberg P, & Roulet N (1958). A study of the development of olfactory preferences. Archives of Neurology & Psychiatry, 80, 264–266. [DOI] [PubMed] [Google Scholar]
  401. Stemmier G (2004). Physiological processes during emotion In Philippot P & Feldman RS (Eds.), The regulation of emotion (pp. 33–70). Mahwah, NJ: Erlbaum & Associates. [Google Scholar]
  402. Stemmier G, Aue T, & Wacker J (2007). Anger and fear: Separable effects of emotion and motivational direction on somatovisceral responses. International Journal of Psychophysiology, 66(2), 141–153. [DOI] [PubMed] [Google Scholar]
  403. Stenberg CR, Campos JJ, & Emde RN (1983). The facial expression of anger in seven-month-old infants. Child Development, 178–184. [DOI] [PubMed] [Google Scholar]
  404. Sterling P (2012). Allostasis: a model of predictive regulation. Physiology & Behavior, 106(1), 5–15. [DOI] [PubMed] [Google Scholar]
  405. Stephens CL, Christie IC, & Friedman BH (2010). Autonomic specificity of basic emotions: evidence from pattern classification and cluster analysis. Biological Psychology, 84, 463–73. [DOI] [PubMed] [Google Scholar]
  406. Stevenson RJ, Oaten MJ, Case TI, Repacholi BM, & Wagland P (2010). Children’s response to adult disgust elicitors: Development and acquisition. Developmental Psychology, 46(1), 165. [DOI] [PubMed] [Google Scholar]
  407. Sullivan MW, & Lewis M (2003). Contextual determinants of anger and other negative expressions in young infants. Developmental Psychology, 39(4), 693. [DOI] [PMC free article] [PubMed] [Google Scholar]
  408. Susskind JM, Lee DH, Cusi A, Feiman R, Grabski W, & Anderson AK (2008). Expressing fear enhances sensory acquisition. Nature Neuroscience, 11(7), 843–850. [DOI] [PubMed] [Google Scholar]
  409. Tassinary LG, & Cacioppo JT (1992). Unobservable facial actions and emotion. Psychological Science, 3(1), 28–33. [Google Scholar]
  410. Tinbergen Niko. (1953). The Herring Gull's World. London: Collins. [Google Scholar]
  411. Todorov A (2017). Face value: The irresistible influence of first impressions. Princeton, NJ: Princeton University Press. [Google Scholar]
  412. Tooby J, & Cosmides L (2008). The evolutionary psychology of the emotions and their relationship to internal regulatory variables In Lewis M, Haviland-Jones JM, & Barrett LF (Eds.), The Handbook of Emotions (3rd ed.). New York, NY: The Guilford Press. [Google Scholar]
  413. Tovote P, Fadok JP, & Lüthi A (2015). Neuronal circuits for fear and anxiety. Nature Reviews Neuroscience, 16(6), 317–331. [DOI] [PubMed] [Google Scholar]
  414. Tracy JL, & Matsumoto D (2008). The spontaneous expression of pride and shame: Evidence for biologically innate nonverbal displays. Proceedings of the National Academy of Sciences of the United States of America, 105(33), 11655–11660. [DOI] [PMC free article] [PubMed] [Google Scholar]
  415. Tracy JL, & Robins RW (2008). The nonverbal expression of pride: evidence for cross21 cultural recognition. Journal of Personality and Social Psychology, 94(3), 516. [DOI] [PubMed] [Google Scholar]
  416. Tracy J, & Randles D (2011). Four models of basic emotions: A review of Ekman and Cordaro, Izard, Levenson, and Panksepp and Watt. Emotion Review, 3(4), 397–405. [Google Scholar]
  417. Turati C (2004). Why Faces Are Not Special to Newborns An Alternative Account of the Face Preference. Current Directions in Psychological Science, 13(1), 5–8. [Google Scholar]
  418. Turati C, Di Giorgio E, Bardi L, & Simion F (2010). Holistic face processing in newborns, 3-month-old infants and adults: Evidence from the composite face effect. Child Development, 81, 1894–1905. [DOI] [PubMed] [Google Scholar]
  419. Turner JR, & Carroll D (1985). Heart rate and oxygen consumption during mental arithmetic, a video game, and graded exercise: Further evidence of metabolically exaggerated cardiac adjustments? Psychophysiology, 22, 261–267. [DOI] [PubMed] [Google Scholar]
  420. Tversky A (1977). Features of similarity. Psychological Review, 84, 327–352. [Google Scholar]
  421. Tversky A; Kahneman D (1974). "Judgment under Uncertainty: Heuristics and Biases" Science, 185 (4157), 1124–1131. [DOI] [PubMed] [Google Scholar]
  422. Vaillant-Molina M, & Bahrick LE (2012). The role of intersensory redundancy in the emergence of social referencing in 5½-month-old infants. Developmental Psychology, 48(1), 1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  423. Vaish A, & Striano T (2004). Is visual reference necessary? Contributions of facial versus vocal cues in 12-month-olds’ social referencing behavior. Developmental Science, 7(3), 261–269. [DOI] [PubMed] [Google Scholar]
  424. Valente D, Theurel A, & Gentaz E (2017). The role of visual experience in the production of emotional facial expressions by blind people: a review. Psychon Bull Rev. doi: 10.3758/s13423-017-1338-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  425. Vallacher RR, & Wegner DM (1987). What do people think their doing? Action identification and human behavior. Psychological Review, 94, 3–15. [Google Scholar]
  426. Valliant-Molina M & Bahrick LE (2012) The role of intersensory redundancy in the emergence of social referencing in 5½-month-old infants. Developmental Psychology, 48, 1–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  427. Valstar M, Zafeiriou S, & Pantic M (2017). Facial actions as social signals In Burgoon JK, Magnenat-Thalmann N, Pantic M & Vinciarelli A (Eds.), Social Signal Processing (p. 123–154). [Google Scholar]
  428. Van den Stock J, Righart R, & De Gelder B (2007). Body expressions influence recognition of emotions in the face and voice. Emotion, 7(3), 487. [DOI] [PubMed] [Google Scholar]
  429. Vick S-J, Waller BM, Parr LA, Smith Pasqualini M, and Bard KA (2007). A cross-species comparison of facial morphology and movement in humans and chimpanzees using FACS. Journal of Nonverbal Behavior, 31, 1–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
  430. Voorspoels W, Vanpaemel W, & Storms G (2011). A formal idea-based account of typicality. Psychon Bull Rev, 18, 1006–1014. [DOI] [PubMed] [Google Scholar]
  431. Wade NJ (2016). Faces and photography in 19th-century visual science. Perception, 45, 1008–1035. [DOI] [PubMed] [Google Scholar]
  432. Wager TD, Kang J, Johnson TD, Nichols TE, Satpute AB, & Barrett LF (2015). A Bayesian model of category-specific emotional brain responses. PLOS Computational Biology. 11(4): e1004066. doi: 10.1371/journal.pcbi.1004066 [DOI] [PMC free article] [PubMed] [Google Scholar]
  433. Walker-Andrews AS (2005). Perceiving social affordances: The development of emotion understanding. The development of social cognition and communication, 93–116. [Google Scholar]
  434. Walker-Andrews AS (1997). Infants' perception of expressive behaviors: differentiation of multimodal information. Psychological bulletin, 121(3), 437. [DOI] [PubMed] [Google Scholar]
  435. Walker-Andrews AS, & Dickson LR (1997). Infants' understanding of affect. The development of social cognition, 161–186. [Google Scholar]
  436. Walker-Andrews AS, & Lennon E (1991). Infants’ discrimination of vocal expressions: Contributions of auditory and visual information. Infant Behavior and Development, 14(2), 131–142. [Google Scholar]
  437. Wallbott HG (1988). In and out of context: Influences of facial expression and context information on emotion attributions. British Journal of Social Psychology, 27, 357–369. [Google Scholar]
  438. Walle EA, & Campos JJ (2014). The development of infant detection of inauthentic emotion. Emotion, 14(3), 488. [DOI] [PMC free article] [PubMed] [Google Scholar]
  439. Wang N, Marsella S and Hawkins T (2008) Individual Differences in Expressive Response: A Challenge for ECA Design, in Proceedings of Autonomous Agents and Multi-Agent Systems (AAMAS). [Google Scholar]
  440. Wang X, Peelen MV, Han Z, He C, Caramazza A, & Bi Y (2015). How Visual Is the Visual Cortex? Comparing Connectional and Functional Fingerprints between Congenitally Blind and Sighted Individuals. J Neurosci, 35(36), 12545–12559. doi: 10.1523/JNEUROSCI.3914-14.2015 [DOI] [PMC free article] [PubMed] [Google Scholar]
  441. Wehrle T, Kaiser S, Schmidt S, & Scherer KR (2000). Studying the dynamics of emotional expression using synthesized facial muscle movements. Journal of Personality and Social Psychology, 78, 105–119. [DOI] [PubMed] [Google Scholar]
  442. Weiss B, & Nurcombe B (1992). Age, clinical severity, and the differentiation of depressive psychopathology: A test of the orthogenetic hypothesis. Development and Psychopathology, 4(1), 113–124. [Google Scholar]
  443. Widen SC & Russell JA (2013). Children’s recognition of disgust in others. Psychological Bulletin, 139, 271–299. [DOI] [PubMed] [Google Scholar]
  444. Widen SC (2016). The development of children's concepts of emotion In Barrett LF, Lewis M, J. M., & Haviland-Jones (Eds.), Handbook of emotions (4th ed., pp. 307–318). New York: Guilford Press. [Google Scholar]
  445. Wierzbicka A (1986). Human emotions: universal or culture-specific? American Anthropologist, 88(3), 584–594. [Google Scholar]
  446. Wierzbicka A (2014). Imprisoned in English The hazards of English as a default language. Oxford University Press. [Google Scholar]
  447. Wieser MJ, & Brosch T (2012). Faces in context: a review and systematization of contextual influences on affective face processing. Front Psychol, 3, 471. doi: 10.3389/fpsyg.2012.00471 [DOI] [PMC free article] [PubMed] [Google Scholar]
  448. Wilson JP & Rule NO (2015). Facial trustworthiness predicts extreme criminal-sentencing outcomes. Psychological Science, 26, 1325–1331. [DOI] [PubMed] [Google Scholar]
  449. Wilson JP & Rule NO (2016). Hypothetical sentencing decisions are associated with actual capital punishment outcomes: The role of facial trustworthiness. Social Psychological and Personality Science, 10.1177/1948550615624142 [DOI] [Google Scholar]
  450. Wilson-Mendenhall C, Barrett LF, & Barsalou LW (2013). Neural evidence that human emotions share core affective properties. Psychological Science, 24, 947–956. PMC4015729 [DOI] [PMC free article] [PubMed] [Google Scholar]
  451. Wilson-Mendenhall CD, Barrett LF, & Barsalou LW (2015). Variety in emotional life: Within-category typicality of emotional experiences is associated with neural activity in large-scale brain networks. Social Cognitive and Affective Neuroscience, 10, 62–71. [DOI] [PMC free article] [PubMed] [Google Scholar]
  452. Wilson-Mendenhall CD, Barrett LF, Simmons WK, Barsalou LW (2011). Grounding emotion in situated conceptualization. Neuropsychologia, 49, 1105–1127. [DOI] [PMC free article] [PubMed] [Google Scholar]
  453. Wilson TD, Houston CE, Etling KM, Brekke N (1996). A new look at anchoring effects: Basic anchoring and its antecedents. Journal of Experimental Psychology: General, 125(4): 387–402. [DOI] [PubMed] [Google Scholar]
  454. Witherington DC, Campos JJ, & Hertenstein MJ (2008). Chapter Sixteen Principles of Emotion and its Development in Infancy Blackwell handbook of infant development, 427. [Google Scholar]
  455. Witherington DC, Campos JJ, Harriger JA, Bryan C, & Margett TE (2010). Emotion and its development in infancy Wiley-Blackwell Handbook of Infant Development, The, Volume 1, Second Edition, 568–591. [Google Scholar]
  456. Wolff PH (1987). The development of behavioral states and the expression of emotions in early infancy: New proposals for investigation. University of Chicago Press. [Google Scholar]
  457. Wood A, Martin JD, Alibali MW, & Niedenthal PM (2018). A sad thumbs up: incongruent gestures and disrupted sensorimotor activity both slow processing of facial expressions. Cognition and Emotion, 1–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  458. Wood A, Rychlowska M, & Niedenthal P (2016). Hetergeneity of long-history migration predictions emotion recognition accuracy. Emotion, 16, 413–420. [DOI] [PubMed] [Google Scholar]
  459. Wörmann V, Holodynski M, Kärtner J, & Keller H (2012). A cross-cultural comparison of the development of the social smile: A longitudinal study of maternal and infant imitation in 6-and 12-week-old infants. Infant Behavior and Development, 35(3), 335–347. [DOI] [PubMed] [Google Scholar]
  460. Wundt W (1998/1897). Outlines of Psychology In: Judd CH. (Trans.). Leipzig: Wilhelm Engelmann. [Google Scholar]
  461. Yan F, Dai SY, Akther N, Kuno A, Yanagihara T, & Hata T (2006). Four-dimensional sonographic assessment of fetal facial expression early in the third trimester. International Journal of Gynecology & Obstetrics, 94(2), 108–113. [DOI] [PubMed] [Google Scholar]
  462. Yik M, & Russell JA (1999). Interpretation of faces: A cross-cultural study of a prediction from Fridlund's theory. Cognition & Emotion, 13(1), 93–104. [Google Scholar]
  463. Yin Y, Nabian M, Ostadabbas S, Fan M, Chou C, & Gendron M (2018). Facial expression and peripheral physiology fusion to decode individualized affective experiences. 2nd IJCAI Workshop on Artificial Intelligence in Affective Computing: Proceedings of Marching Learning Research, 1, 1–16. [Google Scholar]
  464. Yitzhak N, Giladi N, Gurevich T, Messinger DS, Prince EB, Martin K, & Aviezer H (2017). Gently does it: Humans outperform a software classifier in recognizing subtle, nonstereotypical facial expressions. Emotion, 8, 1187–1198. [DOI] [PubMed] [Google Scholar]
  465. Young AW, & Burton AM (2018). Are we face experts? Trends in Cognitive Sciences, 22(2), 100–110. [DOI] [PubMed] [Google Scholar]
  466. Yu H, Garrod OGB, Schyns PG. (2012). Perception-driven facial expression synthesis. Comput. Graph, 36,152–62 [Google Scholar]
  467. Zafeiriou S, Papaioannou A, Kotsia I, Nicolaou M, & Zhao G (2016). Facial Affect‘‘In-The-Wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (pp. 36–47). [Google Scholar]
  468. Zebrowitz LA (1997). Reading faces: Window to the soul? Westview Press. [Google Scholar]
  469. Zebrowitz LA (2017). First impressions from faces. Current Directions in Psychological Science, 26(3), 237–242. [DOI] [PMC free article] [PubMed] [Google Scholar]
  470. Zoll C, Enz S, H. S, Aylett R, & Paiva A. (2006). Fighting Bullying with the Help of Autonomous Agents in a Virtual School Environment. Paper presented at the 7th International Conference on Cognitive Modelling (ICCM-06), Trieste, Italy. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

1

RESOURCES