Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2020 Oct 1.
Published in final edited form as: Emotion. 2018 Nov 26;19(7):1292–1313. doi: 10.1037/emo0000501

Context Facilitates Performance on a Classic Cross-Cultural Emotion Perception Task

Katie Hoemann a, Alyssa N Crittenden b, Shani Msafiri c, Qiang Liu d, Chaojie Li d, Debi Roberson e, Gregory A Ruark f, Maria Gendron a,*, Lisa Feldman Barrett a,g,*
PMCID: PMC6535382  NIHMSID: NIHMS1014210  PMID: 30475026

Abstract

The majority of studies designed to assess cross-cultural emotion perception use a choice-from-array task in which participants are presented with brief emotion stories and asked to choose between target and foil cues. This task has been widely criticized, evoking a lively and prolonged debate about whether it inadvertently helps participants to perform better than they otherwise would, resulting in the appearance of universality. In three studies, we provide a strong test of the hypothesis that the classic choice-from-array task constitutes a potent source of context that shapes performance. Participants from a remote small-scale (the Hadza hunter-gatherers of Tanzania) and two urban industrialized (China and the United States) cultural samples selected target vocalizations that were contrived for six non-English, non-universal emotion categories at levels significantly above chance. In studies of anger, sadness, fear, disgust, happiness, and surprise, above chance performance is interpreted as evidence of universality. These studies support the hypothesis that choice-from-array tasks encourage evidence for cross-cultural emotion perception. We discuss these findings with reference to the history of cross-cultural emotion perception studies, and suggest several processes that may, together, give rise to the appearance of universal emotions.

Keywords: emotion, perception, vocalizations, universality, culture


There has been a lively and prolonged debate about whether or not certain emotion categories are universally expressed and recognized. According to the strongest version of the universality hypothesis, all humans (barring illness) innately produce and perceive anger, disgust, fear, happiness, sadness and surprise in nonverbal behaviors, independent of cultural background and learning.1 The original tests of this hypothesis were conducted in a handful of studies between 1969 and 1975 in two remote, small-scale societies, the Fore and Dani of Papua New Guinea (Ekman, 1972; Ekman & Friesen, 1971; Ekman, Sorenson, & Friesen, 1969; Sorenson, 1975) (for a review, see Gendron et al., in press). The studies employed several versions of a choice-from-array task pioneered by Dashiell (1927): participants heard a brief story about an emotional episode with an emotion word embedded (e.g., “Her child has died and she feels very sad.”) and are then asked to select the target stimulus (e.g., a photograph of a posed facial configuration of a frown) presented alongside one or two foils.2 Participants chose the target stimulus more frequently than chance, leading to the conclusion that anger, sadness, fear, happiness, surprise, and disgust were universal emotions (on the assumption that people infer emotion in certain facial configurations only if those configurations express emotion with a certain degree of fidelity; for discussion, see Jack, Sun, Delis, Garrod, & Schyns, 2016). Since then, hundreds of studies have been published using a choice-from-array task with participants from various urban cultural contexts with some degree of exposure to western cultural practices and norms, such as Brazil, China, Estonia, Ethiopia, France, Greece, India, Indonesia, Israel, Japan, Kenya, Malaysia, Mexico, Singapore, Turkey, and Zambia (Elfenbein & Ambady, 2002; Laukka et al., 2013), providing further support for the universality hypothesis (for discussion, see Ekman, 2017). The strongest evidence continues to come from the original studies of emotion perception in the Fore and Dani of Papua New Guinea (Ekman, 1972; Ekman & Friesen, 1971; Ekman, Heider, Friesen, & Heider, 1972), as well as from choice-from-array tasks used with the Himba of Namibia (Sauter, Eisner, Ekman, & Scott, 2010) and Bhutanese villagers (Cordaro, Keltner, Tshering, Wangchuk, & Flynn, 2016), because participants in these studies had limited exposure to western cultural practices and norms, including media, minimizing alternative explanations for any cross-cultural consistencies that were observed (Norenzayan & Heine, 2005). In the present paper, however, we present three studies to support the hypothesis that the choice-from-array task itself creates evidence for universal emotions. This evidence, we suggest, is not simply the result of demand characteristics or confirmatory bias. The task does not prime innate emotion knowledge. Instead, we suggest that choice-from-array tasks contain psychologically potent features that guide participants as they make emotional meaning of novel stimuli.

Choice-From-Array Tasks

There is now evidence from a variety of experimental studies that choice-from-array tasks provide a potent context that helps participants choose the stimuli that experimenters expect, creating stronger evidence for universal emotions than might occur otherwise (L. F. Barrett, Adolphs, Marsella, Martinez, & Pollak, forthcoming; Russell, 1994; Russell, Bachorowski, & Fernandez-Dols, 2003). Numerous lab-based studies with both children and adults (L. F. Barrett et al., forthcoming), as well as almost a dozen cross-cultural studies published since 2008, which sample a broader range of remote, small-scale populations (see Table S1; also Gendron et al., in press) and ask people to infer emotional meaning in both facial poses and vocalizations, provide evidence that choice-from-array tasks are not psychologically inert. Instead, as reviewed next, these tasks can contain elements that are known to shape participants’ responses (for a broader discussion of contextual factors that influence emotion perception, see Aviezer & Hassin, 2017; L. F. Barrett, Mesquita, & Gendron, 2011; de Gelder, 2016; Gendron, Mesquita, & Barrett, 2013; Hess & Hareli, 2017; Wieser & Brosch, 2012).3

Emotion words.

The emotion words used in choice-from-array tasks may encourage participants to assign emotional meaning to facial actions and vocalizations in ways they would otherwise not. When emotion words are familiar (i.e., associated with known emotion concepts), they directly shape participants’ perceptions, such that facial configurations like scowls and pouts are seen differently when words like “angry” and “sad” are present during the experiment vs. when they are absent (Gendron, Lindquist, Barsalou, & Barrett, 2012). In such contexts, emotion words influence which facial movements are predicted, encoded, and remembered (Chanes, Wormwood, Betz, & Barrett, 2018; Doyle & Lindquist, 2018; Fugate, Gendron, Nakashima, & Barrett, 2017; Fugate, Gouzoules, & Barrett, 2010). These findings suggest the interpretation that emotion words are aids that help perceivers recognize universal emotions more easily. Emotion words may assist perceivers in choosing the target stimulus from an array by strengthening their predictions or enhancing their sensitivity for the target or its perceptual features (Chanes et al., 2018; Mohanty & Sussman, 2013; Vogt, De Houwer, Moors, Van Damme, & Crombez, 2010). Other evidence suggests that the effect of emotion words is more potent than that, potentially creating emotion perceptions that would otherwise not occur. For example, participants label scowling faces as “determined” or “puzzled,” wide-eyed faces as “hopeful” and gasping faces as “pained” when they are provided with stories about those emotions rather than with stories of anger, surprise, and fear (Carroll & Russell, 1996, Study 2). Scowling faces are also more likely to be perceived as fearful when paired with the description of danger (Carroll & Russell, 1996, Study 1).

Words are also particularly potent when people are presented with cues, such as facial movements, that are unfamiliar and have no prior emotional meaning (Fugate et al., 2010), consistent with the broader finding that words support perception for unfamiliar objects (Lupyan, Rakison, & McClelland, 2007). Furthermore, experimental tasks that place fewer constraints on how participants respond, such as asking them to freely label facial poses or vocalizations, rarely provide evidence for the universality hypothesis (see Table S1; L. F. Barrett et al., forthcoming; Gendron, 2017; Gendron et al., in press). Finally, children who do not explicitly possess conceptual knowledge about emotions such as anger, sadness, fear, and who do not understand the meaning of emotion words beyond their affective content, as well as adults who have lost that semantic knowledge from neurodegeneration, are experientially blind to the emotional meaning of facial poses and vocalizations: they only perceive affect (Lindquist, Gendron, Barrett, & Dickerson, 2014; Widen, 2016).

Elaborate manipulation checks.

Choice-from-array tasks have been known to use an elaborate manipulation check procedure that has the potential to teach novel emotion concepts to naïve participants. For example, Sauter et al. (2010) used what they referred to as a “rigorous” manipulation check that did more than ask for verbal confirmation of whether individuals from the Himba culture of northwestern Namibia understood anger, sadness, fear, and the other concepts represented in brief stories used in the experiment. As stated in Sauter et al. (2010), after hearing each emotion story, participants were “asked how the person was feeling” (p. 2411, italics added) to confirm that “they had understood the intended emotion of the story” (p. 2408). In their later commentary, Sauter et al. (2015) elaborated that:

Participants [were allowed] to listen several times to a given recorded story (if needed), until they could explain the intended emotion in their own words, before they proceeded to the experimental trials for that story. The inclusion of a rigorous manipulation check with experimenter verification, rather than reliance on participants’ reports, was thus crucial. (p. 355).

Individuals in the Himba cultural group display opacity of mind, however; they do not make mental state inferences as frequently as more westernized people (H. C. Barrett et al., 2016), focusing on emotion as situated action rather than internal feeling.4 As a consequence, when Sauter et al. asked participants to verbally describe the English emotion concept that was portrayed in the story, they perhaps (unwittingly) encouraged concept learning in their experiment (as hypothesized by Gendron, Roberson, & Barrett, 2015). They did not allow Himba participants to proceed to the experimental trials for a given story until they could conceptualize the emotion stories in a manner consistent with western cultural expectations.

By contrast, a subsequent study using a choice-from-array task with the Himba, but without the elaborate manipulation check, was unable to replicate the Sauter et al. (2010) findings (Gendron, Roberson, van der Vyver, & Barrett, 2014a). Our study used a manipulation check only to verify that participants understood how to perform the task. Participants were given the opportunity to indicate that they understood the emotion stories from their own cultural perspective. We took great pains to ensure participants were attending to the experimental task, understood the action described in the brief stories, and understood what was required of them: we conducted an attention check before every trial, verbally confirmed understanding (yes or no response), and allowed participants to replay the story for any reason. The concern that participants may have missed some of the stories, due to an attentional lapse or a failure of memory, is assuaged by that fact that the stories were repeated many times over the course of the experiment. Analyses of the different foil conditions revealed that Himba participants performed better than chance only when the target and foil differed in valence, such that pleasantness or unpleasantness could be used to distinguish between the two vocalizations. The general pattern of these findings was replicated in a free-labeling task (Gendron et al., 2014a, Study 1).

Repeated trials.

Choice-from-array tasks present the same stimuli over and over. The fact that participants are repeatedly exposed to the same facial configurations and emotion words creates a context for them to learn the intended pairings, even if they do not know them to begin with (Nelson & Russell, 2016a). For example, children learn to label an artificially constructed facial expression (e.g., a blowfish expression) with the word “pax” in a choice-from-array task (Nelson & Russell, 2016b). And participants have been shown to use a process of elimination strategy to complete choice-from-array tasks, boosting agreement levels (DiGirolamo & Russell, 2017). In fact, if the target stimulus is presented against only one foil, all that is required in this task arrangement is to figure out which of two vocalizations or two faces is least expected in the given context. For example, after hearing a story about anger, a participant hears a shout and a laugh, and can choose the shout merely by realizing the laugh is not correct (on the basis of valence). This is similar to selecting the correct answer on a multiple-choice test by eliminating the incorrect alternatives.

Choice-from-array tasks have also been known to block trials by emotion category, which may further encourage perceptual learning and a process of elimination strategy. For example, Sauter et al. (2010) blocked their trials: a Himba participant heard a story, verbalized an English emotion concept, and then heard a series of trials with portrayed vocal bursts for the target emotion category and a foil. For example, after hearing a story about sadness several times and describing the English concept for sadness, a participant would hear a cry (target) and a laugh (foil), then a shout (foil) and a cry (target), and so on. When Himba participants completed a similar task with randomized trials, they did not choose vocalizations for anger, sadness, fear, and the other categories tested in a way that was supportive of the universality hypothesis (Gendron et al., 2014a).

Meaning Making

It is tempting to assume that these design considerations are mere methodological footnotes, but we suggest an alternative hypothesis. Taken together, these findings suggest that the design elements in a choice-from-array task are psychologically potent: they provide a context for perceivers to make meaning of ambiguous physical cues like facial movements and vocalizations, particularly when those cues have no inherent emotional meaning. As demonstrated by a recent review, hundreds of studies of healthy adults across cultures, newborns and young children, and people who are congenitally blind or deaf consistently find that specific facial movements, such as smiling, scowling, and frowning are context-specific in their emotional meaning: instances of the same emotion category, such as instances of anger, are expressed with more variable facial movements than generally acknowledged, and similar facial movements, such as a scowl, can communicate a variety of different emotions, or even carry non-emotional information (L. F. Barrett et al., forthcoming). As a result, perceivers implicitly use situational context to make meaning and infer emotional information in facial movements. The same is likely true of vocalizations (Russell et al., 2003). As a result, choice-from-array tasks may provide a context that allows participants to infer emotional meaning differently than they otherwise would, shaping their emotion perception performance. If this hypothesis is correct, then it provides an alternative explanation for the hundreds of studies that support the hypothesis that people universally recognize emotion in nonverbal behaviors in an automatic and obligatory way.

The hypothesis that choice-from-array tasks facilitate emotion perception fits with the observation that humans are active meaning-makers. Human brains are wired to transduce changes in light, air pressure, and chemicals, and go beyond the information given (Bruner, 1957) to create the sights, sounds, and smells of our surrounds. The same changes in air pressure can be experienced as a person’s laugh in one setting and a sob in another (Belin, Fillion-Bilodeau, & Gosselin, 2008). The psychological literature is full of experiments that manipulate context to examine its effect on meaning. For example, visual context facilitates the recognition of scene-consistent objects (for a review, see Bar, 2004). And it is well known that the acoustic and linguistic context influences which phonemes are heard (Massaro & Cohen, 1983). There are notable examples of how certain experimental design features evoke powerful psychological effects that are hidden contexts for other psychological processes (such as controlled processing; for a review, see Pashler, Johnston, & Ruthruff, 2001). Context effects are well-established in how people perceive emotion in faces, voices and bodies (e.g., Aviezer et al., 2008; Aviezer, Trope, & Todorov, 2012; Calbi, Angelini, Gallese, & Umiltà, 2017; Fernández-Dols & Ruiz-Belda, 1997; Fridlund, 1991; Mobbs et al., 2006; Ruiz-Belda, Fernández-Dols, Carrera, & Barchard, 2003; Van den Stock, Righart, & de Gelder, 2007; Wallbott, 1988).

Our hypothesis here is that choice-from-array tasks contain features that encourage participants to infer particular emotional meaning in vocalizations such as shouts, sighs, and laughs, creating the appearance of universal emotions. Participant performance on tests of cross-cultural emotion perception may be the result of multiple processes, including identifying perceptual similarities or employing process-of-elimination strategies (e.g., Nelson & Russell, 2016a), perceiving affect (e.g., Gendron et al., 2014a), and learning categories online (e.g., Ferry, Hespos, & Waxman, 2010), calling into question the validity of interpreting such performance as direct evidence for the innate universality of certain emotion categories. In addition, it is possible that participants may use conceptual combination to complete cross-cultural emotion perception tasks when faced with unfamiliar emotion categories and exemplars. Conceptual combination is a fundamental cognitive capacity that allows individuals to construct instances of novel categories (such as emotion categories not present in their culture) by flexibly combining previously acquired conceptual knowledge (Barsalou, 1987). Conceptual combination does not imply that all properties of the original concepts will be invoked in a novel instance (Wu & Barsalou, 2009), but only those that are relevant for a particular situation (e.g., a given emotion story). This is the way that people can perceive instances of novel categories, including emotion categories, for which they have no single vocabulary word or even prior experience (for discussion, see L. F. Barrett, 2017a). As such, conceptual combination is a plausible process by which participants leverage features of the experimental context to infer emotional meaning in novel stimuli.

The Present Studies

In the studies that follow, we tested the potency of the classic choice-from-array task to create a context that allows participants to infer emotional meaning for novel vocalizations that have no inherent emotional meaning for them. We present data from one sample of U.S. participants (Study 3) and from samples in two other cultural contexts. In Study 1, we test individuals from a remote, small-scale society with relatively little exposure to western cultural norms, practices, and values (Henrich, Heine, & Norenzayan, 2010; Norenzayan & Heine, 2005) – hunter-gatherers from northern central Tanzania – who may not, on their own, make emotional meaning of vocalizations such as laughs, sighs, and shouts in the same way as U.S. participants. Our recent study of emotion perception in the Hadza indicates that they do not freely label scowls, smiles, frowns and other facial poses in the presumed universal way (Gendron et al., under review). Therefore, Study 1 provides a strong test of our hypothesis. Study 2 tests Chinese participants living in China who have access to western cultural norms, practices, and values, but who also have their own as well, enhancing the variable emotional meanings that are available for physical cues such as facial movements and vocalizations. Recent research with dynamic facial movements comparing U.S. and Chinese participants shows that instances of the same emotion are expressed with multiple sets of facial movements and similar facial movements express different emotion categories (Jack et al., 2016). As a consequence, Study 2 provides a test of the choice-from-array task’s potency for creating a context for emotional meaning making under conditions of enhanced ambiguity. Across three studies, we demonstrate that the choice-from-array task typically used in studies of cross-cultural emotion perception actively (although unintentionally) facilitates task performance, resulting in findings that make non-universal emotion categories appear universal.

In all three studies, the goal was not to separately test the potency of individual elements in the choice-from-array task, such as emotion words, repeated and blocked trials, or elaborate manipulation checks. Rather, we sought to demonstrate that, together, these elements create a highly structured version of the task that encourages emotional meaning making, and therefore task performance, providing evidence for universal emotions that may not otherwise emerge. A more mechanistic approach is not optimal at this stage of our research for two reasons. First, while it may be possible to conduct systematic control conditions in laboratory experiments, it is often not feasible in field studies with remote populations limited in number (such as the Hadza), where researchers must operate under site constraints. Second, it is common practice to establish the replicability and robustness of an effect before investing the resources to undertake the long series of studies required to isolate each feature and testing its unique contribution, or manipulating features to model their synergistic impact.

There are several notable aspects to the studies we report in this paper. First, to provide a stringent test of our hypothesis, we traveled to northern central Tanzania to test emotion perception in members of a remote small-scale hunter-gatherer culture, the Hadza (Study 1). This paper reports the first study of emotion perception ever conducted with the Hadza (the second being our study of emotion perception in faces, reported in Gendron et al., under review). These studies are particularly important because, according to ideas from evolutionary psychology, universal and innate emotional expressions evolved to solve to the recurring fitness challenges of hunting and gathering in small groups on the African savanna (Pinker, 1997; Shariff & Tracy, 2011; Tooby & Cosmides, 2008). Therefore, the Hadza provide the strongest test of whether certain emotion categories are universal. Their cultural isolation is rapidly under assault, becoming contaminated by tourists and assimilation, and so the opportunity to study their emotional lives is rapidly disappearing. We examined the replicability of our findings by conducting the same study in an industrialized culture from the East (China, Study 2) and the West (the United States, Study 3).

Second, we chose to study the impact of the choice-from-array context on six emotion categories that would be novel to our participants: gigil, greng jai, gluckschmerz, itoshii, lajja, and liget (Table 1). These categories are not translatable by a single word in English, meaning that participants likely do not invoke these specific category boundaries in their daily lives. (It does not mean that participants would be unable to understand or relate in some way to the experienced described.) Nor do these emotion categories meet the usual criteria for universality: they do not appear in Darwin’s The Expression of the Emotions in Man and Animals (1872/2005), they do not appear to be evidenced in the behaviors of non-human animals (see Tracy & Randles, 2011), and are they not thought to solve a recurring evolutionary challenge for our Pleistocene ancestors (see Shariff & Tracy, 2011). We provided participants with short descriptive scenarios and stipulated (i.e., made-up) vocalizations for each emotion category (Table 1, see Supplemental Material for details). We verified that the six novel emotion categories were unfamiliar in the cultural samples being tested, in that participants did not consistently and specifically associate the scenarios or vocalizations with pre-existing, emotion categories in their native language. This verification process is reported in the method section of each study.

Table 1.

Novel Emotion Categories

Category Pronunciation Origin Description Scenario
Gigil GHEE-ghil Philippines The overwhelming urge to squeeze or pinch something that is very cute Someone sees a small, chubby, lovely baby and wants to squeeze it tightly. They feel gigil.
Glückschmerz GLOOK-shmairts Germany Displeasure derived from another’s pleasure Someone hears that a bad person had some good fortune, and feels upset about it. They feel glückschmerz.
Greng jai kreng-JAI Thailand The feeling you get when you don’t want someone to do something for you because it would be a pain for them Someone is offered help from others, but does not want it, because it is too much trouble for the others. The person feels greng jai.
Itoshii ee-toe-SHEE Japan Bittersweet longing for an absent loved one Someone thinks pleasant things about their loved one who has moved away (to another camp).* They feel itoshii.
Lajja lah-ZHAH India Respectful restraint or playful shame; pleasant adherence to social norms Someone makes a small mistake that others will notice and feels bad, but also acts playful. They feel lajja.
Liget LI-gut Ilongot (Philippines) Intense focus, passion, and energy associated with actively pursuing a challenge Someone works very hard toward a goal, and feels a rush of energy and intense focus. They feel liget.

Note: Additional testing verified that these categories were unknown to the three cultural samples. Participants were asked to freely label vocalizations and scenarios developed for each category. Examination of the labels confirmed that participants produced neither consistent nor specific labels. See Supplemental Material (p 7–17) for details.

*

The content in parentheses was included in Study 1 for clarification, but not Studies 2 and 3.

On each trial in our studies, participants heard scenarios describing an emotional experience for which they did not have a pre-existing category or word, and then heard two contrived vocalizations, one that was invented as the target for the novel emotion category and the other invented for different novel emotion category (the foil). We hypothesized that the choice-from-array task would aid participants in selecting stipulated target vocalizations for six novel emotion categories from around the world with a level of agreement normally interpreted as evidence for universality. That is, the choice-from-array task would make six non-universal categories look universal in three different samples from around the world.

If certain emotion categories are indeed universal, and the choice-from-array task does not encourage participants to choose the expected answers, we would expect to find support for the null hypothesis: participants would not choose the stipulated vocalizations for these novel emotion categories at levels greater than chance. In contrast, support for our hypothesis would be found if participants perform at above chance levels, choosing the vocalizations we invented for the emotion categories that are not traditionally part of U.S., Chinese, or Hadza culture. Specifically, these results would be consistent with (even if they do not directly test) an account of novel emotion meaning-making based on conceptual combination, in addition to other general processes like affect perception and category learning.

Use of a Bayesian Analytic Approach

We used Bayesian hypothesis testing to quantify the evidence for universality hypothesis (corresponding to the null hypothesis) that participants would select target vocalizations at chance levels for the novel emotion categories, as well as our alternative (task-as-context, henceforth ‘context’) hypothesis that they would select the vocalizations at levels greater than chance. The ability to quantify evidence in favor of both null and alternative hypotheses one of the main advantages of Bayesian hypothesis testing over frequentist approaches such as t-tests and ANOVAs (e.g., Edwards, Lindman, & Savage, 1963; Rouder, Morey, Speckman, & Province, 2012; Wagenmakers et al., 2015) because p-values cannot provide support for the null (Wetzels et al., 2011). Bayesian hypothesis testing allowed us to assess the ratio between the probability of our context hypothesis given the data, as well as the probability of the null (universal) hypothesis given the data.

Another key advantage to Bayesian hypothesis testing is that it yields a de facto power analysis and replaced the need for a separate power analysis (Berger & Mortera, 1999; Berger & Wolpert, 1988; Rouder, 2014; Wagenmakers et al., 2015). The main statistic of interest is the Bayes factor, which expresses the relative evidence for null and alternative hypotheses and consistently trends toward truth as data accumulate, unlike a p-value (Rouder, Speckman, Sun, Morey, & Iverson, 2009). Using this approach, we were able to determine that we had adequate sample sizes to test both hypotheses. We supplemented our Bayesian analyses with hierarchical generalized linear modeling analyses (HGLM; Raudenbush, Bryk, Cheong, Congdon, & du Toit, 2004), as discussed in the results sections of each study.

Study 1: Remote Small-Scale Sample – the Hadza of Tanzania

The Hadza are a semi-nomadic, hunter-gatherer population who live in small bands in Tanzania and remain relatively isolated from other cultural practices and norms. They are egalitarian foragers who bring food back to camp for distribution (Berbesque, Wood, Crittenden, Mabulla, & Marlowe, 2016). They consume a diet that is primarily based on the collection of wild foods (Apicella & Crittenden, 2015); the women gather and the men hunt and collect honey. The Hadza are not a Paleolithic population, but have been continuously hunting and gathering for thousands of years in East Africa, and they are one of the last groups to live in a social and physical context closer to that of our Pleistocene ancestors than our own (Marlowe, 2010). An estimated 150–200 Hadza individuals currently live in remote camps (Jones, 2016). In Study 1, we recruited and tested 55 of these individuals. Prior to our visit to the Hadza in 2016, they had not participated in any studies of emotion perception, although some had previously participated in social cognition research (H. C. Barrett et al., 2016; Bryant et al., 2016). During this visit, we conducted two studies (Study 1 reported here, and a second study with posed emotional faces reported in Gendron et al., under review).

Method

Participants.

Participants were 55 native Hadzane speakers (21 female) from the Hadza ethnic group, recruited from the area surrounding three camps located southeast of Lake Eyasi in northern central Tanzania: camp 1 (n = 20), camp 2 (n = 27), and camp 3 (n = 8) (Figure S7). All three camps were far from the regional towns, and there was no evidence of recent tourism to these areas. During the time of data collection, the majority of the Hadza individuals we tested did not travel except to other Hadza camps, although regional towns and small cities were visited occasionally for access to supplies and medical care. Only three individuals we tested reported venturing further afield to the larger cities of Arusha, Dar es Salaam, or parts of Kenya. Participants ranged in age from 18 to mid-70s, with a median decade of 30–40; Hadza individuals do not keep track of their biological age, so estimates were based on maternal/paternal history and personal knowledge of our translator (the third author, S.M.), who is ethnically Hadza as well as an experienced research assistant.

Participants were tested in their native language, Hadzane, which is a click-based language considered to be a linguistic isolate (Sands, 1998). Almost all Hadza speak some Swahili, however, as they use this language to communicate with neighboring ethnic groups. Twenty adults in our sample reported fair to excellent knowledge of Swahili, whereas the other 35 most reported little knowledge of Swahili. In addition, five individuals reported modest knowledge of another neighboring language, such as Datoga, Iraqw, or Isanzu. Only one participant claimed rudimentary knowledge of English, but did not use it with the experimenters. Sixteen participants in the sample had attended local primary school for between two and seven years. One participant attended the regional secondary school for two years.

Data collection was approved by the Office of Human Subject Research Protection Institutional Review Board at Northeastern University as well as the Tanzanian Commission for Science and Technology (COSTECH). Hadza participants were verbally consented prior to participation and were remunerated with gifts (clothing, cookware, etc.).

Stimuli.

Stimuli were six short scenarios (one per emotion category) and 60 non-word vocalizations (10 per emotion category). For comparison, Sauter et al. (2010) used short scenarios from nine emotion categories (achievement, amusement, anger, disgust, fear, sensual pleasure, relief, sadness, and surprise), each with 10 non-word vocalizations.

Emotion concept selection.

We selected six emotion concepts from the Interdisciplinary Affective Science Laboratory’s database of emotion concepts that have been deemed ‘untranslatable’ into English. We compiled this database (available upon request) using published psychological, anthropological, and linguistic literatures (e.g., Lomas, 2016; Rosaldo, 1980; Russell, 1991), as well as websites and publications in the popular media (e.g., BetterThanEnglish, 2012; De Boinod, 2007; Lin, 2013). We selected six emotion concepts that were easy to portray with static facial poses and vocalizations. None of these concepts are labeled with single words in English. None are frequently described in the English media. And none have been studied in a scientific context or stipulated as universal. We refer to these as ‘novel’ emotion concepts because they are unfamiliar to English and Hadzane speakers alike (i.e., they are not represented as existing emotion categories in either English or Hadzane). Fluent English speakers from the United States on Amazon’s Mechanical Turk (MTurk) rated the six emotion concepts on valence and arousal (n = approx. 15 per concept; 43.5% female; mean age = 32.2, SD = 8.79; see Figure S1).

Emotion scenarios.

We generated six short scenarios, one for each novel emotion concept. The scenarios (Table 1) were developed based on concept descriptions by native speakers of the source languages, and were adapted for cultural fit by the second author (A.N.C.), a nutritional and ecological anthropologist with extensive field experience working among the Hadza. Our translator, the third author (S.M.), a native speaker of Hadzane and a fluent speaker of English, also reviewed the scenarios, and recorded them in Hadzane. He is one of the few Hadza to have attended university, and has previously been involved in anthropological research (e.g., Raichlen et al., 2017). Each scenario was one to two sentences long, and did not differ substantially in structure from those used in Sauter et al. (2010), which were themselves adapted from the scenarios used by Ekman, Sorenson, and Friesen (1969). Each scenario briefly described an emotional situation in concrete terms and concluded with a description of how the protagonist felt using the target emotion word in its original language. For example, “Someone sees a small, chubby, lovely baby and wants to squeeze it tightly. They feel gigil.”

Vocalizations.

We (the authors) invented the short vocal bursts that served as posed vocal expressions of each emotion category (i.e., we stipulated their pitch, tone, duration) based on our folk understanding of the novel emotion concepts in Table 1. We then directed five posers from the United States (three women, two men) to generate these short vocal bursts. Imitating emotion poses is a well-established method for stimulus generation in studies of emotion perception (e.g., Ekman, Levenson, & Friesen, 1983; Schröder, 2003). Posers were first provided with descriptions of the desired vocalizations for each emotion category (Table S3), which were also modeled by the first author. Posers then generated the vocalizations while imagining the emotional situations described in the scenarios. All vocalizations were cleaned for ambient noise and adjusted for mean peak amplitude using Audacity (Audacity Team, 2015).

Next, we examined whether the vocalizations portraying the same emotion category clustered together in multidimensional space. We opted for a multidimensional scaling (MDS) approach (Hout, Papesh, & Goldinger, 2013; Jaworska & Chupetlovska-Anastasova, 2009) because it allowed us to discover, rather than confirm, whether the vocalization associated with the same emotion category are similarly positioned in perceptual space. Prior MDS analyses confirm that vocal stimuli sharing similar acoustic features do cluster together (Gygi, Kidd, & Watson, 2007). Pilot participants (N = 7) heard pairs of vocalizations, one member of the pair played after the other, and then rated their similarity from 1 (very similar) to 9 (very dissimilar). All possible pairwise comparisons were presented in randomized order, with vocalizations randomly assigned to play first or second. A single matrix of mean similarity ratings was computed for all vocalization pairs and subjected to an ALSCAL procedure (Young & Lewyckyj, 1979). Results indicated that vocalizations posed for the same novel emotion category clustered together in multidimensional space (Figures S4 and S5). A stress-by-dimensionality plot (Figure S3) indicated that the three-dimensional solution provided the best fit for the data. The dimension loadings for the vocalizations suggested that the first two dimensions are valence and arousal, replicating prior findings from MDS analyses of affective stimuli (e.g., Russell, 1980; Sauter, Eisner, Calder, & Scott, 2010 also report similar findings using principal component analysis [PCA]). Cluster locations were similar to explicit valence and arousal ratings for the emotion scenarios (Figure S1). A separate sample of MTurk raters (N = 25; 8 female; mean age = 31.56, SD = 7.04) also verified that the posed vocalizations for the same novel emotion category clustered together in multidimensional space (Figure S6). We elicited a second set of vocalizations from five Namibian posers (three women, two men) as part of a separate study, based on these normed vocalizations, for a total of 10 per emotion category.

Procedure

Verifying vocalization novelty.

Before completing the choice-from-array task, nineteen participants freely labeled six of the made-up vocalizations, one for each emotion category under investigation. Participants were asked to describe how the target person making each sound felt using a word or short phrase. Participants who provided a description of a situation or behavior were prompted to describe how the target person was feeling, to replicate the method described by Sauter et al. (2010, 2015). Participant-provided labels were coded according to whether they represented a known emotion category in English, and whether or not this category was subordinate to or synonymous with a purportedly universal emotion category (i.e., anger, disgust, fear, happiness, sadness, surprise; see Supplemental Material for details on the coding procedure). We expected, and found, that the consistency and specificity of labeling each vocalization was low. Participants did not label the vocalizations with the words for the novel emotion concepts, nor did they provide labels such as angry, fearful, sad, etc., and their synonyms. In fact, most participants provided labels for the vocalizations that did not represent a known emotion category in English (for gigil, 79% of responses did not refer to a discrete emotion category; gluckschmerz, 37%; greng jai, 81%; itoshii,78%; lajja, 33%; liget,53%), (panel A of Figure S8). Furthermore, participants labeled each vocalization with words that ranged in their affective features, using both positively or negatively valenced labels, as well as labels that were high or low in arousal, indicating that there was little consistency and specificity in the affective features associated with the vocalizations (panel A of Figures S9 and S10).

Choice-from-array task.

Each participant was tested individually and completed the classic choice-from-array task, following the methods used in Sauter et al. (2010; for additional details, see Sauter et al., 2015). All participants were tested individually using a Dell ATG laptop and headphones. Participants listened to a pre-recorded scenario in their native language (e.g., in English, “Someone sees a small, chubby, lovely baby and wants to squeeze it tightly. They feel gigil.”). Participants then completed an extensive manipulation check (panel A of Figure 1) as described in Sauter et al. (2015). They were asked to describe how the scenario protagonist feels, and a correct answer would contain features other than simply repeating the target emotion label used in the scenario. For example, a response (in Hadzane or Swahili) of “he feels gigil” was not accepted, whereas a response of “he feels good”, “he feels loving”, or “he feels angry” was acceptable. Participants were allowed to listen to the recorded scenario multiple times, and received verbal feedback in their native language on their responses. Because the Hadza are a pre-literate society, we communicated verbal instructions and feedback through our Hadzane translator (S.M.).

Once participants had passed the manipulation check for a given emotion category, they then completed a block of trials for that category (panel B of Figure 1). For example, on every trial within a gigil block, participants listened to the audio recording of the scenario followed by target (an invented gigil sound) and a foil vocalization (e.g., an invented sound for the lajja category). Each trial drew from a list of 10 possible vocalizations for each novel emotion category. Foils were drawn randomly from one of the other (non-target) categories and were matched for sex of poser. As the first vocalization played, an icon appeared on the left side of the computer screen, and then disappeared. While the second vocalization played, the mirror version of the icon appeared on the right side of the screen, and then disappeared. Both icons then appeared simultaneously, and participants pressed the left or right icon to indicate which vocalization matched the emotion portrayed in the scenario. Target position (left or right) was randomized for target and foil. On a given trial, scenarios and vocalizations were repeated for participants who wished to hear them again. Upon completing the first trial, participants heard the gigil scenario again, followed by an invented gigil sound and another foil (e.g., an invented sound for the liget category). Once the block of trials for a given emotion category was complete, participants proceeded to another block of trials for a different emotion category. Emotion category blocks were presented in randomized order. Participants completed six emotion category blocks of five trials each, for a total of 30 trials.

Fig. 1.

Fig. 1.

The classic, highly structured choice-from-array task with novel emotion categories. Pre-recorded material presented over headphones is depicted in light gray boxes; verbal interactions with the experimenter are presented in dark gray boxes. A. Manipulation check: Participants listened to a scenario in their native language and were then asked to describe how the protagonist in the story feels. B. Perception trials blocked by emotion category. Following the manipulation check for a given emotion category, participants completed a block of trials for that category. On every trial, participants listened to the scenario again followed by target and foil vocalizations. Vocalizations were played, one at a time, with an icon presented concurrently on either the left or right side of the screen (no words were present on screen). Participants touched an icon to select a vocalization. Once a trial was complete, participants completed another trial from the same emotion category. Targets were presented randomly on the left or right within a block.

Field site constraints dictated that each participant spend a similar amount of time in testing. As a result, all participants completed all experimental trials, but a participant’s data for a given block of trials were removed prior to analysis if he or she failed the manipulation check for that emotion category three times. The sample size used in the analyses can be found in Figure 3.

Verifying emotion scenario novelty.

Following completion of the choice-from-array task, participants who freely labeled the vocalizations also freely labeled long-form versions of the scenarios that did not have any emotion word embedded. Only six participants were available to do this norming due to time constraints. As expected, participant-provided labels showed little agreement across participants and did not converge with known English concepts anger, disgust, fear, happiness, sadness, and surprise (see panel A of Figure S11). A similar finding was observed for valence and arousal properties (see panel A of Figures S12 and S13).

Results

Bayesian hypothesis testing.

We conducted a Bayesian one sample t-test to directly evaluate the degree of support for the null (universal) hypothesis (i.e., chance-level performance at .5), as well as to evaluate the sequentially accumulating support for the alternative (context) hypothesis (i.e., mean performance greater than chance) (Rouder et al., 2009; Wetzels, Raaijmakers, Jakab, & Wagenmakers, 2009). Analyses were computed in JASP (JASP Team, 2017). Given that the effects sizes that support the context hypothesis might vary, a Bayesian approach evaluates a distribution of expected effect sizes rather than a single estimate. In JASP, we estimated the probability that the context hypothesis was true given a one-sided prior probability distribution where the median effect size was equivalent to Cohen’s δ = .7. This is consistent with strong evidence for the universality of emotion categories in choice-from-array tasks (Haidt & Keltner, 1999).5

A Bayesian one sample t-test conducted on aggregate response data demonstrated robust support for the hypothesis that participants would perform above chance relative to the null hypothesis. That is, Hadza participants performed at a level typically interpreted as evidence of universality (i.e., perceived the intended novel emotions) when tested using the highly structured version of the classic choice-from-array task used by Sauter et al. (2010). As can be seen in panel A of Figure 2, the Bayes factor for our Hadza participants was BF10 = 187.905, indicating that the observed data are 187.905 times more likely under the alternative (context) hypothesis than under the null (universal) hypothesis. This is considered “very strong” (Bayes factor greater than 150; Kass & Raftery, 1995) or “decisive” (Bayes factor greater than 10^2; Jeffreys, 1961) evidence for the context hypothesis.

These findings were robust, even when we varied the parameters of our analysis. For example, we varied the width of the prior probabilities for the effect size under the alternative (context) hypothesis. Narrower priors allow for a smaller range of expected effect sizes, and therefore represent a more stringent test of the context hypothesis when compared to the null (universal) hypothesis (i.e., a prior width very close to zero most favors finding evidence for the null hypothesis). Under these conditions, the Bayes factor still indicates substantial support for the context hypothesis (i.e., BF10 > 3).

We also generated sequential analysis plots to examine the development of the Bayes factors as data accumulated. We observed that cumulative data increasingly provided evidence for the alternative (context) hypothesis when compared to the null (universal) hypothesis, lending confidence that our sample size was large enough and we were sufficiently powered to test our hypotheses (panel D in Figure 2).

Fig. 2.

Fig. 2.

Inferential plots for Study 1, based on Bayesian one-sample t-tests on overall performance of Hadza participants. a. Pizza plot providing a proportional representation of the ratio of evidence for the alternative (context) hypothesis to evidence for null (universal) hypothesis. b. Posterior distribution based on a one-sided prior distribution. Posterior mass to the right of zero indicates that participant performance is consistently above chance-level responding of .5. c. Bayes factor robustness check. Narrower priors indicate a smaller range of expected effect sizes, favoring the null (universal) hypothesis. d. Sequential analysis of evidence accumulated over the course of data collection. All figures adjusted from JASP (JASP Team, 2017).

Hierarchical Generalized Linear Modeling (HGLM).

In addition, we used hierarchical generalized linear modeling (HGLM; Raudenbush et al., 2004) to examine whether participants’ performance was above chance-level responding when analyzed by novel emotion category. We chose HGLM for two reasons. First, the main dependent measure in the choice-from-array task is dichotomous: Did a participant pick the target vocalization on a given trial? (1 = yes, correct; 0 = no, incorrect). Participants’ performance across multiple trials can be modeled as a binomial distribution, bounded at both ends (i.e., has values between 0 and 1). Traditional parametric approaches such as one-sample t-tests and ANOVAs with binomial or categorical data cannot be used to analyze data from the choice-from-array task because binomial response data must be treated as proportions or percentages, resulting in confidence intervals that can extend beyond the interpretable values between 0 and 1, leading to spurious results (e.g., Agresti, 2002; Jaeger, 2008). HGLM accounts for data that are not normally distributed by using a nonlinear link function (Nelder & Wedderburn, 1972).

The second reason we chose HGLM is that non-parametric approaches such as chi-square (e.g., Sauter, Eisner, Ekman, et al., 2010) and exact binomial t-tests (e.g., Cordaro et al., 2016) do not account for the nested, non-independent nature of the data in a repeated measures choice-from-array task. Trials are non-independent because the probability of a correct response on one trial could influence the probability of a correct response on subsequent trials (e.g., due to perceptual similarity of the target vocalizations within a category, which are blocked together in the highly structured version of the task). HGLM is well suited for these dependencies because trials can be grouped into clusters, allowing for the error term to be partitioned and increasing the power of the model to detect the effect (Guo & Zhao, 2000; Kenny, Korchmaros, & Bolger, 2003).

We analyzed the data using a Bernoulli multilevel model, which estimated a log-odds (i.e., the probability of performing above chance of .5) using a log linear link function. Data were structured in a two-level model. Trials on which participants selected either the target or foil vocalization for six emotion categories (level-1) were nested within individuals (level-2). We used an intercept-as-outcome approach with dummy codes for each emotion category (Raudenbush et al., 2004). We have used the intercepts-as-outcomes approach to analyze other repeated measures data (e.g., Anderson, Siegel, & Barrett, 2011; L. F. Barrett & Niedenthal, 2004). We used a random effects model to compute the population-average estimates with robust standard errors, allowing us to generalize the average probability of success beyond those individuals included in the sample. All HGLM analyses were conducted in HLM7 (SSI Inc., Lincolnwood, IL). See Supplemental Material for model specifications.

The results are reported in Figure 3, panel A (see also Table S4 for detailed results). The analysis indicated that participants selected target vocalizations for gigil, itoshii, and liget at levels significantly above chance (p’s range from .005 to .030). Participants selected targets for lajja and glückschmerz at levels approaching conventional levels of statistical significance (m = .58, p = .089 and m = .57, p = .093, respectively); notably, performance means were identical to liget (m = .58). The glückschmerz category obtained a bimodal distribution of responses: while 12 participants performed below chance, 17 participants performed above chance, including 10 who performed at or near ceiling. The lajja category was underpowered because many participants failed the manipulation check after describing the feeling as unpleasant, whereas it is experienced as pleasant in the original Oriya Hindu culture (Menon & Shweder, 1994). Of the 12 participants who passed the manipulation check for lajja, however, eight selected target vocalizations at levels above chance. (See Figure S14 for distribution of performance above and below chance per novel emotion category.)

Fig. 3.

Fig. 3.

Results from Study 1 (panel A) compared against Sauter et al. (2010; panel B). Abbreviations: (panel A) gig, gigil; gre, greng jai; glü, glückschmerz; ito, itoshii; laj, lajja; lig, liget; (panel B) ach, achievement; amu, amusement; ang, anger; dis, disgust; fea, fear; ple, sensual pleasure; rel, relief; sad, sadness; sur, surprise. Results data from Sauter et al. (2010) were originally presented in terms of mean number of correct responses (out of four trials per emotion category). These data have been re-plotted in descending order of proportion correct in order to facilitate direct comparison with the present study. Dashed line indicates chance-level performance (.5). Sample size per emotion category reported in white font at the bottom of each column. Standard error bars (± 1 SE) are provided as distributional information only: the location of error bars above the chance line is not indicative of significant above-chance performance because these data were binomially rather than normally distributed. Effect sizes for Study 1 are reported above each column using the odds ratio (OR), which expresses group difference in probabilities when the outcome is dichotomous and the data are analyzed by logistic regression (Fleiss & Berlin, 2009). The OR is obtained by transforming the b weights using eb (Feingold, 2013). In line with our strong a priori hypotheses, all tests were conducted using one-tailed probability thresholds to avoid Type II errors.

Comparison to Sauter et al. (2010).

We used the highly structured choice-from-array task (i.e., including emotion words, repeated and blocked trials, and elaborate manipulation checks) from Sauter et al. (2010) to assess the ability of Hadza participants to match stipulated vocalizations to descriptions of novel emotion categories. As can be seen in panel B of Figure 4, the overall pattern of results resembled the pattern of findings in Sauter et al. (2010) for anger, sadness, fear, and other emotion categories that have been claimed as universal. In Sauter et al. (2010), individuals from the Himba culture in northwestern Namibia were asked to complete the same choice-from-array task, and chose the expected vocalizations at rates above chance, providing apparent support for the universality hypothesis. These results stand in contrast to those obtained when Himba participants freely labeled the vocalizations, which revealed little evidence that posed vocalizations were perceived universally as anger, fear, etc. (Gendron et al., 2014a).

Studies 2 and 3: Urban Industrialized Samples – China and U.S.A.

In Study 1, we predicted and found support for the hypothesis that the classic choice-from-array task provides an experimental context that helps participants choose the target stimuli for emotion categories. A remote sample of Hadza hunter-gatherers selected stipulated (i.e., made-up) target vocalizations for novel emotion categories at levels significantly above chance, appearing to provide evidence of cross-cultural emotion perception despite verified lack of exposure. In Studies 2 and 3, we replicated the study in samples of Chinese and U.S. participants from urban industrialized cultural contexts.

Method

Participants.

Study 2 participants were 34 native Mandarin speakers (24 female) tested in Dalian, China. Participants were recruited through student networks at Liaoning Normal University, and were required to be native Mandarin Chinese speakers, over 18 years of age, and have normal or corrected-to-normal hearing and vision. All testing was completed at the university. Participants ranged in age from 18 to 27, with a median age of 21 years. Study 2 participants provided written consent prior to participation and were remunerated with ¥20.

Participants for Study 3 were 42 native English speakers (21 female) tested in Boston, Massachusetts. Participants were recruited through both the psychology department and the broader community at Northeastern University, and were required to be native English speakers, over 18 years of age, and have normal or corrected-to-normal hearing and vision. All testing was completed at the university. Participants ranged in age from 18 to 67, with a median age of 19 years. Study 3 participants provided written consent prior to participation and were remunerated with study credit or $10.

As in Study 1, participants completed all experimental trials, even those that were later removed from analysis due to failure to pass a manipulation check. The final sample sizes available for analysis are reported in Figure 6.

Stimuli.

Vocalizations were the same as used for Study 1. Scenarios for Study 2 were translated into Mandarin Chinese by a native speaker of the language who is also fluent in English, and were then back-translated into English by a second bilingual speaker to confirm translational equivalency (as recommended by Brislin, 1970). The fifth author (C.L.) served as the Chinese translator for Study 2. Scenarios were recorded by native speakers of Mandarin Chinese (Study 2) and North American English (Study 3).

Procedure.

The procedure was identical to Study 1, including the assessments of vocalization and emotion concept novelty. We also provided written instructions in addition to verbal instructions and feedback because, unlike the Hadza, both Chinese and English are associated with written text. All participants were literate in their native language. Written instructions for Mandarin Chinese were translated and back-translated.

Verifying vocalization novelty.

All Study 2 and 3 participants completed the vocalization free-labeling task prior to the choice-from-array task. Once again, we found that a substantial percentage of labels did not represent known emotion categories in English (Study 2, 43% of responses; Study 3, 36% of responses). In Study 2, consistency increased as compared to Study 1, but specificity continued to be low: for example, while 47% of participants associated gigil with surprise, this label was also applied to every other vocalization, including 37% of the labels for lajja (panel B of Figure S8). In Study 3, participants used more emotion terms to label the scenarios, although these were wide-ranging and were neither consistently nor specifically associated with anger, disgust, fear, happiness, sadness, surprise or other English emotion categories (e.g., only 11% of labels for gigil were associated with surprise, along with 21% of the labels for lajja; panel C of Figure S6). High overall variation in valence and arousal properties were observed for each novel emotion category, indicating little agreement on their affective features within the cultural samples (panels B and C of Figures S9 and S10).

Verifying emotion category novelty.

All participants from Studies 2 and 3 also completed the scenario free-labeling task following completion of the choice-from-array task. As with the vocalization free-labeling task, many participants provided labels that did not represent a known emotion category (Study 2, 33% of responses; Study 3, 41% of responses). Nonetheless, the labels for the scenarios were more consistent than for the vocalizations. For example, gigil was the most consistently-labeled category in Study 2, with 61% of participants associating it with happiness (although 26% labeled it as another emotion category, and 13% did not associate it with a discrete emotion). In Study 3, itoshii was the most consistently-labeled category, with 35% of participants associating it with sadness (although 48% labeled it as another emotion category, and 16% did not associate it with a discrete emotion). See panels B and C of Figure S11 for the full distribution of results. The affective properties of the scenario labels also evidenced greater consistency in comparison to those for the vocalizations, both within and across cultural groups (panels B and C of Figures S12 and S13). However, valence and arousal did not correspond for the vocalization and the scenario associated each novel emotion category, suggesting that participants did not have a pre-existing association between the two in terms of core affective features (see page 15 of the Supplemental Material for further details and discussion).

Results

Bayesian hypothesis testing.

Bayes factors for one-sample t-tests indicated “extreme” evidence in favor of the alternative (context) hypothesis over the null (universal) hypothesis in both Studies 2 and 3, replicating findings from Study 1, meaning that participants were able to choose the stipulated target vocalization over the foil at significant levels. Proportional representations of the ratio of evidence for the context hypothesis to evidence for the universal hypothesis are provided in panel A of Figures 4 (Study 2) and 5 (Study 3). Once again, we examined the robustness of our conclusions by comparing support for the context hypothesis according to varying widths of the prior for effect size. The Bayes factors continued to indicate extreme support for the context hypothesis even in more stringent tests using prior widths close to zero (panel C of Figures 4 and 5). Further, we observed that cumulative data for Studies 2 and 3 provided clear, increasing support for the context hypothesis (panel D of Figures 4 and 5), once again confirming that our sample sizes provided adequate power.

Fig. 4.

Fig. 4.

Inferential plots for Study 2, based on Bayesian one-sample t-tests on overall performance of Chinese participants. a. Pizza plot providing a proportional representation of the ratio of evidence for the alternative (context) hypothesis to evidence for null (universal) hypothesis. b. Posterior distribution based on a one-sided prior distribution. Posterior mass to the right of zero indicates that participant performance is consistently above chance-level responding of .5. c. Bayes factor robustness check. Narrower priors indicate a smaller range of expected effect sizes, favoring the null (universal) hypothesis. d. Sequential analysis of evidence accumulated over the course of data collection. All figures adjusted from JASP (JASP Team, 2017).

Fig. 5.

Fig. 5.

Inferential plots for Study 3, based on Bayesian one-sample t-tests on overall performance of U.S. participants. a. Pizza plot providing a proportional representation of the ratio of evidence for the alternative (context) hypothesis to evidence for null (universal) hypothesis. b. Posterior distribution based on a one-sided prior distribution. Posterior mass to the right of zero indicates that participant performance is consistently above chance-level responding of .5. c. Bayes factor robustness check. Narrower priors indicate a smaller range of expected effect sizes, favoring the null (universal) hypothesis. d. Sequential analysis of evidence accumulated over the course of data collection. All figures adjusted from JASP (JASP Team, 2017).

Hierarchical Generalized Linear Modeling.

In Study 2, Chinese participants selected the made-up vocalizations at a level significantly above chance for four of six novel emotion categories: gigil, glückschmerz, greng jai, and itoshii (Figure 6, panel A; see also Figure S15 for performance distributions and Table S5 for detailed results). All Chinese participants’ data were excluded from the lajja category because they misunderstood lajja’s valence as unpleasant.

In Study 3, U.S. participants selected the made-up vocalizations at a level significantly above chance for all six novel emotion categories (Figure 6, panel B; see also Figure S16 for performance distributions and Table S6 for detailed results). This is despite the fact that, for the category lajja, only ten U.S. participants were retained for analysis for understanding it as pleasant.

Fig. 6.

Fig. 6.

Results from Study 2 (panel A), Study 3 (panel B), compared against Sauter et al. (2010; panel C). Abbreviations: (panel A) gig, gigil; gre, greng jai; glü, glückschmerz; ito, itoshii; laj, lajja; lig, liget; (panel B) ach, achievement; amu, amusement; ang, anger; dis, disgust; fea, fear; ple, sensual pleasure; rel, relief; sad, sadness; sur, surprise. Results data from Sauter et al. (2010) were originally presented in terms of mean number of correct responses (out of four trials per emotion category). These data have been re-plotted in descending order of proportion correct in order to facilitate direct comparison with the present study. Dashed line indicates chance-level performance (.5). Sample size per emotion category reported in white font at the bottom of each column. Standard error bars (± 1 SE) are provided as distributional information only: the location of error bars above the chance line is not indicative of significant above-chance performance because these data were binomially rather than normally distributed. Effect sizes for Studies 2 and 3 are reported above each column using the odds ratio (OR). In line with our strong a priori hypotheses, all tests were conducted using one-tailed probability thresholds to avoid Type II errors.

Comparing performance across Studies 1 through 3.

Across all three samples, there is no emotion category on which participants performed consistently at chance. Participants in all three studies performed above chance or approaching conventional levels of significance in choosing made-up target vocalizations for three novel emotion categories that they had never before been exposed to: gigil, itoshii, and glückschmerz (Figure 3, panel A; Figure 6, panels A and B).

General Discussion

The choice-from-array task remains widely used in psychological research and is the most common task design in studies of emotion perception. In three studies, we demonstrated support for our hypothesis that a classic, highly structured choice-from-array task creates a context that encourages emotional meaning making, and in so doing may provide stronger evidence of cross-cultural emotion perception than would otherwise be observed. Indeed, such evidence is not observed when using other methods for assessing emotion perception (e.g., Crivelli, Jarillo, Russell, & Fernandez-Dols, 2016; Gendron et al., under review; Gendron et al., 2014a). It has long been known that telling participants an emotion story and asking them to select an emotion cue from a small set of options facilitates more consistent performance than less constrained experimental tasks, such as asking participants to freely label emotion cues. That may be why the method is so popular in the first place (Gendron & Barrett, 2009; Gendron & Barrett, 2017; Widen & Russell, 2013). Dashiell (1927) pioneered the choice-from-array task to overcome the comprehension and compliance issues associated with collecting data in preliterate communities who are unfamiliar with standard laboratory methods and did not provide strong evidence of cross-cultural emotion perception (Russell, 1994). Our studies show for the first time that contrived (i.e., made-up) vocalizations for non-universal emotion categories that are novel in three cultural contexts are made to appear universal when tested using a choice-from-array task.

The present studies did not separately manipulate each psychologically potent feature of the classic choice-from-array task (e.g., the presence of emotion words, blocked trials, and elaborate manipulation checks) to examine their independent or synergistic effects. This could be an avenue for future research. For example, previous evidence indicates that a choice-from-array task without the blocked trial structure still encourages above-chance performance (Gendron et al., 2014a). Our studies did not include explicit control conditions due to field site constraints (e.g., limited access to Study 1 participants). Nonetheless, our procedure for verifying vocalization novelty was a free-label task that was conducted prior to the choice-from-array task and did not include emotion words, blocked trials, or an elaborate manipulation check. Data from this task therefore give an estimate of participants’ perception of the novel vocalizations without any additional experimental context. We found that participants within a cultural sample did not label individual vocalizations with a high degree of agreement, and they often used the same or similar words to label multiple vocalizations (for details, see pages 21–22 and 30 of the Results, and pages 9–13 of Supplemental Material). Thus, participants’ freely generated labels did not provide the evidence of cross-cultural emotion perception observed in the choice-from-array task. These findings parallel prior studies of anger, fear, disgust, happiness, sadness, etc. in which tasks the classic choice-from-array task produce more consistent evidence of cross-cultural emotion perception than free-labeling and other less constrained tasks (see Gendron et al., in press, for a review).

Our findings are consistent with other scientific domains, where it is well-known that the experimental context influences what is observed (e.g., physics: Gleiser, 2015; biology: Lewontin, 2001). Here, as in those scientific domains, the observation is not that context is a contaminating factor that produces demand characteristics, but that contextual factors are authentically part of the phenomena in question. In psychology, emotion perception is typically assumed to be a simple matter of registering or detecting emotional information contained in physical cues such as facial movements and vocalizations. In contrast, our findings, along with recent published evidence, suggest that perceivers are active meaning makers who infer the emotional meaning in faces and voices, and that context is a crucial part of this process. Experimenters may not intend for their choice of task to be a meaningful part of the context, but mounting evidence suggests that a perceiver’s brain treats it this way, nonetheless.

Alternative Interpretations

It may be tempting to interpret our findings as evidence that gigil, glückschmerz, greng jai, itoshii, lajja, and liget are, in fact, universal emotion categories. For example, it has recently been suggested that fiero, an Italian concept similar to liget, may be universal (Ekman & Cordaro, 2011), and the number of putative universal emotion categories is continually on the rise (Cordaro et al., 2016; Sauter, 2017; see also Cowen & Keltner, 2017, but also L. F. Barrett, Khan, Dy, & Brooks, 2018). Participants’ completion of the manipulation check procedure could indeed be seen as demonstrating the universality of these emotion categories. However, such an interpretation fails to consider conceptual combination (Barsalou, 1987), the process by which instances of novel categories can be constructed online by integrating existing knowledge of other emotion concepts that participants possess from their own culture. In addition, the hypothesis that gigil, glückschmerz, greng jai, itoshii, lajja, and liget are universal also fails to explain how participants were able to select the appropriate vocalizations for each category, given that these vocalizations were invented by the experimenters.

It is also possible that our participants have biologically basic emotion concepts for anger, sadness, fear, etc., and that they were combining them to perform well on the choice-from-array task (e.g., creating culture-specific emotion blends; Ekman & Cordaro, 2011; Shao, Doucet, & Caruso, 2015). Yet data collected using less constrained experimental tasks in other small-scale societies (including the same group of Hadza hunter-gatherers) suggests that these emotion categories are not universal (e.g., Crivelli, Jarillo, et al., 2016; Gendron et al., under review; Gendron et al., 2014a). Nevertheless, conceptual combination may allow people to experience and perceive emotions across cultural boundaries: it may be a means of creating universality by sharing (L. F. Barrett, 2017a). For example, it is plausible that participants in the original studies of cross-cultural emotion perception (e.g., Ekman et al., 1969) were combining concepts from their own cultures to complete choice-from-array tasks. While conceptual combination is not inherently at odds with cross-cultural emotion perception, it is inconsistent with the strongest, traditional version of the universality hypothesis. In that view, emotion perception is an inborn or early-to-develop capacity that is independent of emotion concepts (e.g., Izard, 1994). From this perspective, conceptual combination would not be needed for emotion categories whose universality derives from their biologically basicness and innateness. Although recent accounts of universality have discussed it as a graded phenomenon (e.g., Cordaro et al., 2018; Keltner & Haidt, 1999) that can vary based on culture-specific display rules, decoding rules, and dialects of non-verbal behaviors, even discussions that relax the assumptions of universality still assume that culture-specific experience tunes inborn, fixed action programs. A constructionist account, by contrast, posits conceptual combination as a mechanism by which people acquire emotion concepts and become emotionally acculturated (L. F. Barrett, 2017), which guides their expressive behaviros from the outset. This hypothesis awaits experimental testing.

Further, it is important to consider the implications of a hypothesis that gigil, glückschmerz, greng jai, itoshii, lajja, and liget are sufficiently translatable from other emotion concepts that participants possess from their own culture. If we are willing to infer from task performance that emotion categories are ‘sufficiently translatable’, then this interpretation is equally applicable to fear, sadness, anger, disgust, happiness, and surprise. Put plainly, if ‘universal’ emotion categories cannot be distinguished from ‘novel’ emotion categories on the basis of performance on the classic choice-from-array task, then this task is not an adequate method for assessing universality, calling the interpretation of hundreds of prior studies into question.

A related line of interpretation is that participants were familiar enough with the emotional experiences associated with the novel emotion categories that they could successfully complete the task, even though the six concepts we used are not encoded in Hadzane, Mandarin Chinese, or English by unique words. That participants were able to bring their prior experience to bear in completing the task is not at issue: it is precisely through mechanisms such as conceptual combination that, we propose, humans are able to gain a basic understanding of each other’s emotional state. Therefore, the weakest possible interpretation of the current findings is that the classic choice-from-array task faithfully primes (but does not enhance) participants’ ability to make appropriate meaning out of situated non-verbal cues, and that we have only succeeded in extending the range of situations and cues. The history of published studies shows, however, that highly structured choice-from-array tasks provide support for cross-cultural emotion perception when other methods do not, calling into question the robustness and replicability of evidence for universal emotions.

A final alternative interpretation of the current findings is that the six novel emotion categories, along with their stipulated vocalizations, are in fact subordinate members of so-called ‘basic’ emotions (e.g., gigil is a form of happiness). The results of our free-labeling data for vocalizations and long-form scenarios across all three studies do not immediately support this interpretation. Overall, participants in all three cultural samples provided labels for the novel vocalizations that were general affective descriptions such as “good” or “bad”, or offered words such as “love” that do not correspond with anger, sadness, fear, disgust, happiness or surprise. In contrast, a notable proportion of participants consistently provided labels for three novel emotion scenarios that corresponded with presumed universal categories: gigil with happiness (Hadza, 80% of labels; China, 61%; U.S., 29%), itoshii with sadness (Hadza, 70% of labels; China, 47%; U.S., 35%), and glückschmerz with anger (Hadza, 33% of labels; China, 23%; U.S., 24%). These labels were not necessarily used in a specific way, however. Happiness was also frequently associated with other novel emotion scenarios (e.g., Hadza, 83% of liget labels; China, 13% of lajja labels), as was sadness (e.g., Hadza, 40% of greng jai labels; China, 8% of gluckschmerz labels) (see Figure S11 for details). Importantly, this pattern of findings is also consonant with an interpretation of conceptual combination: the labels for novel vocalizations and novel scenarios did not consistently correspond with the same emotion categories, implying that performance on the choice-from-array task may not be driven by conceptual labeling of the scenario alone. To perform well on the task, participants would still have needed to extend any pre-existing categories (perhaps by using conceptual combination) to accommodate novel vocalizations.

Processes Supporting Cross-Cultural Emotion Perception

If the experimental context is full of psychologically potent features that can influence how people infer meaning in vocalizations, then our findings have broader implications for the study of emotion perception. Certainly our findings join others in casting doubt on the claim that cross-cultural emotion perception is “an established axiom of behavioral science” (Izard & Saxton, 1988, pp. 651–652). But beyond the potential inadequacy of the classic choice-from-array task for testing the universality hypothesis, our findings offer indirect support for the hypothesis that emotion perception in the real world is the result of multiple processes, such as identifying similarities between someone’s physical changes (facial movements, vocal acoustics, etc.) in a particular situational context and prior experiences from the past, using knowledge of emotion words and concepts, employing process-of-elimination strategies, as well as perceiving affect and learning categories(for a discussion, see L. F. Barrett, 2017a). These processes may function like ingredients that contribute to different recipes for emotion perception in different cultures.

The current studies do not provide systematic evidence for what these processes are or how they work, individually or synergistically. Instead, they can be thought of as “proof of concept” that such studies are sorely needed and worth the investment. This was also the conclusion of a recent review of scientific evidence on inferring emotion in human facial movements, to be published in Psychological Science in the Public Interest:

The science of emotion expression and emotion perception has been more a science of stereotypes rather than a science of how people actually move their faces to express emotion and the processes by which those movements carry information about emotion to someone else (a perceiver) … In reality, emotions are expressed with facial movements that are more variable and context-dependent … Their context-dependence goes well beyond display rules or cultural accents. As a consequence, the stereotypes … must be replaced by a thriving scientific effort to observe and describe the lexicon of context-sensitive ways in which people move their facial muscles to express emotion, and the discovery of when and how people infer emotions in other people’s facial movements. (L. F. Barrett et al., forthcoming, p. 114 of the manuscript draft)

Following published research, we would suggest the same insights hold true for vocalizations. The present studies, while not conclusively revealing which processes should be the target of empirical focus, do make several suggestions.

Cognitive bootstrapping.

The experimental features of forced-choice designs, including the classic choice-from-array task, can be cognitively bootstrapped (e.g., by identifying perceptual similarities or employing process-of-elimination strategies; Russell, 1994) to promote online category learning. Developmental studies suggest that cognitive bootstrapping underlies successful performance on a wide range of experimental tasks (e.g., Cassels & Birch, 2014; Diesendruck, Hall, & Graham, 2006; Haryu, Imai, & Okada, 2011; Markman & Wachtel, 1988; Waxman & Booth, 2001). In recent studies of emotion perception, participants have been shown to employ process-of-elimination strategies when presented with novel emotion words and facial configurations. These strategies are used in selecting a response option within a given trial (Nelson & Russell, 2016a), in tracking previously selected response options across trials (DiGirolamo & Russell, 2017), and in freely labeling stimuli previously presented in a separate task (Nelson & Russell, 2016b).

Affect perception.

The inference of affective meaning may have contributed to our observed effects and may also be an important feature in emotion perception. There is ample evidence that facial and vocal cues are perceived in terms of the valence and the level of arousal that they communicate (L. F. Barrett & Bliss-Moreau, 2009; Russell & Barrett, 1999). Affect perception is robust across cultures (Russell, 1991; Russell et al., 2003; Russell & Barrett, 1999), in children who do not possess explicit emotion concept knowledge (Widen, 2016), and in patients who have lost emotion knowledge due to semantic dementia (Lindquist et al., 2014). For vocalizations, specific acoustic features (e.g., fundamental frequency and amplitude) are reliably associated with the perception of arousal (Bachorowski, 1999; Bachorowski & Owren, 2008; but see Scherer, Johnstone, & Klasmeyer, 2003 for a review of a discrete emotions account of acoustic features). Recent work on arousal perception suggests that these acoustic features hold across species (Filippi et al., 2017)6.

Category learning.

Participants may also have leveraged conceptual features of the experimental context, such as emotion words, to complete the task of emotion perception. In this way, our findings hold clues to improving cross-cultural emotion communication, in that they suggest a view of emotion perception as culturally dependent upon concepts that are acquired through category learning. A growing body of work demonstrates that words serve as invitations to form abstract categories with limited perceptual regularity across instances (e.g., Ferry, Hespos, & Waxman, 2010). Mounting evidence from studies of the face indicate that emotion categories are abstract in that their instances are highly variable across situations (L. F. Barrett et al., forthcoming), as do studies of psychophysiology and brain imaging (e.g., C. D. Wilson-Mendenhall, L. F. Barrett, & Lawrence W. Barsalou, 2015). Our choice-from-array task, based on that used by Sauter et al. (2010; 2015), contained design features that may have allowed participants to quickly learn novel emotion categories when labeled with words, such that they achieved levels of performance equivalent to those reported in support of cross-cultural emotion perception. This pattern of performance is consistent with a large body of findings from the developmental psychology literature showing that children and even young infants can learn novel, abstract categories with the help of words (e.g., Ferry et al., 2010; Waxman & Booth, 2001; Xu, Cote, & Baker, 2005; Yin & Csibra, 2015).

Conceptual combination.

We hypothesize that category learning may occur in the context of a brief experimental task, as in real life, through the process of conceptual combination. In our study, better performance may have been observed for those categories that were easier to construct via combination of the knowledge and experiences promoted by a given culture. Ease of conceptual combination may also be reflected by the number of participants who passed the manipulation check for a novel emotion category. In such cases, participants will not necessarily understand the concept exactly as a native would; however, conceptual combination may allow for some cross-cultural communication, albeit imperfect. As such, our findings are consistent with the hypothesis that conceptual combination may be the foundation of category learning and, thereby, of cross-cultural emotion communication. This is a hypothesis in need of further scientific investigation.

The specific pattern of performance observed across the three samples suggests that pre-existing cultural knowledge may have played a role in task performance. We speculate that participants were more easily able to learn novel emotion categories that fit local cultural values and practices, consistent with research on cultural fit and emotional values (Richerson & Boyd, 2005; Tamir et al., 2016). For example, the concept of greng jai, which describes the combination of gratitude and social guilt one feels when offered an overly generous gift or burdensome favor, likely does not fit as well with Hadza cultural practices of resource sharing and expectations of communal collaboration (Apicella, Marlowe, Fowler, & Christakis, 2012), as it does with Chinese norms associated with maintaining and saving face (Chang & Holt, 1994). Likewise, we can hypothesize that U.S. participants are anchoring on how the experience of greng jai clashes with the core cultural value of personal independence (Markus & Kitayama, 1991). Future research could further develop a priori hypotheses to test how category learning is impacted by conceptual fit along these and other cultural dimensions.

Conclusion

The patterns of performance we observed in the present studies suggest a new context within which to integrate the hundreds of published studies using a choice-from-array task to test, and ultimately provide support for, the hypothesis that certain emotion categories are universally perceived. Participants from three cultural samples selected stipulated (i.e., made-up) target vocalizations for unfamiliar emotion categories at levels exceeding chance, suggesting that certain experimental design features may facilitate emotional meaning making, even when the emotion concepts and the vocalizations are novel (i.e., not consistently and specifically associated with pre-existing emotion categories). Accordingly, our findings invite discussions about the psychological potency of experimental design features in a task that is pervasive in psychological research, as well as the meaning-making processes that undergird emotion perception. Furthermore, our findings build on previously published studies in suggesting that the variety of processes that contribute to emotion perception, and social perception more generally, may be differentially recruited across cultural contexts.

Our findings also suggest a deeper point about ecological validity. Humans make meaning of their environment, usually as an automatic, effortless, and obligatory consequence of the way they process information (L. F. Barrett, 2017b). This meaning making is not suspended in the context of an experiment; rather, features of this context may facilitate it, becoming psychologically potent in a way that may or may not be representative of everyday life. To acknowledge and account for this, studies of cross-cultural emotion perception must compare findings across multiple methods (e.g., Crivelli, Jarillo, et al., 2016; Crivelli, Russell, Jarillo, & Fernández-Dols, 2017; Gendron et al., 2014a), sample spontaneous behavior from naturally occurring interactions (e.g., Crivelli, Carrera, & Fernández-Dols, 2015; Fernández-Dols & Ruiz-Belda, 1995; Tracy & Matsumoto, 2008), and explore how domain-general processes such as category learning and conceptual combination may influence performance. Together, these steps will lead to a more robust, nuanced, and replicable science of human behavior, including emotion perception.

Supplementary Material

S1

Acknowledgements

The authors are grateful to A. Mabulla, M. Peterson, L. Wang, and J. Zhang for research support, and the Hadza communities and Liaoning Normal University where this research was conducted.

Funding

This research was supported by a U.S. Army Research Institute for the Behavioral and Social Sciences grant (W911NF-15-1-0647) awarded to M. Gendron and L. F. Barrett, as well as a NIH NRSA fellowship (5F32MH105052) awarded to M. Gendron. This research was further supported by the National Natural Science Foundation of China (NSFC 31170982) and Royal Society (IE121122) International Cooperation and Exchanges NSFC (31311130123) awarded to Q. Liu and D. Roberson, and extended to M. Gendron. The views, opinions, and/or findings contained in this paper are those of the authors and shall not be construed as an official Department of the Army position, policy, or decision, unless so designated by other documents.

Footnotes

1

For the purposes of illustrating the theoretical origins of tests of cross-cultural emotion perception, we refer to the strongest, traditional version of the universality hypothesis, and the debate as one between total universality vs. cultural relativism. However, more recent accounts have described universality as a graded phenomenon (e.g., Cordaro et al., 2018; Keltner & Haidt, 1999). This debate can alternatively be framed in terms of sources of cross-cultural (and intra-cultural) consistency vs. diversity (Crivelli & Gendron, 2017a; Gendron, Crivelli, & Barrett, in press; Russell, 1995).

2

An alternative version is to shown photos of facial configurations, one at a time, along with a small set of emotion words and are tasked with selecting the best matching emotion word for each photo.

3

Gendron et al. (2013) was first written and submitted for publication in 2010.

4

Anecdotal evidence is consistent with the idea that people in remote, small-scale villages many not understand emotion concepts in the same way as people from urban, western cultural contexts. Describing his first visit to the Fore tribe in New Guinea, Ekman (2007) wrote “I asked them to make up a story about each facial expression [photograph]. ‘Tell me what is happening now, what happened before to make the person show this expression, and what is going to happen next.’ It was like pulling teeth. I am not certain whether it was the translation process, or the fact that they have no idea what I wanted to hear or why I wanted them to do this. Perhaps making up stories about strangers was just something the Fore didn’t do.” (p. 7) But it is also possible that they just don’t understand emotion as people from the U.S. do, and they may not have the same emotion concepts.

5

In the analysis, we specified a Cauchy prior of r = √2/2 (i.e., .707) (Morey & Rouder, 2015). A Cauchy prior width of r = 1 has also been recommended (Rouder et al., 2009; Wetzels et al., 2009). We used the smaller r value (i.e., r = .707) because it represented a more conservative test of our context hypothesis; however, we generated Bayes factors at r = 1 and r = 1.414 for sake of comparison (panel C in Figures 2, 4, and 5).

6

Note that the vocalizations tested were not normed for actual physiological activation; instead, level of arousal was inferred based on the context in which the vocalizations were produced.

References

  1. Agresti A (2002). Categorical Data Analysis (2nd ed.). New York, NY: John Wiley & Sons. [Google Scholar]
  2. Anderson E, Siegel EH, & Barrett LF (2011). What you feel influences what you see: The role of affective feelings in resolving binocular rivalry. Journal of Experimental Social Psychology, 47(4), 856–860. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Apicella CL, & Crittenden AN (2015). Hunter-gatherer families and parenting In Buss DM (Ed.), The handbook of evolutionary psychology (Vol. IV, pp. 578–597). Hoboken, NJ: John Wiley & Sons, Inc. [Google Scholar]
  4. Apicella CL, Marlowe FW, Fowler JH, & Christakis NA (2012). Social networks and cooperation in hunter-gatherers. Nature, 481(7382), 497–501. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Team Audacity. (2015). Audacity (R): Free Audio Editor and Recorder [Computer software] (Version 2.1.2). Retrieved from http://www.audacityteam.org/
  6. Aviezer H, & Hassin R (2017). Inherently Ambiguous: An Argument for Contextualized Emotion Perception In Fernandez-Dols JM & Russell JA (Eds.), The Science of Facial Expression: Oxford University Press. [Google Scholar]
  7. Aviezer H, Hassin RR, Ryan J, Grady C, Susskind J, Anderson A, . . . Bentin S (2008). Angry, disgusted, or afraid? Studies on the malleability of emotion perception. Psychological science, 19(7), 724–732. [DOI] [PubMed] [Google Scholar]
  8. Aviezer H, Trope Y, & Todorov A (2012). Body cues, not facial expressions, discriminate between intense positive and negative emotions. Science, 338(6111), 1225–1229. doi: 10.1126/science.1224313 [DOI] [PubMed] [Google Scholar]
  9. Bachorowski J-A (1999). Vocal expression and perception of emotion. Current Directions in Psychological Science, 8(2), 53–57. [Google Scholar]
  10. Bachorowski J-A, & Owren MJ (2008). Vocal expressions of emotion In Lewis M, Haviland-Jones JM, & Barrett LF (Eds.), Handbook of emotions (3rd ed., pp. 196–210): The Guilford Press. [Google Scholar]
  11. Bar M (2004). Visual objects in context. Nature Reviews Neuroscience, 5(8), 617–629. [DOI] [PubMed] [Google Scholar]
  12. Barrett HC, Bolyanatz A, Crittenden AN, Fessler DM, Fitzpatrick S, Gurven M, . . . Pisor A (2016). Small-scale societies exhibit fundamental variation in the role of intentions in moral judgment. Proceedings of the National Academy of Sciences, 113(17), 4688–4693. [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Barrett LF (2017a). How emotions are made: The secret life the brain and what it means for your health, the law, and human nature. New York, NY: Houghton Mifflin Harcourt. [Google Scholar]
  14. Barrett LF (2017b). The theory of constructed emotion: An active inference account of interoception and categorization. Social Cognitive and Affective Neuroscience, 1–23. doi: 10.1093/scan/nsw154 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Barrett LF, Adolphs R, Marsella S, Martinez A, & Pollak S (forthcoming). Emotional Expressions Reconsidered: Challenges to Inferring Emotion in Human Facial Movements. Psychological Science in the Public Interest. [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Barrett LF, & Bliss-Moreau E (2009). Affect as a Psychological Primitive. Advances in Experimental Social Psychology, 41, 167–218. doi: 10.1016/S0065-2601(08)00404-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Barrett LF, Khan Z, Dy J, & Brooks D (2018). Nature of Emotion Categories: Comment on Cowen and Keltner. Trends in Cognitive Sciences, 22(2), 97–99. [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Barrett LF, Mesquita B, & Gendron M (2011). Context in emotion perception. Current Directions in Psychological Science, 20(5), 286–290. [Google Scholar]
  19. Barrett LF, & Niedenthal PM (2004). Valence Focus and the Perception of Facial Affect. Emotion, 4(3), 266–274. [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Barsalou LW (1987). The instability of graded structure: Implications for the nature of concepts In Neisser U (Ed.), Concepts reconsidered: The ecological and intellectual bases of categories. New York, NY: Cambridge University Press. [Google Scholar]
  21. Belin P, Fillion-Bilodeau S, & Gosselin F (2008). The Montreal Affective Voices: a validated set of nonverbal affect bursts for research on auditory affective processing. Behavior Research Methods, 40(2), 531–539. [DOI] [PubMed] [Google Scholar]
  22. Berbesque JC, Wood BM, Crittenden AN, Mabulla A, & Marlowe FW (2016). Eat first, share later: Hadza hunter-gatherer men consume more while foraging than in central places. Evolution and Human Behavior, 37(4), 281–286. [Google Scholar]
  23. Berger JO, & Mortera J (1999). Default Bayes factors for nonnested hypothesis testing. Journal of the american statistical association, 94(446), 542–554. [Google Scholar]
  24. Berger JO, & Wolpert RL (1988). The Likelihood Principle (2nd ed.). Haywood, CA: The Institute of Mathematical Sciences. [Google Scholar]
  25. BetterThanEnglish. (2012). Better Than English [Online database]. Retrieved from http://betterthanenglish.com/
  26. Brislin RW (1970). Back-translation for cross-cultural research. Journal of cross-cultural psychology, 1(3), 185–216. [Google Scholar]
  27. Bruner JS (1957). Going beyond the information given. Contemporary approaches to cognition, 1(1), 119–160. [Google Scholar]
  28. Bryant GA, & Barrett HC (2008). Vocal emotion recognition across disparate cultures. Journal of Cognition and Culture, 8(1), 135–148. [Google Scholar]
  29. Bryant GA, Fessler DM, Fusaroli R, Clint E, Aarøe L, Apicella CL, . . . Chavez B (2016). Detecting affiliation in colaughter across 24 societies. Proceedings of the National Academy of Sciences, 113(17), 4682–4687. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Calbi M, Angelini M, Gallese V, & Umiltà M (2017). “ Embodied Body Language”: an electrical neuroimaging study with emotional faces and bodies. Scientific reports, 7(1), 6875–6875. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Carroll JM, & Russell JA (1996). Do facial expressions signal specific emotions? Judging emotion from the face in context. Journal of personality and social psychology, 70(2), 205–218. [DOI] [PubMed] [Google Scholar]
  32. Cassels TG, & Birch SA (2014). Comparisons of an open-ended vs. forced-choice ‘mind reading’task: Implications for measuring perspective-taking and emotion recognition. PLoS One, 9(12), e93653. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Chanes L, Wormwood JB, Betz N, & Barrett LF (2018). Facial expression predictions as drivers of social perception. Journal of personality and social psychology, 114, 380–396. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Chang H-C, & Holt GR (1994). A Chinese perspective on face as inter-relational concern. The challenge of facework: Cross-cultural and interpersonal issues, 95–132. [Google Scholar]
  35. Clore GL, Ortony A, & Foss MA (1987). The psychological foundations of the affective lexicon. Journal of personality and social psychology, 53(4), 751–766. [Google Scholar]
  36. Cordaro DT, Keltner D, Tshering S, Wangchuk D, & Flynn LM (2016). The voice conveys emotion in ten globalized cultures and one remote village in Bhutan. Emotion, 16(1), 117–128. [DOI] [PubMed] [Google Scholar]
  37. Cordaro DT, Sun R, Keltner D, Kamble S, Huddar N, & McNeil G (2018). Universals and cultural variations in 22 emotional expressions across five cultures. Emotion, 18(1), 75. [DOI] [PubMed] [Google Scholar]
  38. Cowen AS, & Keltner D (2017). Self-report captures 27 distinct categories of emotion bridged by continuous gradients. Proceedings of the National Academy of Sciences, 114(38), E7900–E7909. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Crivelli C, Carrera P, & Fernández-Dols J-M (2015). Are smiles a sign of happiness? Spontaneous expressions of judo winners. Evolution and Human Behavior, 36(1), 52–58. [Google Scholar]
  40. Crivelli C, & Gendron M (2017a). Facial expressions and emotions in indigenous societies In Russell JA & Fernández-Dols J-M (Eds.), The Science of Facial Expression. Oxford, UK: Oxford University Press. [Google Scholar]
  41. Crivelli C, & Gendron M (2017b). Facial expressions and emotions in indigenous societies In Fernandez-Dols JM & Russell JA (Eds.), The Science of Facial Expression (pp. 497–515). Oxford: Oxford University Press. [Google Scholar]
  42. Crivelli C, Jarillo S, Russell JA, & Fernandez-Dols JM (2016). Reading emotions from faces in two indigenous societies. J Exp Psychol Gen, 145(7), 830–843. doi: 10.1037/xge0000172 [DOI] [PubMed] [Google Scholar]
  43. Crivelli C, Russell JA, Jarillo S, & Fernández-Dols J-M (2016). The fear gasping face as a threat display in a Melanesian society. Proceedings of the National Academy of Sciences, 113(44), 12403–12407. [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Crivelli C, Russell JA, Jarillo S, & Fernández-Dols J-M (2017). Recognizing spontaneous facial expressions of emotion in a small-scale society of Papua New Guinea. Emotion, 17(2), 337–347. [DOI] [PubMed] [Google Scholar]
  45. Darwin C (1872/2005). The expression of the emotions in man and animais: Digireads.com Publishing. [Google Scholar]
  46. Dashiell JF (1927). A new method of measuring reactions to facial expression of emotion. Psychological bulletin, 24, 174–175. [Google Scholar]
  47. De Boinod AJ (2007). The meaning of tingo: And other extraordinary words from around the world. London: Penguin. [Google Scholar]
  48. de Gelder B (2016). Emotional Body Perception in the Wild In Barrett LF, Lewis M, & Haviland-Jones JM (Eds.), Handbook of emotions (4 ed., pp. 483–494): Guildford Publications. [Google Scholar]
  49. Diesendruck G, Hall DG, & Graham SA (2006). Children’s use of syntactic and pragmatic knowledge in the interpretation of novel adjectives. Child development, 77(1), 16–30. [DOI] [PubMed] [Google Scholar]
  50. DiGirolamo MA, & Russell JA (2017). The emotion seen in a face can be a methodological artifact: The process of elimination hypothesis. Emotion, 17(3), 538–546. [DOI] [PubMed] [Google Scholar]
  51. Doyle CM, & Lindquist KA (2018). When a Word Is Worth a Thousand Pictures: Language Shapes Perceptual Memory for Emotion. Journal of Experimental Psychology: General, 147(1), 62–73. [DOI] [PubMed] [Google Scholar]
  52. Edwards W, Lindman H, & Savage LJ (1963). Bayesian statistical inference for psychological research. Psychological review, 70(3), 193–242. [Google Scholar]
  53. Ekman P (1972). Universals and cultural differences in facial expressions of emotion. Paper presented at the Nebraska Symposium on Motivation, 1971. [Google Scholar]
  54. Ekman P (2007). Emotions revealed: Recognizing faces and feelings to improve communication and emotional life: Macmillan. [Google Scholar]
  55. Ekman P (2017). Facial Expressions In Fernandez-Dols JM & Russell JA (Eds.), The Science of Facial Expression: Oxford University Press. [Google Scholar]
  56. Ekman P, & Cordaro DT (2011). What is meant by calling emotions basic. Emotion Review, 3(4), 364–370. [Google Scholar]
  57. Ekman P, & Friesen WV (1971). Constants across cultures in the face and emotion. Journal of personality and social psychology, 17(2), 124–129. [DOI] [PubMed] [Google Scholar]
  58. Ekman P, Heider E, Friesen WV, & Heider K (1972). Facial expression in a preliterate culture. [Google Scholar]
  59. Ekman P, Levenson RW, & Friesen WV (1983). Autonomic nervous system activity distinguishes among emotions. Science, 221(4616), 1208–1210. [DOI] [PubMed] [Google Scholar]
  60. Ekman P, Sorenson ER, & Friesen WV (1969). Pan-cultural elements in facial displays of emotion. Science, 164(3875), 86–88. [DOI] [PubMed] [Google Scholar]
  61. Elfenbein HA, & Ambady N (2002). On the universality and cultural specificity of emotion recognition: a meta-analysis. Psychological Bulletin, 128(2), 203–235. [DOI] [PubMed] [Google Scholar]
  62. Feingold A (2013). A regression framework for effect size assessments in longitudinal modeling of group differences. Review of General Psychology, 17(1), 111–121. [DOI] [PMC free article] [PubMed] [Google Scholar]
  63. Fernández-Dols JM, & Ruiz-Belda M-A (1995). Are smiles a sign of happiness? Gold medal winners at the Olympic Games. Journal of personality and social psychology, 69(6), 1113–1119. [Google Scholar]
  64. Fernández-Dols JM, & Ruiz-Belda M-A (1997). Spontaneous facial behavior during intense emotional episodes: Artistic truth and optical truth. The psychology of facial expression, 2, 255–274. [Google Scholar]
  65. Ferry AL, Hespos SJ, & Waxman SR (2010). Categorization in 3-and 4-month-old infants: an advantage of words over tones. Child development, 81(2), 472–479. [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Filippi P, Congdon J, Hoang J, Bowling D, Reber S, Pašukonis A, . . . Sturdy C (2017). Humans recognize emotional arousal in vocalizations across all classes of terrestrial vertebrates: evidence for acoustic universals. Proceedings of the Royal Society: Section B (Biological Sciences), 284(1859), 20170990. doi: 10.1098/rspb.2017.0990 [DOI] [PMC free article] [PubMed] [Google Scholar]
  67. Fleiss JL, & Berlin JA (2009). Effect sizes for dichotomous data In Cooper H, Hedges LV, & Valentine JC (Eds.), The handbook of research synthesis and meta-analysis (2nd ed., pp. 237–253). New York, NY: Russell Sage. [Google Scholar]
  68. Fridlund AJ (1991). Sociality of solitary smiling: Potentiation by an implicit audience. Journal of personality and social psychology, 60(2), 229–240. [Google Scholar]
  69. Fridlund AJ (1994). Human facial expression: An evolutionary view: Academic Press. [Google Scholar]
  70. Fugate J, Gendron M, Nakashima S, & Barrett LF (2017). Emotion Words: Adding Face Value. Emotion, Advance online publication. doi: 10.1037/emo0000330 [DOI] [PubMed] [Google Scholar]
  71. Fugate J, Gouzoules H, & Barrett LF (2010). Reading chimpanzee faces: evidence for the role of verbal labels in categorical perception of emotion. Emotion, 10(4), 544–554. [DOI] [PMC free article] [PubMed] [Google Scholar]
  72. Gendron M (2017). Revisiting diversity: cultural variation reveals the constructed nature of emotion perception. Current opinion in psychology, 17, 145–150. [DOI] [PMC free article] [PubMed] [Google Scholar]
  73. Gendron M, & Barrett LF (2009). Reconstructing the Past: A Century of Ideas About Emotion in Psychology. Emotion Review, 1(4), 316–339. doi: 10.1177/1754073909338877 [DOI] [PMC free article] [PubMed] [Google Scholar]
  74. Gendron M, & Barrett LF (2017). Facing the Past: A History of the Face in Psychological Research on Emotion Perception In Fernandez-Dols JM & Russell JA (Eds.), The Science of Facial Expression: Oxford University Press. [Google Scholar]
  75. Gendron M, Crivelli C, & Barrett LF (in press). Universality Reconsidered: Diversity in Meaning Making about Facial Expressions. Current Directions in Psychological Science. [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Gendron M, Hoemann K, Crittenden AN, Msafiri S, Ruark G, & Barrett LF (under review). Emotion perception in Hadza hunter-gatherers: A test of evolutionary hypotheses about universal emotions. [Google Scholar]
  77. Gendron M, Lindquist KA, Barsalou LW, & Barrett LF (2012). Emotion words shape emotion percepts. Emotion, 12(2), 314–325. doi: 10.1037/a0026007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Gendron M, Mesquita B, & Barrett LF (2013). Emotion perception: putting the face in context In Reisberg D (Ed.), Oxford Handbook of Cognitive Psychology (pp. p. 379–389)). New York: Oxford University Press. [Google Scholar]
  79. Gendron M, Roberson D, & Barrett LF (2015). Cultural variation in emotion perception is real: A response to Sauter, Eisner, Ekman, and Scott (2015). Psychol Sci, 26(3), 357–359. [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Gendron M, Roberson D, van der Vyver JM, & Barrett LF (2014a). Cultural relativity in perceiving emotion from vocalizations. Psychological science, 25(4), 911–920. [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Gendron M, Roberson D, van der Vyver JM, & Barrett LF (2014b). Perceptions of emotion from facial expressions are not culturally universal: evidence from a remote culture. Emotion, 14(2), 251–262. doi: 10.1037/a0036052 [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. Gewald J-B (2010). Remote but in contact with history and the world. Proceedings of the National Academy of Sciences, 107(18), E75–E75. [DOI] [PMC free article] [PubMed] [Google Scholar]
  83. Gleiser M (2015). The Island of knowledge: The limits of science and the search for meaning: Basic Books. [Google Scholar]
  84. Guo G, & Zhao H (2000). Multilevel modeling for binary data. Annual review of sociology, 26(1), 441–462. [Google Scholar]
  85. Gygi B, Kidd GR, & Watson CS (2007). Similarity and categorization of environmental sounds. Attention, Perception, & Psychophysics, 69(6), 839–855. [DOI] [PubMed] [Google Scholar]
  86. Haidt J, & Keltner D (1999). Culture and facial expression: Open-ended methods find more expressions and a gradient of recognition. Cognition & Emotion, 13(3), 225–266. [Google Scholar]
  87. Haryu E, Imai M, & Okada H (2011). Object similarity bootstraps young children to action-based verb extension. Child development, 82(2), 674–686. [DOI] [PubMed] [Google Scholar]
  88. Henrich J, Heine SJ, & Norenzayan A (2010). Most people are not WEIRD. Nature, 466(7302), 29. [DOI] [PubMed] [Google Scholar]
  89. Hess U, & Hareli S (2017). The Social Signal Value of Emotions: The Role of Contextual Factors in Social Inferences Drawn from Emotion Displays In Fernandez-Dols JM & Russell JA (Eds.), The Science of Facial Expression: Oxford University Press. [Google Scholar]
  90. Hout MC, Papesh MH, & Goldinger SD (2013). Multidimensional scaling. Wiley Interdisciplinary Reviews: Cognitive Science, 4(1), 93–103. [DOI] [PMC free article] [PubMed] [Google Scholar]
  91. Izard CE (1994). Innate and universal facial expressions: evidence from developmental and cross-cultural research. Psychological bulletin, 115(2), 288–299. [DOI] [PubMed] [Google Scholar]
  92. Izard CE, & Saxton PM (1988). Emotions In Atkinson RC, Hernstein RJ, Lindzey G, & Luce RD (Eds.), Stevens’ handbook of experimental psychology: Perception and motivation; Learning and cognition (2 ed., Vol. 1–2, pp. 627–676). Oxford, England: John Wiley & Sons. [Google Scholar]
  93. Jack RE, Sun W, Delis I, Garrod OG, & Schyns PG (2016). Four not six: Revealing culturally common facial expressions of emotion. Journal of Experimental Psychology: General, 145(6), 708–730. [DOI] [PubMed] [Google Scholar]
  94. Jaeger TF (2008). Categorical data analysis: Away from ANOVAs (transformation or not) and towards logit mixed models. Journal of Memory and language, 59(4), 434–446. [DOI] [PMC free article] [PubMed] [Google Scholar]
  95. JASP Team. (2017). JASP [Computer software] (Version 0.8.1.2). Retrieved from https://jasp-stats.org/
  96. Jaworska N, & Chupetlovska-Anastasova A (2009). A review of multidimensional scaling (MDS) and its utility in various psychological domains. Tutorials in quantitative methods for psychology, 5(1), 1–10. [Google Scholar]
  97. Jeffreys H (1961). Theory of probability (3rd ed.). Oxford: Oxford University Press, Clarendon Press. [Google Scholar]
  98. Jones NB (2016). Demography and Evolutionary Ecology of Hadza Hunter-Gatherers: Cambridge University Press. [Google Scholar]
  99. Kass RE, & Raftery AE (1995). Bayes factors. Journal of the american statistical association, 90(430), 773–795. [Google Scholar]
  100. Keltner D, & Haidt J (1999). Social functions of emotions at four levels of analysis. Cognition & Emotion, 13(5), 505–521. [Google Scholar]
  101. Kenny DA, Korchmaros JD, & Bolger N (2003). Lower Level Mediation in Multilevel Models. Psychological Methods, 8(2), 115–128. [DOI] [PubMed] [Google Scholar]
  102. Laukka P, Elfenbein HA, Söder N, Nordström H, Althoff J, Chui W, . . . Thingujam NS (2013). Cross-cultural decoding of positive and negative non-linguistic emotion vocalizations. Frontiers in Psychology, 4(353), 185–192. [DOI] [PMC free article] [PubMed] [Google Scholar]
  103. Lewontin RC (2001). The triple helix: Gene, organism, and environment: Harvard University Press. [Google Scholar]
  104. Lin P-Y (2013, January 23, 2013). Unspeakableness: An Intervention of Language Evolution and Human Communication. Retrieved from http://uniquelang.peiyinglin.net/01untranslatable.html [Google Scholar]
  105. Lindquist KA, Gendron M, Barrett LF, & Dickerson BC (2014). Emotion perception, but not affect perception, is impaired with semantic memory loss. Emotion, 14(2), 375–387. [DOI] [PMC free article] [PubMed] [Google Scholar]
  106. Lomas T (2016). Towards a positive cross-cultural lexicography: Enriching our emotional landscape through 216 ‘untranslatable’words pertaining to well-being. The Journal of Positive Psychology, 11(5), 546–558. [Google Scholar]
  107. Lupyan G, Rakison DH, & McClelland JL (2007). Language is not just for talking: redundant labels facilitate learning of novel categories. Psychological science, 18(12), 1077–1083. doi: 10.1111/j.1467-9280.2007.02028.x [DOI] [PubMed] [Google Scholar]
  108. Markman EM, & Wachtel GF (1988). Children’s use of mutual exclusivity to constrain the meanings of words. Cognitive psychology, 20(2), 121–157. [DOI] [PubMed] [Google Scholar]
  109. Markus HR, & Kitayama S (1991). Culture and the self: Implications for cognition, emotion, and motivation. Psychological Review, 98(2), 224. [Google Scholar]
  110. Marlowe FW (2010). The Hadza: hunter-gatherers of Tanzania (Vol. 3): Univ of California Press. [Google Scholar]
  111. Massaro DW, & Cohen MM (1983). Phonological context in speech perception. Perception & psychophysics, 34(4), 338–348. [DOI] [PubMed] [Google Scholar]
  112. Menon U, & Shweder RA (1994). Kali’s tongue: Cultural psychology and the power of shame in Orissa, India. Emotion and culture: Empirical studies of mutual influence, 241–284. [Google Scholar]
  113. Mobbs D, Weiskopf N, Lau HC, Featherstone E, Dolan RJ, & Frith CD (2006). The Kuleshov Effect: the influence of contextual framing on emotional attributions. Social Cognitive and Affective Neuroscience, 1(2), 95–106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  114. Mohanty A, & Sussman TJ (2013). Top-down modulation of attention by emotion. Frontiers in Human Neuroscience, 7, 102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  115. Morey RD, & Rouder JN (2015). BayesFactor: Computation of Bayes Factors for Common Designs (Version R package version 0.9.12–2).
  116. Nelder JA, & Wedderburn RWM (1972). Generalized linear models. Journal of the Royal Statistical Society: Series A (General), 135(3), 370–384. [Google Scholar]
  117. Nelson NL, & Russell JA (2016a). Building emotion categories: Children use a process of elimination when they encounter novel expressions. Journal of Experimental Child Psychology, 151, 120–130. doi: 10.1016/j.jecp.2016.02.012 [DOI] [PubMed] [Google Scholar]
  118. Nelson NL, & Russell JA (2016b). A facial expression of pax: Assessing children’s “recognition” of emotion from faces. Journal of Experimental Child Psychology, 141, 49–64. doi: 10.1016/j.jecp.2015.07.016 [DOI] [PubMed] [Google Scholar]
  119. Norenzayan A, & Heine SJ (2005). Psychological universals: What are they and how can we know? Psychological bulletin, 131(5), 763–784. [DOI] [PubMed] [Google Scholar]
  120. Pashler H, Johnston JC, & Ruthruff E (2001). Attention and performance. Annual Review of Psychology, 52(1), 629–651. [DOI] [PubMed] [Google Scholar]
  121. Pinker S (1997). How the mind works. New York: Norton. [DOI] [PubMed] [Google Scholar]
  122. Raichlen DA, Pontzer H, Harris JA, Mabulla AZ, Marlowe FW, Josh Snodgrass J, . . . Wood BM (2017). Physical activity patterns and biomarkers of cardiovascular disease risk in hunter-gatherers. American Journal of Human Biology, 29(2). [DOI] [PubMed] [Google Scholar]
  123. Raudenbush S, Bryk A, Cheong Y, Congdon R, & du Toit M (2004). HLM 6 Hierarchical Linear and Nonlinear Modeling. Lincolnwood, llinois: Scientific Software International: Inc. [Google Scholar]
  124. Raudenbush S, Bryk A, & Congdon R (2011). HLM 7.00 for Windows [Computer software]. Lincolnwood, IL: Scientific Software International, Inc. [Google Scholar]
  125. Richerson PJ, & Boyd R (2005). Not by genes alone: How culture transformed human evolution. Chicago, IL: University of Chicago Press. [Google Scholar]
  126. Rosaldo MZ (1980). Knowledge and Passion: Cambridge University Press. [Google Scholar]
  127. Rouder JN (2014). Optional stopping: No problem for Bayesians. Psychonomic bulletin & review, 21(2), 301–308. [DOI] [PubMed] [Google Scholar]
  128. Rouder JN, Morey RD, Speckman PL, & Province JM (2012). Default Bayes factors for ANOVA designs. Journal of Mathematical Psychology, 56(5), 356–374. [Google Scholar]
  129. Rouder JN, Speckman PL, Sun D, Morey RD, & Iverson G (2009). Bayesian t tests for accepting and rejecting the null hypothesis. Psychonomic bulletin & review, 16(2), 225–237. [DOI] [PubMed] [Google Scholar]
  130. Ruiz-Belda M-A, Fernández-Dols J-M, Carrera P, & Barchard K (2003). Spontaneous facial expressions of happy bowlers and soccer fans. Cognition & Emotion, 17(2), 315–326. [DOI] [PubMed] [Google Scholar]
  131. Russell JA (1980). A circumplex model of affect. Journal of personality and social psychology, 39, 1161–1178. [DOI] [PubMed] [Google Scholar]
  132. Russell JA (1991). Culture and the Categorization of Emotions. Psychological bulletin, 110(3), 426–450. [DOI] [PubMed] [Google Scholar]
  133. Russell JA (1994). Is there universal recognition of emotion from facial expressions? A review of the cross-cultural studies. Psychological Bulletin, 115(1), 102–141. [DOI] [PubMed] [Google Scholar]
  134. Russell JA (1995). Facial expressions of emotion: what lies beyond minimal universality? Psychological Bulletin, 118, 379–391. [DOI] [PubMed] [Google Scholar]
  135. Russell JA, Bachorowski JA, & Fernandez-Dols JM (2003). Facial and vocal expressions of emotion. Annual Review of Psychology, 54, 329–349. doi: 10.1146/annurev.psych.54.101601.145102 [DOI] [PubMed] [Google Scholar]
  136. Russell JA, & Barrett LF (1999). Core affect, prototypical emotional episodes, and other things called emotion: dissecting the elephant. Journal of personality and social psychology, 76(5), 805–819. [DOI] [PubMed] [Google Scholar]
  137. Sands B (1998). The linguistic relationship between Hadza and Khoisan In Schladt M (Ed.), Language, Identity, and Conceptualization among the Khoisan (pp. 265–283): Rudiger Kupper Verlag, Köln, Germany. [Google Scholar]
  138. Sauter DA (2017). The Nonverbal Communication of Positive Emotions: An Emotion Family Approach. Emotion Review, 9(3), 222–234. doi: 10.1177/1754073916667236 [DOI] [PMC free article] [PubMed] [Google Scholar]
  139. Sauter DA, Eisner F, Calder AJ, & Scott SK (2010). Perceptual cues in nonverbal vocal expressions of emotion. Quarterly Journal of Experimental Psychology, 63(11), 2251–2272. [DOI] [PMC free article] [PubMed] [Google Scholar]
  140. Sauter DA, Eisner F, Ekman P, & Scott SK (2010). Cross-cultural recognition of basic emotions through nonverbal emotional vocalizations. Proceedings of the National Academy of Sciences, 107(6), 2408–2412. [DOI] [PMC free article] [PubMed] [Google Scholar]
  141. Sauter DA, Eisner F, Ekman P, & Scott SK (2015). Emotional vocalizations are recognized across cultures regardless of the valence of distractors. Psychological science, 26(3), 354–356. [DOI] [PMC free article] [PubMed] [Google Scholar]
  142. Scherer KR, Johnstone T, & Klasmeyer G (2003). Vocal expression of emotion In Davidson RJ, Scherer KR, & Goldsmith HH (Eds.), Handbook of affective sciences (pp. 433–456): Oxford University Press. [Google Scholar]
  143. Schröder M (2003). Experimental study of affect bursts. Speech communication, 40(1–2), 99–116. [Google Scholar]
  144. Shao B, Doucet L, & Caruso DR (2015). Universality versus cultural specificity of three emotion domains: Some evidence based on the cascading model of emotional intelligence. Journal of cross-cultural psychology, 46(2), 229–251. [Google Scholar]
  145. Shariff AF, & Tracy JL (2011). What are emotion expressions for? Current Directions in Psychological Science, 20(6), 395–399. [Google Scholar]
  146. Shaver P, Schwartz J, Kirson D, & O’Connor C (1987). Emotion knowledge: Further exploration of a prototype approach. Journal of personality and social psychology, 52(6), 1061–1086. [DOI] [PubMed] [Google Scholar]
  147. Sorenson ER (1975). Culture and the expression of emotion In Williams TR (Ed.), Psychological anthropology (pp. 361–372). Chicago, IL: Aldine. [Google Scholar]
  148. Tamir M, Schwartz SH, Cieciuch J, Riediger M, Torres C, Scollon C, . . . Vishkin A (2016). Desired emotions across cultures: A value-based account. Journal of personality and social psychology, 111(1), 67–82. doi: 10.1037/pspp0000072 [DOI] [PubMed] [Google Scholar]
  149. Tooby J, & Cosmides L (2008). The Evolutionary Psychology of Emotions and Their Relationship to Internal Regulatory Variables In Lewis M, Haviland-Jones JM, & Barrett LF (Eds.), Handbook of emotions (3 ed., pp. 114–137): Guildford Press. [Google Scholar]
  150. Tracy JL, & Matsumoto D (2008). The spontaneous expression of pride and shame: Evidence for biologically innate nonverbal displays. Proceedings of the National Academy of Sciences, 105(33), 11655–11660. [DOI] [PMC free article] [PubMed] [Google Scholar]
  151. Tracy JL, & Randles D (2011). Four models of basic emotions: A review of Ekman and Cordaro, Izard, Levenson, and Panksepp and Watt. Emotion Review, 3(4), 397–405. [Google Scholar]
  152. Tracy JL, & Robins RW (2008). The nonverbal expression of pride: Evidence for cross-cultural recognition. Journal of personality and social psychology, 94(3), 516–530. [DOI] [PubMed] [Google Scholar]
  153. Van den Stock J, Righart R, & de Gelder B (2007). Body expressions influence recognition of emotions in the face and voice. Emotion, 7(3), 487–494. doi: 10.1037/1528-3542.7.3.487 [DOI] [PubMed] [Google Scholar]
  154. Vogt J, De Houwer J, Moors A, Van Damme S, & Crombez G (2010). The automatic orienting of attention to goal-relevant stimuli. Acta Psychologica, 134(1), 61–69. [DOI] [PubMed] [Google Scholar]
  155. Wagenmakers E-J, Beek TF, Rotteveel M, Gierholz A, Matzke D, Steingroever H, . . . Sasiadek A (2015). Turning the hands of time again: a purely confirmatory replication study and a Bayesian analysis. Frontiers in Psychology, 6, 494. [DOI] [PMC free article] [PubMed] [Google Scholar]
  156. Wallbott HG (1988). In and out of context: Influences of facial expression and context information on emotion attributions. British Journal of Social Psychology, 27(4), 357–369. [Google Scholar]
  157. Warriner AB, Kuperman V, & Brysbaert M (2013). Norms of valence, arousal, and dominance for 13,915 English lemmas. Behavior Research Methods, 45(4), 1191–1207. [DOI] [PubMed] [Google Scholar]
  158. Waxman SR, & Booth AE (2001). Seeing pink elephants: Fourteen-month-olds’ interpretations of novel nouns and adjectives. Cognitive psychology, 43(3), 217–242. [DOI] [PubMed] [Google Scholar]
  159. Wetzels R, Matzke D, Lee MD, Rouder JN, Iverson GJ, & Wagenmakers E-J (2011). Statistical evidence in experimental psychology: An empirical comparison using 855 t tests. Perspectives on Psychological Science, 6(3), 291–298. [DOI] [PubMed] [Google Scholar]
  160. Wetzels R, Raaijmakers JG, Jakab E, & Wagenmakers E-J (2009). How to quantify support for and against the null hypothesis: A flexible WinBUGS implementation of a default Bayesian t test. Psychonomic bulletin & review, 16(4), 752–760. [DOI] [PubMed] [Google Scholar]
  161. Widen S (2016). The Development of Children’s Concepts of Emotion In Barrett LF, Lewis M, & Haviland-Jones JM (Eds.), Handbook of emotions (4 ed., pp. 307–318): Guildford Publications. [Google Scholar]
  162. Widen S, & Russell JA (2013). Children’s recognition of disgust in others. Psychological bulletin, 139(2), 271–299. [DOI] [PubMed] [Google Scholar]
  163. Wieser MJ, & Brosch T (2012). Faces in context: a review and systematization of contextual influences on affective face processing. Frontiers in Psychology, 3, 471. [DOI] [PMC free article] [PubMed] [Google Scholar]
  164. Wilson-Mendenhall CD, Barrett LF, & Barsalou LW (2015). Variety in emotional life: within-category typicality of emotional experiences is associated with neural activity in large-scale brain networks. Social Cognitive and Affective Neuroscience, 10(1), 62–71. doi: 10.1093/scan/nsu037 [DOI] [PMC free article] [PubMed] [Google Scholar]
  165. Wilson-Mendenhall CD, Barrett LF, & Barsalou LW (2015). Variety in emotional life: within-category typicality of emotional experiences is associated with neural activity in large-scale brain networks. Soc Cogn Affect Neurosci, 10(1), 62–71. doi: 10.1093/scan/nsu037 [DOI] [PMC free article] [PubMed] [Google Scholar]
  166. Wu LL, & Barsalou LW (2009). Perceptual simulation in conceptual combination: Evidence from property generation. Acta Psychologica, 132(2), 173–189. doi:S0001–6918(09)00018–3 [pii], 10.1016/j.actpsy.2009.02.002 [DOI] [PubMed] [Google Scholar]
  167. Xu F, Cote M, & Baker A (2005). Labeling guides object individuation in 12-month-old infants. Psychological science, 16(5), 372–377. [DOI] [PubMed] [Google Scholar]
  168. Yin J, & Csibra G (2015). Concept-based word learning in human infants. Psychological science, 26(8), 1316–1324. [DOI] [PMC free article] [PubMed] [Google Scholar]
  169. Young FW, & Lewyckyj R (1979). ALSCAL User’s Guide: A Guide for Users of ALSCAL: a Nonmetric Multidimensional Scaling and Unfolding Program with Several Individual Differences Options.

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

S1

RESOURCES