Skip to main content
Affective Science logoLink to Affective Science
. 2023 Jan 25;4(2):307–316. doi: 10.1007/s42761-022-00170-1

Labeling and Describing Discrete Emotions in Early Childhood: A Relational Approach

Jennifer M Knothe 1, Eric A Walle 1,
PMCID: PMC10247584  PMID: 37304558

Abstract

Emotion understanding involves appreciating the significance of the relational context; the “aboutness” of the emotion. This study examined how children labeled emotions and described relational elements of discrete emotion contexts. Preschool children (3.5-year-olds, n = 22; 4.5-year-olds, n = 23) described images of 5 emotion contexts (anger, sadness, disgust, fear, and joy). Researchers assessed children’s (1) correct labeling of discrete emotions, and (2) differential mentioning of the emoter (person displaying the emotion) and the referent (the elicitor of the emotion) across discrete emotions. Children’s pattern of accurately labeling discrete emotions was similar to prior research, with both age groups correctly labeled anger, sadness, and joy more often than disgust or fear. Novel to the present study, we found that older children differentially highlighted emotional elements (i.e., the emoter, the referent) when describing discrete emotion contexts. Specifically, 4.5-year-olds emphasized the emoter more when describing anger, sadness, and joy than fear and disgust contexts, and mentioned the referent more in disgust, fear, and joy than anger and sadness contexts. Differential emphasis of relational elements was not observed for 3.5-year-olds. These findings highlight the importance of examining children’s appreciation of relational contexts and indicate important differences in how children differentially emphasize relational elements when viewing discrete emotion contexts. Potential developmental mechanisms, opportunities for further empirical research, and implications for emotion theory are discussed.

Supplementary Information

The online version contains supplementary material available at 10.1007/s42761-022-00170-1.

Keywords: Development, Emotion understanding, Emotional contexts, Facial expressions


A crucial aspect of emotion understanding is appreciating the relational aboutness of the emotional context. As such, individuals must recognize and understand the elements of the emotional context, such as the emotional communication, eliciting object or situation, and likely responses given relation between the two. For example, seeing a child crying while staring at a fallen ice-cream, one would infer that the child is sad about losing their treat. While substantial research has investigated children’s labeling of emotion faces, much less has considered how children appreciate the relational contexts within which emotions occur. This study addressed this gap by examining how children label and talk about distinct elements of emotional contexts.

Emotion Understanding in Young Children

Canonical approaches to studying emotion understanding assess when children label discrete emotions — typically facial expressions. From a relational view of emotion, however, such research falls short of ascertaining children’s appreciation of discrete emotions. Specifically, it is insufficient to label an emotion; rather, an observer must understand the relational “aboutness” of the emotion (Deonna & Teroni, 2012; Gordon, 1974). Examining how children utilize relational contexts of emotion can shed unique light on their emotion understanding.

The Role of Context for Labeling Emotions

Emotion differentiation and labeling has been studied extensively in developmental psychology (e.g., Camras & Allison, 1985; Widen & Russell, 2003, 2008, 2010a, 2010b). Children’s emotion labeling emerges systematically over the first 5 years (see Widen & Russell, 2008), initially including “happy” more than other labels and then gradually adding anger and sadness, followed by fear and surprise, and lastly disgust (e.g., Widen, 2013). This sequential unfolding of emotion labels has been observed across a variety of face labeling and sorting tasks (e.g., Widen & Russell, 2003, 2008, 2010a, 2010b).

Studying emotion labeling within relational contexts provides further insight of early emotion understanding. Indeed, preschoolers’ emotion labeling is influenced by the context, such as hearing a corresponding emotion story (e.g., Camras & Allision, 1985; Denham et al., 1994; Dunn & Hughes, 1998), and their accuracy improves when particular emotions, specifically fear, compassion, embarrassment, disgust, and shame, are described in stories rather than presented solely as facial expressions (Wang et al., 2014; Widen & Russell, 2010b). For example, Leitzke and Pollak (2016) found that 4-year-olds more accurately identified images of disgust facial expressions presented in context (i.e., body holding a dirty object) than the face alone, but this pattern did not hold for anger. Thus, while presenting children with relational contexts may improve emotion labeling, a systematic study of labeling emotional contexts is lacking in the literature.

Attention to Relational Elements in Emotion Contexts

Beyond labeling emotional states, understanding others’ emotions involves appreciating the relational significance between the individual and their environment (Walle, Reschke, Camras et al., 2017; Walle, Reschke, & Knothe, 2017). For example, labeling someone’s facial expression as fear falls short of appreciating that the person’s fear is in relation to an approaching tarantula. Indeed, preschoolers articulate an understanding of emotion-eliciting situations when describing others’ emotions (e.g., Strayer, 1986). From a relational perspective, differential attentional deployment to distinct aspects of the context depending on the communicated emotion is crucial for adaptive functioning (e.g., avoiding a fear-inducing stimulus rather than the fearful person).

A study by Knothe and Walle (2018) examined parents’ emphasis of the emoter and referent when describing emotional contexts to their child. Parents talked about the emoter more for anger and sadness images than those depicting disgust and fear, and emphasized the referent more for disgust, fear, and joy images than anger and sadness images. College students similarly highlighted the emoter and referent when describing discrete emotion contexts (Knothe & Walle, 2019). Behavioral research suggests that children may also differentially prioritize the emoter and the referent across discrete emotion contexts. Heightened focus toward the emoter is observed in contexts of anger and sadness, demonstrated by infants avoiding angry individuals (Camras, 1977; Walle, Reschke, Camras et al., 2017; Walle, Reschke, & Knothe, 2017) and directing help toward a sad experimenter (Svetlova et al., 2010). Conversely, children more readily identify threatening (e.g., snakes) over non-threatening stimuli (e.g., flowers; LoBue & DeLoache, 2008), and increase attention toward (Leitzke & Pollak, 2016) and physical avoidance of (Stevenson et al., 2010) disgust stimuli. Interestingly, joy contexts highlight both the emoter and the referent, with infants increasing proximity to joy-eliciting objects (e.g., Hornik et al., 1987) and preferring individuals expressing positive affect (e.g., Farroni et al., 2007). However, whether children’s verbal descriptions of emotional contexts mirror descriptions by parents and adults remains uninvestigated.

Current Study

We examined how children labeled and highlighted relational elements in discrete emotion contexts. We focused our research 3.5- and 4.5-year-old children given prior research indicating that children accurately identify the emotions of interest between 3- and 5-years-of-age (Widen & Russell, 2008) and may demonstrate differential attention to relational elements at 4-years of age (Leitzke & Pollak, 2016).

First, we observed child spontaneous labeling of discrete emotions when presented with images with a target character expressing the emotion toward a clear referent. Given prior research (e.g., Widen & Russel, 2003), we predicted that children of both age groups would be less likely to correctly label disgust than anger, sadness, fear, and joy. Second, we examined children’s mentioning of the emoter and the referent across discrete emotions. Importantly, it was not our intent to pit the relational elements against one another — after all, the emoter is relating with the referent, not competing with it. Thus, we compared mentioning of the emoter and mentioning of the referent to determine whether children differentially highlighted each element, respectively, across discrete emotion contexts. Based on prior research with parents and college students (Knothe & Walle, 2018, 2019), we predicted that children would mention the emoter more frequently when describing anger and sadness images than disgust and fear images. Conversely, we predicted that children would mention the referent more frequently when describing disgust and fear images than anger and sadness images.

Method

Participants

Forty-six child-parent dyads (40 female caregivers) completed the study across two age groups: 3.5-year-olds (n = 22, 13 female; Mage = 3.53 years, range = 3.23–3.7 years, SD = 0.23) and 4.5-year-olds (n = 24, 11 female; Mage = 4.53 years, range = 4.33–4.87 years, SD = 0.14). Eight additional dyads participated but were excluded because of child inattention (n = 3), a non-English language was spoken (n = 3), or experimenter error (n = 2). Prior research using similar methods and analyses have reported medium to large effect sizes. Thus, we anticipated medium effect sizes (d = 0.50) as a conservative estimate due to the difference in populations (i.e., adult vs. child talk). A power analysis using a medium effect size determined that a sample of at least 21 children for each age group would provide power of 0.80.

Participating families were recruited from the California Central Valley through Department of Public Health birth records and community recruitment events. Most families had an income between $25,000 and $40,000 (range < $25,000 to > $150,000). Child racial demographics reflected those of local area, with parents identifying children as 37% White, 35% mixed-race, 26% Hispanic, and 2% not answering.

All study procedures were in accordance with the Declaration of Helsinki and approved by the Institutional Review Board at the University of California, Merced. The data that support the findings of this study are available from the corresponding author upon reasonable request.

Materials

Stimuli

A picture book was comprised of sixteen 8″ × 10″ color photographs. The first image provided an example image to familiarize the parent with the task. This was followed by 5 images of different children’s faces (gender-matched to the participant) expressing one of five discrete emotions (i.e., anger, sadness, disgust, fear, joy) to warmup the child to the subsequent emotion task of interest. The 5 face images were presented in random order.

The emotion task of interest consisted of 10 emotion context images that depicted an emotional scene featuring a single emoter (i.e., a male or female child for each of the 5 discrete emotions) displaying an emotion and a clear referent related to the emotion. The 10 emotion context images (5 emotions: anger, sadness, disgust, fear, joy; 2 genders) were randomly ordered with the exception that the same emotion context was never presented in succession. A unique child model and contextual scene was featured in each image.

Stimuli Selection and Validation

All context images contained a single emoter expressing (i.e., facially and posturally) one emotion (anger, sadness, disgust, fear, or joy) toward a single referent (an object; e.g., a phone, broccoli, dropped ice cream, puppy). All images included a child expressing the emotion to ensure consistency with images used in prior research (e.g., Knothe & Walle, 2018). The emotion expressions were of moderate intensity, and all images were previously validated as conveying the intended emotion (> 80% agreement for the target emotion; see Knothe & Walle, 2018). All emotional contexts were selected to be familiar events for children (e.g., receiving a gift; disliking a food) and thematically similar to previous vignette studies (e.g., Widen & Russell, 2010b). Example images are presented in Fig. 1, and further detailed descriptions of the images can be found in Knothe and Walle (2018, Fig. 1 and Appendix Tables A1 and A2).

Fig. 1.

Fig. 1

Sample images from the picture book activity (from upper left: Anger, Disgust, Joy, Fear). All images were presented in random order, with exception that the same emotion was not repeated sequentially

Additionally, and unique to the present study, a sample of 28 college-aged participants (21 female) further validated the context images to ensure that the referent of the emotion was clearly identifiable. Adults correctly identified a clear referent in all emotion context images, with an overall agreement of 83% (Anger = 78%, Sadness = 77%, Disgust = 87%, Fear = 85%, Joy = 90%).

Procedure

Each dyad completed the activity in a single campus visit. Upon arrival, a trained researcher provided an overview of the procedures to the parent and answered any questions. Parents then completed consent documents and a demographic questionnaire while the child engaged in a short warm-up period by playing with toys in the room.

Picture Book Activity

The parent and child were seated next to each other in separate chairs. Parents were asked to guide their child through the picture book and provided with the following instructions:

You will be guiding your child through this picture book. There are questions on the opposite page to each image. Please follow these questions but you may use some, or all, in whichever order you choose. Do not ask any leading questions, but asking general follow up phrases such as, ‘tell me more’ or ‘why?’ are fine. However little or much your child wants to say about each image is perfectly fine. Go at your own pace. Please do not point to any of the images but it is okay if your child points. We want their natural, unbiased response to the images.

Warm-up Images

To begin, a non-emotional image (i.e., a kitten batting a flower) was provided with the instructions and used as an example image. After answering any parent questions, the experimenter and any siblings or additional adults left the room for the duration of the activity. To familiarize the child with primary emotions of interest, next the book featured 5 gender-matched images of children’s faces displaying a stereotypical facial configuration associated with the 5 discrete emotions of interest. Each face image was accompanied by a page on the opposite side of the book with the following questions: What do you see? What is she/he feeling? Can you tell me more? The face images were selected from the CAFES image set (LoBue & Thrasher, 2015) and featured a child (same-sex as the participant) displaying a clear emotion. Due to differences in the procedures and composition of the warm-up images and the images of primary interest to the study (i.e., the relational context images), no analyses of the warm-up images were conducted.

Relational Context Images

Each relational context image was accompanied by a page with the following questions: What is going on? What do you see? What is happening? Can you tell me more?

These questions were provided as non-leading, neutral ways for the parent to guide the child and elicit verbal responses. The decision to use a parent rather than an experimenter for the picture book paradigm resulted from extensive piloting indicating that children were more comfortable and more verbal when interacting with their parent than an experimenter. Importantly, all coding and analyses identified and omitted any parent leading questions or off-task discussions (see coding sections below) so as to provide greater standardization across dyads and minimize parent-bias in children’s verbal behavior. The picture book activity lasted approximately 8:21 min (SD = 2:31 min). Two video camcorders on tripods recorded all parent and child behaviors.

Coding

Trained researchers naïve to the study hypotheses transcribed all verbalizations by the parent and child during the picture book activity. Next, the researchers coded each transcript for the variables of interest. A second researcher coded these variables for 25% of the transcripts. Coding disagreements were discussed, but the initial coding of the primary coder was maintained. Interrater reliability is reported below as a Pearson’s correlation coefficient with corresponding mean difference statistic for count variables, and as the percent agreement and corresponding kappa value for binary variables.

Child Talk

The total number of child words pertaining to each image were counted in each transcript to create a measure of on task child words (reliability: r = 0.97, Mdifference = 2.90 words) for each page (i.e., trial). This variable was used to control for child verbosity in the analyses (see Analytic Strategy section below).

Talk relating to off-task topics (e.g., the child asking the parent what they would do after the activity) or responses to parent questions unrelated to the task (e.g., can you sit still?) were excluded from the total amount of child words. Importantly, parent questions or statements deviating from the provided list of parent questions were identified (reliability: 95% agreement, k= 0.89) and all child talk following such instances marked as off-task and not coded for the remainder of that trial. This decision was made due to observed changes of child talk following parent leading questions and statements (e.g., What is she feeling? Why is he scared?). Parent leading utterances did not differ significantly across age groups, ethnicity, or income.

Researchers coded the transcript of each emotion trial for child talk featuring the following:

Correct Emotion Label

Mentioning the target emotion or a related emotion term was considered an emotion label (Reliability: 94% agreement, k = 0.88). Emotion labels were coded dichotomously as either correct (1) or incorrect/absent (0). Related emotion words indicating the target emotion for anger (e.g., mad, frustrated), sadness (e.g., depressed, down, blue), disgust (e.g., gross, yucky, icky), fear (e.g., afraid, scared, frightened), and joy (e.g., happy, joyful) were coded as labeling the emotion.

Emoter

The emoter in each image was classified as the individual displaying the emotion. Words indicating the emoter included but were not limited to: he, she, him, her, boy, and girl. Researchers coded the frequency of child mentioning the emoter on each page (Reliability: r = 0.91, Mdifference = 0.79).

Referent

The object or situation toward which the emotion was directed was considered the referent of the emotional display. Words indicating the referent included but were not limited to: green juice, broccoli, dog, puppy, ice cream, spider, and homework. Researchers coded the frequency of child mentioning the referent on each page (Reliability: r = 0.76, Mdifference = 0.21).1

Data Availability

The datasets generated during and/or analyzed during the current study are not publicly available due to privacy concerns relating to the participants being minors but are available from the corresponding author on reasonable request.

Results

General Analytic Strategy

Of central interest was whether each age group differentially labeled and described the images as a function of the discrete emotion. Given prior research examining child labeling as a function age and as a function of discrete emotion, planned analyses of emotion labeling examined differences in correct labeling across emotions, between age groups, and within age groups (i.e., the Emotion x Age interaction term). Analyses of child mentioning the emoter and mentioning the referent were primarily interested in whether children in each group differentially mentioned each relational element across discrete emotions, not comparisons between age groups. Specifically, we were curious if each age group mentioned the emoter (or referent) more for one emotion versus another emotion, not whether a particular age group mentioned the emoter (or referent) more than another age group. Thus, while the Emotion x Age interaction term was included in these models, only planned comparisons within age group across discrete emotions were conducted.

Analyses were conducted using Generalized Mixed Linear Modeling using the Statistical Package for Social Sciences (SPSS) Version 23. All models controlled for child gender, trial number, on-task child words, and the gender of the child in the image. Child gender was included in the analyses given prior research with this age range finding gender differences in parent–child emotion talk (e.g., Adams et al., 1995; Cervantes & Callanan, 1998), and on-task child words was included in all models to control for child verbosity. We also controlled for the gender of the child in the image and the sizes of the emoter and the referent (for emoter and referent analyses, respectively) to ensure that any relevant effects could be attributed to the discrete emotion rather than compositional features of the images. Pairwise comparisons incorporated a Benjamini–Hochberg correction (Benjamini & Hochberg, 1995).

Emotion Labeling

Children’s spontaneous correct emotion labeling was analyzed with a binomial distribution, Satterthwaite approximation, robust estimation, compound symmetry covariance matrix, and identity link function. The estimated means and standard errors of child correct emotion labeling are presented in Table 1. The raw proportions of all emotion label responses for each emotion image are provided in the full confusion matrix in Supplemental Table 1 and 2.

Table 1.

Estimated means and standard errors of correct emotion labeling

Anger Sadness Disgust Fear Joy
Variable M (SE) M (SE) M (SE) M (SE) M (SE)
Overall

.67 D F

(.06)

.68 D F

(.05)

.21 A S J

(.05)

.29 A S J

(.06)

.57 D F

(.07)

3.5-year-olds

.56 4.5 D F

(.09)

.56 D F

(.07)

.13 4.5 A S J

(.06)

.21 A S

(.07)

.41 4.5 D

(.09)

4.5-year-olds

.76 3.5 D F

(.08)

.78 D F

(.06)

.32 3.5 A S J

(.07)

.40 A S J

(.08)

.72 3.5 D F

(.08)

Estimated means with standard errors in parentheses. Emotion label values represent the proportion of children providing a label for that emotion. Letters next to each mean (S sadness, F fear, A anger, D disgust, J joy) designate significant pairwise comparisons between emotions for each variable. For example, a significantly greater proportion of children labeled Anger (0.61) images than in Disgust (0.16) and Fear (0.07) images. Numbers next to each mean designate significant pairwise comparisons between age groups for each emotion

The analysis revealed significant main effects of picture emotion, F(4, 208) = 12.39, p < 0.001, η2 = 0.19, and child age group, F(1,48) = 12.85, p < 0.001, η2 = 0.21. However, the Picture Emotion x Age interaction was not significant, F(4, 213) = 0.15, p = 0.96, η2 < 0.01.

Differences Across Discrete Emotions

Comparison of correct emotion labeling across discrete emotions revealed that overall, children were less likely to correctly label disgust than anger, t(134) = 5.29, p < 0.001, d = 0.91, sadness, t(231) = 6.63, p < 0.001, d = 0.87, and joy, t(173) = 4.52, p < 0.001, d = 0.69. Additionally, children were less likely to correctly label fear than anger, t(191) = 4.78, p < 0.001, d = 0.69, sadness, t(296) = 5.47, p < 0.001, d = 0.64, and joy, t(187) = 3.57, p < 0.001, d = 0.55.

Differences Across Age Group

Examination of the main effect of age group revealed that a greater proportion of 4.5-year-olds provided a correct emotion label than did 3.5-year-olds, t(48) = 3.76, p < 0.001, d = 1.09.

Differences Between Discrete Emotions and Age Groups

A priori hypotheses and prior research suggested the likelihood of age-related differences in children’s labeling of discrete emotions across age groups (e.g., Widen & Russell, 2003). Thus, further planned analyses compared age groups for each discrete emotion, as well as differences across discrete emotions within each age group.

Differences Between Age Groups for Discrete Emotions

Pairwise comparisons tested differences between age groups for each discrete emotion. The older age group was more likely than the younger age to correctly label sadness, t(317) = 2.37, p = 0.02, d = 0.27, and joy, t(103) = 2.56, p = 0.01, d = 0.50. However, the difference in older children correctly labeling disgust compared to younger children did not survive the Benjamini–Hochberg correction, t(210) = 2.07, p = 0.039, d = 0.29). Additionally, a trend indicated that 4.5-year-olds were more likely to correctly label fear than 3.5-year-olds, t(246) = 1.83, p = 0.068, d = 0.23).

Differences Within Age Groups for Discrete Emotions
  • 3.5-year-olds

    Analyses of the younger age group revealed that 3.5-year-olds were more likely to correctly label anger than disgust, t(139) = 3.80, p < 0.001, d = 0.64, or fear, t(177) = 3.09, p = 0.002, d = 0.46. The younger children were also more likely to correctly label sadness than disgust, t(298) = 4.80, p < 0.001, d = 0.56, or fear, t(246) = 3.59, p < 0.001, d = 0.46, though the difference with joy did not survive the Benjamini–Hochberg correction, t(419) = 1.94, p = 0.05, d = 0.19. Lastly, 3.5-year-olds were more likely to correctly label joy than disgust, t(265) = 3.05, p = 0.003, d = 0.37, though the difference with fear did not survive the correction, t(195) = 1.99, p = 0.048, d = 0.29.

  • 4.5-year-olds

    Analyses examining differences between discrete emotion within the older children revealed that anger were more likely to be correctly labeled than disgust, t(125) = 3.77, p < 0.001, d = 0.67, or fear, t(267) = 3.78, p < 0.001, d = 0.46. This age group was also more likely to correctly label sadness than disgust, t(219) = 4.81, p < 0.001, d = 0.65, or fear, t(428) = 4.28, p < 0.001, d = 0.41. Finally, 4.5-year-olds were more likely to correctly label joy than disgust, t(124) = 3.40, p = 0.001, d = 0.61, or fear, t(222) = 3.14, p = 0.002, d = 0.42.

Relational Context Descriptions

Descriptions of relational contexts were analyzed with a Poisson distribution with a Satterthwaite approximation, robust estimation, compound symmetry covariance matrix, and log link function. The estimated means and standard errors of child mentioning the emoter and referent are presented in Table 2.

Table 2.

Estimated means and standard errors of frequency of mentioning relational elements by age group

Anger Sadness Disgust Fear Joy
Variable M (SE) M (SE) M (SE) M (SE) M (SE)
3.5-year-olds
Emoter

2.10

(0.20)

1.96

(0.20)

1.74

(0.22)

1.58

(0.19)

1.80

(0.22)

Referent

1.10 F J

(0.15)

1.46

(0.18)

1.51

(0.17)

1.69 A

(0.16)

1.65 A

(0.25)

4.5-year-olds
Emoter

2.46

(0.28)

2.39 F

(0.22)

2.38 F

(0.23)

1.74 D S J

(0.23)

2.43 F

(0.15)

Referent

1.35 F J

(0.16)

1.03 D F J

(0.13)

1.61 S

(0.20)

2.03 A S

(0.20)

2.13 A S

(0.22)

Estimated means with standard errors in parentheses. Values represent the frequency with which children mentioned the emoter and the referent, respectively. Letters next to each mean (S sadness, F fear, A anger, D disgust, J joy) designate significant pairwise comparisons between emotions for each variable. For example, 4.5-year-olds mentioned the emoter significantly more in sadness images (2.39) than fear images (1.74)

Emoter

Analysis of children’s mentioning the emoter in relational contexts revealed a significant main effect of picture emotion, F(4, 179) = 2.72, p = 0.03, η2 = 0.06, but no significant effect of child age, F(1, 55) = 0.05, p = 0.83, η2 < 0.01, or child gender, F(1, 39) = 2.02, p = 0.16, η2 = 0.05. The Emotion x Age interaction term was also not significant, F(4, 186) = 0.74, p = 0.57, η2 = 0.02.

Of central interest was whether children within each age group differed in mentioning the emoter across discrete emotions. Thus, planned comparisons tested differences across emotions within each age group.

3.5-Year-Olds

Planned pairwise comparisons revealed no significant differences in 3.5-year-olds’ mentioning of the emoter across discrete emotion images (all ps > 0.08).

4.5-Year-Olds

Planned pairwise comparisons of picture emotion revealed that 4.5-year-old children mentioned the emoter significantly less often for fear context images than images depicting sadness, t(190) = 3.13, p = 0.002, d = 0.45, disgust, t(273) = 3.14, p = 0.002, d = 0.38, or joy, t(139) = 3.09, p = 0.002, d = 0.52, though the difference with anger did not survive the correction, t(80) = 2.17, p = 0.03, d = 0.49.

Referent

Analysis of children’s mentioning the referent in relational contexts revealed a significant main effect of picture emotion, F(4, 199) = 6.92, p < 0.001, η2 = 0.12, but no significant effect of child age, F(1, 38) = 0.55, p = 0.46, η2 = 0.01, or child gender, F(1, 39) = 0.01, p = 0.94, η2 < 0.01. The Emotion x Age interaction term also was not significant, F(4, 152) = 1.22, p = 0.30, η2 = 0.03.

Again, we were primarily interested in whether children within each age group differed in mentioning the referent across discrete emotions. Thus, planned comparisons tested differences across emotions within each age group.

3.5-Year-Olds

Planned pairwise comparisons of picture emotion revealed that 3.5-year-old children mentioned the referent significantly less often for anger context images than images depicting fear, t(271) = 3.80, p < 0.001, d = 0.46, or joy, t(327) = 3.54, p < 0.001, d = 0.39.

4.5-Year-Olds

Planned pairwise comparisons of picture emotion indicated that 4.5-year-old children mentioned the referent significantly less often for sadness context images than images depicting disgust, t(118) = 2.40, p = 0.02, d = 0.44, fear, t(105) = 3.54, p < 0.001, d = 0.69, or joy, t(327) = 3.75, p < 0.001, d = 0.41. Older children also mentioned the referent significantly less when describing anger images than images of fear, t(52) = 2.35, p = 0.02, d = 0.65, or joy, t(242) = 3.36, p < 0.001, d = 0.43.

Discussion

Appreciating the relational aboutness of emotions is crucial for understanding others’ emotions. We found that 3.5- and 4.5-year-old children varied in accurately labeling emotions and differentially emphasized relational elements across emotion contexts. This underscores the importance of considering the relational context when examining children’s emotion understanding.

Emotion Labeling of Relational Contexts

Children’s spontaneous emotion labeling was generally in line with prior research (e.g., Widen & Russell, 2003). Children in both age groups were more likely to correctly label anger, sadness, and joy more than disgust and fear. It is worth noting that our use of emotional context images was distinct from most prior research of children’s emotion labeling. Specifically, children’s accuracy in the present study was higher than observed in prior research. For example, 4.5-year-old children’s correct labeling of disgust images (0.32) was noticeably higher than previously reported, such as 0.03 to 0.10 by Widen and Russell (2003, Study 2), 0.11 by Widen and Russell (2008), and 0.09 by Widen and Russell (2010a, 2010b, Study 2). Contextual elements provided in the present images may have provided more cues about the relational significance, and thus facilitated more accurate labeling of disgust expressions. However, it is also possible that relational information could make attention more diffuse or generate confusion when appreciating relational contexts. For example, one child described a girl expressing disgust toward a piece of broccoli on her fork as her blowing on the broccoli to “cool it off” before eating it. Our findings underscore the importance of considering emotional development as the child’s ability to appreciate the relational significance between the individual and their environment.

Differential Highlighting of Relational Elements in Emotion Contexts

Novel to the present study, we found that children in both age groups differentially highlighted relational elements across discrete emotion contexts, though this differentiation was more prevalent in the older age group. Specifically, 4.5-year-old children mentioned the emoter more frequently when describing anger, sadness, and joy images than fear images. Conversely, these children mentioned the referent of the emotion more when describing fear and joy images than those depicting anger and sadness, and more when describing disgust images than sad images. Older children’s differential emphasis on the emoter across discrete emotions is similar to that observed with parents (Knothe & Walle, 2018) and college students (Knothe & Walle, 2019). However, differences in mentioning the referent only partially supported our predictions, with disgust failing to differentiate from anger. This may indicate that children’s attention toward relational elements particularly relevant for disgust contexts is still developing at this age.

Interestingly, 3.5-year-olds demonstrated some differential mentioning of the referent, but not the emoter, across discrete emotion images. Like the older age group, 3.5-year-olds mentioned the referent less when describing anger images than those depicting fear or joy. However, this age group did not demonstrate as many significant differences across discrete emotions and had similar mentioning of the emoter across discrete emotion contexts. This may indicate that differential attention to relational elements of discrete emotion contexts is continuing to develop in early childhood, as even the older children failed to demonstrate the nuanced patterns of attention shown by adults (Knothe & Walle, 2018, 2019). However, it is also worth noting that while these children did not differentially describe elements of discrete emotional contexts, infants and toddlers do demonstrate differentiated behavioral responses to discrete emotions (e.g., Hornik et al., 1987; Walle, Reschke, Camras et al., 2017; Walle, Reschke, & Knothe,  2017). Such behavioral research suggests that younger children can appreciate referential specificity in emotional contexts, but verbal task demands in the present study may have obscured these capacities.

Additionally, a possible limitation was that our images were not standardized. For example, some images featured object-referents, whereas others featured agentic-referents. However, we believe it is unlikely that differences in referent agency accounted for the findings for two reasons. First, specifying whether the image had an agentic referent was not significant (p = 0.68) in models examining mentioning the referent. Second, comparison of the joy images, 1 of which featured an agentic referent and 1 that did not, indicated no significant difference in highlighting the referent between the two joy images, t(430) = 0.43, p = 0.67, d = 0.09. Even so, future research including more standardized images that control or tease apart such effects is recommended.

Future Directions

This study extends prior research of children’s emotion labeling and is the first to demonstrate that children differentially describe elements of discrete emotional contexts. Further empirical work and theoretical considerations for a relational perspective of emotion are described below.

Developmental Mechanisms

Our findings were generally consistent with parents’ descriptions of emotional contexts to their infants (Knothe & Walle, 2018). Differential highlighting of specific aspects of emotional contexts may be socialized by the parent to the child. Indeed, prior research has demonstrated that parent talk about emotions predicts infants’ goal-directed behavior in emotional situations (e.g., Brownell et al., 2013). Additional underlying mechanisms for consideration include the child’s personal experience with emotion contexts (e.g., Pollak et al., 2009), observing others’ attention and responding to emotional contexts (e.g., Repacholi et al., 2008), and cross-cultural differences in perceiving contextual elements (e.g., Masuda et al., 2008). Understanding mechanisms, such culture, parent expressivity, and family dynamics (see Halberstadt & Lozada, 2011; Halberstadt et al., 2013), through which differences in appreciating emotions arise would bolster researchers’ understanding of emotion and emotional development.

Visual Attention and Memory

Investigating children’s visual attention to emotion contexts would lessen the verbal demand on young children (e.g., Scott & Roby, 2015), complement findings from verbal tasks (the present study), and extend prior work on children’s visual scanning patterns of emotion faces and faces with postures (e.g., Nelson & Mondloch, 2017). Additionally, future work could examine how attention influences behavioral responses toward relational elements following a delay. For example, infants may have greater memory for a disgusting referent and subsequently avoid that object, whereas an angry person may be more memorable than the object that elicited the anger (see Repacholi et al., 2016). Such research would help investigators understand how infants attend to and retain specific aspects of emotion contexts (e.g., time looking at the emoter vs. the referent), how parent emotion talk may underlie such patterns, at what age differential processing is present, and whether the ontogeny of such differences is rooted in socialization or biological/evolutionary mechanisms (see Öhman & Mineka, 2001).

A Relational View of Emotion

These findings emphasize the importance of relational elements when perceiving emotions and the importance of examining differences in infants’, children’s, and adults’ processing of emotional contexts (see Camras & Shutter, 2010). Recent research with adults highlights the importance of context when perceiving emotional expressions (e.g., Hassin et al., 2013). Research indicating the influence of facial expressions (Mumenthaler & Sander, 2012), postures (Aviezer et al., 2008), and scenes (Reschke & Walle, 2021) on emotion perception may result from differential attentional focus to emotional elements. We encourage future research that considers the relational nature of emotions, those features that make them functionally distinct from one another, and how such aspects of emotion understanding change throughout the lifespan.

Supplementary Information

Below is the link to the electronic supplementary material.

Additional Information

Funding

Not applicable.

Conflict of Interest

The authors declare no competing interests.

Data availability

Data and syntax for this article can be found at https://osf.io/a9u6y/.

Author contributions

Not applicable.

Ethics approval

Not applicable.

Consent to participate

Not applicable.

Consent for publication

Not applicable.

Footnotes

1

The somewhat lower reliability for the referent variable was likely due to the variety of words that could be used to denote the referent in an image. In particular, instances of “it” required inference by the coder to determine what “it” referred to.

Handling Editor: Ruth Feldman

References

  1. Adams S, Kuebli J, Boyle PA, Fivush R. Gender differences in parent– child conversations about past emotions: A longitudinal investigation. Sex Roles. 1995;33:309–323. doi: 10.1007/BF01954572. [DOI] [Google Scholar]
  2. Aviezer, H., Hassin, R. R., Ryan, J., Grady, C., Susskind, J., Anderson, A., & Bentin, S. (2008). Angry, disgusted, or afraid? Studies on the malleability of emotion perception. Psychological Science, 19, 724–732. 10.1111/j.1467-9280.2008.02148.x [DOI] [PubMed]
  3. Barrett KC, Campos JJ. Perspectives on emotional development II: A functionalist approach to emotions. In: Osofsky J, editor. Handbook of infant development. 2. Wiley; 1987. pp. 55–578. [Google Scholar]
  4. Benjamini, Y., & Hochberg, Y. (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society, Series B (Methodological)77(1), 289–300.
  5. Brownell CA, Svetlova M, Anderson R, Nichols SR, Drummond J. Socialization of early prosocial behavior: Parents’ talk about emotions is associated with sharing and helping in toddlers. Infancy. 2013;18:91–119. doi: 10.1111/j.1532-7078.2012.00125.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Camras LA. Facial expressions used by children in a conflict situation. Child Development. 1977;48:1431–1435. doi: 10.2307/1128503. [DOI] [PubMed] [Google Scholar]
  7. Camras LA, Allison K. Children’s understanding of emotional facial expressions and verbal labels. Journal of Nonverbal Behavior. 1985;9:84–94. doi: 10.1007/BF00987140. [DOI] [Google Scholar]
  8. Camras LA, Shutter JM. Emotional facial expressions in infancy. Emotion Review. 2010;2:120–129. doi: 10.1177/1754073909352529. [DOI] [Google Scholar]
  9. Cervantes CA, Callanan MA. Labels and explanations in mother-child emotion talk: Age and gender differences. Developmental Psychology. 1998;34(1):88–98. doi: 10.1037/0012-1649.34.1.88. [DOI] [PubMed] [Google Scholar]
  10. Denham SA, Cook M, Zoller D. ‘Baby looks very sad’: Implications of conversations about feelings between mother and preschooler. British Journal of Developmental Psychology. 1992;10(3):301–315. doi: 10.1111/j.2044-835X.1992.tb00579.x. [DOI] [Google Scholar]
  11. Denham SA, Zoller D, Couchoud EA. Socialization of preschoolers’ emotion understanding. Developmental Psychology. 1994;30:928–936. doi: 10.1037/0012-1649.30.6.928. [DOI] [Google Scholar]
  12. Deonna JA, Teroni F. Homing in on the emotions. In: Deonna JA, Teroni F, editors. The emotions: A philosophical introduction. Routledge; 2012. pp. 1–13. [Google Scholar]
  13. Dunn J, Hughes C. Young children's understanding of emotions within close relationships. Cognition & Emotion. 1998;12:171–190. doi: 10.1080/026999398379709. [DOI] [Google Scholar]
  14. Farroni T, Menon E, Rigato S, Johnson MH. The perception of facial expressions in newborns. European Journal of Developmental Psychology. 2007;4:2–13. doi: 10.1080/17405620601046832. [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Gordon RM. The aboutness of emotions. American Philosophical Quarterly. 1974;11:27–36. [Google Scholar]
  16. Halberstadt AG, Lozada FT. Emotion development in infancy through the lens of culture. Emotion Review. 2011;3:158–168. doi: 10.1177/1754073910387946. [DOI] [Google Scholar]
  17. Halberstadt, A. G., Parker,A. E., & Castro, V. L. (2013). Nonverbal communication: Developmental perspectives. In J. A. Hall & M. L. Knapp (Eds.), Nonverbal Communication (Vol. 2, pp. 93–127). de Gruyter Mouton.
  18. Hassin RR, Aviezer H, Bentin S. Inherently ambiguous: Facial expressions of emotions, in context. Emotion Review. 2013;5:60–65. doi: 10.1177/1754073912451331. [DOI] [Google Scholar]
  19. Hornik R, Risenhoover N, Gunnar M. The effects of maternal positive, neutral, and negative affective communications on infant responses to new toys. Child Development. 1987;58:937–944. doi: 10.2307/1130534. [DOI] [Google Scholar]
  20. Knothe JM, Walle EA. Parental communication about emotional contexts: Differences across discrete categories of emotion. Social Development. 2018;27:247–261. doi: 10.1111/sode.12276. [DOI] [Google Scholar]
  21. Knothe JM, Walle EA. Functions of emotional communication: The role of affective expressions in social interactions. In: Hess U, Hareli S, editors. What Emotions Can Tell Us About the World. Springer; 2019. pp. 83–101. [Google Scholar]
  22. Leitzke BT, Pollak SD. Developmental changes in the primacy of facial cues for emotion recognition. Developmental Psychology. 2016;52:572–581. doi: 10.1037/a0040067. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. LoBue V, DeLoache JS. Detecting the snake in the grass: Attention to fear-relevant stimuli by adults and young children. Psychological Science. 2008;19:284–289. doi: 10.1111/j.1467-9280.2008.02081.x. [DOI] [PubMed] [Google Scholar]
  24. LoBue V, Thrasher C. The Child Affective Facial Expression (CAFÉ) set: Validity and reliability from untrained adults. Frontiers in Psychology. 2015;5:1–8. doi: 10.3389/fpsyg.2014.01532. [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Masuda T, Ellsworth PC, Mesquita B, Leu J, Tanida S, Van de Verrdonk E. Placing the face in context. Journal of Personality and Social Psychology. 2008;94:365–381. doi: 10.1037/0022-3514.94.3.365. [DOI] [PubMed] [Google Scholar]
  26. Mumenthaler C, Sander D. Social appraisal influence recognition of emotions. Journal of Personality and Social Psychology. 2012;102(6):1118–1135. doi: 10.1037/a0026885. [DOI] [PubMed] [Google Scholar]
  27. Nelson N, Mondloch C. Adults’ and children’s perception of facial expressions is influenced by body postures even for dynamic stimuli. Visual Cognition. 2017;25:1–12. doi: 10.1080/13506285.2017.1301615. [DOI] [Google Scholar]
  28. Öhman A, Mineka S. Fears, phobias, and preparedness: Toward an evolved module of fear and fear learning. Psychological Review. 2001;108(3):483–522. doi: 10.1037//0033-295X.108.3.483. [DOI] [PubMed] [Google Scholar]
  29. Pollak SD, Messner M, Kistler DJ, Cohn JF. Development of perceptual expertise in emotion recognition. Cognition. 2009;110:242–247. doi: 10.1016/j.cognition.2008.10.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Repacholi BM, Meltzoff AN, Hennings TM, Ruba AL. Transfer of social learning across contexts: Exploring infants’ attribution of trait-like emotions to adults. Infancy. 2016;21(6):785–806. doi: 10.1111/infa.12136. [DOI] [Google Scholar]
  31. Repacholi BM, Meltzoff AN, Olsen B. Infants’ understanding of the link between visual perception and emotion: “If she can’t see me doing it, she won’t get angry”. Developmental Psychology. 2008;44:561–574. doi: 10.1037/0012-1649.44.2.561. [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Reschke PJ, Walle EA. The unique and interactive effects of faces, postures, and scenes on emotion categorization. Affective Sciences. 2021;2(4):468–483. doi: 10.1007/s42761-021-00061-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Scott RM, Roby E. Processing demands impact 3-year-olds’ performance in a spontaneous-response task: New evidence for the processing-load account of early false-belief understanding. PLoS ONE. 2015;10:1–20. doi: 10.1371/journal.pone.0142405. [DOI] [PMC free article] [PubMed] [Google Scholar]
  34. Stevenson RJ, Oaten MJ, Case TI, Repacholi BM, Wagland P. Children’s response to adult disgust elicitors: Development and acquisition. Developmental Psychology. 2010;46:165–177. doi: 10.1037/a0016692. [DOI] [PubMed] [Google Scholar]
  35. Strayer J. Children’s attributions regarding the situational determinants of emotion in self and others. Developmental Psychology. 1986;22:649–654. doi: 10.1037/0012-1649.22.5.649. [DOI] [Google Scholar]
  36. Svetlova M, Nichols SR, Brownell CA. Toddlers’ prosocial behavior: From instrumental to empathic to altruistic helping. Child Development. 2010;81:1814–1827. doi: 10.1111/j.1467-8624.2010.01512.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Walle, E. A., Reschke, P. J., Camras, L. A., & Campos, J. J. (2017). Infant differential behavioral responding to discrete emotions. Emotion,17, 1078–1091. 10.1037/emo0000307 [DOI] [PMC free article] [PubMed]
  38. Walle, E. A., Reschke, P. D., & Knothe, J. M. (2017). Social referencing: Defining and delineating a basic process of emotion. Emotion Review,9, 245–252.
  39. Wang Z, Lü W, Zhang H, Surina A. Free-labeling facial expressions and emotional situations in children aged 3–7 years: Developmental trajectory and a face inferiority effect. International Journal of Behavioral Development. 2014;38:487–498. doi: 10.1177/0165025414531462. [DOI] [Google Scholar]
  40. Widen SC. Children’s interpretation of facial expressions: The long path from valence based to specific discrete categories. Emotion Review. 2013;5:72–77. doi: 10.1177/1754073912451492. [DOI] [Google Scholar]
  41. Widen SC, Russell JA. A closer look at preschoolers’ freely produced labels for facial expressions. Developmental Psychology. 2003;39:114–128. doi: 10.1037/0012-1649.39.1.114. [DOI] [PubMed] [Google Scholar]
  42. Widen SC, Russell JA. Children acquire emotion categories gradually. Cognitive Development. 2008;23:291–312. doi: 10.1016/j.cogdev.2008.01.002. [DOI] [Google Scholar]
  43. Widen SC, Russell JA. Differentiation in preschooler’s categories for emotion. Emotion. 2010;10:651–661. doi: 10.1037/a0019005. [DOI] [PubMed] [Google Scholar]
  44. Widen SC, Russell JA. Children’s scripts for social emotions: Causes and consequences are more central than are facial expressions. British Journal of Developmental Psychology. 2010;28:565–581. doi: 10.1348/026151009X457550d. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Data Availability Statement

The datasets generated during and/or analyzed during the current study are not publicly available due to privacy concerns relating to the participants being minors but are available from the corresponding author on reasonable request.

Data and syntax for this article can be found at https://osf.io/a9u6y/.


Articles from Affective Science are provided here courtesy of Springer

RESOURCES