Skip to main content
Scientific Reports logoLink to Scientific Reports
. 2017 Nov 14;7:15525. doi: 10.1038/s41598-017-15091-4

Dogs and humans respond to emotionally competent stimuli by producing different facial actions

Cátia Caeiro 1,2,, Kun Guo 1, Daniel Mills 2
PMCID: PMC5686192  PMID: 29138393

Abstract

The commonality of facial expressions of emotion has been studied in different species since Darwin, with most of the research focusing on closely related primate species. However, it is unclear to what extent there exists common facial expression in species more phylogenetically distant, but sharing a need for common interspecific emotional understanding. Here we used the objective, anatomically-based tools, FACS and DogFACS (Facial Action Coding Systems), to quantify and compare human and domestic dog facial expressions in response to emotionally-competent stimuli associated with different categories of emotional arousal. We sought to answer two questions: Firstly, do dogs display specific discriminatory facial movements in response to different categories of emotional stimuli? Secondly, do dogs display similar facial movements to humans when reacting in emotionally comparable contexts? We found that dogs displayed distinctive facial actions depending on the category of stimuli. However, dogs produced different facial movements to humans in comparable states of emotional arousal. These results refute the commonality of emotional expression across mammals, since dogs do not display human-like facial expressions. Given the unique interspecific relationship between dogs and humans, two highly social but evolutionarily distant species sharing a common environment, these findings give new insight into the origin of emotion expression.

Introduction

The common origin of emotions has long been a subject of scientific interest1 with different emotional responses producing a diverse range of communicative elements, especially through the face. Facial expressions are also correlates of internal state in both humans2 and other animals36 and so may be used, in part, to infer emotion alongside other component processes, such as physiological activation and behavioural tendencies7. Many studies (e.g.8,9) use an holistic approach (i.e. categorizing the whole face as angry, happy, etc.) to classify the target facial expressions, which reflects the way the human brain processes faces7,10, but can be problematic when examining the underlying mechanism of emotion perception across species. For instance, there is a diverse range of smiling faces with different visual characteristics and different emotional meanings in humans11,12. As a classic example, the Duchenne smile (felt, true enjoyment) differs by one muscle contraction from the non-Duchenne smile (unfelt, usually produced in formal greetings). Moreover, during laughter and depending on the context, a blend of both Duchenne and non-Duchenne smiles is often observed13. Hence, simply classifying a facial expression as “happy” is too simplistic and less meaningful for cross-species comparison. Furthermore, the same ‘holistic’ facial morphological configuration could have different functional meanings (i.e. result in distinctly different behavioural consequences) depending on the species14,15. For example, the Play Face (PF) and the Full Play Face (FPF) are variants of the same facial expression, where the former presents an open mouth with lower teeth exposed, and the latter incorporates visible upper teeth. Both the PF and the FPF represent different degrees of playful expression in great apes (humans included)1618. Conversely, in crested macaques, mandrills and geladas, the FPF is not just a more intense version of the PF, but instead is derived from convergence between the PF and the silent-bared teeth display (SBT), a facial expression observed in affiliative settings such as grooming19. Additionally, the SBT indicates submission and appeasement in Barbary macaques14, signals affinity and benign intentions in humans, and, in chimpanzees, is present in a range of situations from response to aggression to affinity contexts15.

As an alternative to an holistic descriptive approach, the decomposition and objective description of distinct anatomical regions of facial features, such as occurs with the Facial Action Coding System (FACS20), has been the golden standard to study human facial expressions of emotion across individuals of different races and cultures for several decades21,22. Each of the discrete facial movements identified (Action Units, AUs) is the result of an independent facial muscle contraction that can produce several changes in appearance to the face, which in turn are used to identify which AUs are activated. Thus, FACS codes facial movements from a purely anatomical basis, avoiding circular reasoning or a priori assumptions of emotion meaning. Recently, FACS has been adapted to several non-human species2330, such as chimpanzees and orangutans, following the original methodology20 and has proven to be a successful tool for objectively investigating and comparing facial expressions of closely related species31,32. For example, chimpanzees and humans share an identical facial muscle plan (differing by only one muscle)20,33, but chimpanzees display both homologous (e.g. play face and human laugh) and species-specific expressions (e.g. pant-hoot)32,34.

While the human prototypical facial expressions of emotion are well established, little is known about the quantitative and empirical nature of the emotional facial displays of the domestic dog, an evolutionarily remote, but socially complex species which often shares the human social environment and frequently engages in interspecific communication with an emotional content (e.g.35). To date, functional facial expressions in dogs have been largely discussed holistically in relation to their approach-avoidance value, for example, the “threat gape” in fight-flight situations3, and the PF or the Relaxed Open Mouth (ROM) as a social communicative signal for play solicitation and within play bouts3,36,37. With the development of the FACS for the domestic dog24, it becomes possible to apply a bottom-up technique to investigate the composition and meaning of dogs’ facial expressions and, more importantly, to establish possible analogies with humans, with whom they socially interact.

Dogs and humans, like other mammals, have a homologous facial anatomy plan38,39 even though they belong to phylogenetically distant groups. Additionally, both species share a common mammalian neuroanatomy for the basic emotions such as fear and happiness4044, typically live in a common social and physical environment, are very facially expressive (e.g.2,3,20,24), and respond to the same or similar conspecific and heterospecific social cues (e.g.45). Consequently, the facial cues and expression of emotion in home-dwelling pet dogs provide a unique comparative model for the study of phylogenetic inertia (i.e. absence of expected change and/or adaptation to an optimal state given specific selection pressures in the current environment)4648 versus evolutionary divergence (i.e. a set of changes brought about by selection pressures from a common ancestor resulting in homologies)49,50 versus evolutionary convergence (i.e. a set of changes from selection pressures acting in independent lineages to create similarity in the resulting analogies)47,50.

Here, we investigated the mechanistic basis of facial expressions in humans and dogs, by objectively measuring their video recorded facial actions during immediate reactions to emotionally-competent stimuli. The FACS20 and the DogFACS24 were applied in a range of contexts associated with four categories of emotional responses: a) happiness, b) positive anticipation, c) fear, and d) frustration (Table 1). Instead of selecting the basic emotions that are known to produce universal facial signals in humans51, we focused on emotions that are defined by evolutionary and biologically consistent criteria: 1) essential for solving adaptive problems in mammals (e.g. fear of a threat prompts flight increasing survival)52, 2) arise from corresponding physiological markers (e.g. elevated opioid levels can increase playfulness53), and 3) correlate with specific neuroanatomical regions (e.g. Nucleus accumbens neurons activate before a positive event leading to positive anticipation44,54). This approach reduces anthropomorphic and anthropocentric bias in the selection and comparison of emotions, i.e. instead of trying to identify stereotypically human emotions in dogs, we focused on examining shared underlying mammalian homologies55. Furthermore, for each category of emotion (e.g. fear), we used a range of contexts to generate the emotional response (thunderstorms, specifically avoided objects, etc.). This increased the likelihood of identifying the general facial responses to the emotional category of stimulus (e.g. general facial actions of fear), instead of behavioural motivations (e.g. facial actions displayed for thunderstorms, but not in other fear contexts). We only analysed spontaneous emotional reactions because posed responses could differ from spontaneous ones in duration, intensity, symmetry and form5658.

Table 1.

Emotion, brain system44, definition of emotion, trigger stimuli and context analysed for both species. Different triggers were selected for humans and dogs to allow functional equivalence between species93,94.

Emotion Brain system Definition Humans Dogs
Trigger Context Trigger Context
Fear Fear/anxiety Aversion/avoidance to a stimulus perceived as a threat, leading to flight, fight, freeze and/or distress response40,44. Visualisation of dangerous animal, experiencing high/fast moving vehicles. During presentation of animal or experience in a park ride. Experience of thunderstorms, visualisation of specific objects. During thunderstorms or presentation of objects.
Frustration Rage/Anger Result of denial of a resource by presence of a physical or social barrier, omission/delay of an expected reward, a barrier to success or to achieve a goal42,103. Possibility of gaining a high monetary reward with its subsequent loss(es). After loss of game/life from the UK TV show game The Cube till start of another attempt/return next to the presenter after game ends. Visualisation of a desired resource (toy, food, space) that is or becomes inaccessible. After first attempt to gain access to the resource and during its subsequent denials.
Positive anticipation Seeking/expectancy Expectation of potential gain. Desire of a known and predictable resource/goal. Period from signalling of a reward till moment before receiving reward99,104. Visualisation of food, unwrapping a gift. From visual presentation of food item till moment before eating; From visual presentation of gift till revelation of item. Visualisation of food or hearing meal/food related word(s); Visualisation of leash, hearing walk related word(s). After trigger presentation till moment before obtaining food or leaving home.
Happiness Play/Joy After/during a positive activity/situation that has a positive outcome for the individual and is intrinsically hedonistic/pleasurable. Only observed in the absence of immediate fitness threats and is highly dependent on all proximate needs being fulfilled105,106. Gain of a high monetary reward. After win from the UK TV show game The Cube till moment that contestant returns next to the presenter. Initiation of a play bout; visualisation of owner. During play with conspecifics or owner; reunion with owner after long period of separation.
Relaxation Absence of any emotionally-linked stimuli and response.

Given the common origin of mammalian facial musculature59 and its link to emotion2, and the nature of the long association between humans and dogs, it is plausible that similar emotional reactions might share common facial correlates in these two species (i.e. that the same emotions are closely associated with the same core facial muscle movements). Therefore, we tested two hypotheses: 1) Do dogs produce specific facial actions in response to different categories of emotionally-competent stimuli (i.e. stimuli that produce an emotion cascade resulting in a distinct brain response60,61)? If so, this would provide evidence that dogs possess an adaptive facial behavioural repertoire (sensu Schmidt and Cohn62) with expressive and/or communicative functions associated with the underlying emotional response, as a result of evolutionary pressures. This is a precondition for the main hypothesis: 2) Do dogs use similar facial muscles to humans to express similar emotions? A lack of significant differences between humans and dogs would potentially reflect phylogenetic inertia and be consistent with the shared basic origin of emotional expression as proposed by Darwin1 or reflect convergent evolution. On the other hand, widespread significant differences would indicate that facial expressions of emotion are not highly conserved features across different species.

Results

Human facial actions of emotion

Our study showed convergent validity with previous studies (Table 2) for the core AUs63 associated with each emotion (Table 1). Humans showed a differential use of their facial musculature to produce a higher rate (i.e. in comparison to the relaxed condition representing neutral facial expression) of specific prototypical facial actions during an emotional response, while using various AUs flexibly between contexts. The comparable core AUs identified in our study included AU5 and AU20 for fear, AU14, AU17 and AU24 for frustration, and AU6 for happiness, confirming our human videos as a suitable baseline to compare with the dog videos. This also allowed us to verify that the naturalistic videos used to extract our data can still produce robust results, with facial signals strongly linked to corresponding emotional responses. Additionally, the human FACS coding created a baseline to compare with the dogs’ responses, based on human-dog muscular homologies.

Table 2.

Unique human facial movements displayed during different emotional contexts. Significantly different to relaxed context in one context only (p < 0.05, Kruskal-Wallis with post-hoc pairwise comparisons, Dunn-Bonferroni corrected). Unique facial actions found in previous studies are also reported here for comparison purposes, with the respective literature source. H: Test statistic, SE: Standard Error.

Emotion Humans Previous studies Source
Rate H SE
Fear AU5 AU7 AU20 AD38 19.700 19.900 24.700 15.500 5.285 5.028 5.285 4.161 AU5, AU20 (54)
Frustration AU14 AU17 AU24 AU28 AD84 21.100 20.100 21.900 19.600 16.950 5.985 5.920 5.849 5.162 4.555 AU14, AU17, AU24 (20,55,56)
Positive Anticipation NS
Happiness AU6 33.100 6.308 AU6 (54)

Specifically, our analysis showed that compared to the relaxed condition, humans displayed a significantly higher rate of specific facial actions, in response to stimuli associated with happiness, fear and frustration, but not positive anticipation. For the fear context, the rate of AU5, AU7, AU20 and AD38 were significant; for the frustration context, the rate of AU14, AU17, AU24, AU28 and AD84 were significant; for the happiness context, the rate of AU6 was significant. In the Supplementary Table S4, the facial actions that humans displayed during two or more emotions, but not during the relaxed context are also reported.

Dog facial actions of emotion

In support of our first hypothesis, we found significant differences between each emotional response and relaxation for particular facial and ear actions: Dogs consistently showed a higher rate of AD126 during fear contexts; AD37, AD137 and EAD102 during positive anticipation; and AU27 during happiness. However, frustrated dogs did not display higher rates of any of the facial actions (Fig. 1, Table 3). The higher rates of these specific facial actions are thus strongly linked to the respective emotional contexts. In the Supplementary Table S4, we report the display of significantly higher rates of those facial actions common between emotions (but absent during relaxation); like humans, dogs made a clear, flexible use of their facial musculature.

Figure 1.

Figure 1

Examples of visual representations of unique dog facial actions displayed during emotional reactions. (Individual images composed for illustration of muscular movements only, found in Pixabay 2016, free for commercial use, https://pixabay.com/en/herder-action-dog-animal-dog-plays-1726406/, https://pixabay.com/en/dogs-cute-pet-animal-canine-1181868/, https://pixabay.com/en/animal-canine-cute-dog-friendly-1837003/).

Table 3.

Unique dog facial movements displayed during different emotional contexts. Significantly different than relaxed context in one context only (p < 0.05, Kruskal-Wallis with post-hoc pairwise comparisons, Dunn-Bonferroni corrected). H: Test statistic, SE: Standard Error.

Emotion Dogs
Rate H SE
Fear AD126 29.300 7.945
Frustration NS
Positive Anticipation AD37 AD137 EAD102 25.200 30.825 24.300 7.263 8.716 8.281
Happiness AU27 27.600 7.436

Comparison of human and dog facial actions of emotion

To investigate our second and main hypothesis, direct comparison between human and dog facial actions for each emotion revealed significant differences in specific actions for all the examined emotions (Table 4), demonstrating the differential use of facial musculature between humans and dogs when facing equivalent emotional situations. When compared with humans, dogs displayed higher rates of AD19 in a fearful situation, AU45 during frustration, AD19 and AD37 during positive anticipation, AD19 during happiness, and AU1 in relaxed situations. We also report in Table 4 the facial movements that were significantly higher for humans compared with dogs in the same context. These results show that, in equivalent contexts, dogs and humans mostly activate different facial actions. Out of all the facial actions entered in the analysis, only two Action Units (AU25, AU26) and two Action Descriptors (AD33 and AD40) showed no significant differences between the two species in any of the contexts. This indicates that the majority of facial actions which both dogs and humans are able to produce were used differently for all emotional contexts. Effect sizes were mostly intermediate to large and are reported in the Supplementary Table S5.

Table 4.

Facial movements that differ between species, within each emotional context (p < 0.05). Bold: higher for dogs, U: Test statistic, SE: Standard Error.

Emotion Rate U SE
Fear AD19 AD50 AU10 AU16 AU27 AU43 40.000 178.000 172.000 163.000 170.000 150.000 22.425 20.564 22.517 22.680 20.132 14.759
Frustration AU43 AU45 AU10 160.000 167.000 156.500 15.884 22.728 22.517
Positive Anticipation AD19 AD37 AU10 AU18 30.000 45.000 168.000 147.000 20.941 19.638 22.588 19.638
Happiness AD19 AU10 AU12 36.000 184.000 147.000 21.995 22.588 22.728
Relaxation AU1 35.000 20.564

Control variables

We found no significant differences between different categories of cephalic type, ear morphology and breed for any of the significant facial actions in our main analysis of dogs, i.e. these variables had no significant impact on the production of facial actions in the emotional contexts analysed in this study. For jowl length, the rate of AU10 during happiness was indistinguishable in dogs with short and long jowls, but was higher than in dogs with no jowls (Kruskal-Wallis, with Dunn-Bonferroni correction, H2 = 6.736, p = 0.050); furthermore AU12 had a significantly higher rate in dogs with short jowls than in dogs with long jowls (Kruskal-Wallis, with Dunn-Bonferroni correction, H2 = 9.889, p = 0.036). However, this observation might be a coding artefact as long or no jowls make these AU movements less conspicuous and thus harder to recognise. Finally, we found significantly higher levels of arousal for all emotions when compared with the relaxed referent (Kruskal-Wallis, with Dunn-Bonferroni correction, Fear: H4 = 45.20, p = 0.0001, frustration: H4 = 69.650, p = 0.0001, positive anticipation: H4 = 69.857, p = 0.0001, happiness: H4 = 96.500, p = 0.0001) and for happiness when compared with fear (Kruskal-Wallis, with Dunn-Bonferroni correction, H4 = 51.300, p = 0.0001). This supports the validity of the relaxed context as a point of reference, since this represents, by definition, absence of/very low arousal.

Discussion

This study provides the first empirical evidence to address two important research questions in comparative emotion and interspecific emotion understanding: dogs display specific discriminatory facial movements in response to different emotional stimuli, but do not display similar facial movements to humans when reacting to emotionally comparable contexts.

Dogs tended to make more frequent use of AD126 in a fearful context; AD37, AD137 and EAD102 during positive anticipation; and AU27 in a happiness context (Fig. 1). Such observation is in agreement with a widely held belief that dogs facially react in a common way as a species when confronted with particular challenges or situations in their environment (e.g. when fearful64,65). This also supports Darwin’s suggestions that facial expressions are correlates of internal states in animals1, and are based on homologous facial musculature with the potential to produce similar facial information for a given species. Given previous studies mentioning varied facial cues in frustrated dogs (also described as displacement or conflict behaviours35,66), it was surprising that we did not observe distinctive facial actions for frustration in dogs. However, it might be that the frustration signals previously reported are not specific to frustration and are instead used flexibly in association with other behaviours and/or in response to different contexts (e.g. stress67, discomfort68, fear65), and/or that facial expressions of frustration are more context specific, without a common dynamic anatomical structure. Analysing more videos or featuring a wider range of contexts, to account for potential flexibility, might identify specific facial actions for this emotion in certain common contexts.

Regarding our second hypothesis, our results revealed that humans and dogs produced distinct facial expressions of emotion and little commonality was found in the facial actions triggered, i.e., dogs did not show human-like facial expressions. This observation refutes the idea of continuity of emotional expression between different species. Given the clear facial muscular homology and comparable anatomical facial plan between humans and dogs38,39, their shared social environment and long history of intensive mutual social interaction69,70, it might have been expected that dogs and humans would at least share some facial responses within a comparable emotion context.

Most of the basic facial muscles are similar between humans and dogs, and share primary functions unrelated to facial expression (e.g. the orbicularis oculi muscle that closes the eye to protect the eye or the zygomaticus major that draws the lips back to facilitate feeding). However, when it comes to emotional expression, dogs do not seem to make use of their muscles in the same way that humans do. This might be due to the inability of the dog’s muscles to produce particular movements present in humans because of their different facial morphology (e.g. lack of localised fat deposits). This is the case, for example, with AU6 (produced by the orbicularis oculi muscle), which is a fundamental AU present in all Duchenne happy faces in humans, but in dogs it was never observed even though the same muscle is present and functional. The human face has a more complex facial musculature, i.e. a few muscles are not present in dogs, such as the levator palpebrae that produces AU5 or the risorius that produces AU20. Both AU5 and AU20 (i.e. eyes widely open, lips stretched laterally and downwards) are critical facial actions of a fearful face in humans, but it is impossible for dogs to produce the equivalent facial change.

Another example of muscular variance that might reflect the reported difference in the displayed facial actions relates to the human-dog frustration comparison. Dogs lack a zygomaticus minor muscle, which produces AU14 in humans. This specific movement is one of the core action units of human frustration facial expression. Given the lack of these specific muscles in dogs, it is possible that other muscles could be activated to produce reliable fear or frustration cues in this species. This appears to be the case for fear, since AD126 production had unique characteristics in this context. However, we did not identify an equivalent in the case of frustration, as discussed earlier.

Given the low number of specific facial actions produced in association with each emotion, we suggest that dogs do not display a composed facial expression with several facial actions being integrated in a stereotypical display, as is observed in humans. Instead, dogs seem to produce isolated actions in response to specific emotionally-competent stimuli.

Due to well-known human perception biases (e.g. anthropomorphic tendencies71), it is essential to determine in an objective and standardized way exactly how dogs produce emotional cues. The results in relation to our second hypothesis illustrate the error of broad descriptions and over-reliance on holistic categorization of facial behaviour, which can lead to biases in the perception of facial expressions of a different species (e.g.14,15). It has been argued that an alternative explanation for why humans perceive non-human animals’ emotions as if conspecifics is because of evolutionary homologies between the species, based on shared mammalian brain structures and behavioural expression patterns42,71. However, our study refutes this hypothesis, as the homologies between humans and dogs seem to be limited to the underlying musculature, rather than their activation. In other domestic animals, such as horses72 and sheep73, the basic facial musculature also appears to be well conserved26,74, but the emotional displays appear to diverge from what is observed in humans. It is worth noting that in both of these species (and arguably most domestic animals) the ears appear to be widely used as part of emotional signals, while in humans the ears have more vestigial muscles. Close analysis of the aural signals may therefore be a fruitful line of enquiry in the future. In this respect, it is also worth noting that greater similarities are seen between different species of primates’ (including human) facial expressions75, with them showing a continuity not only at a muscular level but also at the level of expression production. In the case of the dog-human comparison, and unlike the chimpanzee-human comparison, facial expressions seem to be exclusive to the species. Phylogenetic inertia was likely involved in the development of the well-conserved facial musculature of dogs and humans47, but there are clearly differences in the way this musculature is used to produce facial expressions associated with the intentional or unintentional communication of emotion.

Our findings of humans and dogs displaying distinctively different facial expressions have important theoretical and practical implications on human-dog emotion understanding and its underlying cognitive mechanisms. There is little theoretical understanding of interspecific emotional appraisal, but we examine the extension of two competing theories of human-human emotion perception into the field of human-dog emotion perception: “The mental simulation strategy” (i.e. if the same facial movements are produced in response to the same emotional stimuli in both species, individuals could form an understanding of the emotions of others through their own facial actions), and “The implicit social learning strategy” (i.e. if facial movements produced are different between species, an associative learning between external events and motivational states is likely needed for understanding the emotions of other species)76,77. Given that in our study facial movement production differed between humans and dogs, it is unlikely that a mental simulation strategy could be useful or adopted by both species when interpreting heterospecific expressive facial signals. Instead, an implicit social learning strategy would be a more meaningful way to establish a functional human-dog emotion understanding. This would at least partly explain why untrained humans do not seem proficient in reading dogs’ facial and body language7880, particularly subtle cues such as head turning or nose licking81. This is further supported by the neurocognitive evidence that people read dogs’ and humans’ social cues using overlapping brain areas82,83 and similar cognitive mechanisms76,84. Indeed, humans represent non-human animals’ affective space similarly to that of conspecifics’71,85 and incorrectly identify emotions in dogs that have been shown to be a direct result of anthropomorphic subjective judgements (e.g. guilt86). In our study, most emotional facial actions produced by dogs were in the mouth area, using the tongue or the opening of the mouth, and none were in the eye area, which is an important difference from humans that produce eye region AUs in many of the prototypical facial expressions. Thus this might be another reason why humans seem to find it very hard to read dogs’ communication.

The preceding studies strongly suggest that humans read dog communication as they would read other humans (own-species bias87), and our results indicate this is potentially problematic when it comes to evaluation of the emotional content of these signals. This leads to another important issue: all studies to date on human perception of dogs are missing bias-free empirical validation (e.g.8890). For example, when showing a “happy dog” static picture, the only “validation” is the agreement of other subjective human judges, without ever specifying why a “happy dog” is identified as happy, leading to subjective anthropomorphic assessments and circular reasoning, i.e. different human dog “experts” agree with each other on what a “happy dog” should look like, instead of examining what does a “happy dog” actually look like. For the first time, our study demonstrates how to validate dogs’ facial expressions by applying an anatomical and bias-free method based on the presence of empirical stimuli of relevance.

It remains to be seen to what extent the current findings can be generalised across all dog breeds and other emotional contexts, given the variation in facial morphology and possibly muscular functionality in different breeds. Nonetheless, we did establish that features like variation in general head shape may be less important than changes in jowl structure, given the location of key AUs used to express the emotions studied. Despite this study’s focus on facial expressions alone, it is important to mention that emotional expression is clearly multimodal in many species including those studied here (e.g. vocalisations, body posture, and odours may be important moderators of the message sent). While the stimuli were selected for their fundamental emotional equivalence between the two species, the resulting expression might have been moderated by other factors associated with their processing, which were not controlled for (e.g. social vs non-social stimulus source91). Nevertheless, the critical feature is that they should largely share a final common emotional process, whose expression we set out to measure. Thus, future studies might aim to compare dogs of different breeds and morphologies, as well as incorporate the multimodality of emotional signals, and test these hypotheses in different settings (e.g. the more controlled laboratory) to expand on these conclusions.

Methods

Video source and individuals

The videos used in this study were selected from online databases (www.youtube.com and AM-FED database92). Only one individual per video was coded. Whenever there was more than one individual present in the same video that fulfilled all the necessary criteria (see section b), one of the individuals was randomly selected and coded.

The sample consisted of 150 individuals in total, 50 humans and 100 family dogs, distributed equally between emotional contexts (happiness, positive anticipation, fear, frustration and relaxation, Table 1). For detailed information on individuals, please see Supplementary material S1.

Criteria for video selection

The four categories of emotion were defined according to literature sources (Table 1). As a control, we selected videos of neutral/relaxed individuals, i.e., where any specific emotionally-triggering stimuli or overt behavioural reactions were absent. The videos were chosen on the basis of stimulus quality (e.g. source) and its clear association with an evoked response, with different stimuli selected for humans and dogs to allow functional equivalence93,94 (e.g., most humans are not afraid of thunderstorms, while this is a common fear trigger in dogs9597). Videos of dog training, clinical/veterinary settings or with music covering the audio were excluded. Only videos with minimum editing, high image quality (at least 720p), good lighting and visibility of faces were selected. The putative emotion eliciting stimulus had to be apparent and clearly identifiable for at least part of the video. By including homemade/amateur videos with naturalistic and spontaneous behaviour we ensured that the responses were ecologically valid, less constricted, and more robust, especially when compared to laboratory studies on emotions98.

To ensure that the emotionally-competent stimuli were present in the dog videos and were impacting their behaviour, the first author of this study (CC) selected, blinded and randomised the video order. Another author (DM, specialist in veterinary behavioural medicine) relabelled all videos according to the emotion likely triggered by the stimulus99. Only videos that had full agreement were included in the study.

FACS, control variables and reliability coding

All human and dog videos were coded with FACS20 and DogFACS24, respectively, by a certified coder (CC) via the open source software BORIS (Behavioural Observation Research Interactive Software) V2.96100. Each video duration was determined by the individual’s emotional response to the trigger, i.e. from the starting of the behavioural display to the ending time point (neutral behaviour). For example, for the fear trigger of a thunderstorm, the trigger was present throughout the video, for all thunderstorm videos analysed, and the duration of the video was considered while the response (flight/freeze, etc.) of the individual was observed. Before starting the FACS coding, one or more neutral face frames were selected for all individuals. The number of facial actions was identified by watching the videos frame-by-frame, logged in BORIS and extracted for analysis. During speech/vocalizations, facial actions were coded only if they were above level A intensity (as per FACS investigator guide recommendation101). To ensure an unbiased coding, an independent DogFACS certified coder, unaware of the aim of the study or the emotional categorization of the videos, re-coded a random 30% of the dog videos. Inter-coder reliability was thus obtained with 84% mean agreement in the Intraclass Correlation Coefficient (ICC(3,k), CI range of 0.80–0.87).

Since morphological differences could potentially impact upon the production of facial actions in dogs, we used cephalic type (brachycephalic, mesaticephalic, dolichocephalic), jowls length (none, short, long), ear morphology (floppy, partially erect, erect) and breed (as per the British Kennel Club) as control variables. We only analysed the effects of these control variables in the facial actions that were significant in our main analysis. As an additional control, we categorized all the videos for the perceived level of arousal (five point Likert scale: 1 – very low, 2 – low, 3 – moderate, 4 – high, 5 – very high) to ensure arousal was not a confounding variable for emotion. These control variables allowed us to increase the internal validity of our study, since it excludes alternative explanations for the causal relationship between independent and dependent variables102. Different blinded coders recoded the whole sample for each of the control variables to assess mean agreement of inter-rater reliability (ICC(3,k)), obtaining 94% (CI range of 0.91–096) for ears, 85% (CI range of 0.78–0.90), for jowls, and 87% (CI range of 0.81–0.90) for arousal. Human videos were not coded for control or reliability purposes, instead, our results were compared with the respective facial movements reported in the established literature.

Statistical analysis

Because the videos varied in duration (62 ± 45 s, mean ± s.d.) due to differences in the individuals’ responses to the triggers and the emotions examined, the number of facial actions (4.59 ± 8.97, mean ± s.d.) was normalised into rates of visible time, i.e., the number of each AU was divided by the duration of the video where the face is visible. Variables with very low rates (<0.001) were excluded from analysis (Supplementary material S3). Thus, we included 30 human facial actions and 22 dog facial actions in the statistical analysis (Supplementary Table S2). We performed Kolmogorov–Smirnov tests for normality and non-parametric Levene’s test for homoscedasticity. Since all variables violated both assumptions in at least one of the groups, non-parametric tests were used throughout the analysis. To investigate if dogs produced differential facial actions in response to specific emotional triggers, we compared the facial actions displayed between emotions and between each emotion and the control relaxation videos with Kruskal-Wallis (adjusted for ties) followed by post-hoc pairwise multiple comparisons tests, with Dunn-Bonferroni correction. For our second hypothesis, where we were interested in assessing if dogs were using the same facial actions as humans in response to an equivalent emotional trigger, we performed Mann-Whitney tests, with exact p-value for the homologous facial actions between species. For both hypotheses, the rate of facial movements was used as the dependent variable and the emotion category was the independent variable. Additionally, we calculated effect sizes to consider and minimize the risk of type 2 error in our interpretation of the results.

To analyse the potential effect of the control variables cephalic type, ear morphology, jowl length, breed, and arousal level on the facial actions, we used Kruskal-Wallis tests (adjusted for ties) and post-hoc pairwise multiple comparisons tests, with Dunn-Bonferroni correction. Here, we compared the rate of each facial action (dependent variable) between the categories (independent variables) of the control variables, in order to ensure the distribution of facial actions was the same regardless of the dogs’ morphological variation. For ear morphology and jowl length, we only performed the analysis for the relevant facial actions (e.g. jowl length only analysed for lower face actions). All statistical analyses were performed using IBM SPSS v22, except for the effect sizes that were computed with GPower v3.1.

Ethics

This study was approved by the School of Psychology Research Ethics Committee and follows the Ethics Guidelines for Internet-mediated Research by the British Psychological Society.

Data availability

The datasets generated and analysed during the current study are available from the corresponding author on request.

Electronic supplementary material

Supplementary Information (393.9KB, pdf)
Supplementary Info File 1 (47.3KB, xlsx)

Acknowledgements

Mónica Costa, Satoko Matsubara, Pearl Gillum and Jane Harris for inter-reliability coding. School of Psychology technicians for software support. Daniel Pincheira Donoso for his comments and discussion of points in earlier versions of the manuscript. This study was funded by a University of Lincoln Research Investment Fund studentship, awarded to C.C.C.

Author Contributions

C.C., K.G. and D.M. developed the study concept and design. C.C. and D.M. selected videos. C.C. coded videos and analysed the data. C.C., K.G. and D.M. interpreted the data, and C.C. drafted the initial manuscript. All authors provided critical revisions and approved the final version.

Competing Interests

The authors declare that they have no competing interests.

Footnotes

A correction to this article is available online at https://doi.org/10.1038/s41598-018-28651-z.

Electronic supplementary material

Supplementary information accompanies this paper at 10.1038/s41598-017-15091-4.

Publisher's note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Change history

7/5/2018

A correction to this article has been published and is linked from the HTML and PDF versions of this paper. The error has been fixed in the paper.

References

  • 1.Darwin, C. The expression of the emotions in man and animals. (D. Appleton and Company, 1896).
  • 2.Ekman P, Oster H. Facial expressions of emotion. Annu. Rev. Psychol. 1979;30:527–554. doi: 10.1146/annurev.ps.30.020179.002523. [DOI] [Google Scholar]
  • 3.Fox MW. A comparative study of the development of facial expressions in canids: wolf, coyote and foxes. Behaviour. 1970;36:49–73. doi: 10.1163/156853970X00042. [DOI] [Google Scholar]
  • 4.Chevalier-Skolnikoff, S. Facial expression of emotion in nonhuman primates. In Darwin and Facial Expression: A Century of Research InReview (ed. Ekman, P.) (Academic Press, 1973).
  • 5.Tate AJ, Fischer H, Leigh AE, Kendrick KM. Behavioural and neurophysiological evidence for face identity and face emotion processing in animals. Philos. Trans. R. Soc. B Biol. Sci. 2006;361:2155–2172. doi: 10.1098/rstb.2006.1937. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Leopold DA, Rhodes G. A comparative view of face perception. J. Comp. Psychol. 2010;124:233–251. doi: 10.1037/a0019460. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Scherer KR. What are emotions? And how can they be measured? Soc. Sci. Inf. 2005;44:695–729. doi: 10.1177/0539018405058216. [DOI] [Google Scholar]
  • 8.Dimberg U, Petterson M. Facial reactions to happy and angry facial expressions: Evidence for right hemisphere dominance. Psychophysiology. 2000;37:693–696. doi: 10.1111/1469-8986.3750693. [DOI] [PubMed] [Google Scholar]
  • 9.Surguladze SA, et al. Recognition accuracy and response bias to happy and sad facial expressions in patients with major depression. Neuropsychology. 2004;18:212–218. doi: 10.1037/0894-4105.18.2.212. [DOI] [PubMed] [Google Scholar]
  • 10.Etcoff NL, Magee JJ. Categorical perception of facial expressions. Cognition. 1992;44:227–240. doi: 10.1016/0010-0277(92)90002-Y. [DOI] [PubMed] [Google Scholar]
  • 11.Ekman P, Friesen WV. Felt, false, and miserable smiles. J. Nonverbal Behav. 1982;6:238–252. doi: 10.1007/BF00987191. [DOI] [Google Scholar]
  • 12.Gervais M, Wilson DS. The evolution and functions of laughter and humor: A synthetic approach. Q. Rev. Biol. 2005;80:395–430. doi: 10.1086/498281. [DOI] [PubMed] [Google Scholar]
  • 13.Harris C, Alvarado N. Facial expressions, smile types, and self-report during humour, tickle, and pain. Cogn. Emot. 2005;19:655–669. doi: 10.1080/02699930441000472. [DOI] [Google Scholar]
  • 14.Preuschoft S. “Laughter” and “Smile” in Barbary Macaques (Macaca sylvanus) Ethology. 1992;91:220–236. doi: 10.1111/j.1439-0310.1992.tb00864.x. [DOI] [Google Scholar]
  • 15.Waller BM, Dunbar RIM. Differential behavioural effects of silent bared teeth display and relaxed open mouth display in chimpanzees (Pan troglodytes) Ethology. 2005;111:129–142. doi: 10.1111/j.1439-0310.2004.01045.x. [DOI] [Google Scholar]
  • 16.Waller BM, Cherry L. Facilitating play through communication: Significance of teeth exposure in the gorilla play face. Am. J. Primatol. 2012;74:157–164. doi: 10.1002/ajp.21018. [DOI] [PubMed] [Google Scholar]
  • 17.Palagi E. Social play in bonobos (Pan paniscus) and chimpanzees (Pan troglodytes): Implications for natural social systems and interindividual relationships. Am. J. Phys. Anthropol. 2006;129:418–426. doi: 10.1002/ajpa.20289. [DOI] [PubMed] [Google Scholar]
  • 18.Palagi E, Antonacci D, Cordoni G. Fine-tuning of social play in juvenile lowland gorillas (gorilla gorilla gorilla) Dev. Psychobiol. 2007;49:433–445. doi: 10.1002/dev.20219. [DOI] [PubMed] [Google Scholar]
  • 19.Palagi E, Mancini G. Playing with the face: Playful facial “chattering” and signal modulation in a monkey species (Theropithecus gelada) J. Comp. Psychol. 2011;125:11–21. doi: 10.1037/a0020869. [DOI] [PubMed] [Google Scholar]
  • 20.Ekman, P., Friesen, W. & Hager, J. Facial Action Coding System (FACS): manual. (Research Nexus, 2002).
  • 21.Ekman P, Friesen WV. A new pan-cultural facial expression of emotion. Motiv. Emot. 1986;10:159–168. doi: 10.1007/BF00992253. [DOI] [Google Scholar]
  • 22.Ekman, P. & Rosenberg, E. L. What the face reveals: Basic and applied studies of spontaneous expression using the Facial Action Coding System (FACS). (Oxford University Press, 1997).
  • 23.Caeiro CC, Waller BM, Zimmermann E, Burrows AM, Davila-Ross M. OrangFACS: A muscle-based facial movement coding system for orangutans (Pongo spp.) Int. J. Primatol. 2013;34:115–129. doi: 10.1007/s10764-012-9652-x. [DOI] [Google Scholar]
  • 24.Waller BM, et al. Paedomorphic facial expressions give dogs a selective advantage. PLOS ONE. 2013;8:e82686. doi: 10.1371/journal.pone.0082686. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Caeiro CC, Burrows A, Waller BM. Development and application of CatFACS: Are human cat adopters influenced by cat facial expressions? Appl. Anim. Behav. Sci. 2017;189:66–78. doi: 10.1016/j.applanim.2017.01.005. [DOI] [Google Scholar]
  • 26.Wathan J, Burrows AM, Waller BM, McComb K. EquiFACS: The Equine Facial Action Coding System. PLOS ONE. 2015;10:e0131738. doi: 10.1371/journal.pone.0131738. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Waller BM, Lembeck M, Kuchenbuch P, Burrows AM, Liebal K. GibbonFACS: A muscle-based facial movement coding system for hylobatids. Int. J. Primatol. 2012;33:809–821. doi: 10.1007/s10764-012-9611-6. [DOI] [Google Scholar]
  • 28.Parr LA, Waller BM, Burrows AM, Gothard KM, Vick SJ. Brief communication: MaqFACS: A muscle-based facial movement coding system for the rhesus macaque. Am. J. Phys. Anthropol. 2010;143:625–630. doi: 10.1002/ajpa.21401. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Julle-Danière É, et al. MaqFACS (Macaque Facial Action Coding System) can be used to document facial movements in Barbary macaques (Macaca sylvanus) PeerJ. 2015;3:e1248. doi: 10.7717/peerj.1248. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Vick SJ, Waller BM, Parr LA, Pasqualini MCS, Bard KA. A cross-species comparison of facial morphology and movement in humans and chimpanzees using the Facial Action Coding System (FACS) J. Nonverbal Behav. 2007;31:1–20. doi: 10.1007/s10919-006-0017-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Waller BM, Misch A, Whitehouse J, Herrmann E. Children, but not chimpanzees, have facial correlates of determination. Biol. Lett. 2014;10:20130974–20130974. doi: 10.1098/rsbl.2013.0974. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Parr LA, Waller BM, Vick SJ. New developments in understanding emotional facial signals in chimpanzees. Curr. Dir. Psychol. Sci. 2007;16:117–122. doi: 10.1111/j.1467-8721.2007.00487.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Burrows AM, Waller BM, Parr LA, Bonar CJ. Muscles of facial expression in the chimpanzee (Pan troglodytes): descriptive, comparative and phylogenetic contexts. J. Anat. 2006;208:153–167. doi: 10.1111/j.1469-7580.2006.00523.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Parr LA, Waller BM, Vick SJ, Bard KA. Classifying chimpanzee facial expressions using muscle action. Emotion. 2007;7:172–181. doi: 10.1037/1528-3542.7.1.172. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Kuhne F, Hößler JC, Struwe R. Emotions in dogs being petted by a familiar or unfamiliar person: Validating behavioural indicators of emotional states using heart rate variability. Appl. Anim. Behav. Sci. 2014;161:113–120. doi: 10.1016/j.applanim.2014.09.020. [DOI] [Google Scholar]
  • 36.Bekoff M. Social play in coyotes, wolves, and dogs. BioScience. 1974;24:225–230. doi: 10.2307/1296803. [DOI] [Google Scholar]
  • 37.Cordoni G, Nicotra V, Palagi E. Unveiling the “secret” of play in dogs (Canis lupus familiaris): Asymmetry and signals. J. Comp. Psychol. 2016;130:278–287. doi: 10.1037/com0000035. [DOI] [PubMed] [Google Scholar]
  • 38.Bolwig N. Facial expression in primates with remarks on a parallel development in certain carnivores (a preliminary report on work in progress) Behaviour. 1964;22:167–192. doi: 10.1163/156853964X00012. [DOI] [Google Scholar]
  • 39.Diogo R, Wood BA, Aziz MA, Burrows A. On the origin, homologies and evolution of primate facial muscles, with a particular focus on hominoids and a suggested unifying nomenclature for the facial muscles of the Mammalia. J. Anat. 2009;215:300–319. doi: 10.1111/j.1469-7580.2009.01111.x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Lang PJ, Davis M, Öhman A. Fear and anxiety: animal models and human cognitive psychophysiology. J. Affect. Disord. 2000;61:137–159. doi: 10.1016/S0165-0327(00)00343-8. [DOI] [PubMed] [Google Scholar]
  • 41.Berridge, K. C. Comparing the emotional brains of humans and other animals. In Handbook of AffectiveSciences (eds. Davidson, R. J., Scherer, K. R. & Goldsmith, H. H.) 25–51 (Oxford University Press, 2003).
  • 42.Panksepp J. Affective consciousness: Core emotional feelings in animals and humans. Conscious. Cogn. 2005;14:30–80. doi: 10.1016/j.concog.2004.10.004. [DOI] [PubMed] [Google Scholar]
  • 43.Phelps EA, LeDoux JE. Contributions of the amygdala to emotion processing: From animal models to human behavior. Neuron. 2005;48:175–187. doi: 10.1016/j.neuron.2005.09.025. [DOI] [PubMed] [Google Scholar]
  • 44.Panksepp J. The basic emotional circuits of mammalian brains: Do animals have affective lives? Neurosci. Biobehav. Rev. 2011;35:1791–1804. doi: 10.1016/j.neubiorev.2011.08.003. [DOI] [PubMed] [Google Scholar]
  • 45.Guo K, Meints K, Hall C, Hall S, Mills D. Left gaze bias in humans, rhesus monkeys and domestic dogs. Anim. Cogn. 2009;12:409–418. doi: 10.1007/s10071-008-0199-3. [DOI] [PubMed] [Google Scholar]
  • 46.Hansen, T. F., Pienaar, J. & Orzack, S. H. A comparative method for studying adaptation to a randomly evolving environment. Evolution 1965–1977, 10.1111/j.1558-5646.2008.00412.x (2008). [DOI] [PubMed]
  • 47.Shanahan T. Phylogenetic inertia and Darwin’s higher law. Stud. Hist. Philos. Sci. Part C Stud. Hist. Philos. Biol. Biomed. Sci. 2011;42:60–68. doi: 10.1016/j.shpsc.2010.11.013. [DOI] [PubMed] [Google Scholar]
  • 48.Blomberg SP, Garland T. Tempo and mode in evolution: phylogenetic inertia, adaptation and comparative methods. J. Evol. Biol. 2002;15:899–910. doi: 10.1046/j.1420-9101.2002.00472.x. [DOI] [Google Scholar]
  • 49.Darwin, C. On the origin of species. (John Murray, 1859).
  • 50.Hall BK. Descent with modification: the unity underlying homology and homoplasy as seen through an analysis of development and evolution. Biol. Rev. 2003;78:409–433. doi: 10.1017/S1464793102006097. [DOI] [PubMed] [Google Scholar]
  • 51.Ekman P. An argument for basic emotions. Cogn. Emot. 1992;6:169–200. doi: 10.1080/02699939208411068. [DOI] [Google Scholar]
  • 52.Panksepp J, Fuchs T, Iacobucci P. The basic neuroscience of emotional experiences in mammals: The case of subcortical FEAR circuitry and implications for clinical anxiety. Appl. Anim. Behav. Sci. 2011;129:1–17. doi: 10.1016/j.applanim.2010.09.014. [DOI] [Google Scholar]
  • 53.Trezza V, Baarendse PJJ, Vanderschuren LJMJ. The pleasures of play: pharmacological insights into social reward mechanisms. Trends Pharmacol. Sci. 2010;31:463–469. doi: 10.1016/j.tips.2010.06.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Berridge KC. Reward learning: Reinforcement, incentives, and expectations. Psychol. Learn. Motiv. 2000;40:223–278. doi: 10.1016/S0079-7421(00)80022-5. [DOI] [Google Scholar]
  • 55.Charland LC. The natural kind status of emotion. Br. J. Philos. Sci. 2002;53:511–537. doi: 10.1093/bjps/53.4.511. [DOI] [Google Scholar]
  • 56.Ekman P, Hager JC, Friesen WV. The symmetry of emotional and deliberate facial actions. Psychophysiology. 1981;18:101–106. doi: 10.1111/j.1469-8986.1981.tb02919.x. [DOI] [PubMed] [Google Scholar]
  • 57.Cohn JF, Schmidt KL. The timing of facial motion in posed and spontaneous smiles. Int. J. Wavelets Multiresolution Inf. Process. 2004;2:121–132. doi: 10.1142/S021969130400041X. [DOI] [Google Scholar]
  • 58.Raheja, J. L. & Gupta, J. Distinguishing Facial Expressions: Genuine Vs Fake. Int. J. Recent Trends Eng. Technol. 3 (2010).
  • 59.Diogo R, et al. The head and neck muscles of the serval and tiger: Homologies, evolution, and proposal of a mammalian and a veterinary muscle ontology. Anat. Rec. Adv. Integr. Anat. Evol. Biol. 2012;295:2157–2178. doi: 10.1002/ar.22589. [DOI] [PubMed] [Google Scholar]
  • 60.Damasio, A. Emotions and feelings: A neurobiological perspective. In Feelings and Emotions: The Amsterdam Symposium (eds. Manstead, A. S. R., Frijda, N. & Fischer, A.) 49–57 (Cambridge University Press, 2004).
  • 61.Damasio A. Fundamental feelings. Nature. 2001;413:781. doi: 10.1038/35101669. [DOI] [PubMed] [Google Scholar]
  • 62.Schmidt KL, Cohn JF. Human facial expressions as adaptations: Evolutionary questions in facial expression research. Am. J. Phys. Anthropol. 2001;116:3–24. doi: 10.1002/ajpa.20001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Waller BM, Cray JJ, Burrows AM. Selection for universal facial emotion. Emotion. 2008;8:435–439. doi: 10.1037/1528-3542.8.3.435. [DOI] [PubMed] [Google Scholar]
  • 64.Beerda B, Schilder MB, van Hooff JA, de Vries HW, Mol JA. Behavioural, saliva cortisol and heart rate responses to different types of stimuli in dogs. Appl. Anim. Behav. Sci. 1998;58:365–381. doi: 10.1016/S0168-1591(97)00145-7. [DOI] [Google Scholar]
  • 65.Stellato AC, Flint HE, Widowski TM, Serpell JA, Niel L. Assessment of fear-related behaviours displayed by companion dogs (Canis familiaris) in response to social and non-social stimuli. Appl. Anim. Behav. Sci. 2017;188:84–90. doi: 10.1016/j.applanim.2016.12.007. [DOI] [Google Scholar]
  • 66.Lund JD, Jørgensen MC. Behaviour patterns and time course of activity in dogs with separation problems. Appl. Anim. Behav. Sci. 1999;63:219–236. doi: 10.1016/S0168-1591(99)00011-8. [DOI] [Google Scholar]
  • 67.Beerda B, Schilder MB, van Hooff JA, de Vries HW. Manifestations of chronic and acute stress in dogs. Appl. Anim. Behav. Sci. 1997;52:307–319. doi: 10.1016/S0168-1591(96)01131-8. [DOI] [Google Scholar]
  • 68.Hecht, J. & Horowitz, A. Introduction to dog behaviour. In Animal Behavior for Shelter Veterinarians and Staff (eds. Weiss, E., Heather Mohan-Gibbons & Stephen Zawistowski) (Wiley-Blackwell, 2015).
  • 69.Vilà C, et al. Multiple and ancient origins of the domestic dog. Science. 1997;276:1687–1689. doi: 10.1126/science.276.5319.1687. [DOI] [PubMed] [Google Scholar]
  • 70.Hare B, Tomasello M. Human-like social skills in dogs? Trends Cogn. Sci. 2005;9:439–444. doi: 10.1016/j.tics.2005.07.003. [DOI] [PubMed] [Google Scholar]
  • 71.Konok V, Nagy K, Miklósi Á. How do humans represent the emotions of dogs? The resemblance between the human representation of the canine and the human affective space. Appl. Anim. Behav. Sci. 2015;162:37–46. doi: 10.1016/j.applanim.2014.11.003. [DOI] [Google Scholar]
  • 72.Hintze S, Smith S, Patt A, Bachmann I, Würbel H. Are eyes a mirror of the soul? What eye wrinkles reveal about a horse’s emotional state. PLOS ONE. 2016;11:e0164017. doi: 10.1371/journal.pone.0164017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73.Boissy A, et al. Cognitive sciences to relate ear postures to emotions in sheep. Anim. Welf. 2011;20:47. [Google Scholar]
  • 74.Swielim, G. E. A. Atlas - Anatomy of sheep. (The Academic Bookshop, Egyptian Joint-Stock Co., 2006).
  • 75.Parr LA, Waller BM. Understanding chimpanzee facial expression: insights into the evolution of communication. Soc. Cogn. Affect. Neurosci. 2006;1:221–228. doi: 10.1093/scan/nsl031. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Keysers C, Perrett DI. Demystifying social cognition: a Hebbian perspective. Trends Cogn. Sci. 2004;8:501–507. doi: 10.1016/j.tics.2004.09.005. [DOI] [PubMed] [Google Scholar]
  • 77.Bruce, V. & Young, A. W. Face perception. (Psychology Press, 2012).
  • 78.Tami G, Gallagher A. Description of the behaviour of domestic dog (Canis familiaris) by experienced and inexperienced people. Appl. Anim. Behav. Sci. 2009;120:159–169. doi: 10.1016/j.applanim.2009.06.009. [DOI] [Google Scholar]
  • 79.Colombo ES, Prato-Previde E. Empathy and recognition of dogs’ (canis familiaris) emotions: a pilot focusing on vets. J. Vet. Behav. Clin. Appl. Res. 2014;9:e18. doi: 10.1016/j.jveb.2014.09.061. [DOI] [Google Scholar]
  • 80.Kerswell KJ, Bennett P, Butler KL, Hemsworth PH. Self-Reported Comprehension Ratings of Dog Behavior by Puppy Owners. Anthrozoos Multidiscip. J. Interact. People Anim. 2009;22:183–193. [Google Scholar]
  • 81.Mariti C, et al. Perception of dogs’ stress by their owners. J. Vet. Behav. Clin. Appl. Res. 2012;7:213–219. doi: 10.1016/j.jveb.2011.09.004. [DOI] [Google Scholar]
  • 82.Stoeckel LE, Palley LS, Gollub RL, Niemi SM, Evins AE. Patterns of brain activation when mothers view their own child and dog: An fMRI study. PLoS ONE. 2014;9:e107205. doi: 10.1371/journal.pone.0107205. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83.Törnqvist H, et al. Comparison of dogs and humans in visual scanning of social interaction. R. Soc. Open Sci. 2015;2:150341. doi: 10.1098/rsos.150341. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84.Kujala MV, Kujala J, Carlson S, Hari R. Dog Experts’ Brains Distinguish Socially Relevant Body Postures Similarly in Dogs and Humans. PLoS ONE. 2012;7:e39145. doi: 10.1371/journal.pone.0039145. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Morris PH, Doe C, Godsell E. Secondary emotions in non-primate species? Behavioural reports and subjective claims by animal owners. Cogn. Emot. 2008;22:3–20. doi: 10.1080/02699930701273716. [DOI] [Google Scholar]
  • 86.Horowitz A. Disambiguating the “guilty look”: Salient prompts to a familiar dog behaviour. Behav. Processes. 2009;81:447–452. doi: 10.1016/j.beproc.2009.03.014. [DOI] [PubMed] [Google Scholar]
  • 87.Pascalis O, Kelly DJ. The origins of face processing in humans: Phylogeny and ontogeny. Perspect. Psychol. Sci. 2009;4:200–209. doi: 10.1111/j.1745-6924.2009.01119.x. [DOI] [PubMed] [Google Scholar]
  • 88.Rugaas, T. On Talking Terms with Dogs: Calming Signals. (Dogwise Publishing, 2005).
  • 89.Shepherd, K. Development of behaviour, social behaviour and communication in dogs. In BSAVA Manual of Canine and Feline Behavioural Medicine (eds. Horwitz, D. & Mills, D. S.) 13–16 (2009).
  • 90.Bloom T, Friedman H. Classifying dogs’ (Canis familiaris) facial expressions from photographs. Behav. Processes. 2013;96:1–10. doi: 10.1016/j.beproc.2013.02.010. [DOI] [PubMed] [Google Scholar]
  • 91.Davidson JR. Use of benzodiazepines in social anxiety disorder, generalized anxiety disorder, and posttraumatic stress disorder. J. Clin. Psychiatry. 2004;65:29–33. [PubMed] [Google Scholar]
  • 92.McDuff, D. et al. Affectiva-MIT facial expression dataset (AM-FED): Naturalistic and spontaneous facial expressions collected ‘in-the-wild’. In 2013 IEEE Conference on Computer Vision and Pattern Recognition Workshops 881–888 10.1109/CVPRW.2013.130 (2013).
  • 93.Tomasello M, Call J. Assessing the validity of ape-human comparisons: A reply to Boesch (2007) J. Comp. Psychol. 2008;122:449–452. doi: 10.1037/0735-7036.122.4.449. [DOI] [PubMed] [Google Scholar]
  • 94.Green, G. & Saunders, R. Stimulus equivalence. In Handbook of Research Methods in Human Operant Behavior (eds. Lattal, K. & Perone, M.) (Springer Science & Business Media, 2013).
  • 95.McCobb E, Brown E, Damiani K, Dodman N. Thunderstorm phobia in dogs: an Internet survey of 69 cases. J. Am. Anim. Hosp. Assoc. 2001;37:319–324. doi: 10.5326/15473317-37-4-319. [DOI] [PubMed] [Google Scholar]
  • 96.Overall KL, Love M. Dog bites to humans—demography, epidemiology, injury, and risk. J. Am. Vet. Med. Assoc. 2001;218:1923–1934. doi: 10.2460/javma.2001.218.1923. [DOI] [PubMed] [Google Scholar]
  • 97.Blackwell EJ, Bradshaw JWS, Casey RA. Fear responses to noises in domestic dogs: Prevalence, risk factors and co-occurrence with other fear related behaviour. Appl. Anim. Behav. Sci. 2013;145:15–25. doi: 10.1016/j.applanim.2012.12.004. [DOI] [Google Scholar]
  • 98.Fernández-Dols J-M, Crivelli C. Emotion and expression: Naturalistic studies. Emot. Rev. 2013;5:24–29. doi: 10.1177/1754073912457229. [DOI] [Google Scholar]
  • 99.Mills, D. S., Dube, M. B. & Zulch, H. Stress and pheromonatherapy in small animal clinical behaviour. (John Wiley & Sons, 2012).
  • 100.Friard O, Gamba M. BORIS: a free, versatile open-source event-logging software for video/audio coding and live observations. Methods Ecol. Evol. 2016;7:1325–1330. doi: 10.1111/2041-210X.12584. [DOI] [Google Scholar]
  • 101.Ekman, P., Friesen, W. & Hager, J. FACS investigator’s guide. (Research Nexus, 2002).
  • 102.Brewer, M. B. & Crano, W. D. Research design and issues of validity. in Handbook of Research Methods in Social and PersonalityPsychology (eds. Reis, H. T. & Judd, C. M.) 11–26 (Cambridge University Press, 2014).
  • 103.Broom, D. M. & Johnson, K. G. Stress and Animal Welfare. (Springer Science & Business Media, 1993).
  • 104.O’Doherty JP, Deichmann R, Critchley HD, Dolan RJ. Neural responses during anticipation of a primary taste reward. Neuron. 2002;33:815–826. doi: 10.1016/S0896-6273(02)00603-7. [DOI] [PubMed] [Google Scholar]
  • 105.Boissy A, et al. Assessment of positive emotions in animals to improve their welfare. Physiol. Behav. 2007;92:375–397. doi: 10.1016/j.physbeh.2007.02.003. [DOI] [PubMed] [Google Scholar]
  • 106.Held SDE, Špinka M. Animal play and animal welfare. Anim. Behav. 2011;81:891–899. doi: 10.1016/j.anbehav.2011.01.007. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Information (393.9KB, pdf)
Supplementary Info File 1 (47.3KB, xlsx)

Data Availability Statement

The datasets generated and analysed during the current study are available from the corresponding author on request.


Articles from Scientific Reports are provided here courtesy of Nature Publishing Group

RESOURCES