Abstract
Appropriate response to companions’ emotional signals is important for all social creatures. The emotional expressions of humans and non-human animals have analogies in their form and function, suggesting shared evolutionary roots, but very little is known about how animals other than primates view and process facial expressions. In primates, threat-related facial expressions evoke exceptional viewing patterns compared with neutral or positive stimuli. Here, we explore if domestic dogs (Canis familiaris) have such an attentional bias toward threatening social stimuli and whether observed emotional expressions affect dogs’ gaze fixation distribution among the facial features (eyes, midface and mouth). We recorded the voluntary eye gaze of 31 domestic dogs during viewing of facial photographs of humans and dogs with three emotional expressions (threatening, pleasant and neutral). We found that dogs’ gaze fixations spread systematically among facial features. The distribution of fixations was altered by the seen expression, but eyes were the most probable targets of the first fixations and gathered longer looking durations than mouth regardless of the viewed expression. The examination of the inner facial features as a whole revealed more pronounced scanning differences among expressions. This suggests that dogs do not base their perception of facial expressions on the viewing of single structures, but the interpretation of the composition formed by eyes, midface and mouth. Dogs evaluated social threat rapidly and this evaluation led to attentional bias, which was dependent on the depicted species: threatening conspecifics’ faces evoked heightened attention but threatening human faces instead an avoidance response. We propose that threatening signals carrying differential biological validity are processed via distinctive neurocognitive pathways. Both of these mechanisms may have an adaptive significance for domestic dogs. The findings provide a novel perspective on understanding the processing of emotional expressions and sensitivity to social threat in non-primates.
Introduction
Appropriate reading and responding to companions’ emotional signals is important to social creatures. For humans, facial expressions are the main channel for communicating emotional states and have thus been at the center of scientific interest for decades [1,2]. Most mammalian species have musculature for producing facial movements that resemble human expressions [3,4]. The analogies in the form and function of human and non-human animal emotional expressions suggest shared evolutionary roots [2,4,5]. Nevertheless, little is known about how non-primates view and process facial expressions. Full understanding of the complexity and cognitive mechanisms of facial expressions requires a wider approach, taking into account various examples from the animal kingdom (for reviews, see [4,6,7]).
Humans, chimpanzees (Pan troglodytes) and macaques (Macaca mulatta) exhibit exceptional viewing behavior toward threat-related expressions compared with neutral or positive stimuli. Typically the threatening stimuli accelerate detection [8,9] or prolong attention [9–16], but sometimes they evoke aversive reactions [9,17–19]. The phenomenon has been extensively studied because it is particularly evident in anxiety and mood disorders [20–22]. The attentional bias toward threat is considered to originate from a phylogenetically old adaptive mechanism: the sensitivity to detect and avoid threats represents a survival advantage [21]. Studying sensitivity to social threat in animals other than humans can promote understanding of the origins of phenomenon [15].
During viewing of emotional faces different facial features attract attention differently [23,24]. In humans and other primates the viewed emotional expression alter the distribution of gaze fixations among inner key structures, i.e. eyes, nose and mouth [10,25–31], probably because these features convey different emotional information for a viewer [24,27–29]. It is not known how other animals than primates scan facial expressions and what facial features are relevant for them.
Here, we chose domestic dogs (Canis familiaris) as study subjects because canids use facial cues richly in social interactions and their facial expressions share similarities with primates [5,32,33]. In addition, domestic dogs are known for their well-developed social communication skills (for reviews, see [34,35]) and they respond flexibly to human emotional signals [36–39]. The facial expressions of dogs can be linked to affective situations (e.g. anticipation of the reward, social isolation or reunion with the owner) [40–43] and categorized according to basic emotion classification used for humans [41]. Certain characteristics of facial communication may have contributed to the relation between dogs and humans, and even given a selective advantage to dogs during domestication [43]. However, only scarce information exists considering how dogs process facial expressions of conspecifics’ and humans. To date we know that dogs can distinguish human facial expressions from images after discrimination training [44,45] and their viewing behavior is sensitive to the facial expressions of both conspecifics and humans [46].
The approaches in which human and non-human emotional perception have been studied have differed greatly both in methodological constraints and theoretical aspects [4]. In humans and other primates, eye gaze tracking has been a widely used method for assessing affective responses. To narrow the methodological gap between the studies of primates and non-primates, we employed the rapidly advanced eye tracking methodology [47–50] to measure the voluntary eye gaze of domestic dogs during viewing of emotional expressions of dogs and humans. Eye tracking findings to date show that the gazing behavior of dogs has analogies with that of primates: dogs spontaneously focus attention on informative objects, like faces [49] and eyes [50]; they prefer images representing conspecifics over images of human or inanimate objects [49], upright faces over inverted ones [50] and familiar faces over strange ones [50]. While watching dynamic social stimuli, dogs gazing behavior also resembles their behavior in real communicative situations [48]. Together these findings suggest that dogs comprehend two-dimensional representations and their gaze targets are altered by the seen contextual information.
The aim of the present study was to explore how dogs scan facial expressions of conspecifics (social and phylogenetic stimuli) compared with the species that is most likely emotionally relevant for them, humans (social, but non-phylogenetic stimuli). We were particularly interested to discover whether dogs exhibit specific attentional biases toward potentially threatening social stimuli versus non-threatening social stimuli, and whether their gaze distribution among the inner facial features is dependent on the seen emotional content.
Materials and Methods
Ethical statement
This study was performed in strict accordance with the Finnish Act of Animal Experimentation (62/2006) in which the European convention for the protection of vertebrate animals used for experimental and other scientific purposes (Directive 86/609/EEC) is fully implemented. All protocols were approved by the Ethical Committee for the Use of Animals in Experiments at the University of Helsinki (2/2010) and the Finnish national Animal Experiment Board (approval#STH367A/ESLH-2008-04236/Ym-23). Animals were not harmed in any way during the experiments and neither mechanical nor manual restraint was applied. Only non-invasive and reward-based methods were used during training and measurement.
Animals and pre-training
All experiments were conducted at the Veterinary Faculty of the University of Helsinki. Participant dogs (25 privately owned pet dogs and 8 kennel-housed beagles) were on average 4.6 years old (SD 2.2), and represented 13 different breeds and mongrels (8 beagles, 6 border collies, 3 hovawarts, 3 beauce shepherds, 2 rough collies, 2 smooth collies, 1 great Pyrenees, 1 Welsh corgi, 1 Australian kelpie, 1 lagotto romagnolo, 1 Manchester terrier, 1 Swedish vallhund, 1 Finnish lapphund, 2 mongrels). The daily routines of the dogs were kept similar to those in their regular life. Pet dogs (14 intact females, 4 sterilized females, 5 intact males and 2 castrated males) lived in their home environment, were fed once or twice a day and taken outdoors three to five times for 0.5–2 hours at a time. Kennel dogs (2 sterilized females and 6 castrated males) lived in the kennel facilities of the University of Helsinki. They were fed twice a day and released into an exercise enclosure once a day for two hours.
All the dogs had previously participated in an eye-tracking experiment [50] but were naïve regarding images used in the present study. Prior the previous eye-tracking experiment, all subjects were clicker-trained to lie still in front of a monitor and lean their jaw on a specially designed chin rest [49]. The criterion for passing the training period was that the dog took the pre-trained position without being commanded to do so and remained in that position for at least 30 seconds while the owner and experimenters were positioned behind an opaque barrier. During the training, the dogs were not encouraged to fix their eyes on a monitor or images and they were not restrained or forced to perform the task.
Stimuli
Digital color photographs of unfamiliar dog and human faces with direct gaze and three emotional expressions were used as stimulus images (Fig 1). Expressions were Threatening (10 aggressive dog faces; 10 angry human faces), Pleasant (10 positively valenced dog faces; 10 smiling human faces) and Neutral (10 expressionless dogs; 10 expressionless humans) (Table 1). The dog faces represented a total of 25 different breeds. Half of the human faces were images of women, half of men, balanced over expressions. Human portraits were acquired from royalty-free online stocks (e.g. MS ClipArt, 123RF®, BigStock). Canine photographs were acquired from authors’ and photographer Aino Pikkusaari’s collections, online image services (Google, Wikimedia Commons, Flickr) and royalty-free online stocks (123RF®, BigStock).
Table 1. Facial actions of Dog and Human images with three emotional expressions.
Dog faces | DogFACS1 | Human faces | FACS2 | |
---|---|---|---|---|
Threatening | Brows lowered and drawn together; lips parted3; upper lip raised; nose wrinkled; tongue shown3; ears forward, flattened or rotated backwards4 | AU4; AU25, AU109; AU110; AU193; EAD101, EAD103 or EAD1044 | Brows lowered and drawn together; upper eyelids raised; nasolabial deepened; lips parted3; upper lip raised; mouth stretched; tongue shown3 | AU4; AU5; AU11; AU25; AU10; AU27; AU193 |
Pleasant | Lip corners pulled upwards; lips parted; tongue shown; ears forward | AU12; AU25; AU19; EAD101 | Cheeks raised; lip corners pulled upwards; lips parted; | AU6; AU12; AU25 |
Neutral | Eyes, brows and cheeks relaxed; ears forward | AU0; EAD | Eyes, brows and cheeks relaxed | AU0 |
2 Human facial action coding system [69].
3 Teeth visible; five of the models were teeth together and five were teeth parted with tongue shown.
4 Three of the models were ears forward; three ears backward and four ears sideward.
The low-level image properties (luminance, kurtosis, skewness, standard deviation and root mean square contrast) were assessed using MatLab 7.12 (MathWorksInc, USA). In addition, arousal and valence of the stimuli were independently ranked afterwards by 22 adult humans (with varying expertise and experience about dogs), naïve regarding the study protocol (11 females, 11 males, aged 25–46). Mean valence ratings with 7-point scales (1 = negative, 4 = neutral, 7 = positive) were as follows (mean ± SD): Threatening dog 1.79 ± 0.5, Threatening human2.47 ± 0.6, Pleasant dog 4.99 ± 0.6, Pleasant human 5.57 ± 0.4, Neutral dog 3.82 ± 0.5, Neutral human 3.61 ± 0.2. Arousal ratings (1 = not arousing, 7 = highly arousing) for each stimulus category were the following: Threatening dog 5.65 ± 0.6, Threatening human 5.02 ± 0.7, Pleasant dog 3.25 ± 0.9, Pleasant human 4.25 ± 0.7, Neutral dog 2.78 ± 1.0, Neutral human 2.48 ± 0.8.
Equipment
The binocular eye movements of dogs were recorded with an infrared-based contact-free eye-tracker (iView X™ RED250, SensoMotoric Instruments GmbH, Germany), which was integrated below a 22” LCD monitor placed at an average distance of 0.70 m from the dogs’ eyes (SD 0.03 m, ranging between 0.61 and 0.73 m). The distance was adjusted individually for each dog for optimal detection of the eyes with the eye tracker. The monitor, the eye tracker and the chin rest were placed in a cardboard cabin (h = 1.5 m, w = 0.9 m, d = 0.9 m) with three walls and a roof.
The eye tracker was calibrated for each dog’s eyes using a five-point procedure (for details, see [49,50]. The calibrated area was the whole monitor, equal to a visual field of 40.5° x 24.4° from the average distance of 0.70 m. After a successful calibration, the calibration was saved and the accuracy of the calibration was checked twice. Dogs were free to change their position between the calibration and calibration check runs. The average calibration accuracy was 97% (SD 7%, ranging between 80 and 100%), scored as a proportion of fixated points out of five calibration points within 1° margin. Two of the 25 pet dogs were excluded from the experiment because of inadequate calibration accuracy (under 80%). To maximize the attentiveness and motivation of the dogs, the calibration and image viewing sessions were run on separate days. Illumination and the position of the chin rest, monitor, and eye tracker were kept the same during the calibration and the image viewing, and the accuracy of the central point fixations was re-assessed visually immediately before the image viewing. In our previous experiment, the calibrations of six dogs were tested four to six times, resulting in an average accuracy of 84% [49]. Thus, according to our previous findings, the stored calibration can be used repeatedly during separate days.
Procedure
The stimulus presentation and the eye gaze tracking were conducted using a similar protocol as described in [50]. Dogs viewed a block of 8–12 images (1500 ms per image with 500 ms inter-stimulus-interval) independently while the experimenters and the owners waited behind the opaque barrier (S1 Movie). The stimulus block was followed by a break when the dog was rewarded; in total, six stimulus blocks were recorded per single measurement session. The length of a stimulus block and the order of the stimuli were randomized in order to prevent anticipatory behavior. Two measurement sessions were carried out on separate testing days. The time delays between the testing days were 3–12 days (average 5.2 days, SD 2.2).
Data processing
A total of 3,479 stimulus images yielded sufficient gaze data (Threatening dog N = 583; Threatening human N = 575; Pleasant dog N = 574; Pleasant human N = 589; Neutral dog N = 592; Neutral human N = 566; S1 Dataset). A total of 241 images were excluded from analysis because of missing data lasting over 750 ms during a single stimulus presentation (interrupted eye tracking due to technical problems or a dog’s behavior). The raw gaze data were further processed using BeGaze 2.4™ software (SensoMotoric Instruments GmbH, Berlin, Germany) for defining of gaze fixations. The fixation was encoded if the minimum gaze duration was 75 ms and the maximum dispersion value D = 250 px [49].
Fixations were collected from five areas of interest (AOIs): eyes, midface (i.e. nose / muzzle), mouth (applied from [25,31]) and inner face (covering the combination of eyes, midface and mouth; applied from [11,18]) and the whole face (Fig 1). As the total viewing times vary for different pictures, and the size of the facial features varies for different depicted species, the looking duration was measured as a fixation score, which was normalized by taking into account both the relative size of the facial feature and the total face viewing duration (applied from [31,51,52]). For example, fixation score for the eye area was normalized by subtracting the relative AOI size (e.g. the size of the eyes divided by the size of the whole face) from the relative fixation duration (e.g. the total fixation time targeted to the eyes divided by the total fixation time targeted to the whole face area). Thus, if the value of fixation score is zero, fixations are distributed randomly. Positive fixation scores indicate that the AOI receives longer looking duration than would be expected according to the relative size of the area, and negative fixation scores indicate the opposite [51].
As a measure of speed of the attention, the entry time (i.e. the duration from stimulus onset to the first fixation targeted at certain AOI, ms) was calculated for all AOIs. In addition, the orientation of initial attention (i.e. the spatial distribution of the first fixations) was measured as the 1st fixation probability (%) (applied from [25,31]), which was calculated by counting the number of images where the first fixations hit the particular AOI (inner face, eyes, midface or mouth) and then dividing this value by the number of images where the first fixations hit the whole face area. Further, this proportion was normalized by subtracting the relative AOI size from it.
Statistical analysis
The statistical analysis was performed on the stimulus category means (S2 Dataset). Comparisons among expression conditions, depicted species and target facial features were conducted with a repeated measures linear mixed-effects model (MIXED) using a first order autoregressive (AR1) covariance structure. In the final model for analyzing the gazing behavior toward the inner face, the fixed effects were depicted species (dog or human), expression (Threatening, Pleasant or Neutral) and an interaction between depicted species and expression. In the final model for analyzing the fixation distribution among the facial features, the fixed effects were depicted species, expression, target facial feature (eyes, midface or mouth) and an interaction between species and facial feature. Other interactions were omitted as being statistically non-significant. The subject (i.e. the tested dog) and the testing day were included as random effects. The group (kennel dog or pet dog) and the gender of the subject were tested as random effects and omitted as being redundant for the models. In addition, the age of the subject, the calibration accuracy, low-level image properties and in the case of entry times, the actual AOI size and the AOI size in relation to the size of the whole face were tested as covariates, but omitted as being not statistically significant. The normality and homogeneity assumptions of the models were checked with normal probability plots of residuals against fitted values. The natural logarithm transformation was used for variables in which the residuals were not normally distributed and the values were back-converted for reporting. The covariance type of the linear mixed-effects models was selected by using the Akaike Information Criteria. Post hoc analyses with Bonferroni test, based on Student's t statistic, were included in the MIXED procedure. The alpha-level was set at 0.05, and the p-values are reported as adjusted values, considering values p < 0.05 as statistically significant (*) and p < 0.08 be a tendency (#). All the results are reported as estimated means with standard errors or means (Mean ± SD). All the statistical analyses were conducted with SPSS statistics 22.0 (IBM, New York, USA).
Results
Gazing behavior toward inner face area
Fixation score
The fixation scores toward the inner face area (i.e. the eyes, midface and mouth combined) did not differ significantly among the different facial expressions (main effect of the MIXED, p = 0.994). However, the interaction between the depicted species and expressions was statistically significant (p < 0.001). The post hoc tests revealed that dogs looked longer at the inner face of Threatening dogs compared with the inner face of Pleasant or Neutral dogs (Fig 2A), suggesting heightened attention toward the threatening signals of conspecifics. In contrast, when human faces were presented, the response was reversed (Fig 2A). Examples of the looking patterns of dogs are shown in Fig 1 and S2 Movie.
The fixation scores toward the inner face area differed significantly between the depicted species (main effect, p < 0.001; dog 0.124 ± 0.017, human 0.063 ± 0.017). Considering the significant interaction between the depicted species and expressions (p < 0.001), the post hoc tests revealed that dogs looked at inner faces of dogs longer compared with the inner faces of humans only in the Threatening expression (p < 0.001, Fig 2A).
1st fixation probability
The probability of the 1st fixation to land on the inner face area did not differ significantly among the facial expressions (main effect, p = 0.173). However, the interaction between the depicted species and expressions was statistically significant (p = 0.003). The post hoc tests revealed that the inner faces of Threatening humans received lower likelihood to be the targets of the first fixations compared with the inner faces of Pleasant or Neutral humans (Fig 2B).
The probability of the 1st fixation to land on the inner face area differed significantly between the depicted species (main effect, p = 0.008; dog 0.213 ± 0.023, human 0.149 ± 0.023). Considering the significant interaction between the depicted species and expressions (p = 0.003), the post hoc tests revealed that the inner faces of Threatening dogs received higher likelihood to be the targets of the first fixations compared with the inner faces of Threatening humans (p < 0.001, Fig 2B).
Entry time
The average entry time toward the inner face was 435.2 ms ± 22.0 ms with no statistically significant main effects for the expressions (p = 0.845) or the depicted species (p = 0.598). The interaction between the depicted species and expressions was not statistically significant (p = 0.893).
Gazing behavior toward eyes, midface and mouth
Fixation score
The fixation scores toward different target facial features differed significantly from each other (main effect, p < 0.001; eyes 0.069 ± 0.007, midface 0.031 ± 0.007 mouth -0.031 ± 0.007). In addition, the interaction among the depicted species, expressions and target facial features was statistically significant (p < 0.001). The post hoc tests revealed that dogs looked longer at the eye area compared with the mouth area irrespective of the depicted species or expression (Fig 3).
The fixation scores toward the target facial features did not differ significantly among the expressions (main effect, p = 0.940). However, the interaction among the depicted species, expressions and target facial features was statistically significant (p < 0.001). The post hoc tests revealed that dogs looked longer at the mouths of Threatening and Neutral dogs compared with the mouths of Pleasant dogs (Fig 3). In addition, dogs looked longer at the midfaces of Threatening dogs compared with the midfaces of Pleasant dogs (Fig 3).
The fixation scores toward the target facial features differed significantly between the depicted species (main effect, p = 0.003; dog 0.032 ± 0.006, human 0.014 ± 0.006). In addition, the interaction among the depicted species, expressions and target facial features was statistically significant (p < 0.001). The post hoc tests revealed that dogs looked longer at threatening dogs compared with threatening humans within all facial feature areas (within the eyes p < 0.001; midface p = 0.019; mouth p = 0.027, Fig 3). Further, dogs looked at the eyes of Pleasant dogs longer than at the eyes of Pleasant humans (p = 0.032, Fig 3) and the mouths of Pleasant dogs shorter than the mouths of Pleasant humans (p = 0.030, Fig 3).
1st fixation probability
The 1st fixation probabilities differed significantly among the target facial features (main effect, p < 0.001; eyes 0.092 ± 0.009, midface 0.056 ± 0.009, mouth -0.023 ± 0.009). The post hoc test revealed that eyes received higher likelihood to be targets of the first fixations compared with the midface (p = 0.016) or the mouth areas (p < 0.001) and the midface received higher likelihood compared with the mouth (p < 0.001). Furthermore, the depicted species differed significantly from each other in the 1st fixation probabilities (main effect, p = 0.017; dog 0.051 ± 0.007, human 0.030 ± 0.008). The main effect for the expressions was not statistically significant (p = 0.312), neither was the interaction among the depicted species, expressions and target facial features (p = 0.069). Examples of the first fixation targets are shown in Fig 4.
Entry time
The gaze entry times to the different target facial features differed significantly (main effect, p = 0.002; eyes 503.4 ms ± 19.3 ms, midface 501.0 ms ± 20.3 ms, mouth 583.9 ms ± 24.2 ms). The post hoc test revealed that dogs fixated at the eyes and the midface areas faster than the mouth area (eyes vs. mouth p = 0.005 and midface vs. mouth p = 0.005). The main effects for the expression or the depicted species were not statistically significant (p = 0.576 and p = 0.628, respectively), neither was the interaction among the depicted species, expressions and target facial features (p = 0.966).
Discussion
In this study, we broadened the scope of facial expression research to include the gazing behavior of domestic dogs (Canis familiaris). The present evidence confirms previous behavioral observations suggesting that dogs differentiate among depicted facial expressions [44–46] and adjust their viewing behavior spontaneously according to the observed emotion [46]. To our knowledge, this is the first evidence of emotion-related gaze patterns in non-primates. Despite primates and canids being distant in the phylogenetic tree, and their facial expressions having species-specific characteristics, the gazing behavior of dogs resembles the visual strategies of humans: their fixations are distributed among facial features systematically and their scanning behavior is modulated by facial expressions, especially that of threat. These phenomena may be based on phylogenetically old mechanisms.
Dogs scanned the facial features of conspecifics and humans in a similar manner from the initial attention: they fixated faster the eyes and midface areas than the mouth areas, and eyes received the higher probability to be the targets of the first fixations than the midfaces and the mouths. Likewise, humans, chimpanzees and macaque monkeys typically direct their first gaze preferentially on eyes [25,26,28,31,53,54] or just below them [23,31,55,56]. The initial focus is probably guided by a pre-attentive mechanism, which maximizes perceptual encoding performance [28,56].
Generally, dogs gazed at eyes overall longer than mouth areas regardless of the depicted species. Correspondingly, humans [24,26,28–31,57], chimpanzees [25,52,57], gorillas (Gorilla gorilla), orangutans (Pongo abelii) [57] and macaques [10,27,51,53,54] exhibit such fixation distribution among facial features (eyes > midface > mouth), even during viewing other species’ faces [10,25,31,53,57], suggesting that face-like configurations trigger systematic scanning behavior automatically [31]. The present results confirm the previous assumption that eyes play a significant role in the face perception of dogs [50]. In humans the intenseness to look at eyes may have developed to serve language-like functions, as joint attentional and communicative interactions, which requires collecting eye information actively [25]. Taking into account that domestic dogs resemble humans in their gaze communication (for reviews, see [34,35]), the analogous explanation may also apply to dogs.
Dogs’ gaze fixation distribution among the facial features was altered by the facial expression. In humans, expression-specific conspicuous visual properties (i.e. contrast, luminance, shape etc.) facilitate orienting toward particular facial structures [23]. However, physical saliency does not solely explain the locations of fixations [24], but attention is rather focused on the cues that are most relevant for encoding the certain emotion [24,26,28,29]. For example, humans view the mouths of pleasant faces most, probably because the mouth area is a key element for the detection and recognition of the human smile [23,24,29]. Dogs instead looked longer at the mouths of threatening and neutral dogs compared with the mouths of pleasant dogs. Likewise, chimpanzees [25] and rhesus monkeys [27] show heightened attention toward threatening mouths compared to pleasant mouths. For these animals the mouth area may be highly important for the interpretation of threatening expressions. Also the midface of threatening conspecifics’ seems to be salient for dogs. Instead macaques attend the midface area most in pleasant (lipsmack) monkey faces [27], and humans in disgust and sad human faces [29]. Although these expressions represent different emotions, they all include distinctive wrinkled muzzle/nose gestures, which may attract attention. The fixations of dogs could accumulate in midface area also because dogs might covertly observe the eyes of potentially threatening individuals by fixating near to eyes, but not directly on them. However, dogs did not actually avoid the eyes of threatening conspecifics. Instead, during exposure to human faces dogs tended to look at angry eyes less than neutral eyes, which could be due to the eye contact aversion, as in humans [18]. Overall, the features which are characteristic of species specific expressions appear to attract attention. Despite the variation in fixation distributions among the expressions, dogs keep the main focus on the eyes irrespective of the seen expression, corresponding to gazing behavior of humans [24,26,28–30].
The examination of the inner facial features as a whole revealed that the threatening faces evoked unique gazing patterns compared with the pleasant or neutral faces, referring to distinct visuo-cognitive processing strategies for threat-related facial expressions [11]. The result suggests that dogs do not base their perception of facial expressions on viewing of single structures, but the interpretation of the composition formed by eyes, midface and mouth. Dogs may process facial expressions holistically, at least threatening expressions. Further research is needed to clarify what perceptional strategy (purely visual saliency, featural or holistic [23,24,28,29]) dogs use in the recognition of facial expressions.
Notably, the attentional bias was dependent on the depicted species: dogs looked at the threatening dog faces for longer than at the other expressions of conspecifics, but scanned the threatening human faces in a contrary way, apparently avoiding looking at them. In primates including humans, threatening stimuli commonly enhance attention toward threatening conspecifics, which has been documented as accelerated initial orientation toward threat [8,9] or overall prolonged attention [9,11,12,16] which is the consequence of delay in disengaging the attention away from threat [13–15,58]. In dogs, we did not find evidence of faster orientation toward threatening faces: in terms of the entry time, dogs fixated on all expressions equally quickly. Instead, the sustained attention may arise from great interest toward threatening conspecific faces (interest guides dogs’ gaze fixations; [49,50]), or difficulties in disengaging the attention away from threat [13,58]. Such “stuck attention” has been suggested to be linked to a fear-induced freezing behavior, an evolutionarily important mechanism controlled by the amygdala [21,59]. The amygdala guides attention to emotionally significant stimuli, such as potentially harmful external events. The emotional response is further modulated by the prefrontal cortex, which enables more flexible behavioral outcomes. In arousing situations, the amygdala overrides prefrontal modulation and leads to failure to disengage attention away from the threatening stimuli (for reviews, see [21,22,59]). Arousal and fear relevance tend to be positively correlated with heightened attention [16,20,27,58]. Thus, it is possible that the images of aggressive dogs held attention because dogs find them more arousing than the images of angry humans. However, without physiological measurements we are not able to draw any conclusions about the subjective emotional state of dogs.
Dogs appeared to deem depicted angry human faces aversive, as has been proposed by the behavioral observations [45]. The avoidance of threatening stimuli is typical of anxious humans (for reviews, see [20,22]) but has also been observed in healthy humans [17,18] and monkeys during stressful emotional states [9]. Dogs avert their gaze from threatening human faces as early as the first fixations, which suggests that dogs can recognize the valence of the expression by their covert attention [60] even before the first fixations. Rapid gaze avoidance has been linked to submissive behavior [19]. Considering dogs’ flexible inter-species social communication [34,35] and their sensitivity to human threat signals [36,61], the gaze aversion from the threatening human faces could be a manifestation of appeasing behavior. This may be a learned mannerism, as in humans [18], or domestication may have equipped dogs with a propensity to display conflict-avoiding signals toward humans. In general, dogs show more common and severe aggressive reactions toward other dogs than people [62–63]. Selection for reduced human-directed aggression has likely shaped dogs to be highly tolerant to humans [64]. Flexible utilization of conflict-resolving behaviors may be a prominent element of domestic dogs’ social competence, which has enabled them to form peaceful mixed-species groups with humans [61,64].
Differential scanning of own-species’ and other-species’ threat signals may arise also from biological relevance of the stimuli [15,65,66]. Dog and human faces are both social and likely also emotional stimuli for dogs, but faces of conspecifics may carry more biological significance for them. Biological emotional stimuli (e.g. object relevant to survival or reproduction) and social emotional stimuli (e.g. object relevant to social life) are perhaps processed in different pathways and thus they alter attention contrarily [65–67]. In humans, biologically relevant images are detected rapidly and they facilitate attention engagement because of the strong amygdala activity and interconnection between the amygdala and the visual cortex. Instead, socially emotional images evoke greater activity in the prefrontal cortex-based circuit [59,65,67]. We propose that in the present study the initial attention was driven by the earlier limbic pathways during the viewing of both canine and human faces. Further, the biologically more relevant threatening conspecific faces were mainly processed via this route, which led to heightened attention. Threatening human faces received more complex prefrontal processing instead, which led to the avoidance response. The enhanced prefrontal modulation could be linked to submissive behavior [19] and/or memory-based social evaluations [66].
The subjects’ current emotional state (e.g. stress) and personality (e.g. anxiousness and neuroticism) [9,20,22,59], and also the stimulus types and paradigm may have affected the results in our study. In future studies the interaction among phylogenetic, biological and personal validity of the stimuli should be taken into account. For example, do dogs exhibit attentional bias toward biological stimuli that evoke a phylogenetic fear response (e.g. snakes) versus non-biological stimuli, which evoke an ontogenetic fear response (e.g. nail clippers; for examples of paradigms, see [15,58,65–67])? The physical saliency of the stimuli should be controlled by using saliency models [68] and stimuli that are not socially relevant, but morphologically similar to the dog faces (e.g. other animal faces or artificial stimuli). In addition, the investigation should be broadened to other emotionally informative facial structures, including their movements, for example ears [27,42].
In conclusion, during viewing emotional faces of conspecifics and humans, domestic dogs’ gaze fixations spread systematically among facial features: regardless of the viewed expression, eyes were the most probable targets of the first fixations and gathered overall more attention than mouth area. The primacy of eyes suggests that eyes are physically salient and/or highly informative facial feature for dogs. However, dogs did not base their perception of facial expressions on viewing single facial structures, but the interpretation of the composition formed by eyes, midface and mouth. They evaluated social threat signals rapidly, and this evaluation led to attentional bias, which emerged as enhanced attention or avoidance, depending on the biological validity of the stimuli. We propose that the biologically more relevant threatening conspecific faces are processed mainly by the earlier limbic pathways, which results in enhanced attention. In contrast, the avoidance response induced by the threatening signals of humans may be more modulated by the prefrontal cortex. These two mechanisms may both have an adaptive significance for domestic dogs. The tolerant behavior strategy of domestic dogs toward humans may partially explain the results. The present findings bring a new perspective to understanding the processing of emotional expressions in non-primate animals. Exploring further the sensitivity to social threat in canids provides a novel comparative approach to unravel the etiology of neurocognitive-affective disorders that exist in both humans and dogs, such as social phobias and anxiety.
Supporting Information
Acknowledgments
We thank Aino Pikkusaari for wonderful dog photographs, Reeta Törne for assisting in the experiments and Aija Koskela for the data preparation. Special thanks go to all pet dog owners for dedication to the training of their dogs.
Data Availability
All relevant data are within the paper and its Supporting Information files.
Funding Statement
Academy of Finland (http://www.aka.fi) project #137931 to OV; Emil Aaltonen Foundation (http://www.emilaaltonen.fi/) to SS; The Brain Research at Aalto University and University of Helsinki BRAHE consortium (http://www.brain-mind.fi) to MVK. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Ekman P. Emotion in the human face New York: Cambridge University Press; 1982. [Google Scholar]
- 2.Matsumoto D, Keltner D, Shiota MN, O’Sullivan M, Frank M. Facial expressions of emotion In: Lewis M, Haviland-Jones J, Feldman Barret L, editors. Handbook of Emotion. New York: The Guilford Press; 2008. pp. 211–234. [Google Scholar]
- 3.Diogo R,Wood BA, Aziz MA, Burrows A. On the origin, homologies and evolution of primate facial muscles, with a particular focus on hominoids and a suggested unifying nomenclature for the facial muscles of the mammalia. J Anat 2009;215: 300–319. 10.1111/j.1469-7580.2009.01111.x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Waller BM, Micheletta J. Facial expression in nonhuman animals. Emot Rev 2013;1: 54–59. [Google Scholar]
- 5.Darwin CR. The expression of the emotions in man and animals. London: John Murray; 1872. [Google Scholar]
- 6.Tate AJ, Fischer H, Leigh AR, Kendrick KM. Behavioural and neurophysiological evidence for face identity and face emotion processing in animals. Philos Trans R Soc Lond B Biol Sci. 2006;361: 2155–2172. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Brecht M, Freiwald WA. The many facets of facial interactions in mammals. Curr Opin Neurobiol 2012;22: 259–266. 10.1016/j.conb.2011.12.003 [DOI] [PubMed] [Google Scholar]
- 8.Fox E, Lester V, Russo R, Bowles RJ, Pichler A, Dutton K. Facial expressions of emotion: Are angry faces detected more efficiently? Cogn Emot 2000;14: 61–92. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Bethell EJ, Holmes A, MacLarnon A, Semple S. Evidence that emotion mediates social attention in rhesus macaques. PLoS ONE 2012;7(8): e44387 10.1371/journal.pone.0044387 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Nahm FKD, Perret A, Amaral DG, Albright TD. How do monkeys look at faces? J Cogn Neurosci 1997;9: 611–623. 10.1162/jocn.1997.9.5.611 [DOI] [PubMed] [Google Scholar]
- 11.Green MJ, Williams LM, Davidson D. In the face of danger: Specific viewing strategies for facial expressions of threat? Cogn Emot 2003;17: 779–786. [Google Scholar]
- 12.Kano F, Tomonaga M. Attention to emotional scenes including whole-body expressions in chimpanzees (Pan troglodytes). J of Comp Psychol 2010;124: 287–294. [DOI] [PubMed] [Google Scholar]
- 13.Belopolsky AV, Devue C, Theeuwes J. Angry faces hold the eyes. Vis Cogn 2011;19: 27–36. [Google Scholar]
- 14.King HM, Kurdziel LB, Meyer JS, Lacreuse A. Effects of testosterone on attention and memory for emotional stimuli in male rhesus monkeys. Psychoneuroendocrinology 2012;37: 396–409. 10.1016/j.psyneuen.2011.07.010 [DOI] [PubMed] [Google Scholar]
- 15.Lacreuse A, Schatz K, Strazzullo S, King HM, Ready R. Attentional biases and memory for emotional stimuli in men and male rhesus monkeys. Anm Cogn 2013;16: 861–871. [DOI] [PubMed] [Google Scholar]
- 16.Kret ME, Roelofs K, Stekelenburg JJ, de Gelder B. Emotional signals from faces, bodies and scenes influence observers' face expressions, fixations and pupil-size. Front Human Neurosci 2013;7: 810 10.3389/fnhum.2013.00810 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Becker MW, Detweiler-Bedell B. Early detection and avoidance of threatening faces during passive viewing. Q J Exp Psychol 2009;62: 1257–1264. [DOI] [PubMed] [Google Scholar]
- 18.Hunnius S, de Wit TCJ, Vrins S, von Hofsten C. Facing threat: Infants' and adults' visual scanning of faces with neutral, happy, sad, angry, and fearful emotional expressions. Cogn Emot 2011;25: 193–205. 10.1080/15298861003771189 [DOI] [PubMed] [Google Scholar]
- 19.Terburg D, Aarts H, van Honk J. Memory and attention for social threat: Anxious hypercoding-avoidance and submissive gaze aversion. Emotion 2012;12: 666–672. 10.1037/a0027201 [DOI] [PubMed] [Google Scholar]
- 20.Mogg K, Bradley PB. A cognitive-motivational analysis of anxiety. Behav Res Ther 1998;36: 809–848. [DOI] [PubMed] [Google Scholar]
- 21.Öhman A, Mineka S. Fears, phobias, and preparedness: Toward an evolved module of fear and fear learning. Psychol Rev 2001;108: 483–522 [DOI] [PubMed] [Google Scholar]
- 22.Armstrong T, Olatunji B. Eye tracking of attention in the affective disorders: A meta-analytic review and synthesis. Clin Psychol Rev 2012;32: 704–723. 10.1016/j.cpr.2012.09.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Calvo MG, Nummenmaa L (2008) Detection of emotional faces: Salient physical features guide effective visual search. Journal Exp Psychol Gen 137: 471–494. [DOI] [PubMed] [Google Scholar]
- 24.Beaudry O, Roy-Charland A, Perron M, Cormier I, Tapp R. Featural processing in recognition of emotional facial expressions. Cogn Emot 2014;28: 416–432. 10.1080/02699931.2013.833500 [DOI] [PubMed] [Google Scholar]
- 25.Kano F, Tomonaga M. Face scanning in chimpanzees and humans: Continuity and discontinuity. Anim Behav 2010;79: 227–235. [Google Scholar]
- 26.Eisenbarth H, Alpers GW. Happy mouth and sad eyes: Scanning emotional facial expressions. Emotion 2011;11: 860–865. 10.1037/a0022758 [DOI] [PubMed] [Google Scholar]
- 27.Mosher CP, Zimmerman PE, Gothard KM. Videos of conspecifics elicit interactive looking patterns and facial expressions in monkeys. Behav Neurosci. 2011;125: 639–652. 10.1037/a0024264 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Scheller E, Büchel C, Gamer M. Diagnostic features of emotional expressions are processed preferentially. PloS ONE 2012;7(7): e41792 10.1371/journal.pone.0041792 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Guo K. Holistic gaze strategy to categorize facial expression of varying intensities. PLoS ONE 2012;7(8): e42585 10.1371/journal.pone.0042585 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Cangoz B, Altun A, Askar P, Baran Z, Mazman SG. Examining the visual screening patterns of emotional facial expressions with gender, age and lateralization. J Eye Mov Res 2013;6(4):3, 1–15. [Google Scholar]
- 31.Guo K, Tunnicliffe D, Roebuck H. Human spontaneous gaze patterns in viewing of faces of different species. Perception 2010;39: 533–542 [DOI] [PubMed] [Google Scholar]
- 32.Bolwig N. Facial expression in primates with remarks on a parallel development in certain carnivores (a preliminary report on work in progress). Behaviour 1964;22: 167–193. [Google Scholar]
- 33.Fox MW. A comparative study of the development of facial expressions in canids: Wolf, coyote and foxes. Behaviour 1970;36: 49–73. [Google Scholar]
- 34.Miklósi A, Topál J. What does it take to become 'best friends'? Evolutionary changes in canine social competence. Trends Cogn Sci (Regul Ed) 2013;17: 287–294 [DOI] [PubMed] [Google Scholar]
- 35.Kaminski J, Nitzschner M. Do dogs get the point? A review of dog-human communication ability. Learn Motiv 2013;44: 294–302. [Google Scholar]
- 36.Vas J, Topál J, Gácsi M, Miklósi A, Csányi V. A friend or an enemy? Dogs' reaction to an unfamiliar person showing behavioural cues of threat and friendliness at different times. Appl Anim Behav Sci 2005; 94: 99–115. [Google Scholar]
- 37.Merola I, Prato-Previde E, Marshall-Pescini S. Dogs' social referencing towards owners and strangers. PLoS ONE 2012;7(10): e47653 10.1371/journal.pone.0047653 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Custance D, Mayer J. Empathic-like responding by domestic dogs (Canis familiaris) to distress in humans: An exploratory study. Anim Cogn 2012;15: 851–859. 10.1007/s10071-012-0510-1 [DOI] [PubMed] [Google Scholar]
- 39.Buttelmann D, Tomasello M. Can domestic dogs (Canis familiaris) use referential emotional expressions to locate hidden food? Anim Cogn 2013;16: 137–145. 10.1007/s10071-012-0560-4 [DOI] [PubMed] [Google Scholar]
- 40.Schirmer A, Seow CS, Penney TB. Humans process dog and human facial affect in similar ways. PLoS ONE 2013;8(9): e74591 10.1371/journal.pone.0074591 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Bloom T, Friedman H. Classifying dogs’ (Canis familiaris) facial expressions from photographs. Behav Processes 2013;96: 1–10. 10.1016/j.beproc.2013.02.010 [DOI] [PubMed] [Google Scholar]
- 42.Nagasawa M, Kawai M, Mogi K, Kikusui T. Dogs show left facial lateralization upon reunion with their owners. Behav Processes 2013;98: 112–116. 10.1016/j.beproc.2013.05.012 [DOI] [PubMed] [Google Scholar]
- 43.Waller BM, Peirce K, Caeiro CC, Scheider L, Burrows AM, McCune S et al. Paedomorphic facial expressions give dogs a selective advantage. PLoS ONE 2013;8(12): e82686 10.1371/journal.pone.0082686 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Nagasawa M, Murai K, Mogi K, Kikusui T. Dogs can discriminate human smiling faces from blank expressions. Anim Cogn 2011;14: 525–533. 10.1007/s10071-011-0386-5 [DOI] [PubMed] [Google Scholar]
- 45.Müller CC, Schmitt K, Barber ALA, Huber L. Dogs can discriminate emotional expressions of human faces. Curr Biol. 2015; Available: 10.1016/j.cub.2014.12.055 [DOI] [PubMed] [Google Scholar]
- 46.Racca A, Guo K, Meints K, Mills DS. Reading faces: Differential lateral gaze bias in processing canine and human facial expressions in dogs and 4-year-old children. PLoS ONE 2012;7(4): e36076 10.1371/journal.pone.0036076 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Williams FJ, Mills DS, Guo K. Development of a head-mounted eye-tracking system for dogs. J Neurosci Methods, 2011;194: 259–265. 10.1016/j.jneumeth.2010.10.022 [DOI] [PubMed] [Google Scholar]
- 48.Téglás E, Gergely A, Kupán K, Miklósi A, Topál J. Dogs' gaze following is tuned to human communicative signals. Curr Biol 2012;22: 209–212. 10.1016/j.cub.2011.12.018 [DOI] [PubMed] [Google Scholar]
- 49.Somppi S, Törnqvist H, Hänninen L, Krause C, Vainio O. Dogs do look at images: Eye tracking in canine cognition research. Anim Cogn 2012;15: 163–174 10.1007/s10071-011-0442-1 [DOI] [PubMed] [Google Scholar]
- 50.Somppi S, Törnqvist H, Hänninen L, Krause C, Vainio O. How dogs scan familiar and inverted faces: An eye movement study. Anim Cogn 2014;17: 793–803. 10.1007/s10071-013-0713-0 [DOI] [PubMed] [Google Scholar]
- 51.Dahl CD, Wallraven C, Bülthoff HH, Logothetis NK. Humans and macaques employ similar face-processing strategies. Curr Biol 2009;19: 509–513. 10.1016/j.cub.2009.01.061 [DOI] [PubMed] [Google Scholar]
- 52.Hirata S, Fuwa K, Sugama K, Kusunoki K, Fujita S. Facial perception of conspecifics: Chimpanzees (Pan troglodytes) preferentially attend to proper orientation and open eyes. Animal Cognition. 2010;13: 679–688. 10.1007/s10071-010-0316-y [DOI] [PubMed] [Google Scholar]
- 53.Guo K, Robertson RG, Mahmoodi S, Tadmor Y, Young MP. How do monkeys view faces?—a study of eye movements. Exp Brain Res 2003;150: 363–374. [DOI] [PubMed] [Google Scholar]
- 54.Gothard KM, Erickson CA, Amaral DG. How do rhesus monkeys (Macaca mulatta) scan faces in a visual paired comparison task? Anim Cogn 2004;7: 25–36. [DOI] [PubMed] [Google Scholar]
- 55.Hsiao JH, Gottrell G. Two fixations suffice in face recognition. Psychol Sci 2008;19: 998–1006. 10.1111/j.1467-9280.2008.02191.x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Peterson MF, Eckstein MP. Looking just below the eyes is optimal across face recognition tasks. Proc Natl Acad Sci USA 2012;109: E3314–E3323. 10.1073/pnas.1214269109 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Kano F, Call J, Tomonaga M. Face and eye scanning in gorillas (Gorilla gorilla), orangutans (Pongo abelii), and humans (Homo sapiens): Unique eye-viewing patterns in humans among hominids. J Comp Psychol. 2012;126: 388–398. 10.1037/a0029615 [DOI] [PubMed] [Google Scholar]
- 58.Brosch T, Sharma D. The role of fear-relevant stimuli in visual search: A comparison of phylogenetic and ontogenetic stimuli. Emotion 2005;5: 360–364. [DOI] [PubMed] [Google Scholar]
- 59.Kim MJ, Loucks RA, Palmer AL, Brown AC, Solomon KM, Marchante AN et al. The structural and functional connectivity of the amygdala: From normal emotion to pathological anxiety. Behav Brain Res 2011;223: 403–410. 10.1016/j.bbr.2011.04.025 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Moore T, Armstrong KM, Fallah M. Visuomotor origins of covert spatial attention. Neuron 2003;40: 671–683. [DOI] [PubMed] [Google Scholar]
- 61.Gyori B, Gácsi M, Miklósi A. Friend or foe: Context dependent sensitivity to human behaviour in dogs. Appl Anim Behav Sci 2010;128: 69–77. [Google Scholar]
- 62.Duffy DL, Hsu A, Serpell JA. Breed differences in canine aggression. Appl Anim Behav Sci 2008;114: 441–460. [Google Scholar]
- 63.Casey RA, Loftus B, Bolster C, Richards GJ, Blackwell EJ. Human directed aggression in domestic dogs (Canis familiaris): Occurrence in different contexts and risk factors. Appl Anim Behav Sci 2013;152: 52–63. [Google Scholar]
- 64.Virányi Z, Range F. On the way to a better understanding of dog domestication: Aggression and cooperativeness in dogs and wolves In: The Social Dog. Behavior and Cognition. eds. Kaminski J, Marshall-Pescini. 2014. pp. 35–62. [Google Scholar]
- 65.Sakaki M, Niki M, Mather M. Beyond arousal and valence: The importance of the biological versus social relevance of emotional stimuli. Cogn Affect Behav Ne 2012;12: 115–139. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Cao Z, Zhao Y, Tan T, Chen G, Ning X, Zhan L et al. Distinct brain activity in processing negative pictures of animals and objects—the role of human contexts. Neuroimage 2014;84: 901–910. 10.1016/j.neuroimage.2013.09.064 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Yang J, Bellgowan PS, Martin A. Threat, domain-specificity and the human amygdala. Neuropsychologia 2012;50: 2566–2572. 10.1016/j.neuropsychologia.2012.07.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Itti L, Koch C. A saliency-based search mechanism for overt and covert shifts of visual attention. Vision Res. 2000; 40: 1489–1506. [DOI] [PubMed] [Google Scholar]
- 69.Ekman P, Friesen WV, Hager JC, editors. Facial action coding system. Salt Lake City: Research Nexus; 2002. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
All relevant data are within the paper and its Supporting Information files.