Abstract
Background
This study examines how dogs observe images of natural scenes containing living creatures (wild animals, dogs and humans) recorded with eye gaze tracking. Because dogs have had limited exposure to wild animals in their lives, we also consider the natural novelty of the wild animal images for the dogs.
Methods
The eye gaze of dogs was recorded while they viewed natural images containing dogs, humans, and wild animals. Three categories of images were used: naturalistic landscape images containing single humans or animals, full body images containing a single human or an animal, and full body images containing a pair of humans or animals. The gazing behavior of two dog populations, family and kennel dogs, were compared.
Results
As a main effect, dogs gazed at living creatures (object areas) longer than the background areas of the images; heads longer than bodies; heads longer than background areas; and bodies longer than background areas. Dogs gazed less at the object areas vs. the background in landscape images than in the other image categories. Both dog groups also gazed wild animal heads longer than human or dog heads in the images. When viewing single animal and human images, family dogs focused their gaze very prominently on the head areas, but in images containing a pair of animals or humans, they gazed more at the body than the head areas. In kennel dogs, the difference in gazing times of the head and body areas within single or paired images failed to reach significance.
Discussion
Dogs focused their gaze on living creatures in all image categories, also detecting them in the natural landscape images. Generally, they also gazed at the biologically informative areas of the images, such as the head, which supports the importance of the head/face area for dogs in obtaining social information. The natural novelty of the species represented in the images as well as the image category affected the gazing behavior of dogs. Furthermore, differences in the gazing strategy between family and kennel dogs was obtained, suggesting an influence of different social living environments and life experiences.
Keywords: Scene, Family dogs, Kennel dogs, Animal images, Human images, Biological information, Cognition, Eye gaze tracking, Full body images
Introduction
Visual attention, often measured as ‘looking preferences’, is commonly used to assess animal behavior and cognition (Winters, Dubuc & Higham, 2015). Eye gaze tracking provides more quantifiable and specific information, and during the past decade, it has been successfully applied when investigating dogs’ responses towards still images. Previously, dogs have been found to spontaneously discriminate human and dog faces from each other and from inanimate objects (Racca et al., 2010; Somppi et al., 2012; Gergely et al., 2019) and even facial expressions of humans and dogs (Nagasawa et al., 2011; Somppi et al., 2016; Somppi et al., 2017). Dogs can learn to differentiate facial images of dogs from the facial images of other animal species (Autier-Dérian et al., 2013) and to discriminate between naturalistic images that contain a dog from those that do not (Range et al., 2008). However, most of the studies using looking-time paradigms have used facial images and unlike primate studies (Kano & Tomonaga, 2009; Kano, Call & Tomonaga, 2012) little canine research has used naturalistic whole-body images of humans and animals (Törnqvist et al., 2015), or different kinds of images (e.g., landscapes and faces) embedded in the same study design (Range et al., 2008; Autier-Dérian et al., 2013). Extending the research of canine gazing behavior to more naturalistic images provides a unique opportunity to study both the effect of image composition and object or category content at the same time.
The capacity to respond appropriately to biologically relevant stimuli (prey, predators and social partners) is crucial for animal survival, and it has been suggested to be an unlearned social predisposition (for reviews, Rosa-Salva, Mayer & Vallortigara, 2015; Di Giorgio et al., 2017). The primate visual system is specialized in detecting living creatures from complex natural scenes (Fabre-Thorpe, Richard & Thorpe, 1998; Thorpe et al., 2001), and typically humans and monkeys focus their gaze specifically on the head area of these creatures (e.g., Yarbus, 1967; Kano & Tomonaga, 2009; Kano, Call & Tomonaga, 2012). Biologically relevant stimuli such as animals attract attention more than inanimate objects and social information draws more attention than non-social (e.g., in non-human primates: Machado et al., 2011; Hanazuka et al., 2012; in humans: New, Cosmides & Tooby, 2007; Simion, Regolin & Bulf, 2008; in dogs: Somppi et al., 2012; Törnqvist et al., 2015). Dogs’ spontaneous preferences toward social information has also been recently demonstrated with biological motion videos (Kovács et al., 2016; Ishikawa et al., 2018).
The amount of visual experience a person has in a particular domain can promote an expert cognitive strategy in processing visual information (e.g., Diamond & Carey, 1986; Kundel et al., 2008; Kujala et al., 2012). For example, when viewing images of two dogs interacting with or facing away from one another, dog experts gaze relatively longer at the dogs’ bodies than at their heads when compared with non-experts (Kujala et al., 2012). Dog owners assess the approachability of dogs by assessing their facial expression more rapidly and with fewer fixations than non-dog owners, but when assessing human and monkey faces, both groups have similar viewing behavior (Gavin, Houghton & Guo, 2017). Also human infants, who have previous experience with pets, focus their attention more on the head areas in cat and dog images than infants without pet experience (Kovack-Lesh, McMurray & Oakes, 2014). When humans and non-human primates observe social videos of other animals demonstrating their species-typical behaviors, the distinct viewing patterns exhibited by these observers are influenced by their previous experiences of that species (Kano et al., 2018). Likewise, the amount of previous exposure also affects dogs’ gazing behavior during the presentation of human and dog faces (Somppi et al., 2014; Barber et al., 2016) and natural images of social interaction (Törnqvist et al., 2015).
In the present study, we measured the gazing behavior of family and kennel dogs while they viewed natural images of dogs, humans and wild exotic animals. Three categories of images were used: landscape images containing single humans or animals; full body images containing a human or an animal; and full body images containing a pair of humans or animals of the same species. We hypothesized that when dogs viewed naturalistic images, dogs would focus more attention on living creatures—heads and bodies of animals and humans—even when the figure was small in relation to the image. In addition, we assumed that the composition of the images would affect dogs’ gazing behavior. We also anticipated that dogs would gaze at the heads and bodies of living creatures, even when viewing unfamiliar species of animals; and that dogs would focus more on their conspecifics than non-conspecifics. Furthermore, we assumed that family and kennel dogs’ social experiences would affect the way in which they viewed biological information.
Materials & Methods
Ethical statement
This study was conducted in strict accordance with the Finnish Act of Animal Experimentation (62/2006) and with fully implemented European convention for the protection of vertebrate animals used for experimental and other scientific purposes (Directive 86/609/EEC). The study was approved by the Ethical Committee for the Use of Animals in Experiments at the University of Helsinki (2/2010) and the Finnish national Animal Experiment Board (approval#STH367A/ESLH-2008-04236/Ym-23). Dogs were not harmed in any way and their movements were not restricted during the training or the study. Only non-invasive methods were used during the measurements.
Subjects
A total of 16 privately owned family dogs and eight kennel dogs participated in the study. Family dogs (11 females and five males) represented nine different breeds and two mongrels. They were 1–8 years old [4.3 ± 2.2 years (mean ± standard deviation)] and lived with their owners. Their daily routine consisted of food provision once or twice a day and being taken outdoors three to five times daily. Family dogs regularly saw dogs and other animal species (cats, birds etc.) in their living environment. They lived in the homes of their owners and constantly interacted with their human families. Family dogs did not have any specific behavioral disturbances (e.g., anxiety or aggression). Kennel dogs (two females and six males) were purpose-bred beagles that lived together as a social group in a kennel-like environment at the facilities of University of Helsinki. During the experiment, the kennel dogs were 4 years old. They were given food twice a day and were taken to an outside exercise area for 2 h once a day. Kennel dogs did not meet dogs of other breeds or any other animal species, but they regularly saw other beagle dogs living in the same building. They were rarely taken outside the kennel and seldom met other humans except the caretakers and the researchers with whom they were familiar. Compared to family dogs, kennel dogs were quite fearful and cautious in training and experimental situations. All the dogs had participated in eye tracking experiments prior to the present experiment, but they were naïve to stimuli image content.
Training of the dogs
Before the experiment, dogs were trained to lie still and lean their heads on a U-shaped chin rest during the presentation of the images, as in our previous experiments (e.g., Somppi et al., 2012; Törnqvist et al., 2015). Dogs were trained with the positive operant conditioning method (clicker). Owners trained their family dogs as instructed by the experimenters and kennel dogs were trained by the experimenters. Dogs were neither encouraged to fixate on a screen or images during the training nor trained for visual discrimination tasks. The dogs were not physically restrained, and they took the pre-trained position of their own volition without commands (for more details, see Somppi et al., 2012).
Eye gaze tracking and calibration
The eye gaze measurements were conducted at the Faculty of Veterinary Medicine at the University of Helsinki. Dogs’ binocular eye movements were measured with an infrared contact-free eye tracker (iView XTM RED250, SensoMotoric Instruments GmbH, Berlin, Germany) integrated under liquid-crystal display (LCD). The data were collected with a sampling rate of 250 Hz. The distance between dog’s eyes and the screen was 0.61–0.73 m, and the distance depended on the size of the dog (smaller dogs were closer to the screen). The screen, eye tracker and the chin rest were placed in a cardboard cabin (height = 1.5 m, width = 0.9 m, depth = 0.9 m) that had three walls and a roof (see Somppi et al., 2012) for a schematic picture of the setup). In front and above the screen were fluorescent lamps, and the intensity of illumination was 4,200–13,400 lx (11,000 ± 2,300 lx) measured on top of the chin rest. Calibration of the eye tracker for each dog’s eyes was done using a five-point procedure (see Somppi et al., 2012). The screen was replaced with a plywood wall with five 30-mm holes in the calibration point positions, and the experimenter lifted up a flap covering the hole and showed a treat in the hole to catch the dog’s attention. Two additional calibration check trials were conducted after the calibration for verification. Calibration were repeated one –13 times (4 ± 0.6) for each dog in order to obtain an optimal calibration. The average calibration accuracy was 95%, which was calculated as a proportion of the fixated points out of five calibration points over two calibration checks. To maintain dogs’ ideal vigilance and to prevent frustration, calibration and the eye gaze measurements were done on separate days. During the calibration, calibration checks and eye gaze measurements the position of the head, illumination and eye tracker position were maintained the same.
Stimuli
Altogether, 36 different stimulus images were presented in this study. The stimuli were color photographs of dogs, humans, and wild animals (mammals, e.g., elephants, tigers, pandas). Three categories of images were used: (1) landscape images containing a human or an animal, (2) full body images of a single human or an animal, (3) full body images of two paired humans or animals (12 images per each category, see examples in Fig. 1). Each image category (landscape, single, paired) contained images of dogs, humans, and wild animals (four images of each subtype in all categories). For example, the landscape image category consisted of images of dogs, humans, and wild animals within a landscape. The physical dimensions of the images were (19.2 × 14.6 cm (height × width). The pixel resolution of the images was 725 × 550 pixels, overlaid on a grey background of 1,680 × 1,050 pixels. Stimuli were obtained from 123RF, Shutterstock, Adobe stock, Wikimedia commons, Microsoft clipart and Bigstock databases.
Experimental procedure
In the test room, the dog was unleashed and allowed to settle at the pre-trained position. During the experiment, the experimenter and the dog’s owner remained behind an opaque barrier, out of sight of the dog. The dog’s behavior was monitored with a webcam (Labtec Webcam 2200), which was on top of the monitor. Experiment center version. 3.0™ software (SensoMotoric Instruments GmbH, Germany) was used to show the images on a 22 inch (47.4 × 29.7 cm) LCD screen. The images were shown on two separate days (the same 36 images were repeated) in a pseudorandomized order. The stimuli were presented in six blocks of seven to nine images (a total of 36 images per dog per day), 3 s per stimulus and with a 500 ms inter-stimulus-interval. The dog was rewarded with a treat after every image block. Thereafter, the dog was free to return to the pre-trained position.
Data processing
Successful eye gaze recordings were obtained from all 24 dogs, on average 65 ± 2 images per dog. In total, 128 images were excluded from the analysis owing to eye-tracker software problems, the dog lift its head from the chin rest or the dog left its’ place. BeGaze v. 2.4™ software (SensoMotoric Instruments GmbH, Germany) was used to analyze the raw eye movement data. The fixation was scored with a low-speed event detection algorithm calculating potential fixations with a moving window spanning consecutive data points. To be coded as fixation, the minimum fixation duration was 75 ms and the maximum dispersion value D = 250 px {D=[max(x) − min(x)] + [max(y) − min(y)]}. Elsewise, the measured sample was defined as part of a saccade. Binocular gaze data was averaged before analyzes.
Statistical analyses
Each stimulus image was divided into four areas of interest (AOI): (1) the object area, comprising the heads and bodies of the animals or humans depicted in the images (2) the background area comprising the whole image but not the object area (3) the head area comprising the face and neck areas, and (4) the body area comprising the torso, arms and legs (AOI are illustrated in Fig. 1).
Total gaze times (sum of durations of all fixations and saccades) were calculated from the raw binocular data for each AOI (object, background, head and body). Because the sizes of the AOI varied between image categories and species represented in the images, the gaze time was measured as a normalized score (applied from Dahl et al., 2009; Guo, Tunnicliffe & Roebuck, 2010; Somppi et al., 2016), henceforth referred to as “proportional gazing time”. The score was calculated by subtracting the relative AOI size (e.g., the size of the head divided by the size of the whole object) from the relative gaze time (e.g., the total gaze time of the head divided by the total gaze time of the whole object area). Positive gaze score means that the AOI is looked at longer than would be expected according to the size of the area, and negative scores indicate the opposite. The statistical analyses were conducted using SPSS Statistics version 24.0 (IBM, NY, USA). The total gaze time of the head, body, object and background area in family and kennel dogs were compared with repeated-measures analysis of variance (ANOVA) with a between-subjects factor ‘group’ (family, kennel) and within-subjects factors ‘image category’ (landscape, single, paired), ‘species’ (dog, human, wild animal) and ‘AOI area’ (head, body, object, background). Paired samples t-tests were conducted to clarify the ANOVA results. All results are reported as mean (± standard error of the mean) of proportional gazing times.
Results
The total gaze time did not differ between family and kennel dogs (between-subjects factor group, F1,22 = 0.024, p = 0.877). Instead, main effect of AOI area (F2,38 = 38.9, p = 0.001) and interaction effects between AOI area × group (F2,38 = 4.6, p = 0.020), AOI area × species (F3,62 = 2.9, p = 0.046), AOI area × image category (F3,69 = 10.1, p = 0.001) and AOI area × image category x group (F3,69 = 3.2, p = 0.028) were found.
As a main effect, both dog groups gazed at the head area longer than the body (0.10 ± 0.03 and−0.10 ± 0.03, respectively; t23 = 3.3, p = 0.003) or background area (0.10 ± 0.03 and −0.26 ± 0.03, respectively; t23 = 8.6, p = 0.001). Furthermore, the body area was gazed longer than the background area (−0.10 ± 0.03 and −0.26 ± 0.03, respectively; t23 = 3.4, p = 0.002). In addition, the object area was gazed longer than the background area (0.27 ± 0.03 and −0.26 ± 0.03, respectively; t23 = 8.3, p = 0.001; Fig. 2).
Within-groups comparisons showed that family dogs gazed at the head area longer than the body (0.15 ± 0.03 and −0.15 ± 0.03, respectively; t15 = 4.6, p = 0.001) or background (0.15 ± 0.03 and −0.31 ± 0.04, respectively; t15 = 11.0, p = 0.001) areas. In addition, gazing time was longer for the object area than for the background (0.31 ± 0.04 and −0.31 ± 0.04, respectively; t15 = 8.0, p = 0.001) areas in family dogs. Furthermore, the body area gazing time was longer than the background area gazing time (−0.15 ± 0.03 and −0.31 ± 0.04, respectively; t15 = 2.7, p = 0.018). Also in kennel dogs, the gazing time was longer for the head than the background (0.01 ± 0.05 and −0.18 ± 0.04, respectively; t7 = 3.3, p = 0.012) area. Furthermore, kennel dogs gazed at the object area longer than the background area (0.18 ± 0.04 and −0.18 ± 0.04, t7 = 4.1, p = 0.005; Fig. 3 & Table 1).
Table 1. AOI area gazing times in family and kennel dogs.
AOI | Dog group | Proportional gaze time mean (SEM) |
AOI | AOI | AOI | |||
---|---|---|---|---|---|---|---|---|
Body | Object | Background | ||||||
t | p | t | p | t | p | |||
Head | Family | 0.15 (0.03) | 4.6 | 0.001* | – | – | 11.0 | 0.001* |
Kennel | 0.01 (0.05) | 0.1 | 0.930 | – | – | 3.3 | 0.012* | |
Body | Family | −0.15 (0.03) | – | – | – | – | 2.7 | 0.018* |
Kennel | −0.01 (0.05) | – | – | – | – | 2.1 | 0.075 | |
Object | Family | 0.31 (0.04) | – | – | – | – | 8.0 | 0.001* |
Kennel | 0.18 (0.04) | – | – | – | – | 4.1 | 0.005* | |
Background | Family | −0.31 (0.04) | 2.7 | 0.018* | 8.0 | 0.001* | – | – |
Kennel | −0.18 (0.04) | 2.1 | 0.075 | 4.1 | 0.005* | – | – |
Notes.
Significant pairwise comparison (p < 0.05).
Both dog groups gazed longer at the head area in wild animal images vs. dog images (0.18 ± 0.04 and 0.06 ± 0.06, respectively; t23 = −2.1, p = 0.050, statistical trend) and likewise longer in wild animal vs. human images (0.18 ± 0.04 and 0.07 ± 0.04, respectively; t23 = −2.1, p = 0.043). The body area was gazed longer in images containing dogs vs. wild animals (−0.06 ± 0.06 and −0.18 ± 0.04, respectively; t23 = 2.1, p = 0.050, statistical trend), and also in images containing humans vs. wild animals (−0.07 ± 0.04 and −0.18 ± 0.04, respectively; t23 = 2.1, p = 0.043). In addition, the background was gazed longer in images containing dogs vs. wild animals (−0.22 ± 0.04 and −0.31 ± 0.03, respectively; t23 = − 2.1, p = 0.048; Fig. 4 & Table 2).
Table 2. Different species AOI area gazing times in both dog groups.
AOI | Proportional gaze time mean (SEM) |
t | p | Proportional gaze time mean (SEM) | t | p | Proportional gaze time mean (SEM) |
t | p | |||
---|---|---|---|---|---|---|---|---|---|---|---|---|
Dog | Human | Dog | Wild animal | Human | Wild animal | |||||||
Head | 0.06 (0.06) | 0.07 (0.04) | −0.1 | 0.952 | 0.06 (0.06) | 0.18 (0.04) | −2.1 | 0.050# | 0.07 (0.04) | 0.18 (0.04) | −2.1 | 0.043* |
Body | −0.06 (0.06) | −0.07 (0.04) | 0.1 | 0.952 | −0.06 (0.06) | −0.18 (0.04) | 2.1 | 0.050# | −0.07 (0.04) | −0.18 (0.04) | 2.1 | 0.043* |
Object | 0.23 (0.04) | 0.26 (0.05) | −0.7 | 0.495 | 0.23 (0.04) | 0.31 (0.03) | −1.9 | 0.069 | 0.26 (0.05) | 0.31 (0.03) | −1.3 | 0.217 |
Background | −0.22 (0.04) | −0.26 (0.05) | 1.0 | 0.340 | −0.22 (0.04) | −0.31 (0.03) | −2.1 | 0.048* | −0.26 (0.05) | −0.31 (0.03) | 1.3 | 0.215 |
Notes.
Significant pairwise comparison (p < 0.05).
Statistical trend (p = 0.05).
Family dogs gazed longer at the head area in images containing a single vs. paired human or animal images (0.22 ± 0.06 and 0.07 ± 0.02, respectively; t15 = 2.7, p = 0.017). Family dogs also gazed longer at the body area in paired vs. single human or animal images (−0.07 ± 0.02 and −0.22 ± 0.06, respectively; t15 = −2.7, p = 0.017; Fig. 5). Furthermore, they gazed longer at objects in single human or animal vs. landscape images (0.49 ± 0.08 and 0.11 ± 0.03, respectively; t15 = −4.9, p = 0.001) and likewise longer in paired human or animal vs. landscape images (0.31 ± 0.06 and 0.11 ± 0.03, respectively; t15 = −3.2, p = 0.007) and also longer in single human or animal vs. paired images (0.49 ± 0.08 and 0.31 ± 0.06, respectively; t15 = 2.4, p = 0.030). Instead, family dogs gazed longer at the background area in landscape vs. single (−0.11 ± 0.03 and −0.49 ± 0.08, respectively; t15 = 4.9, p = 0.001) or paired images (−0.11 ± 0.03 and −0.28 ± 0.05, respectively; t15 = 3.0, p = 0.010). Likewise, they gazed longer at the background areas in paired vs. single images (−0.28 ± 0.05 and −0.49 ± 0.08, respectively; t15 = −3.5, p = 0.003; Fig. 6). For kennel dogs, the gazing times of object area were also longer for paired human or animal vs. landscape images (0.29 ± 0.09 and 0.08 ± 0.06, respectively; t7 = −4.0, p = 0.005). Kennel dogs also gazed longer at the background area in landscape vs. paired human or animal images (−0.08 ± 0.06 and –0.26 ± 0.07, respectively; t7 = 3.5, p = 0.009; Fig. 6 & Table 3).
Table 3. Different image categories AOI area gazing times in family and kennel dogs.
AOI | Dog group | Proportional gaze time mean (SEM) |
t | p | Proportional gaze time mean (SEM) |
t | p | Proportional gaze time mean (SEM) |
t | p | |||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Image category | Image category | Image category | |||||||||||
Landscape | Single | Landscape | Paired | Single | Paired | ||||||||
Head | Family | 0.08 (0.05) | 0.22 (0.06) | −1.9 | 0.084 | 0.08 (0.05) | 0.07 (0.02) | 0.1 | 0.947 | 0.22 (0.06) | 0.07 (0.02) | 2.7 | 0.017* |
Kennel | −0.02 (0.07) | −0.05 (0.07) | 0.5 | 0.600 | −0.02 (0.07) | −0.03 (0.03) | 0.2 | 0.832 | −0.05 (0.07) | −0.03 (0.03) | −0.3 | 0.809 | |
Body | Family | −0.08 (0.05) | −0.22 (0.06) | 1.9 | 0.084 | −0.08 (0.05) | −0.07 (0.02) | −0.1 | 0.947 | −0.22 (0.06) | −0.07 (0.02) | −2.7 | 0.017* |
Kennel | 0.02 (0.07) | 0.05 (0.07) | −0.5 | 0.600 | 0.02 (0.07) | 0.03 (0.03) | −0.2 | 0.832 | 0.05 (0.07) | 0.03 (0.03) | 0.3 | 0.809 | |
Object | Family | 0.11 (0.03) | 0.49 (0.08) | −4.9 | 0.001* | 0.11 (0.03) | 0.31 (0.06) | −3.2 | 0.007* | 0.49 (0.08) | 0.31 (0.06) | 2.4 | 0.030* |
Kennel | 0.08 (0.06) | 0.20 (0.05) | −2.1 | 0.074 | 0.08 (0.06) | 0.29 (0.09) | −4.0 | 0.005* | 0.20 (0.05) | 0.29 (0.09) | −1.0 | 0.341 | |
Background | Family | −0.11 (0.03) | −0.49 (0.08) | 4.9 | 0.001* | −0.11 (0.03) | −0.28 (0.05) | 3.0 | 0.010* | −0.49 (0.08) | −0.28 (0.05) | −3.5 | 0.003* |
Kennel | −0.08 (0.06) | −0.20 (0.05) | 2.1 | 0.069 | −0.08 (0.06) | −0.26 (0.07) | 3.5 | 0.009* | −0.20 (0.05) | −0.26 (0.07) | 0.7 | 0.495 |
Notes.
Significant pairwise comparison (p < 0.05).
Discussion
In this work, we studied dogs’ visual scanning of images of natural scenes containing dogs, humans, and exotic wild animals. The present work was the first canine eye tracking study using stimuli species other than dogs or humans, and using images, which varied in composition (landscapes and close-up images of humans and animals). Generally, dogs identified the living creatures from the images and focused more attention on their heads relative to the background, which is very much in line with the previous eye-tracking studies (e.g., Téglás et al., 2012; Somppi et al., 2012; Barber et al., 2016). Human and non-human primates are known to concentrate their gaze on the human and animal figures with particular attention paid to the faces of those figures (e.g., Yarbus, 1967; Henderson & Hollingworth, 1999; Kano & Tomonaga, 2009; Machado et al., 2011). Furthermore, human and non-human primates are able to detect living creatures from a landscape image quickly and accurately, even if they are partially hidden and relatively small compared to the background (Fabre-Thorpe, Richard & Thorpe, 1998; Thorpe et al., 2001; New, Cosmides & Tooby, 2007). Our current results suggest a similar, spontaneous detection of living creatures by dogs. The dogs’ preference for biologically relevant information within visual scenes is consistent with the “life detector mechanism”; a basic visual method for focusing attention towards and detecting other animals (reviewed in Rosa-Salva, Mayer & Vallortigara, 2015). Dogs’ preference for biological information is also highlighted in the recent studies using point-light animations of biological motion (Kovács et al., 2016; Ishikawa et al., 2018). In addition to the whole biological creatures, dogs generally fixated more on the heads of animals and humans rather than backgrounds of the images, which further emphasizes the importance of faces in social animals’ visual processing (for a review, Leopold & Rhodes, 2010).
Contrary to our expectations, dogs’ gazing behavior was not greatly altered by the species depicted in the images. Dogs gazed at the head areas in wild animal images longer than in human or dog images. Possibly, the naturally novel wild animals attracted dogs’ attention to a higher degree than humans or dogs (Fantz, 1964), and this kind of natural novelty might reflect the general visual experience of the observer. Previously, rhesus monkeys showed an increased pupillary response, when they viewed videos containing non-primate animals (e.g., marine mammals) and landscapes (e.g., polar), for which they have had limited exposure during their lives (Machado et al., 2011). The ability to detect and respond quickly to novel events is essential for survival in the nature (Ranganath & Rainer, 2003), which is also highlighted in our current results. In previous studies using facial images, dogs preferred to look at conspecific faces over those of humans (e.g., Somppi et al., 2012; Somppi et al., 2014; Bognár, Iotchev & Kubinyi, 2018). However, when they viewed photographs presenting whole bodies within a scene, they have gazed more at humans than dogs (Törnqvist et al., 2015). This difference means that factors other than merely the species influences viewing behavior (Dahl, Logothetis & Hoffman, 2007). For example, human infants show spontaneous gazing preferences toward dog versus human pictures (Gergely et al., 2019), probably owing to the novelty of the stimuli. In our study, no great differences were found between the species depicted, possibly because the images were emotionally neutral full-body figures with scant social gestures and three different types of images were used. In rhesus macaques, gaze distribution in viewing unfamiliar non-conspecific animals varied systematically according to their phylogenetic distance from the viewed species. It was also dependent upon the social and emotional contexts of the image (McFarland et al., 2013).
The present study showed, for the first time, that the image composition affects the dogs’ gazing behavior. In general, dogs spent less time gazing at the object areas in landscape images, where the size of the target was small relative to the image size, which contrasts from the other image types. From this result we infer that dogs may have had difficulty in locating objects within the landscape images. The visual acuity and the ability to detect brightness of the canine visual system is lower than in humans (Pongrácz et al., 2017; for a review, Byosiere et al., 2018). Therefore, the size of stimuli that dogs can see clearly may vary greatly (e.g., review in Byosiere et al., 2018; Tanaka et al., 2000; Lind et al., 2017). In this study, dogs were shown color images, which may also have affected their ability to perceive these images. Dogs have dichromatic color vision, but the data determining which colors dogs can discriminate are controversial (e.g., Neitz, Geist & Jacobs, 1989; Miller & Murphy, 1995; for a review, Byosiere et al., 2018). Recent studies suggest that dogs are unable to distinguish red from green color and that color information might be more important than brightness when dogs are required to choose between stimuli (Kasparson, Badridze & Maximov, 2013; Siniscalchi et al., 2017).
Additionally, family dogs gazed at the head areas of single animals or humans longer than images of paired animals or humans. Conversely, family dogs gazed at the body areas of animals or humans longer in the paired rather than the single image conditions. Thus, family dogs’ gazing preference for the head area was clearly shown in images containing single animals, likely indicating that fixations may have been spread more widely where there were two head areas i.e., in the paired animal images. In these images, body areas may have drawn the attention of family dogs’, because of the social bodily gestures that are present in these images. In the paired images, two animals or humans were standing or sitting close to each other. Interestingly, kennel dogs’ gazing behavior did not differ between head and body areas in the single or paired images. This result is in line with our previous studies, where kennel dogs gazed at scenes of social interaction and facial images less than family dogs (Törnqvist et al., 2015; Somppi et al., 2014). Since family dogs have had more opportunities for social interactions than kennel dogs, they may be more interested and have a greater ability to search whether there are social gestures presented in those images. However, the differences between family and kennel dogs’ gazing behavior may be more evident with socially richer stimuli than the ones used in the current study.
Other than the latter difference between the two study groups, family and kennel dogs gazed at the images in a similar manner, focusing their gaze on the head and object areas of the images. This result suggests that the basic viewing processes are the same regardless of the social living environment and life experiences of the participants. Several previous studies have shown that humans and some other newborn vertebrates have visual predispositions for the head region of living creatures (e.g., Johnson et al., 1991; for a review, Morton & Johnson, 1991). Monkey infants and newly hatched chicks show visual preference for faces, even when there has been no previous exposure to faces (Sugita, 2008; Rosa-Salva et al., 2011). Despite their more restricted social living environment, kennel dogs had daily contact with humans and other beagle dogs. This might have diminished the differences in gazing behavior between family and kennel dogs. In addition, the study populations were not equal in number between family and kennel dogs (2:1), which may have affected our results.
In this study, dogs were trained to remain still during image presentation, because spontaneous movement during the eye tracking recordings can cause serious artifacts and loss of data. However, dogs were not trained to target their gaze at the images or screen. During the experiments, dogs were rewarded with a treat after every image block, which was considered necessary in order to prevent frustration and to ensure that dogs were motivated. It is unlikely that the dogs learnt to associate rewards with certain presented images or that a particular viewing strategy would have been reinforced. Dogs were rewarded only when the screen was blank, and the last image of the block was always different making it hard to learn any specific gazing strategy.
Conclusions
In summary, when viewing photographs of natural scenes, dogs focus their gaze on the informative areas such as the bodies and heads of living creatures, as previously shown for primates. Gazing behavior was dependent on the size and species of the target animals in the images for both dogs living in a socially rich and stimulating family environment and dogs living in a more restricted kennel environment. Nevertheless, there were minor differences in viewing behavior of these dog groups, suggesting that social experiences modulate dogs’ spontaneous viewing patterns when looking at natural scenes.
Supplemental Information
Acknowledgments
We thank Christina M. Krause for providing the eye tracking system for the usage in this study and Reeta Törne for assisting in the experiments and data preparation.
Funding Statement
This study was supported by the Academy of Finland (project #137931) and Emil Aaltonen foundation. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Additional Information and Declarations
Competing Interests
The authors declare there are no competing interests.
Author Contributions
Heini Törnqvist conceived and designed the experiments, performed the experiments, analyzed the data, prepared figures and/or tables, authored or reviewed drafts of the paper, and approved the final draft.
Sanni Somppi conceived and designed the experiments, performed the experiments, authored or reviewed drafts of the paper, and approved the final draft.
Miiamaaria V. Kujala analyzed the data, authored or reviewed drafts of the paper, and approved the final draft.
Outi Vainio conceived and designed the experiments, authored or reviewed drafts of the paper, and approved the final draft.
Animal Ethics
The following information was supplied relating to ethical approvals (i.e., approving body and any reference numbers):
The research was approved by the Ethical Committee for the Use of Animals in Experiments at the University of Helsinki (2/2010) and the Finnish national Animal Experiment Board (approval#STH367A/ESLH-2008-04236/Ym-23).
Data Availability
The following information was supplied regarding data availability:
The raw eye tracking data is available in the Supplemental File.
References
- Autier-Dérian et al. (2013).Autier-Dérian D, Deputte BL, Chalvet-Monfray K, Coulon M, Mounier L. Visual discrimination of species in dogs (Canis familiaris) Animal Cognition. 2013;16:637–651. doi: 10.1007/s10071-013-0600-8. [DOI] [PubMed] [Google Scholar]
- Barber et al. (2016).Barber ALA, Randi D, Müller CA, Huber L. The processing of human emotional faces by pet and lab dogs: evidence for lateralization and experience effects. PLOS ONE. 2016;11:e0152393. doi: 10.1371/journal.pone.0152393. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Bognár, Iotchev & Kubinyi (2018).Bognár Z, Iotchev IB, Kubinyi E. Sex, skull length, breed, and age predict how dogs look at faces of humans and conspecifics. Animal Cognition. 2018;21:447–456. doi: 10.1007/s10071.018.1180.4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Byosiere et al. (2018).Byosiere SE, Chouinard PA, Howell TJ, Bennett PC. What do dogs (Canis familiaris) see? A review of vision in dogs and implications for cognition research. Psychonomic Bulletin and Review. 2018;25:1798–1813. doi: 10.3758/s13423-017-1404-7. [DOI] [PubMed] [Google Scholar]
- Dahl, Logothetis & Hoffman (2007).Dahl CD, Logothetis NK, Hoffman KL. Individuation and holistic processing of faces in rhesus monkeys. Proceedings of the Royal Society B: Biological Sciences. 2007;274:2069–2076. doi: 10.1098/rspb.2007.0477. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Dahl et al. (2009).Dahl CD, Wallraven C, Bülthoff HH, Logothetis NK. Humans and macaques employ similar face-processing strategies. Current Biology. 2009;19:509–513. doi: 10.1016/j.cub.2009.01.061. [DOI] [PubMed] [Google Scholar]
- Di Giorgio et al. (2017).Di Giorgio E, Loveland JL, Mayer U, Rosa-Salva O, Versace E, Vallortigara G. Filial responses as predisposed and learned preferences: early attachment in chicks and babies. Behavioural Brain Research. 2017;325:90–104. doi: 10.1016/j.bbr.2016.09.018. [DOI] [PubMed] [Google Scholar]
- Diamond & Carey (1986).Diamond R, Carey S. Why faces are and are not special: an effect of expertise. Journal of Experimental Psychology: General. 1986;115:107–117. doi: 10.1037/0096-3445.115.2.107. [DOI] [PubMed] [Google Scholar]
- Fabre-Thorpe, Richard & Thorpe (1998).Fabre-Thorpe M, Richard G, Thorpe SJ. Rapid categorization of natural images by rhesus monkeys. Neuroreport. 1998;9:303–308. doi: 10.1097/00001756-199801260-00023. [DOI] [PubMed] [Google Scholar]
- Fantz (1964).Fantz RL. Visual experience in infants: decreased attention to familiar patterns relative to novel ones. Science. 1964;146:668–670. doi: 10.1126/science.146.3644.668. [DOI] [PubMed] [Google Scholar]
- Gavin, Houghton & Guo (2017).Gavin CJ, Houghton S, Guo K. Dog owners show experience-based viewing behaviour in judging dog face approachability. Psychological Research. 2017;81:75–82. doi: 10.1007/s00426-015-0718-1. [DOI] [PubMed] [Google Scholar]
- Gergely et al. (2019).Gergely A, Petró E, Oláh K, Topál J. Auditory–visual matching of conspecifics and non-conspecifics by dogs and human infants. Animals. 2019;9:17. doi: 10.3390/ani9010017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Guo, Tunnicliffe & Roebuck (2010).Guo K, Tunnicliffe D, Roebuck H. Human spontaneous gaze patterns in viewing of faces of different species. Perception. 2010;39:533–542. doi: 10.1068/p6517. [DOI] [PubMed] [Google Scholar]
- Hanazuka et al. (2012).Hanazuka Y, Shimizu M, Kurotori H, Midorikawa A. Visual discrimination in an orangutan (Pongo pygmaeus): measuring visual preference. Perceptual and Motor Skills. 2012;114:429–432. doi: 10.2466/22.24. [DOI] [PubMed] [Google Scholar]
- Henderson & Hollingworth (1999).Henderson JM, Hollingworth A. High-level scene perception. Annual Review of Psychology. 1999;50:243–271. doi: 10.1146/annurev.psych.50.1.243. [DOI] [PubMed] [Google Scholar]
- Ishikawa et al. (2018).Ishikawa Y, Mills D, Willmott A, Mullineaux D, Guo K. Sociability modifies dogs’ sensitivity to biological motion of different social relevance. Animal Cognition. 2018;21:245–252. doi: 10.1007/s10071-018-1160-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Johnson et al. (1991).Johnson MH, Dziurawiec S, Ellis H, Morton J. Newborns’ preferential tracking of face-like stimuli and its subsequent decline. Cognition. 1991;40:1–19. doi: 10.1016/0010-0277(91)90045-6. [DOI] [PubMed] [Google Scholar]
- Kano, Call & Tomonaga (2012).Kano F, Call J, Tomonaga M. Face and eye scanning in gorillas (Gorilla gorilla), orangutans (Pongo abelii), and humans (Homo sapiens): Unique eye-viewing patterns in humans among hominids. Journal of Comparative Psychology. 2012;126:388–398. doi: 10.1037/a0029615. [DOI] [PubMed] [Google Scholar]
- Kano et al. (2018).Kano F, Shepherd SV, Hirata S, Call J. Primate social attention: species differences and effects of individual experience in humans, great apes, and macaques. PLOS ONE. 2018;13:e0193283. doi: 10.1371/journal.pone.0193283. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kano & Tomonaga (2009).Kano F, Tomonaga M. How chimpanzees look at pictures: a comparative eye-tracking study. Proceedings of the Royal Society of London B: Biological Sciences. 2009;276:1949–1955. doi: 10.1098/rspb.2008.1811. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kasparson, Badridze & Maximov (2013).Kasparson AA, Badridze J, Maximov VV. Colour cues proved to be more informative for dogs than brightness. Proceedings of the Royal Society of London B: Biological Sciences. 2013;280:20131356. doi: 10.1098/rspb.2013.1356. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kovack-Lesh, McMurray & Oakes (2014).Kovack-Lesh KA, McMurray B, Oakes LM. Four-month-old infants’ visual investigation of cats and dogs: relations with pet experience and attentional strategy. Developmental Psychology. 2014;50:402–413. doi: 10.1037/a0033195. [DOI] [PubMed] [Google Scholar]
- Kovács et al. (2016).Kovács K, Kis A, Kanizsár O, Hernádi A, Gácsi M, Topál J. The effect of oxytocin on biological motion perception in dogs (Canis familiaris) Animal Cognition. 2016;19:513–522. doi: 10.1007/s10071-015-0951-4. [DOI] [PubMed] [Google Scholar]
- Kujala et al. (2012).Kujala MV, Kujala J, Carlson S, Hari R. Dog experts’ brains distinguish socially relevant body postures similarly in dogs and humans. PLOS ONE. 2012;7:e39145. doi: 10.1371/journal.pone.0039145. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Kundel et al. (2008).Kundel HL, Nodine CF, Krupinski EA, Mello-Thomas C. Using gaze-tracking data and mixture distribution analysis to support a holistic model for the detection of cancers on mammograms. Academic Radiology. 2008;15:881–886. doi: 10.1016/j.acra.2008.01.023. [DOI] [PubMed] [Google Scholar]
- Leopold & Rhodes (2010).Leopold DA, Rhodes G. A comparative view of face perception. Journal of Comparative Psychology. 2010;124:233–251. doi: 10.1037/a0019460. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Lind et al. (2017).Lind O, Milton I, Andersson E, Jensen P, Roth LS. High visual acuity revealed in dogs. PLOS ONE. 2017;12:e0188557. doi: 10.1371/journal.pone.0188557. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Machado et al. (2011).Machado CJ, Bliss-Moreau E, Platt ML, Amaral DG. Social and nonsocial content differentially modulates visual attention and autonomic arousal in rhesus macaques. PLOS ONE. 2011;6:e26598. doi: 10.1371/journal.pone.0026598. [DOI] [PMC free article] [PubMed] [Google Scholar]
- McFarland et al. (2013).McFarland R, Roebuck H, Yan Y, Majolo B, Li W, Guo K. Social interactions through the eyes of macaques and humans. PLOS ONE. 2013;8:e56437. doi: 10.1371/journal.pone.0056437. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Miller & Murphy (1995).Miller PE, Murphy CJ. Vision in dogs. Journal of American Veterinary Medical Association. 1995;207:1623–1634. [PubMed] [Google Scholar]
- Morton & Johnson (1991).Morton J, Johnson MH. CONSPEC and CONLERN: a two-process theory of infant face recognition. Psychological Review. 1991;98:164–181. doi: 10.1037/0033-295X.98.2.164. [DOI] [PubMed] [Google Scholar]
- Nagasawa et al. (2011).Nagasawa M, Murai K, Mogi K, Kikusui T. Dogs can discriminate human smiling faces from blank expressions. Animal Cognition. 2011;14:525–533. doi: 10.1007/s10071-011-0386-5. [DOI] [PubMed] [Google Scholar]
- Neitz, Geist & Jacobs (1989).Neitz J, Geist T, Jacobs GH. Color vision in the dog. Visual Neuroscience. 1989;3:119–125. doi: 10.1017/S0952523800004430. [DOI] [PubMed] [Google Scholar]
- New, Cosmides & Tooby (2007).New J, Cosmides L, Tooby J. Category-specific attention for animals reflects ancestral priorities, not expertise. Proceedings of the National Academy of Sciences of the United States of America. 2007;104:16598–16603. doi: 10.1073/pnas.0703913104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Pongrácz et al. (2017).Pongrácz P, Ujvári V, Faragó T, Miklósi Á, Péter A. Do you see what I see? The difference between dog and human visual perception may affect the outcome of experiments. Behavioural Processes. 2017;140:53–60. doi: 10.1016/j.beproc.2017.04.002. [DOI] [PubMed] [Google Scholar]
- Racca et al. (2010).Racca A, Amadei E, Ligout S, Guo K, Meints K, Mills D. Discrimination of human and dog faces and inversion responses in domestic dogs (Canis familiaris) Animal Cognition. 2010;13:525–533. doi: 10.1007/s10071-009-0303-3. [DOI] [PubMed] [Google Scholar]
- Ranganath & Rainer (2003).Ranganath C, Rainer G. Cognitive neuroscience: neural mechanisms for detecting and remembering novel events. Nature Reviews Neuroscience. 2003;4:193–202. doi: 10.1038/nrn1052. [DOI] [PubMed] [Google Scholar]
- Range et al. (2008).Range F, Aust U, Steurer M, Huber L. Visual categorization of natural stimuli by domestic dogs. Animal Cognition. 2008;11:339–347. doi: 10.1007/s10071-007-0123-2. [DOI] [PubMed] [Google Scholar]
- Rosa-Salva et al. (2011).Rosa-Salva O, Farroni T, Regolin L, Vallortigara G, Johnson MH. The evolution of social orienting: evidence from chicks (Gallus gallus) and human newborns. PLOS ONE. 2011;6:e18802. doi: 10.1371/journal.pone.0018802. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Rosa-Salva, Mayer & Vallortigara (2015).Rosa-Salva O, Mayer U, Vallortigara G. Roots of a social brain: developmental models of emerging animacy-detection mechanisms. Neuroscience & Biobehavioral Reviews. 2015;50:150–168. doi: 10.1016/j.neubiorev.2014.12.015. [DOI] [PubMed] [Google Scholar]
- Simion, Regolin & Bulf (2008).Simion F, Regolin L, Bulf H. A predisposition for biological motion in the newborn baby. Proceedings of the National Academy of Sciences of the United States of America. 2008;105:809–813. doi: 10.1073/pnas.0707021105. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Siniscalchi et al. (2017).Siniscalchi M, d’Ingeo S, Fornelli S, Quaranta A. Are dogs red-green colour blind? Royal Society Open Science. 2017;4:170869. doi: 10.1098/rsos.170869. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Somppi et al. (2012).Somppi S, Törnqvist H, Hänninen L, Krause CM, Vainio O. Dogs do look at images –Eye tracking in canine cognition research. Animal Cognition. 2012;15:163–174. doi: 10.1007/s10071-011-0442-1. [DOI] [PubMed] [Google Scholar]
- Somppi et al. (2014).Somppi S, Törnqvist H, Hänninen L, Krause CM, Vainio O. How dogs scan familiar and inverted faces: an eye movement study. Animal Cognition. 2014;17:793–803. doi: 10.1007/s10071-013-0713-0. [DOI] [PubMed] [Google Scholar]
- Somppi et al. (2016).Somppi S, Törnqvist H, Kujala MV, Hänninen L, Krause CM, Vainio O. Dogs evaluate threatening facial expressions by their biological validity-evidence from gazing patterns. PLOS ONE. 2016;11:e0143047. doi: 10.1371/journal.pone.0143047. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Somppi et al. (2017).Somppi S, Törnqvist H, Topál J, Koskela A, Hänninen L, Krause CM, Vainio O. Nasal oxytocin treatment biases dogs’ visual attention and emotional response toward positive human facial expressions. Frontiers in Psychology. 2017;8:1854. doi: 10.3389/fpsyg.2017.01854. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Sugita (2008).Sugita Y. Face perception in monkeys reared with no exposure to faces. Proceedings of the National Academy of Sciences of the United States of America. 2008;105:394–398. doi: 10.1073/pnas.0706079105. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Tanaka et al. (2000).Tanaka T, Ikeuchi E, Mitani S, Eguchi Y, Uetake K. Studies on the visual acuity of dogs using shape discrimination learning. Nihon Chikusan Gakkaiho. 2000;71:614–620. doi: 10.2508/chikusan.71.614. [DOI] [Google Scholar]
- Téglás et al. (2012).Téglás E, Gergely A, Kupán K, Miklósi Á, Topál J. Dogs’ gaze following is tuned to human communicative signals. Current Biology. 2012;22:209–212. doi: 10.1016/j.cub.2011.12.018. [DOI] [PubMed] [Google Scholar]
- Thorpe et al. (2001).Thorpe SJ, Gegenfurtner KR, Fabre-Thorpe M, Bulthoff HH. Detection of animals in natural images using far peripheral vision. European Journal of Neuroscience. 2001;14:869–876. doi: 10.1046/j.0953-816x.2001.01717.x. [DOI] [PubMed] [Google Scholar]
- Törnqvist et al. (2015).Törnqvist H, Somppi S, Koskela A, Krause CM, Vainio O, Kujala MV. Comparison of dogs and humans in visual scanning of social interaction. Royal Society Open Science. 2015;2:150341. doi: 10.1098/rsos.150341. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Winters, Dubuc & Higham (2015).Winters S, Dubuc C, Higham JP. Perspectives: the looking time experimental paradigm in studies of animal visual perception and cognition. Ethology. 2015;121:625–640. doi: 10.1111/eth.12378. [DOI] [Google Scholar]
- Yarbus (1967).Yarbus AL. Eye movements and vision. New York: Plenum Press; 1967. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The following information was supplied regarding data availability:
The raw eye tracking data is available in the Supplemental File.