Skip to main content
PLOS One logoLink to PLOS One
. 2021 Mar 29;16(3):e0249137. doi: 10.1371/journal.pone.0249137

Sclera color in humans facilitates gaze perception during daytime and nighttime

Jessica L Yorzinski 1,*, Amy Harbourne 2, William Thompson 3
Editor: Enkelejda Kasneci4
PMCID: PMC8006985  PMID: 33780503

Abstract

Species vary widely in the conspicuousness of their eye morphology and this could influence gaze perception. Eyes with conspicuous morphology can enhance gaze perception while eyes with camouflaged morphology may hinder gaze perception. While evidence suggests that conspicuous eye morphology enhances gaze perception, little is known about how environmental conditions affect this interaction. Thus, we investigated whether environmental light conditions affect gaze perception. Human subjects (Homo sapiens) were instructed to find direct-gaze faces within arrays of averted-gaze faces or to find averted-gaze faces within arrays of directed-gaze faces. The faces were displayed under conditions simulating nighttime or daytime conditions. Furthermore, the faces had naturally-colored sclera (white) or modified sclera (same color as the iris). Participants were fastest and most accurate in detecting faces during the daytime and nighttime conditions when the sclera were naturally-colored. Participants were worst at detecting faces with modified sclera during the nighttime conditions. These results suggest that eyes with conspicuous morphology enhance gaze perception during both daytime and nighttime conditions.

Introduction

Gaze perception allows individuals to determine where others are directing their overt attention [1]. Individuals can evaluate if others are looking directly toward them or looking elsewhere within the environment [2, 3]. Furthermore, they can follow the gaze of others to learn about the location of salient objects. Humans and other species follow the gaze of conspecifics to distant locations (reviewed in [4]). Complex forms of social cognition, such as theory of mind, also likely involve gaze perception [1]. Even though gaze perception is vital to many forms of social cognition, we still have much to learn about the factors that affect it.

One factor that could influence gaze perception is eye morphology. Eyes with conspicuous morphology may have evolved to enhance gaze perception while eyes with camouflaged morphology may have evolved to hinder gaze perception [5]. The amount of exposed sclera, sclera color, and iris color are aspects of eye morphology that could contribute to conspicuousness. Humans have unusually large sclera [5] while other primates often have sclera that are less exposed [6, 7]. Humans also have white sclera but many other primates have pigmented sclera [5, 8 but see 7, 9]. Furthermore, human iris color ranges from black to blue [10] while the majority of other primates have irises with shades of brown or yellow [5, 11]. Previous work in humans has demonstrated that eye morphology influences gaze perception [1219].

Environmental conditions likely interact with eye morphology in influencing gaze perception. This interaction is important because humans often need to function in both daytime and nighttime environments [20, 21]. In fact, preindustrial human populations frequently engaged in nighttime activities (such as foraging, rituals, and socializing; [22, 23]) and these nighttime activities likely involve gaze perception. When humans view objects in low-light conditions, their vision is limited: the threshold visibility, color appearance, sensitivity over time, and visual acuity change compared to viewing the same objects in conditions with bright light [2426]. For example, humans’ visual acuity (ability to resolve spatial detail) is substantially worse during nighttime compared to daytime [24, 27]. Given human’s limited vision under low-light conditions, we would expect their nocturnal communication signals, including gaze signals, to exhibit properties that facilitate visibility at night (such as high contrast; [28]). We are unaware of any previous work investigating the interaction between environmental light conditions and gaze perception.

We therefore examined human gaze perception under simulated daylight and nighttime conditions. We recorded the eye movements of human (Homo sapiens) participants as they searched for directed-gaze faces within arrays of averted-gaze faces or searched for averted-gaze faces within arrays of directed-gaze faces. We used a similar methodology as Yorzinski and Miller [19]. The faces had naturally-colored sclera (white sclera) or modified sclera (sclera color matched the iris color). The faces with modified sclera are a phenotype that is similar to many other primate species; in many primate species, the sclera has a similar color as the iris color (but see [79]) or the sclera is minimally exposed [5]. The faces were displayed under conditions simulating nighttime and daytime conditions. The faces that were large and upright, small and upright, or inverted and upright. The large faces were intended to simulate close-range interactions and the small faces were intended to simulate more distant interactions. The inverted faces altered facial configuration but maintained low-level visual features, including contrast and luminance [29]. Our previous work found that humans were faster and more accurate in detecting faces with naturally-colored sclera compared to faces with sclera that matched the iris color under daytime conditions [19]. This study expands upon these results by testing whether participants are also faster and more accurate in detecting faces with naturally-colored sclera (versus faces with sclera that match the iris color) under nighttime conditions. We expected that participants would be fastest and most accurate in detecting faces with naturally-colored sclera under both daytime and nighttime conditions.

Materials and methods

Participants

Thirty women and 30 men (between the ages of 18 and 30 years old) participated in this study at Texas A&M University from September 2018 through May 2019. They were of Caucasian ethnicity, and had normal or corrected-to-normal vision. We recruited the participants using emails and flyers, and they earned $20 for their participation. We told participants that the goal of the study was to examine gaze detection. The Institutional Review Board of Texas A&M University (#2016-0575D) approved this study and all participants provided written consent before participating.

Equipment

We recorded the eye movements of participants using a Tobii Pro screen-based eye-tracker (X2-60; Tobii Technology, Inc., Danderyd, Sweden; accuracy: 0.4 degrees; data rate: 60 Hz; binocular tracking). The eye-tracker was controlled by a laptop computer (Dell Mobile Precision 7510). The experimental stimuli were displayed on a 25” monitor (Dell UltraSharp UP2516D, 2560 x 1440 pixels; Dell Computer Corporation, Round Rock, TX) using Tobii Studio (version: 3.4, Tobii Technology). The eye-tracker noninvasively recorded the precise location of where the participants were directing their overt gaze. In order to minimize head movements, the participants rested their chins on a chin cup (UHCOTech HeadSpot; positioned approximately 60 cm from the screen). Participants were initially told that we were recording the size of their pupils (we debriefed them at the end of the trial to explain that we were recording their eye movements). The luminance of the monitor (sensor positioned in the middle of the monitor and directed toward the screen) displaying a white background was 190 cd/m2 and the illuminance of the testing room (sensor positioned in the middle of the monitor and directed toward the participant) was 30 lux (Spectra Cine PhoRad Meter, SC-820, Burbank, California). Before trials began, the eye-tracker was calibrated by having participants fixate 5 points, which appeared sequentially on the monitor. We employed the Tobii Velocity-Threshold Identification filter to classify fixations and saccades based on the velocity of eye movements using the following parameters: I-VT filter,75 ms gap fill-in; average eye selection; 20 ms velocity calculator window, 30 degrees/s I-VT classifier threshold, 75 ms merge adjacent time, and 0.5 degrees merge adjacent angle. Eye movements below the velocity threshold (30 degrees/s) were classified as fixations and those above were classified as saccades.

Experimental stimuli

The faces we used as stimuli were obtained from the Oslo Face Database [30]. The faces depicted people that were facing the camera (their eyes were either directed or averted) and had neutral expressions. We manipulated the photographs using Adobe Photoshop (Adobe Systems, San Jose, California) and Matlab (Mathworks, Inc., Natick, MA). We removed corneal reflections by replacing them with adjacent colors.

We created three blocks of faces each with four sets for two treatments (Figs 1 and 2). The three blocks included faces that were (1) upright and subtended 5.7° (‘large and upright’), (2) faces that were upright and subtended 2.85° (‘small and upright’), and (3) faces that were inverted and subtended 5.7° (‘large and inverted’). Within each of these three blocks, there were four sets of 8-array photographs: (1) face with a direct gaze and natural color within an array of faces with averted gaze and natural color (‘Target Directed Natural’), (2) face with a direct gaze and modified sclera color within an array of faces with averted gaze and modified sclera color (‘Target Directed Modified’), (3) face with an averted gaze and natural color within an array of faces with direct gaze and natural color (‘Target Averted Natural’), and (4) face with an averted gaze and modified sclera color within an array of faces with direct gaze and modified sclera color (‘Target Averted Modified’; Figs 1 and 2). In the faces with modified sclera color, the color of the sclera matched the mean color of the iris. We used the CIELAB color space to make these color modifications because CIELAB is designed to be perceptually uniform for human viewers [31]; it has three axes consisting of L* (lightness), a* (green to red), and b* (blue to yellow). Within each set, 40 different faces were shown. Half of these 40 faces had light-colored irises (subjectively categorized as blue or green; a*: less than two; b*: less than eight; mean ± SE L*: 48.8 ± 0.99) and the other half of these faces had dark-colored irises (subjectively categorized as brown; a*: greater than four; b*: greater than seventeen; mean ± SE L*: 35.5 ± 1.4).

Fig 1. Example stimuli when gaze is direct or averted for a face with light irises for the daytime and nighttime stimuli.

Fig 1

The man in the stimuli below gave informed consent for his photograph to be used for scientific purposes [30]. Reprinted from [30] under a CC BY license, with permission from Siri Leknes, original copyright 2014.

Fig 2. Example of array used in the search task for the Target Directed Natural set at nighttime.

Fig 2

The man in the stimuli below gave informed consent for his photograph to be used for scientific purposes [30]. Reprinted from [30] under a CC BY license, with permission from Siri Leknes, original copyright 2014.

The two treatments included faces that simulated daytime and nighttime conditions. The original images from the face database were photographed under bright lights and were therefore used as the daytime faces. The nighttime faces were created by modifying the daytime faces to simulate nighttime viewing at mesopic luminance levels using the methods outlined in [25]. The use of simulated nighttime viewing was chosen over actual viewing in the dark because it provided us with a good balance of ecological validity and experimental control directly comparable to the daytime viewing cases. Visual sensitivity changes at low light levels due to factors that include differing spectral and magnitude sensitivity of the rods and cones, pupil size, accommodative error, several different neural effects, and others [32]. Perceptually, as luminance levels transition from photopic (only cones active) to mesopic (a mixture of cone and rod activity) to photopic (only rods active), acuity decreases, the ability to differentiate colors decreases, and sensitivity to colors towards the red end of the visible spectrum increases relative to colors towards the blue end of the visible spectrum [24]. Several other subjective effects also occur, including a bluish appearance, an apparent sharpness inconsistent with the decrease in acuity, and noise [25]. For simulated nighttime viewing, we modified luminance and color sensitivity [33].

Because the blue shifting, addition of uncorrelated Gaussian noise, and darkening of display images described in [25] are highly dependent on the experimental conditions, we empirically determined these parameters. Forty images, each with 27 filtered face images (varying in blue shift, darkening, and σnoise), were shown to 15 participants (seven women and eight men). For each image, the participants were asked to indicate the top three faces that appeared most similar to how faces appear in moonlight. The parameters were set based on the face that participants selected most often (53% of participants most often selected faces with these parameters as one of their top three choices).

Typical lunar illuminance is around 0.05 to 0.1 lux [34]. Assuming an average skin reflectance of 0.6 for faces and assuming the visible portion of the faces is in direct moonlight, this yields a luminance of the faces of 0.01–0.02 cd/m2. We determined the parameterization needed to simulate perception of detail at this light level in ten participants (five men and five women), using the familiar eye chart character recognition task. One of the few studies to empirically measure character legibility under low light levels found a visual acuity of logMAR 0.3 (20/40 Snellen) at 0.4 cd/m2 (Lin et al., 2015). We chose to simulate a visual acuity of logMAR 0.4 (20/50 Snellen), since we were simulating a slightly dimmer light level. The parameter controlling visual acuity in the method described in [25] is σblur. Based on informal testing, filtered logMAR chart images were generated for values of σblur equal to 2.0, 2.5, and 3.0. Each original chart used a different random selection of Sloan characters. For each chart, participants were asked to say the letters of the lowest line that was completely legible to them. They were then asked to read the next lowest line, with the instructions that they should make a reasonable guess so long as they were reasonably certain. The acuity assigned to a particular filtering corresponded to the logMAR value associated with the lowest row in which they could correctly identify a majority of the characters. The σblur for which the smallest readable line was most often logMAR 0.4 was 2.5. Based on the analyses above, we parameterized a Matlab implementation of the method described in [25] and converted all of the daytime faces to nighttime faces. The Matlab scripts used to create these nighttime images are available at https://github.com/jyorzinski/CreateNightImages.

In the daytime and nighttime faces with natural sclera color, the contrast between the sclera and iris are similarly high (Weber’s contrast mean ± SE: daytime 0.47 ± 0.02; nighttime 0.48 ± 0.02; calculated using the method outlined in [35]); likewise, the contrast between the sclera and pupil are also similarly high (Weber’s contrast mean ± SE: daytime 0.79 ± 0.01; nighttime 0.81 ± 0.02). As expected, in the daytime and nighttime faces with sclera color that matches the iris color, there is minimal contrast between the sclera and iris (Weber’s contrast mean ± SE: daytime 0.10 ± 0.01; nighttime 0.13 ± 0.01); the contrast between the sclera and pupil are also similar but high (Weber’s contrast mean ± SE: daytime 0.64 ± 0.01; nighttime 0.61 ± 0.01). The contrast between the iris and pupil are similar in the daytime and nighttime faces (Weber’s contrast mean ± SE: daytime 0.61 ± 0.01; nighttime 0.65 ± 0.02).

Experimental procedure

Participants initially performed a practice trial that used photographs of domestic cat faces (Felis catus). They were shown photographic arrays that consisted of seven cat faces with directed eyes and one cat face with averted eyes. They were asked to find the cat face with averted eyes as quickly and accurately as possible, and then indicate their response using a custom keypad (model: CP24-USBHID; Genovation, Irvine, CA.). The keypad had eight keys that were arranged in the same configuration as the photographic arrays (a 3 x 3 grid with no key in the middle). The participants also saw photographic arrays in which there were seven cat faces with averted eyes and one cat face with directed eyes (and they had to find the cat with the directed eyes). After indicating each response, a fixation screen briefly appeared (1 s; black dot on a white background) and the next array became visible.

After finishing the practice trial, participants began the experimental task that used photographs of human faces rather than cat faces. During the experimental task, participants were asked to find the face with directed gaze among arrays that consisted of one face with a directed gaze and seven faces with averted gaze (Target Directed Natural & Target Directed Modified). Conversely, participants were also asked to find the face with averted gaze among arrays that consisted of one face with an averted gaze and seven faces with directed gaze (Target Averted Natural & Target Averted Modified). They indicated each response using the custom keypad and a fixation screen briefly appeared (1 s; black dot on a white background) before the next array became visible. Many other studies examining attention use this type of visual search task [3640]. A given subject completed both treatments (daytime and nighttime; randomized across participants). The blocks and sets were randomized within participants.

Measurements and statistical analysis

We calculated the amount of time that elapsed between the array appearing and participants fixating on the target (Latency to Fixate Target). We also calculated the amount of time that elapsed between participants fixating on the target and indicating their response (via a keypress; Latency to Press Key After Fixate Target). We performed these calculations by determining whether each fixation coordinate was directed within a rectangular region of interest surrounding the target faces, distractor faces, or neither of them. For each participant, we calculated the mean value of the metrics within each of the four sets (Target Directed Natural, Target Directed Modified, Target Averted Natural, and Target Averted Modified) of the three blocks (large, small, and inverted) for each treatment (daytime and nighttime). We excluded a given array from the analysis if participants never looked at the target or if more than 10% of the gaze data was missing (11% of arrays were excluded using this criteria); this exclusion criteria was set to eliminate cases in which the participants were not performing the task correctly (never looking at the target) or the eye-tracker was malfunctioning (the eye-tracker did not correctly record the participant looking at the target).

We analyzed our data using linear mixed-effects models with repeated measures in SAS (PROC MIXED; Version 9.4; SAS Institute Inc., Cary, NC). The dependent variable was the latency to fixate the target face. The independent variables were the block (large, small, or inverted), set (Target Directed Natural, Target Directed Modified, Target Averted Natural, and Target Averted Modified), treatment (daytime and nighttime), the interaction among block, set and target as well as iris color (light or dark), age, and gender of the participants. Participant identity was included within the models to account for repeated measures. We performed a priori contrasts to compare the latency to detect the target face in daytime versus nighttime as well as between the modified sclera and the natural sclera; we performed 24 comparisons and used the false discovery rate correction to evaluate statistical significance [41]. We performed similar repeated-measures mixed linear models using the latency to indicate their response (via a key press) after fixating the target as the dependent variable. In addition, we performed a similar generalized linear mixed model (PROC GLIMMIX; Poisson distribution) using the percentage of correct responses as the dependent variable.

Results

Overall, participants were slower to fixate target faces under nighttime versus daytime conditions (F1,59 = 202.99, p<0.0001; Table 1: Latency to Fixate Target; Fig 3a). Participants were slower to detect target faces with modified sclera under nighttime compared to daytime conditions. This was the case for faces that were large and upright (directed: F1,1003 = 5.80, p<0.0001; averted: F1,1003 = 6.40, p<0.0001), small and upright (directed: F1,1003 = 4.45, p<0.0001; averted: F1,1003 = 5.98, p<0.0001) and large and inverted (directed: F1,1003 = 8.35, p<0.0001; averted: F1,1003 = 7.94, p<0.0001). When the target faces had naturally-colored sclera, participants were slower to detect the large and upright faces in the nighttime versus the daytime conditions (directed: F1,1003 = 2.49, p = 0.013; averted: F1,1003 = 2.08, p = 0.038) but they detected small and upright faces at similar latencies irrespective of daytime or nighttime conditions (directed: F1,1003 = 0.89, p = 0.37; averted: F1,1003 = 1.54, p = 0.12). They were also slower to detect the large and inverted faces in the nighttime versus daytime but only for directed gaze (directed: F1,1003 = 2.09, p = 0.037; averted: F1,1003 = 1.35, p = 0.18). Participants were faster to detect target faces with light rather than dark iris colors (F1,59 = 63.24, p<0.0001).

Table 1. The effect of block, set, treatment, iris color, age and gender on the latency to fixate the target and latency to press a key after fixating the target.

Numerator df, Denominator df Latency to Fixate Target Latency to Press Key After Fixate Target
Overall model
 Block 2, 118 101.80 (<0.0001)* 109.72 (<0.0001)*
 Set 3, 177 275.61 (<0.0001)* 425.98 (<0.0001)*
 Treatment 1, 59 202.99 (<0.0001)* 514.66 (<0.0001)*
 Block*Set*Treatment 17, 1003 5.46 (<0.0001)* 23.93 (<0.0001)*
 Iris Color 1, 59 63.24 (<0.0001)* 386.63 (<0.0001)*
 Age 1, 57 43.92 (<0.0001)* 17.20 (0.0001)*
 Gender 1, 57 86.31 (<0.0001)* 2.67 (0.11)
Comparisons
Large and Upright
 Target Directed Natural Daytime vs. Directed Natural Nighttime 1, 1003 2.49 (0.013)* 1.13 (0.26)
 Target Directed Modified Daytime vs. Directed Modified Nighttime 1, 1003 5.80 (<0.0001)* 7.72 (<0.0001)*
 Target Averted Natural Daytime vs. Averted Natural Nighttime 1, 1003 2.08 (0.038)* 0.18 (0.86)
 Target Averted Modified Daytime vs. Averted Modified Nighttime 1, 1003 6.40 (<0.0001)* 10.24 (<0.0001)*
 Target Directed Natural Daytime vs. Directed Modified Daytime 1, 1003 5.74 (<0.0001)* 5.07 (<0.0001)*
 Target Averted Natural Daytime vs. Averted Modified Daytime 1, 1003 5.39 (<0.0001)* 2.05 (0.041)*
 Target Directed Natural Nighttime vs. Directed Modified Nighttime 1, 1003 9.05 (<0.0001)* 11.66 (<0.0001)*
 Target Averted Natural Nighttime vs. Averted Modified Nighttime 1, 1003 9.71 (<0.0001)* 12.11 (<0.0001)*
Small and Upright
 Target Directed Natural Daytime vs. Directed Natural Nighttime 1, 1003 0.89 (0.37) 1.63 (0.10)
 Target Directed Modified Daytime vs. Directed Modified Nighttime 1, 1003 4.45 (<0.0001)* 10.49 (<0.0001)*
 Target Averted Natural Daytime vs. Averted Natural Nighttime 1, 1003 1.54 (0.12) 1.17 (0.24)
 Target Averted Modified Daytime vs. Averted Modified Nighttime 1, 1003 5.98 (<0.0001)* 12.25 (<0.0001)*
 Target Directed Natural Daytime vs. Directed Modified Daytime 1, 1003 7.08 (<0.0001)* 8.40 (<0.0001)*
 Target Averted Natural Daytime vs. Averted Modified Daytime 1, 1003 6.76 (<0.0001)* 8.72 (<0.0001)*
 Target Directed Natural Nighttime vs. Directed Modified Nighttime 1, 1003 10.63 (<0.0001)* 17.25 (<0.0001)*
 Target Averted Natural Nighttime vs. Averted Modified Nighttime 1, 1003 11.2 (<0.0001)* 19.80 (<0.0001)*
Large and Inverted
 Target Directed Natural Daytime vs. Directed Natural Nighttime 1, 1003 2.09 (0.037)* 2.78 (0.0055)*
 Target Directed Modified Daytime vs. Directed Modified Nighttime 1, 1003 8.35 (<0.0001)* 13.37 (<0.0001)*
 Target Averted Natural Daytime vs. Averted Natural Nighttime 1, 1003 1.35 (0.18) 2.99 (0.0028)*
 Target Averted Modified Daytime vs. Averted Modified Nighttime 1, 1003 7.94 (<0.0001)* 14.63 (<0.0001)*
 Target Directed Natural Daytime vs. Directed Modified Daytime 1, 1003 6.01 (<0.0001)* 4.66 (<0.0001)*
 Target Averted Natural Daytime vs. Averted Modified Daytime 1, 1003 4.24 (<0.0001)* 3.37 (0.0008)*
 Target Directed Natural Nighttime vs. Directed Modified Nighttime 1, 1003 12.27 (<0.0001)* 15.25 (<0.0001)*
 Target Averted Natural Nighttime vs. Averted Modified Nighttime 1, 1003 10.82 (<0.0001)* 15.01 (<0.0001)*

F values are displayed for the overall model and t values are displayed for the comparisons; p-values are indicated in parentheses and statistically significant comparisons are indicated with an asterisk.

Fig 3. The (a) latency to initially fixate the target, (b) latency to indicate choice (via key press) after initially fixating the target, and (c) percentage of correct responses for large and upright faces, small and upright faces, and large and inverted faces for the daytime and nighttime stimuli.

Fig 3

Means and standard errors are shown; horizontal lines indicate planned comparisons that were statistically significant.

After fixating the correct target face, nighttime and daytime viewing conditions also impacted how quickly participants indicated (via a key press) that they had found the target face (F1,59 = 514.66, p<0.0001; Table 1: Latency to Press Key After Fixate Target; Fig 3b). Similar to above, when the target faces had modified sclera, participants were slower to indicate that they found the target under nighttime compared to daytime conditions when the faces were large and upright (directed: F1,1003 = 7.72, p<0.0001; averted: F1,1003 = 10.24, p<0.0001), small and upright (directed: F1,1003 = 10.49, p<0.0001; averted: F1,1003 = 12.25, p<0.0001), or large and inverted (directed: F1,1003 = 13.37, p<0.0001; averted: F1,1003 = 14.63, p<0.0001). Participants were also slower to indicate that they found the target when the faces with naturally-color sclera were large and inverted in the nighttime versus the daytime conditions (directed: F1,1003 = 2.78, p = 0.0055; averted: F1,1003 = 2.99, p = 0.0028). However, they indicated their responses with similar latencies when the target faces were large and upright (directed: F1,1003 = 1.13, p = 0.26; averted: F1,1003 = 0.18, p = 0.86;) as well as small and upright (directed: F1,1003 = 1.63, p = 0.10; averted: F1,1003 = 1.17, p = 0.24). Participants were faster to indicate that they detected the target when the target faces had light rather than dark iris colors (F1,59 = 386.63, p<0.0001).

Participants were most accurate in finding the correct target face during daytime compared to nighttime conditions (F1,59 = 138.23, p<0.0001; Table 2; Fig 3c). When the target faces had modified sclera, participants were less accurate in identifying the target under nighttime compared to daytime conditions for large and upright faces (directed: F1,1003 = 3.74, p<0.0001), small and upright faces (directed: F1,1003 = 8.08, p<0.0001; averted: F1, 1003 = 6.92, p<0.0001), and large and inverted faces (directed: F1,1003 = 10.89, p<0.0001; averted: F1,1003 = 8.71, p<0.0001) except when the faces were large and upright with averted gaze (F1,1003 = 2.02, p = 0.044). When the target faces had sclera that were naturally colored, participants were equally accurate in identifying the correct target under nighttime and daytime conditions irrespective of whether the faces were large and upright (directed: F1,1003 = 0.20, p = 0.84; averted: F1,1003 = 0.30, p = 0.76), small and upright (directed: F1,1003 = 0.27, p = 0.78; averted: F1,1003 = 0.22, p = 0.83) or large and inverted (directed: F1,1003 = 0.36, p = 0.72; averted: F1,1003 = 0.86, p = 0.39). They were also more accurate in finding target faces with light versus dark irises (F1,59 = 155.65, p<0.0001).

Table 2. The effect of block, set, treatment, iris color, age and gender on the percentage of correct responses.

Numerator df, Denominator df Correct Response
Overall model
 Block 2, 118 14.36 (<0.0001)*
 Set 3, 177 81.70 (<0.0001)*
 Treatment 1, 59 138.23 (<0.0001)*
 Block*Set*Treatment 17, 1003 13.19 (<0.0001)*
 Iris Color 1, 59 155.65 (<0.0001)*
 Age 1, 57 2.96 (0.091)
 Gender 1, 57 0.01 (0.91)
Comparisons
Large and Upright
 Target Directed Natural Daytime vs. Directed Natural Nighttime 1, 1003 0.20 (0.84)
 Target Directed Modified Daytime vs. Directed Modified Nighttime 1, 1003 3.74 (0.0002)*
 Target Averted Natural Daytime vs. Averted Natural Nighttime 1, 1003 0.30 (0.76)
 Target Averted Modified Daytime vs. Averted Modified Nighttime 1, 1003 2.02 (0.044)
 Target Directed Natural Daytime vs. Directed Modified Daytime 1, 1003 0.96 (0.34)
 Target Averted Natural Daytime vs. Averted Modified Daytime 1, 1003 0.37 (0.71)
 Target Directed Natural Nighttime vs. Directed Modified Nighttime 1, 1003 4.90 (<0.0001)*
 Target Averted Natural Nighttime vs. Averted Modified Nighttime 1, 1003 2.69 (0.0072)*
Small and Upright
 Target Directed Natural Daytime vs. Directed Natural Nighttime 1, 1003 0.27 (0.78)
 Target Directed Modified Daytime vs. Directed Modified Nighttime 1, 1003 8.08 (<0.0001)*
 Target Averted Natural Daytime vs. Averted Natural Nighttime 1, 1003 0.22 (0.83)
 Target Averted Modified Daytime vs. Averted Modified Nighttime 1, 1003 6.92 (<0.0001)*
 Target Directed Natural Daytime vs. Directed Modified Daytime 1, 1003 1.9 (0.057)
 Target Averted Natural Daytime vs. Averted Modified Daytime 1, 1003 0.71 (0.48)
 Target Directed Natural Nighttime vs. Directed Modified Nighttime 1, 1003 10.25 (<0.0001)*
 Target Averted Natural Nighttime vs. Averted Modified Nighttime 1, 1003 7.41 (<0.0001)*
Large and Inverted
 Target Directed Natural Daytime vs. Directed Natural Nighttime 1, 1003 0.36 (0.72)
 Target Directed Modified Daytime vs. Directed Modified Nighttime 1, 1003 10.89 (<0.0001)*
 Target Averted Natural Daytime vs. Averted Natural Nighttime 1, 1003 0.86 (0.39)
 Target Averted Modified Daytime vs. Averted Modified Nighttime 1, 1003 8.71 (<0.0001)*
 Target Directed Natural Daytime vs. Directed Modified Daytime 1, 1003 0.68 (0.50)
 Target Averted Natural Daytime vs. Averted Modified Daytime 1, 1003 1.59 (0.11)
 Target Directed Natural Nighttime vs. Directed Modified Nighttime 1, 1003 11.93 (<0.0001)*
 Target Averted Natural Nighttime vs. Averted Modified Nighttime 1, 1003 9.44 (<0.0001)*

F values are displayed for the overall model and t values are displayed for the comparisons; p-values are indicated in parentheses and statistically significant comparisons are indicated with an asterisk.

During daytime and nighttime viewing, participants were slower to fixate the correct target face when the face had modified sclera compared to naturally-colored sclera (p<0.0001; Table 1: Latency to Fixate Target; Fig 3a). Participants were also slower to indicate they found the target face when the face had modified sclera rather than naturally-colored sclera (p<0.0001; Table 1: Latency to Press Key After Fixate Target; Fig 3b). Participant accuracy in identifying the correct target was similar (mean accuracy above 97%) in the daytime conditions regardless of gaze direction and sclera color (p>0.05; Table 2; Fig 3c). However, participants were least accurate in identifying the correct target face (large and upright, small and upright, and large and inverted) when the sclera matched the iris color (compared to when the sclera was naturally-colored; p<0.0001; Table 2; Fig 3c).

Discussion

During daytime and nighttime conditions, participants were faster at fixating target faces with naturally-colored sclera (white sclera) compared to target faces with modified sclera (sclera color that matched the iris color). The target faces simulated both close-up and distant interactions. In addition, after initially finding the correct faces, participants took less time to indicate their responses (via a key press) when the faces had naturally-colored sclera rather than modified sclera, suggesting that they were most certain of their choices when viewing faces with white sclera [42]. We found similar patterns for the inverted faces [43]. These findings suggest that faces with naturally-colored sclera enhance gaze perception in both daytime and nighttime. These results replicate our previous findings that humans are fastest at fixating and indicating their responses when searching for target faces with naturally-colored sclera (versus faces with sclera that match the iris color) in daytime conditions [19].

While faces with modified sclera limit gaze perception during day and night, this effect was particularly pronounced at night. On average, participants took 1.17 times longer to find the faces with modified sclera during the nighttime versus daytime but only 1.04 times longer to find the faces with naturally-colored sclera during the nighttime versus daytime. Similarly, participants spent 1.94 and 1.09 more time to indicate (via a key press) that they found the faces with modified sclera rather than naturally-colored sclera, respectively, during the nighttime rather than daytime. Even though participants were highly accurate in gaze perception (mean accuracy over 90%), they performed the worst when faces had modified sclera during the night. Previous work has also shown that humans are less accurate in judging gaze direction when faces have darkened sclera color [12, 19]. Our results demonstrate that humans are much slower and less accurate in detecting faces with modified sclera during nighttime compared to daytime. Future studies could investigate how subtle changes in sclera color (i.e., sclera colors that are manipulated in small increments) impact gaze detection in both daytime and nighttime.

Faces with naturally-colored sclera may therefore be especially useful in communicating effectively at night. This finding would be expected when considering human night vision [24, 25, 27]. Humans vision is worse under low-light conditions. In fact, a person with 20/10 vision during daylight has 20/300 vision under starlight [24, 27]. Furthermore, visual acuity also depends on contrast such that visual acuity is better when contrast is higher [24]. Given that naturally-colored sclera exhibit higher contrast (white sclera versus black pupil) compared to the modified sclera (blue/brown sclera versus black pupil), we would expect that naturally-colored sclera are easier to resolve. The ability to effectively communicate at night is likely valuable in many contexts. For example, humans are successful hunters at night (Hames, 1979; Hawkes et al., 1991) and conspicuous sclera may be important during these nocturnal hunts for humans to noiselessly communicate with each other [44, 45].

An important limitation of this study is that it simulated nighttime viewing conditions rather than using actual nighttime viewing conditions. While simulated nighttime conditions cannot perfectly mimic actual nighttime conditions, the simulated nighttime condition allowed us to approximate differences in gaze perception between day and night. The simulated nighttime conditions adjusted for luminance and color sensitivities associated with nighttime viewing [24, 25, 33]. Future studies could investigate gaze perception of people when they are outdoors under varying levels of sunlight and moonlight. The combination of highly-controlled laboratory experiments with naturalistic experiments would provide greater insight into the influence of light levels on gaze perception.

Overall, we provided evidence indicating that eyes with naturally-colored sclera enhance gaze perception during both daytime and nighttime conditions. Eye morphology varies widely across species [59] and we have a limited understanding of how it impacts gaze perception [44, 46]. Furthermore, inter- and intraspecific differences in visual systems could impact gaze perception relative to eye morphology [47, 48]. Additional work that investigates how environmental conditions, such as lighting or air quality, affect gaze perception across species will provide insight into the factors shaping the evolution of eye morphology.

Supporting information

S1 Data

(XLSX)

Data Availability

Supplemental material for this article is available online: Scripts for creating the nighttime images are available here: https://github.com/jyorzinski/CreateNightImages.

Funding Statement

This study was funded by the National Science Foundation (BCS #1926327), the College of Agriculture and Life Sciences at Texas A&M University and Texas A&M AgriLife Research to J.L.Y. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Emery NJ. The eyes have it: the neuroethology, function and evolution of social gaze. Neuroscience & Biobehavioral Reviews 2000; 24: 581–604. 10.1016/s0149-7634(00)00025-7 [DOI] [PubMed] [Google Scholar]
  • 2.Coss RG, Marks S, Ramakrishnan U. Early environment shapes the development of gaze aversion by wild bonnet macaques (Macaca radiata). Primates 2002; 43: 217–222. 10.1007/BF02629649 [DOI] [PubMed] [Google Scholar]
  • 3.Taglialatela JP, Russell JL, Pope SM, Morton T, Bogart S, Reamer LR, et al. Multimodal communication in chimpanzees. American Journal of Primatology 2015; 77(11): 1143–1148. 10.1002/ajp.22449 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Shepherd SV. Following gaze: Gaze-following as a window into social cognition. Frontiers in Integrative Neuroscience 2010; 4: 1–13. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Kobayashi H, Kohshima S. Unique morphology of the human eye and its adaptive meaning: comparative studies on external morphology of the primate eye. Journal of Human Evolution 2001; 40: 419–435. 10.1006/jhev.2001.0468 [DOI] [PubMed] [Google Scholar]
  • 6.Kaplan G, Rogers LJ. Patterns of gazing in orangutans (Pongo pygmaeus). International Journal of Primatology 2002; 23: 501–526. [Google Scholar]
  • 7.Mayhew JA, Gómez JC. Gorillas with white sclera: A naturally occurring variation in a morphological trait linked to social cognitive functions. American Journal of Primatology 2015; 77: 869–877. 10.1002/ajp.22411 [DOI] [PubMed] [Google Scholar]
  • 8.Perea-García JO, Kret ME, Monteiro A, Hobaiter C. Scleral pigmentation leads to conspicuous, not cryptic, eye morphology in chimpanzees. Proceedings of the National Academy of Sciences of the United States of America 2019; 116: 19248–19250. 10.1073/pnas.1911410116 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Perea-García JO. Quantifying ocular morphologies in extant primates for reliable interspecific comparisons. Journal of Language Evolution 2016; 1: 151–158. [Google Scholar]
  • 10.Sturm RA, Larsson M. Genetics of human iris colour and patterns. Pigment Cell & Melanoma Research 2009; 22: 544–562. 10.1111/j.1755-148X.2009.00606.x [DOI] [PubMed] [Google Scholar]
  • 11.Meyer WK, Zhang S, Hayakawa S, Imai H, Przeworski M. The convergent evolution of blue iris pigmentation in primates took distinct molecular paths. American Journal of Physical Anthropology 2013; 151: 398–407. 10.1002/ajpa.22280 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Ricciardelli P, Baylis G, Driver J. The positive and negative of human expertise in gaze perception. Cognition 2000; 77: B1–B14. 10.1016/s0010-0277(00)00092-5 [DOI] [PubMed] [Google Scholar]
  • 13.Langton SR, Watt RJ, Bruce II. Do the eyes have it? Cues to the direction of social attention. Trends in Cognitive Sciences 2000; 4: 50–59. 10.1016/s1364-6613(99)01436-9 [DOI] [PubMed] [Google Scholar]
  • 14.Ando S. Luminance-induced shift in the apparent direction of gaze. Perception 2002; 31, 657–674. 10.1068/p3332 [DOI] [PubMed] [Google Scholar]
  • 15.Ando S. Perception of gaze direction based on luminance ratio. Perception 2004; 33: 1173–1184. 10.1068/p5297 [DOI] [PubMed] [Google Scholar]
  • 16.Jenkins S. Last but not least. Perception 2007; 36: 1266–1268. [DOI] [PubMed] [Google Scholar]
  • 17.Olk B, Symons LA, Kingstone A. Take a look at the bright side: Effects of contrast polarity on gaze direction judgments. Perception & psychophysics 2008; 70(7): 1298–1304. [DOI] [PubMed] [Google Scholar]
  • 18.Anderson NC, Risko EF, Kingstone A. Motion influences gaze direction discrimination and disambiguates contradictory luminance cues. Psychonomic Bulletin & Review 2016; 23: 817–823. 10.3758/s13423-015-0971-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Yorzinski JL, Miller J. Sclera color enhances gaze perception in humans. PLoS ONE 2020; 15(2): e0228275. 10.1371/journal.pone.0228275 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Hames RB. A comparison of the efficiencies of the shotgun and the bow in neotropical forest hunting. Human Ecology 1979; 7: 219–252. [Google Scholar]
  • 21.Hawkes K, O’Connell JF, Jones NG. Hunting income patterns among the Hadza: big game, common goods, foraging goals and the evolution of the human diet. Philosophical Transactions of the Royal Society B 1991; 334: 243–251. 10.1098/rstb.1991.0113 [DOI] [PubMed] [Google Scholar]
  • 22.Worthman CM, Melby MK. Toward a comparative developmental ecology of human sleep (pp. 69–117). Cambridge, MA: 2002; Cambridge University Press. [Google Scholar]
  • 23.Worthman CM. After dark: The evolutionary ecology of human sleep (pp. 291–313). Oxford: 2008; Oxford University Press. [Google Scholar]
  • 24.Ferwerda JA, Pattanaik SN, Shirley P, Greenberg DP. A model of visual adaptation for realistic image synthesis. In Proceedings of SIGGRAPH 1996, pp. 249–258.
  • 25.Thompson WB, Shirley P, Ferwerda JA. A spatial post-processing algorithm for images of night scenes. Journal of Graphics Tools 2002; 7: 1–12. [Google Scholar]
  • 26.Paulun VC, Schütz AC, Michel MM, Geisler WS, Gegenfurtner KR. Visual search under scotopic lighting conditions. Vision Research 2015; 113: 155–168. 10.1016/j.visres.2015.05.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Shaler S. The relation between visual acuity and illumination. Journal of General Physiology 1937; 21: 165–188. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Penteriani V, Delgado MM. Living in the dark does not mean a blind life: bird and mammal visual communication in dim light. Philosophical Transactions of the Royal Society B 2017; 372: 20160064. 10.1098/rstb.2016.0064 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Yokoyama T, Sakai H, Noguchi Y, Kita S. Perception of direct gaze does not require focus of attention. Scientific Reports 2014; 4: 3858. 10.1038/srep03858 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Chelnokova O, Laeng B, Eikemo M, Riegels J, Løseth G, Maurud H, et al. Rewards of beauty: The opioid system mediates social motivation in humans. Molecular Psychiatry 2014; 19: 746–747. 10.1038/mp.2014.1 [DOI] [PubMed] [Google Scholar]
  • 31.Pazda AD, Thorstenson CA, Elliot AJ, Perrett DI. Women’s facial redness increases their perceived attractiveness: mediation through perceived healthiness. Perception 2016; 45: 739–754. 10.1177/0301006616633386 [DOI] [PubMed] [Google Scholar]
  • 32.Lin RL, Ng JS, Nguyen AL. Determinants and standardization of mesopic visual acuity. OPTOM Vision Science 2015; 92: 559–565. 10.1097/OPX.0000000000000584 [DOI] [PubMed] [Google Scholar]
  • 33.Larson GW, Rushmeier H, Piatko C. A visibility matching tone reproduction operator for high dynamic range scenes. IEEE Trans Vis Comput Graph 1997; 3: 291–306. [Google Scholar]
  • 34.Kyba C, Mohar A, Posch T. How bright is full moonlight? Astronomy & Geophysics 2017; 58: 1.31–1.32. [Google Scholar]
  • 35.Yorzinski JL, Platt ML, Adams G. Eye-spots in Lepidoptera attract attention in humans. Royal Society Open Science 2015; 2: 150155. 10.1098/rsos.150155 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Öhman A, Flykt A, Esteves F. Emotion drives attention: Detecting the snake in the grass. Journal of Experimental Psychology: General 2001; 130: 466–478. 10.1037//0096-3445.130.3.466 [DOI] [PubMed] [Google Scholar]
  • 37.LoBue V. And along came a spider: An attentional bias for the detection of spiders in young children and adults. Journal of Experimental Child Psychology 2010; 107: 59–66. 10.1016/j.jecp.2010.04.005 [DOI] [PubMed] [Google Scholar]
  • 38.Penkunas MJ, Coss RG. Rapid detection of visually provocative animals by preschool children and adults. Journal of Experimental Child Psychology 2013; 114: 522–536. 10.1016/j.jecp.2012.10.001 [DOI] [PubMed] [Google Scholar]
  • 39.Yorzinski JL, Penkunas MJ, Platt ML, Coss RC. Dangerous animals capture and maintain attention in humans. Evolutionary Psychology 2014; 12: 534–548. [PubMed] [Google Scholar]
  • 40.Kawai N, Kubo K, Masataka N, Hayakawa S. Conserved evolutionary history for quick detection of threatening faces. Animal Cognition 2016; 19(3): 655–660. 10.1007/s10071-015-0949-y [DOI] [PubMed] [Google Scholar]
  • 41.Benjamini Y, Hochberg Y. Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society 1995; 57: 289–300. [Google Scholar]
  • 42.Brunyé TT, Gardony AL. Eye tracking measures of uncertainty during perceptual decision making. International Journal of Psychophysiology 2017; 120: 60–68. 10.1016/j.ijpsycho.2017.07.008 [DOI] [PubMed] [Google Scholar]
  • 43.Tanaka JW, Simonyi D. The “parts and wholes” of face recognition: A review of the literature. Quarterly Journal of Experimental Psychology 2016; 69: 1876–1889. 10.1080/17470218.2016.1146780 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Ueda S, Kumagai G, Otaki Y, Yamaguchi S, Kohshima S. A comparison of facial color pattern and gazing behavior in canid species suggests gaze communication in gray wolves (Canis lupus). PLoS ONE 2014; 9: e98217. 10.1371/journal.pone.0098217 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Grossmann T. The eyes as windows into other minds: an integrative perspective. Perspectives on Psychological Science 2017; 12: 107–121. 10.1177/1745691616654457 [DOI] [PubMed] [Google Scholar]
  • 46.Tomasello M, Hare B, Lehmann H, Call J. Reliance on head versus eyes in the gaze following of great apes and human infants: the cooperative eye hypothesis. Journal of Human Evolution 2007; 52: 314–320. 10.1016/j.jhevol.2006.10.001 [DOI] [PubMed] [Google Scholar]
  • 47.Higham JP, Hughes KD, Brent LJN, Dubuc C, Engelhardt A, Heistermann M, et al. Familiarity affects assessment of facial signals of female fertility by free-ranging male rhesus macaques. Proceedings of the Royal Society of London B 2011; 278: 3452–3458. 10.1098/rspb.2011.0052 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Troscianko J, Stevens M. Image calibration and analysis toolbox–a free software suite for objectively measuring reflectance, colour and pattern. Methods in Ecology and Evolution 2015; 6: 1320–1331. 10.1111/2041-210X.12439 [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

S1 Data

(XLSX)

Data Availability Statement

Supplemental material for this article is available online: Scripts for creating the nighttime images are available here: https://github.com/jyorzinski/CreateNightImages.


Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES