Abstract
An infant’s everyday visual environment is composed of a complex array of entities, some of which are well integrated into their surroundings. Although infants are already sensitive to some categories in their first year of life, it is not clear which visual information supports their detection of meaningful elements within naturalistic scenes. Here we investigated the impact of image characteristics on 8-month-olds’ search performance using a gaze contingent eye-tracking search task. Infants had to detect a target patch on a background image. The stimuli consisted of images taken from three categories: vegetation, non-living natural elements (e.g., stones), and manmade artifacts, for which we also assessed target background differences in lower- and higher-level visual properties. Our results showed that larger target-background differences in the statistical properties scaling invariance and entropy, and also stimulus backgrounds including low pictorial depth, predicted better detection performance. Furthermore, category membership only affected search performance if supported by luminance contrast. Data from an adult comparison group also indicated that infants’ search performance relied more on lower-order visual properties than adults. Taken together, these results suggest that infants use a combination of property- and category-related information to parse complex visual stimuli.
Introduction
During their first year of life, human infants explore their visual environment in an increasingly selective manner [1, 2], relying on an increasing array of visual properties and category information [3, 4]. Visual abilities develop early and rapidly (e.g., [5]), although some development continues into adolescence (e.g., [6–8]). By about six months of age, basic low-level visual capabilities have emerged which enable infants to distinguish visual pattern within their environment [9–11]. These include grating acuity (i.e., the finest stripes of varying size which can be resolved; e.g., [12]), contrast-sensitivity at higher spatial frequencies (i.e., more narrow changes between light and dark regions; [13–15]), and orientation [10, 16, 17]. These basic functions become more detailed and refined during infancy and early childhood [8, 18–20]. They also provide the basis for higher level visual abilities including the integration of contour segments [21], and the perception of fine detail (i.e., letter acuity; e.g., [22]), all of which support the organization of visual scenes.
Higher order visual competencies, such as visual categorization, also have their onset within the first year of life (e.g., [23–25]; for a review, see [26]) Infants’ attentional deployment is modulated by some categorical distinctions by three months of age (e.g., [23, 24, 27, 28]) and they already show sensitivities to particular naturalistic stimuli in the first year of life, including faces [29, 30] and signals of ancestrally recurrent threats like snakes, spiders, and potentially toxic plants [31–33]). This rapid development occurs in the context of varied and cluttered visual scenes consisting of diverse textures, colors, and lighting gradients. However, the previous literature has typically employed stimuli in which such entities are presented in isolation or in well-delineated ways that are not typical of everyday environments. Frequently, entities are well integrated into their surroundings, like books scattered across a child’s colorful carpet, or fallen leaves on the playground’s sand. The detection of entities within such a scene relies on the ability to perceptually organize its visual information, making additional visual abilities necessary than those necessary for categorization (for a discussion of this problem see [34]). To date, there have been few studies investigating infants’ responses to images of naturalistic environments. These studies point to two key factors that affect infants’ processing of such scenes: These were (i) sensitivity to naturally appearing visual regularities (i.e., statistically assessed vs. manipulated visual properties; [35, 36]), and (ii) sensitivity to entities with ecological significance (e.g., faces, natural scenes; [37–39]). The aim of the current study is to investigate infants’ sensitivity to naturalistic visual information by assessing the effect of different kinds of information (i.e., visual properties, category membership) on visual search performance in a gaze-continent eye-tracking search task.
Visual regularities affecting segmentation of real-world scenes
Similar to adults, infants are likely to orient towards locations of a visual scene that stick out from their surrounding due to cues of low-level salience (e.g., high contrasts in lighting or color; [37, 40, 41]). Next to this sensitivity to low-level visual information, contour integration (e.g., [42]), and texture segregation (i.e., the effortless segregation of texture patches, for a review see: [43]) are seen as a major mechanisms determining the successful visual organization and identification of scene elements [34, 44, 45]. Scene segmentation is necessary to detect a target in a real-world scene and can be seen as a very basic level of categorization. However, the structures of a naturalistic scene differ in various ways, and their segregation might still pose difficulties for immature visual abilities [46]. Indeed, infants’ gaze within their first year of life is still strongly affected by differences in luminance compared to higher-order information such as discontinuities in orientation when viewing artificial stimuli [47] or photographs (e.g., [38, 48]). Still, infants’ ongoing gathering of visual information makes it likely that they also base their visual responses on abstract or complex visual properties of the environment [49]. Infants’ attention to objects or visual patterns is strongly guided by visual information that provides learning opportunities [50, 51]. Attention to the environment’s countless opportunities to receive, organize, differentiate, and accumulate visual information supports the development of visual functions, which are in return adaptive to the visual tasks provided by the environment [22, 29]. This claim is supported by Balas and colleagues [35, 36], who showed that by 9 months of age, infants are sensitive to contrasts between the appearances of naturalistic textures and their statistical transformations. Moreover, infants are surprisingly proficient at processing depth and surface properties [11, 52, 53] suggesting that infants are able to integrate pictorial depth cues into tasks like perceptual organization and action planning [54]. Further, naturalistic entities within a scene, such as natural and human made entities, differ in their visual properties [55–57], and the presence of certain types of entities or general categories with significance to humans may also affect scene segmentation (e.g., [58]).
The significance of category information
Research with adults provides many examples of facilitated processing of characteristics that can be encountered in natural environments. For example, the physiology of the visual system responds efficiently to particular aspects of the distribution of spatial frequencies (i.e., changes between light and dark image regions of different amplitudes) in natural environments (e.g., [59–61]). Faster visual processing of, and increased visual memory for, images depicting natural environments compared to images with manipulated properties or non-natural environments (e.g., [62]) suggests ways in which the human visual system adapted to significant aspects of the environment over evolutionary time (e.g., [63]). For example, adults are very fast at detecting animals compared to non-living objects (e.g., [64, 65]). Moreover, information such as social signals (e.g., human faces, [37, 66]), or signals indicating threat [67] are detected fast and in a privileged way.
Importantly, sensitivity to some of these entities and signals is also already evident during infancy. These include categorical distinctions which refer to the animate-inanimate distinction [68–70], and responses to threat signals from spiders or snakes [31, 32, 71]. Moreover, infants showed distinct behavioral responses (i.e., avoidance, increased social information seeking, enhanced learning) when they were confronted with plants compared to other entities [33, 72, 73]. Vegetation plays an important role for humans such that while it provides both food and raw materials, it can be hazardous (e.g., toxins, thorns; [74]). For this reason, the categorization of plants for subsistence strategies was an integral part of ancestral human life [74–76].
These examples indicate that infants’ reactions to visual stimuli in their first year are increasingly driven by the ecological or cultural relevance of a stimulus’ category [74, 77], suggesting that their visual organization of scenes is also determined by content-related aspects. Nevertheless, it is not yet understood how early sensitivities to certain naturalistic categories relate to the visual properties of these categories. So far, there are only few studies which used real-world scenes to investigate infants’ ability to distinguish significant content-related visual information. They frequently included target stimuli with well defined visual characteristics (e.g., scenes with face-targets: [37–39]). The detection of such iconic cues might rely on other mechanisms (e.g., [32]) than the ability to distinguish between the heterogeneous appearances of particular naturalistic categories.
The current investigation
Here we investigate infants’ sensitivity to particular categories and their visual properties using stimuli representing extracts of real-world human environments. The visual structures depicted on the stimuli differed such that they belong to categories with distinct roles over human evolution: vegetation, non-living natural elements (e.g., rocks, water surfaces), and human-made artifacts. These categories have accompanied humans over different timeframes (e.g., categories of the natural environment vs. the younger category of artifacts; (e.g., [78, 79]), and pose visual tasks with different significance on humans (e.g., vegetation as a substantial source of food [80] vs. non-living natural elements such as stones or water (e.g., their role as landmarks during navigation, [81, 82]).
We tested infants’ ability to distinguish visual structures by comparing the predictive values of different image attributes on their visual search performance. To do this, we conducted a visual search task with 8-month-olds including images of real-world structures to investigate the effect of visual properties and categories on scene segmentation (i.e., the detection of a target structure on a background structure). In contrast to stimuli frequently used in studies of categorization and visual development (i.e., faces, objects, or graphics; for overviews see e.g., [11, 24]; but: [35, 36]), we included photographs depicting homogeneous assemblies of natural entities and artifacts. Such visual structures characterize an important proportion of the human environment and their inherent visual properties are relevant for categorization in adults [56, 57].
We used photographs of our three chose superordinate categories: vegetation, non-living natural elements, and artifacts. We chose these categories because they cover important aspects of human environments that are of ecological and social significance, and have been so over evolutionary time. They differ in the role they played during human evolution, the tasks they imposed to humans, and their presence in modern living spaces. For instance, these three categories are part of either a natural or a manmade world (e.g., [55, 83, 84]), they determine the quality and behavioral affordances of a surrounding (e.g., [85–87]), and they can provide organic or mineral material, represent tools, or provide food (e.g., [74, 79]). Moreover, infants typically have visual contact with a variety of instances of each of these categories which provide learning opportunities for some of their aspects.
We used an eye-tracking visual search task in which infants had to find a patch of an image presented on a discrepant background image. This task allowed us to test whether aspects of the images (i.e., visual properties, category membership) affect detection. Infants received a reward (i.e., a colorful butterfly and sound) when their gaze landed on the visual target patch. The reward was included to stimulate visual search. By the age of eight months, infants are able to perform eye movements in order to trigger a reward [88], and gaze-contingent rewards motivate infants’ search in eye-tracking experiments if there is no clear pop-out effect for the target [89, 90].
We assessed visual properties selected from research on adult visual categorization of naturalistic entities (e.g., [56, 91, 92]). The current selection of visual properties was chosen based on their potential to discriminate between the general categories of the images used in the current and a previous study of Schlegelmilch and Wertz [93]. The selected visual properties include basic statistical properties that were assessed computationally from pixel greyscale values of the stimulus images, and higher-order characteristics that were assessed by adult raters. Statistical and rated properties are both integral parts of visual categorization in adults (e.g., [94]). Statistical properties differentiate low-level characteristics between general categories and surface properties (e.g., [95, 96]), whereas rated properties are based on perceptual judgments and experience, and are processed in higher visual areas [97]. To date, it is not known what role they play for infants when scenes are visually segmented as a first step in categorization. The code is available on https://osf.io/uyg76/?view_only=0b1446f6b6504b7193b58fae3a8cb7a3.
In addition to these more structure-related visual properties, we computed a measure of low-level target salience, namely target-background differences in luminance. Luminance contrasts strongly predict infants’ gaze (e.g., [37]), so we included luminance differences as a control variable in the analysis. Furthermore, we included variables that quantified the perceived dissimilarity of the images assessed with preschool children and adults in the previous study [93]. Importantly, that study showed that dissimilarity judgments were affected by lower- and higher-level characteristics of the image-structures, and to some extent by the individuals’ assumptions about their category membership. Thus, including these dissimilarity judgments allowed us to investigate whether infants’ ability to distinguish real-world structures is related to higher-order perceptual judgments in older age groups.
Taken together, the variables we expected to predict infants’ search performance belonged to three groups: (a) content-related visual information (i.e., category-congruency and dissimilarity judgments), (b) structure-related visual properties (statistical and rated), (c) low-level salience (i.e., differences in luminance; see Table 1 for descriptions of the variables).
Table 1. Definitions of the visual properties.
Name | Definition | Relevance |
---|---|---|
Similarity judgments a | ||
Perceived dissimilarity | Judgments of visual similarity between the images included in target-background image combinations, assessed in sorting tasks with 4–5-year-olds and adults [93]. Transformed to dissimilarity values. | Subjective judgments of the similarity between the images made by preschoolers and adults rely on perception and assumed category membership of the depicted entities [93]. Related to higher-level image characteristics. |
Computational b | ||
Luminance | Mean pixel luminance. | The overall lightness or luminance of a structure. Differences in luminance. |
Alpha | Steepness of the distribution of energy across spatial frequencies (SF) (1/f alpha), referring to the proportion of larger changes to more narrow changes between light and dark image regions. | In natural scenes, alpha values are found to lie in a typical range. The adult visual cortex is tuned to these typical ranges of alpha (e.g., [62]). |
Deviation | Deviation (i.e., area under the curve) of an image’s actual SF distribution from the line fitted to this distribution defined by Alpha [98]. Deviation distinguishes images in which some SF dominate from images with more evenly distributed SF. | Deviation differs between artifacts, plants, and natural scenes [98]. In naturalistic scenes, low values of deviation relate to higher scaling-invariance (e.g., [99]), in that movement towards the scene does not change its SF-distribution. |
Entropy | Shannon entropy of pixel luminance values [100]. | Measure of magnitude and predictability of informational content and differentiation. Low values of entropy refer to only a few shades of grey, whereas high values include more differentiated shades. |
Skew | Skew of the pixel luminance histogram. | Related to impressions of shading and lighting [101,102]. |
Rated c | ||
Curvature | Angular vs. curved. | Perceived curvature supports classification between animate and inanimate objects [92,103]. |
Regularity | Regular vs. chaotic. | Important characteristic for texture and surface discrimination [91,104]. |
Symmetry | Symmetrical vs. asymmetrical. | Symmetry attracts attention in natural scenes (e.g., [105]). Characterizes organic or living things [92]. |
Depth | Plane vs. three-dimensional. | Indicates spatial arrangement of scene elements. Significant for scene segmentation and action planning (e.g., [5]) |
a Assessed in sorting tasks with 4–5-year-olds and adults [93].
b Computational properties were assessed with functions implemented in Matlab (version R2017b) or provided by literature on image processing [106]. The code is available at https://osf.io/uyg76/?view_only=0b1446f6b6504b7193b58fae3a8cb7a3.
c Rated properties were formulated as opposites and judged on a continuous scale by adult participants.
We expected that the selected visual properties could influence infants’ search performance in three non-exclusive ways: a) properties of a background image might hinder the detection of the target if they attract infants’ attention, b) a property that exceeds the visual abilities of the infant can cause the infant to look-away from the stimulus due to self-regulatory attentional processes [107, 108]; c) stronger differences between a visual property in the target patch and the background image might increase detectability of the target.
Our predictions for the current study were that infants would detect a target patch more easily (a) if it depicts a category which is distinct from the background category, rather than belonging to the same category, (b) if the target’s visual properties differ more strongly from the background properties. In addition, infants’ search performance can be better interpreted in comparison to adults performing the same task [109–111]. If different patterns of significant predictors occur in adults than in infants, these differences can be related to the infants’ visual and cognitive abilities. Thus, we added data of an adult comparison group to the analysis who had subsequently performed the same experiment.
Methods
The Ethic Commity of the Max-Planck-Institut für Bildungsforschung has approved the study i2018-06.1, former MatSoC 2018/03 by written consent.
Participants
The final infant sample was N = 39 eight-month-olds (age: M = 8 months, 11 days; range = 8 months, 0 days to 8 months, 29 days; 18 female), recruited from urban and suburban regions of a large European city. We chose 8-month-olds for the current investigation given their successful performance on gaze-contingent search tasks in previous studies [88, 90, 112], and early evidence for distinctions between general categories within the second half of the first year of life [25]. An additional two infants were tested but excluded because no data could be assessed due to problems with the eye-tracker. All infant participants had normal vision.
In response to a helpful comment from a reviewer, we decided to include a comparison dataset of adults (N = 20, age: M = 32 years; range = 23 to 58 years; 12 female) who performed the identical experiment as the infants. All adults had normal or corrected to normal vision by wearing contact lenses.
Participants were recruited from our internal participant database and tested in the Max Planck Institute for Human Development, Berlin, Germany. Parents gave written consent for their child’s participation. Participation was compensated with 10 Euros and infants additionally received a participation certificate.
Stimuli
The 27 images which comprised the search stimuli of the current study were selected from a set of 60 greyscale images used in the study of Schlegelmilch and Wertz [93] that investigated the impact of visual properties on categorization in preschool children and adults. The images depict extracts of real-world structures representing one of the three superordinate categories of either vegetation (e.g., foliage, bark, grass), non-living natural elements (e.g., water surfaces, rocks), or artifacts (e.g., cloth, office supplies; the 27 images used in the study are shown in Fig 1A). Each entity occupies the full size of the image. They were photographed by the first author, or downloaded from license-free online image repositories (pixabay.com, pxhere.com, gettyimages.de).
Fig 1. Creation of the search stimuli used in the study.
(A) The 27 images used in this study, grouped by category, in the format in which they were used as backgrounds in the search task. Within each category, the images are arranged left to right according to decreasing values in rated depth. (B) Examples of the five target patches sampled from each of the 27 images. Targets could appear at one of 10 possible target locations, see Fig 2. (C) Stimuli with moderate target salience were created by placing different target samples at each of the 10 locations, respectively, and comparing the results with a salience algorithm (GBVS; [113]). Targets that were salient (indicated in the two examples by orange to red overlay) without being the only salient region of the stimulus at the respective location were selected. The 261 stimuli identified through this process were converted to three monochromatic colors (green, blue, red) each. From this set, we produced the eight versions of the experiment with 36 test trials each following these restrictions: The stimuli (i) were balanced over target and background categories, (ii) were balanced over the crossed factors category- and depth-congruency, and (iii) no image should appear more than twice as either target or background in any of the eight versions. Taken together, the eight versions of the full experiment included 288 test-trials. The resulting frequencies of the defining factors in these trials are reported in S1 Table in S1 Text.
The stimuli for the search task each consisted of one background image into which a circular patch of a different image was inserted as target. The size of the background image measured 1280 × 1024 pixels, leading to 32° × 25.5° of visual angle during presentation (vis), the target patch measured 235 pixel (6° vis) in diameter. Targets were placed at one of 10 possible locations arranged in a circle with ca. 710 pixel (18° vis) distance to the screen center (Fig 2). Each target had a blurred border to prevent that the circular contour of the target was used as cue. With the same intention, a pattern of blurred circles along the outer contours of the 10 possible target locations was included in each background image. In order to obtain stimuli with moderate target saliency, we applied a salience algorithm to all possible target-background combinations and locations, using the statistical software Matlab (version R2017b, http://www.mathworks.com). We chose the Graph-Based Visual Saliency algorithm (GBVS; [113]) specified for discontinuities in luminance and orientation. GBVS had reflected infant gaze patterns well (for a discussion of salience applied in infant research see: [41]). We then chose target-background combinations in which a target was quantified as at least moderately salient, but not as the only salient region of the stimulus (Fig 1C). Note that the GBVS salience map differs from the computation of the property diff_luminance in that the salience map uses various factors to predict gaze, whereas diff_luminance solely assesses the effect of target luminance on the overall variability of luminance within the image.
Fig 2. The arrangement of target locations and blurred contours.
Central pixel coordinates of the ten possible target locations, in equal distance to the background center (first value: horizontal from top left, second value vertical). The contours around all of the possible target locations—arranged as circles in a ring—were blurred on each background image, and the target patch was included in one of the circles, respectively.
Target patches and background images either depicted the same or a different category, leading to stimuli with congruent categories (e.g., vegetation target on vegetation background), or incongruent categories (e.g., artifact target on vegetation background). In the previous study with preschool children and adults, depth cues were an important predictor of categorization decisions. In order to prevent depth from being a confound in the analysis of category-congruency, we balanced images with high and low rated depth within the respective categories. We then crossed category-congruency (congruent vs. incongruent) with a control variable we termed depth-congruency (similarly high ratings of depth vs. contrasting ratings of depth in the target and the background image). In addition to its role as control variable, depth-congruency contains higher order information about the target-background relationship in depth, and thus complements the property depth, which refers only to the background image.
The computationally-assessed properties were included in our analysis as difference variables: We first partitioned the background images without the target patch as well as the background image including the target patch into squares (size = 256 px by 256 px) which fitted the size of the target patch. Next, we calculated a property’s variance a) between the partitions of the background, and b) between the partitions of the background including the target patch. Finally, we subtracted the background’s variance from the target-plus-background’s variance. This procedure was applied to the 261 stimuli covering all target-background image combinations and their target locations included in the study. The obtained difference variables represented the impact of a target property on the variability of this property in the whole stimulus (termed diff_*property-name*). High values of difference variables were obtained if a target exhibited very high or very low levels of a property, which exceeded the range of the background-image’s property-levels in either direction. Low values of difference variables resulted from backgrounds in which the levels of a property varied between high and low extremes, so that the property level of the target could not substantially increase the background’s variance. If an infant’s detection performance was predicted by a difference variable, the infant must have been sensitive to discontinuities of this property—either within the background image (background difficulty) or between background and target (detection facilitation). Note that the impact of a difference variable does not provide information about an infant’s respective sensitivity for particular high or low levels of the property. In contrast, the visual properties based on human ratings had been assessed for entire images [93], so these ratings could not validly represent the small regions of the images used as target patch. Consequently, we only included the background’s rated properties in our analysis without assessing target-background differences. To make the experiment more engaging for infants, we used three alternating monochromatic colors for the search stimuli. This was done by transforming the greyscale images to HSL color space with the hues: 90° (green), 210° (blue), 330° (red), using the software Adobe Photoshop (Adobe Photoshop CC, Version 2017.0.0). The target and background always shared the same color within a stimulus. Increasing infants’ attention generally reduces movement, increases the periods of recorded gaze and leads to better eye-tracking data quality [114].
In sum, the target-background combinations of 27 images on 10 possible locations and presented in three different colors led to 261 different stimuli that crossed category congruency and pictorial depth congruency. The frequencies of the factors defining the image combinations are provided in S1 Table in S1 Text, their respective visual properties are accessible on https://osf.io/uyg76/?view_only=0b1446f6b6504b7193b58fae3a8cb7a3.
Experimental design and procedure
First, a target sticker was placed on the infant’s forehead and the infant was seated in a dimmed room in front of the eye-tracker (EyeLink 1000 Plus; SR Research Ltd. 2013–2015) either in a baby chair (N = 37) with the caregiver right behind, or on their caregiver’s lap (N = 2). A welcome video was played during the set-up of the eye-tracking camera (EyeLink 1000 Plus High-speed Camera with a 16 mm lens), which was placed approximately 60 cm in front of the target sticker as recommended by the manufacturer [115]. Monocular pupil and corneal reflections were assessed in a sampling rate of 500 Hz. The presentation monitor (50” display, with 1280 by 1024 pixel resolution, and 400Hz CMR refresh rate) was set at a distance of 140 cm away from the infants’ eyes to approximately fit the trackable area of 32° vis by 26° vis in accordance with the manufacturer’s suggestion. After set-up, the experimenter stepped behind a curtain from where the infant and caregiver could be monitored on a video screen and started the experiment.
At the beginning of the experiment and after at least each eighth trial, five-point calibrations were conducted with calibration targets alternating in color, form and sound. Changes in these characteristics increase infants’ interest in the calibration procedure, leading to higher data quality [114]. The trial started when the average calibration error was below 1° vis (see Fig 3).
Fig 3. Trial example.
This example shows one trial with a monochromatic blue search stimulus, and one of the five butterflies that were alternated as rewards. Each trial included a central attention grabber of 5° vis in diameter. As soon as the infant’s gaze rested on its central area (2.5° vis in diameter) for 100 ms, one of the search stimuli was shown for a maximum of 4500 ms. If the infant’s gaze rested on the target patch before timeout for at least 100 ms, rewarding music started to play, then a colorful butterfly loomed out of the target’s center and moved to the center of the screen, and the trial ended. If a target was missed, the butterfly was only shown for a shorter time, accompanied by a neutral sound. Directly after the butterflies disappeared, a new trial started. Every fifth trial, an attention grabber was shown at a peripheral location in addition to the central location. If the infant’s gaze was not recorded within the central region of the attention grabber—for example because of inattentiveness or changes in the distance between the eyes and the eye-tracking camera—an additional calibration was initiated and the camera set-up was corrected if necessary. The length of the full trial varied due to when infants fixated on the attention grabber, and whether they detected the target.
The first five trials of the experiment were practice trials. They started with an easy-to-detect target patch on a simple background and gradually increased in difficulty. Then, 36 test trials were presented in randomized order. In each trial, the color of the stimulus was randomly altered (green, blue, and red). Additional easy-to-detect practice trials were initiated if it seemed that the infant became unaware of the task after several misses without receiving a reward. If the infant became inattentive, showed fatigue, or if the caregiver requested a break, we paused the experiment for a few minutes, or terminated the experiment prematurely. There were eight versions of the experiment that alternated between participants. Each version included a different selection of 36 target-background image combinations taken from our 261 stimulus variants. No target or background image was included more than twice in one version. To avoid memory effects when an image was repeated, its second occurrence was part of a different target-background image combination and used a different target location and color.
Results
Preliminary analysis and data reduction
Infants completed a median of 34 trials (range = 26 to 36 trials), with a median of 88% of gaze recorded by the eye-tracker per trial (range: 1% to 100%). In the following analysis, we included only trials in which infants attended to the stimulus long enough to have the opportunity to detect it. Thus, we decided on distinct criteria for hit and miss trials. This had the advantage of reducing noise in the latency analysis (e.g., [116]), while respecting infants’ attention spans during unsuccessful search in a realistic way. Using an identical proportional inclusion criterion in the hit and miss trials would have unnecessarily excluded trials in which infants may have searched for the target for a considerable amount of time but nevertheless failed to detect it within the 4500 ms trial. We therefore defined our criteria as follows: Trials in which infants detected a target (hit) were accepted if they had a minimum of 80% of recorded gaze. Trials in which a target was not detected (miss) were accepted if they included at least 1240 ms of recorded gaze, which was the median of the hit latency for the whole sample (for studies defining minimum periods of recorded gaze see e.g., [37, 41]). Applying these inclusion criteria led to Mdn = 32 valid trials per infant (range = 23 to 36).
Adults completed 1439 trials (range = 35 to 36). The inclusion criteria that we applied in the adult sample was identical to that of the infants, and led to 1429 valid trials. Due to recruitment issues during the COVID-19 pandemic, we recruited 20 adult participants who performed two runs of different experiment versions each. This led to N = 40 runs and approximately equals the number of all runs presented to the infant sample (N = 39).
Infants detected a target in Md = 39% of the trials, range = 17% to 55%. In contrast, adults detected the target in almost all the trials (Md = 100%, range = 92% to 100%, equating to only 27 misses and 1402 hits). The adults’ ceiling effect was the result of using a trial duration that was originally adjusted to infants. Because of the small number of misses, we only analyzed detection latency for the adult data. Table 2 shows gaze and performance characteristics of infants and adults. Longer fixation durations, later initiation of first saccades, and shorter total durations of gaze spend on the scene are common for infants compared to adults (reviewed in: [117]). We will report the results of the adults together with the infants’ results for each of the respective models. The data of the infant and adult samples is publicly available at (https://osf.io/uyg76/?view_only=0b1446f6b6504b7193b58fae3a8cb7a3).
Table 2. Gaze and performance characteristics of infants and adults.
Infants | Adults | |||||
---|---|---|---|---|---|---|
Md | SD | range | Md | SD | range | |
Hit-ratea | .39 | .1 | [.17, .55] | .98 | .02 | [.92, 1] |
Latency until hit (ms)a | 1471 | 353 | [991, 2127] | 632 | 106 | [238, 4152] |
N fixations until hitab | 8 | 3.1 | [2, 16] | 2.5 | 1.1 | [1, 9.2] |
Rate 1st hit-fixationb | .09 | .06 | [0, .26] | .26 | .18 | [0, .69] |
Proportion of recorded gazec | .84 | .8 | [.6, .97] | 1 | .02 | [.89, 1] |
Mean fixation duration (ms)ab | 467 | 100 | [276,717] | 230 | 25 | [197,311] |
Note. Infants’ data is averaged by id (N = 39), adults’ data by id and run (N = 40).
a Assessed from data included in the analysis, with recorded gaze ≥ .8 in hits, ≥ 1240ms in misses.
b Fixations > 50 ms are included.
c Proportion of registered gaze points within trials in raw data.
Statistical analysis of search performance
In order to investigate the effect of image characteristics on search performance, we assessed the binarily coded dependent variable (DV) success (hit, miss) and the continuous DV latency, which represented the time until a target was detected if it was a hit. These two DVs covered infants’ reactions to aspects of stimulus salience, detection difficulty and background complexity. To account for individual differences, non-normality and unbalanced conditions which are common in infant and eye-tracking data [118, 119], we conducted mixed effect models with the R-package lme4 [120]. For the generalized linear effect models (GLMM) on the DV success, we used the function glmer and specified a binomial error structure. The units of analysis included as random effects on success were participant, background image, and target location. On the DV latency, linear mixed effect models (LMM) were conducted with the function lmer. For latency, the random effects participant and background image were defined, whereas target location did not improve the model fit and was not included (χ2(1) = 0.04, n.s.). Residual and specification diagnostics were carried out with the R package DHARMa [121] and by inspection of residual plots. Influential cases were diagnosed with regard to DFBetas (function influence; R-package lme4 [120]). The significance of predictors was assessed by comparing the current model with a model reduced by the respective predictor in chi-square likelihood-ratio tests (LRT) with the R-function Anova (package car; [122]).
To avoid problems of interdependencies between IVs (see e.g., [123]), we reduced the number of IVs in each comparison by conducting separate models for different research questions (e.g., the impact of computationally assessed visual properties). For these models, we estimated the effect of collinearity by Variance Inflation Factors (VIF; [124]) with the function vif (R-package car; [122]) and only combined IVs in models if VIF values remained below 2.5.
Effects of covariates
We tested if stimulus color [green, red, blue] generally affected search performance. The factor color [red, green, blue] did not predict infants’ search performance (success: χ2(2) = 1.6, n.s.; latency: χ2(2) = 3.7, n.s.), nor performance of adults (all χ2(2) < 1.4, n.s.), confirming that the colors we chose to enhance infants’ interest in the study did not lead to differences in the detectability of the target, see also Section B in S1 Text.
Because movement during remote-mode eye-tracking substantially affects data quality [114, 125], we calculated the covariate movement as the maximum of absolute change in head-camera distance within fixations during each presentation of a search stimulus (for details see Section A in S1 Text). Movement was included as a covariate in all models.
In our analysis, we were interested in the impact of the predictor variables beyond target-background differences in luminance. We therefore included diff_luminance as fixed effect in all models. We also included the interaction term between diff_luminance and the other predictor variables if it significantly improved the model compared to a model with fixed effects only, as assessed in a LRT (R function anova, package stats; [126]). Significant interactions between diff_luminance and another predictor variable indicate that the other predictor’s impact on target detection is affected by the targets’ high luminance contrast.
The two runs of the experiment executed by the adult comparison sample included different stimuli. Increasing experience with the task might have led to better detection performance in the second run. We therefore conducted an analysis of the effect of the factor run [1, 2] together with the covariates diff_luminance and movement on adults’ detection latency. This showed that run did not significantly contribute to the model fit (χ2(1) = 0.8, p = .378), whereas the covariates diff_luminance (χ2(1) = 7.1, p = .008) and movement (χ2(1) = 135, p < .001) both affected latency. This confirms that performance between the two runs did not differ in the adult data. Yet, in order to rule out any practice-effect, we nevertheless included the factor run as additional covariate in the adult models.
The impact of content-related visual information on detection performance
Category-congruency
Here, we investigated whether differences between the background category and the target category affected search performance. Category-congruency and the control variable depth-congruency were included in the models as predictors.
Infants’ GLMM on success was improved if it included the interaction terms between category- and depth-congruency and diff_luminance compared to the same model with only main effects, Δχ2(1) = 8.4, p = .004. LRTs on success indicated significant contributions of the fixed effect category-congruency (χ2(1) = 7.3, p = .007) and the interaction term between category-congruency and luminance (χ2(1) = 10.5, p = .001). Congruent categories led to a higher probability to detect a target if combined with greater diff_luminance. Incongruent categories led to better detection performance than congruent categories if the full range of diff_luminance is taken into account, and they were affected less by differences in luminance. Therefore, incongruent categories differed more strongly from congruent categories when combined with greater than with lower differences in luminance, see top row in Fig 4C.
Fig 4. Infants’ and adults’ performance as functions of target-background congruency and image dissimilarity.
Stimuli examples (A, B) and marginal effects of the models (C, D) conducted with independent variables relating to categories and higher-level image characteristics. (A) Category-congruency and depth-congruency of target and background images. Two out of the four possible combinations are shown in greyscale. The stimuli alternated between three monochromatic colors during presentation, see Stimuli paragraph in Methods section. (B) Stimuli including target and background images with low or high dissimilarity, judged by preschool children and adults in a previous study [93]. (C) Top: Interactions between diff_luminance and the congruency variables, diff_luminance low = 5%, high = 95% percentile. High diff_luminance is related to higher detection success for target categories that were congruent with the background categories (red), compared to incongruent categories (blue), whereas category congruency does not affect performance if differences in luminance are low. Depth congruency was not significantly affected by diff_luminance.
In the LMM on latency, category-congruency did not contribute to the model, χ2(1) = 3.6, p = .059. The control variable depth-congruency neither improved the model fit on success (main effect: χ2(1) = 0, p = .960; interaction: χ2(1) = 3.3, p = .069), nor on latency, χ2(1) = 2, p = .160, Fig 3 and Table 3.
Table 3. Results for congruency and image dissimilarity.
Property | GLMM on success | LMM on latency | ||||||
---|---|---|---|---|---|---|---|---|
Log-Odds | 95% CI | z | p a | b (ms) | 95% CI | t | p a | |
Infants | ||||||||
Category-congruency | ||||||||
Diff_luminance | 0.22 | [-0.22, 0.65] | 0.99 | .322 | -88 | [–193, 17] | -1.64 | .103 |
Category-congruency | -0.61 | [-1.05, -0.17] | -2.70 | .007 | -157 | [–320, 6] | -1.89 | .059 |
Depth-congruency | 0.01 | [-0.47, 0.45] | -0.05 | .960 | -130 | [–310, 51] | -1.41 | .159 |
Diff_luminance: Category-congruency | 0.83 | [0.33, 1.32] | 3.25 | .001 | ||||
Diff_luminance: Depth-congruency | 0.49 | [-.04, 1.02] | 1.82 | .069 | ||||
Image dissimilarity | ||||||||
Diff_luminance | 0.71 | [0.44, 0.98] | 5.13 | < .001 | -90 | [-192, 13] | -1.73 | .085 |
Child_dissimilarity | 0.01 | [-0.17, 0.16] | -.08 | .941 | -89 | [-174, -3] | -2.04 | .042 |
Adult_dissimilarity | 0.11 | [-0.06, 0.27] | 1.28 | .199 | -7 | [-91, 76] | -0.17 | .863 |
Adults | ||||||||
Category-congruency | ||||||||
Diff_luminance | -- | -- | -- | -- | -.51 | [-95, -6] | -2.23 | .026 |
Category-congruency | -- | -- | -- | -- | -12 | [-63, 39] | -0.46 | .646 |
Depth-congruency | -- | -- | -- | -- | 137 | [155, 217] | 11.82 | < .001 |
Image dissimilarity | ||||||||
Diff_luminance | -- | -- | -- | -- | -62 | [-107, -17] | -2.68 | .007 |
Child_dissimilarity | -- | -- | -- | -- | -15 | [-43, 12] | -1.08 | .281 |
Adult_dissimilarity | -- | -- | -- | -- | -48 | [-75, -20] | -3.44 | < .001 |
a P-values obtained by chi-square likelihood-ratio tests.
The adults’ LMM on latency was not improved by the inclusion of an interaction term between diff_luminance and the congruency variables, Δχ2(2) = 4.9, p = .084. LRTs indicated a significant contribution of depth-congruency (χ2(1) = 21.2, p < .001), whereas category-congruency did not contribute to the model (χ2(1) = 0.2, p = .646), Fig 4C, Table 3.
High-level image dissimilarity
With the continuous variables child_dissimilarity and adult_dissimilarity, which were taken from ratings of the target and background images in a previous study [93], we examined whether detection success was influenced by the higher-level similarity judgments of older age groups.
In the infants’ GLMM on success, neither child_dissimilarity nor adult_dissimilarity improved the model fit, as indicated by the LRT (both χ2(1) < 1, n.s.). However, infants’ detection latency was predicted by child_dissimilarity (χ2(1) = 4.2, p = .042), but not by adult_dissimilarity, indicating that infants were faster at detecting targets if they were perceived as more dissimilar by preschoolers, but not by adults, Fig 4D top rows.
In adults, the variable adult_dissimilarity improved the model fit on latency (χ2(1) = 11.8, p < .001), whereas child_dissimilarity did not contribute to the model (χ2(1) = 1.2, p = .281), Fig 4 and Table 3. Adults were faster at detecting the target if adults in another study [93] had judged it as more dissimilar to the background, whereas preschoolers’ judgments from that study did not affect adults’ detection.
The effect of visual properties on detection performance
Target-background differences in computational properties (i.e., diff_deviation, diff_alpha, diff_entropy, and diff_ skew) and rated background properties (i.e., curvature, depth, regularity and symmetry) were analyzed in separate models. None of the models included interaction terms with diff_luminance. This led to four analyses conducted to assess the impact of visual properties on infants’ search performance.
In the infant GLMM of computational properties on success, diff_luminance contributed to the model fit with χ2(1) = 19.8, p < .001. Of the structure-related predictors, only diff_deviation contributed with (χ2(1) = 22.2, p < .001), in that higher values of both variables lead to a higher probability to detect the target. Latency was predicted by the structure-related property diff_entropy, which contributed to the fit of the LMM with χ2(1) = 8.5, p = .004. Stronger target-background differences in diff_entropy led to a faster detection of the targets, see Fig 5 and Table 4.
Fig 5. Structure-related visual properties and their effect on search performance.
(A) Examples of search stimuli with the respective low and high values of visual properties that contributed significantly to any of the models. (B) Visual properties as functions of search performance and participant group, estimated as marginal effects in the models. Asterisks indicate significant contributions of the variable to the full models, see Table 4. Diff_luminance is estimated as single variable together with the covariates of the respective model groups (i.e., infants: movement; adults: movement, run).
Table 4. Results of structure-related properties that predicted detection performance.
Property | GLMM on success | LMM on latency | ||||||
---|---|---|---|---|---|---|---|---|
Log-Odds | 95% CI | z | p a | b (ms) | 95% CI | t | p a | |
Infants | ||||||||
Computational target-background difference b | ||||||||
Diff_luminance | 0.62 | [0.35, 0.90] | 4.45 | < .001 | -42 | [-148, 65] | -0.76 | .446 |
Diff_alpha | -0.14 | [-0.33, 0.05] | -1.4 | .160 | -2 | [-93, 90] | -0.04 | .968 |
Diff_deviation | 0.55 | [0.32, 0.78] | 4.71 | < .001 | -37 | [-124, 49] | -0.84 | .402 |
Diff_entropy | 0.19 | [-0.06, 0.45] | 1.5 | .134 | -128 | [-214, -42] | -2.92 | .004 |
Diff_skew | 0.09 | [-0.1, 0.28] | 0.92 | .356 | -42 | [-118, 34] | -1.08 | .283 |
Rated background property | ||||||||
Diff_luminance | 0.70 | [0.44, 0.96] | 5.28 | < .001 | -95 | [-189, -1] | -1.97 | .054 |
Curvature | 0.07 | [-0.2, 0.34] | 0.51 | .611 | -11 | [-117, 95] | -0.2 | .844 |
Depth | -0.31 | [-0.61, -0.01] | -2 | .046 | 187 | [77, 297] | 3.38 | .004 |
Regularity | 0.19 | [-0.22, 0.6] | 0.9 | .370 | -66 | [-220, 88] | -0.84 | .412 |
Symmetry | 0.11 | [-0.28, 0.5] | 0.54 | .590 | 128 | [-20, 276] | 1.7 | .106 |
Adults | ||||||||
Computational target-background difference b | ||||||||
Diff_luminance | -- | -- | -- | -- | -32 | [-78, 14] | -1.36 | .175 |
Diff_alpha | -- | -- | -- | -- | -31 | [-62, 0] | -1.98 | .048 |
Diff_deviation | -- | -- | -- | -- | -16 | [-54, 22] | -0.84 | .403 |
Diff_entropy | -- | -- | -- | -- | -25 | [-65, 16] | -1.2 | .231 |
Diff_skew | -- | -- | -- | -- | -22 | [-54, 11] | -1.34 | .187 |
Rated background property | ||||||||
Diff_luminance | -- | -- | -- | -- | -44 | [-88, -1] | -2.02 | .044 |
Curvature | -- | -- | -- | -- | -33 | [-97, 31] | -1.02 | .320 |
Depth | -- | -- | -- | -- | 55 | [-16, 125] | 1.52 | .143 |
Regularity | -- | -- | -- | -- | -37 | [-133, 60] | -0.74 | .46 |
Symmetry | -- | -- | -- | -- | -48 | [-140, 43] | -1.04 | .312 |
Note. Visual properties were included together with the covariate movement as fixed effects.
a P-values obtained by chi-square likelihood-ratio tests.
b Assessed as difference between the properties’ variance within the background image alone and within the background including the target patch, see Stimuli in Method section.
In the infant GLMM on success including the rated background properties, diff_luminance contributed to the model with χ2(1) = 27.9, p < .001 and depth affected the model fit with (χ2(1) = 4, p = .046), in that higher values of diff_luminance, but lower values of background depth, lead to a higher probability that a target was detected. Depth also contributed to the fit of the LMM on latency (χ2(1) = 11.1, p < .001), with higher values of depth leading to longer detection latencies.
LRTs indicated that no other visual properties affected infants’ detection performance, see Table 4 for all results and Fig 5 for stimuli examples of significant properties.
Adults’ latency was predicted by the structure-related property diff_alpha in the LMM including computational properties (χ2(1) = 5, p = .024) such that stronger target-background differences in alpha led to a faster target detection. In the LMM on latency with rated properties, the control variable diff_luminance was the only variable that contributed to the model fit (χ2(1) = 4.1, p = .043), see Fig 5B and Table 4.
Did infants detect the targets by coincidence?
In order to investigate whether infants may have fixated on the targets by coincidence, we compared the number of fixations during the presentation of the search-stimulus on each of the 10 possible target locations without a target to the number of fixations on the target. We considered fixations to the 10 locations on which targets could occur to be areas of interest (AOIs) and not all fixations because infants might have learned that targets are located a certain distance from the screen center. A GLMM specified for count data on the numbers of fixations with the predictor location (the target contrasted to the 10 AOIs) and participant as random effect indicated that there were less fixations to any AOI without a target than to the target itself, LRT on the IV location (χ2(10) = 599, p < .001), all contrasts p < .005.
We then compared the number of first fixations on the target to the mean number of first fixations on any of the 9 AOIs without the target. First fixations to targets occurred more frequently (Md = 3, range = 0–8) than first fixations to non-target AOIs on average (Md = .89, range = .22–2.56), as confirmed by a t-test (t(44) = 6.81, CI(95%) = [-2.8, -1.5], p < .001). The proportion of first fixations on the target, relative to the number of first fixations to the 10 possible target locations is shown in Fig 6, in addition to the chance-level of hitting any specific one of the 10 locations (p = .1, red line). Only three of the 39 infants fell below chance-level. These results confirm that infants’ target detection was non-accidental.
Fig 6. First fixations on areas of interest including the target.
Points indicate the rate of an infant’s number of first fixations on targets relative to the number of the first fixations on any of the 10 possible target locations (AOIs), independent of their including the target or not (rates: Md = .25, sd = .16, range = 0 to .67). The red line indicates the chance-level of a first fixation on a particular target location, if only fixations within the AOIs are considered. Points are scattered sideways to avoid overlap.
Discussion
Here we investigated which image characteristics (i.e., category, luminance, structure-related visual property) affected 8-month-olds’ ability to detect a discrepant image patch on a complex background image using a gaze-contingent eye-tracking search task. The images depicted one of the three superordinate categories: vegetation, non-living natural elements or artifacts.
Our results indicate that infants attended to combinations of higher- and lower-level visual image characteristics to distinguish complex naturalistic patterns. Consistent with the previous literature (e.g., [37]), detection performance was affected to a large extent by differences in luminance. However, going beyond the results of previous studies, we found that structure-related visual properties of the images, such as deviation, entropy, and rated depth, predicted detection performance independently of luminance. Furthermore, judgments of image-dissimilarity by preschoolers—but not by adults—predicted infants’ detection performance. Yet, the impact of categorical information on infants’ detection performance depended on the stimulus’ target-background difference in luminance.
In the current study, targets were detected non-accidentally, indicating that infants learned to search for the targets and were able to direct their attention to discontinuities in the appearance of the structures. The current findings differ from earlier eye-tracking search tasks with infants (e.g., [39, 89]) in that targets did not represent a delineated object. Instead, photographs of complex naturalistic surfaces or assemblies of elements alternated as targets and backgrounds. A target was only defined by it being a discrepant structure patch to the background and by leading to a visual reward if looked at. Therefore, our results are relevant for research on image segmentation and visual search (e.g., [49, 127, 128]).
Structure-related visual properties affected detection performance
Target-background differences in deviation and entropy explained a similar or even higher amount of variance in detection performance than differences in luminance (R2marginal = .042 for diff_deviation vs. R2marginal = .019 for diff_luminance on success; R2marginal = .033 for diff_entropy vs. R2marginal = .003 for diff_luminance on latency; R-function r.squaredGLMM, package MuMIn; [129]). This is intriguing because infants’ attention at this age is strongly affected by luminance (e.g., [37]) and several aspects of these visual properties may have facilitated infants’ detection of the discrepant target structure.
It is possible that differences in the amount of shades of grey, and in the amount of spatial scales that vary in the properties entropy and deviation, affected infants’ detection performance. A structure defined by high deviation values is dominated by only some spatial frequency scales, providing similarly shaded regions of repetitive sizes, whereas a structure defined by low deviation includes all spatial frequency scales. Structures with high values of entropy include similar numbers of each of the 256 possible shades of grey of an 8 bit image, while in low entropy high proportions of only some shades lead to less differentiated shading or more monotonous structure regions. Accordingly, greater differences in deviation and entropy between the target and background were associated with more fine-grained, cluttered patterns and differentiated contrasts, versus less detailed, smoother, or repetitive image regions. It is interesting that infants were sensitive to these image properties because infants’ immature processing of fine detail and lower sensitivity to contrast may lead to uncertainty and make it difficult for them to detect variability in visual structures that differ in these ways. Sensitivity to uncertain visual information might nevertheless be beneficial for the infant for various reasons. It can lead to strategic behavioral reactions such as further exploration, avoidance, or social referencing (e.g., [27, 77]) and support basic distinctions of significant categories. Sensitivity to uncertain visual information might also underlie infants’ novelty preference [29] and in young children’s choice of actions that resolve the greatest amount of uncertainty [130–132]. For example, if infants move towards a visual structure with low deviation, the distribution of small to large spatial pattern will remain the same due to its scaling invariance. It is plausible that infants become sensitive to such phenomena without being fully able to process every visual detail. Sensitivities to these phenomena are particularly advantageous, since differences in scaling invariance distinguish important general categories (e.g., artifacts vs. natural elements; [98]). Perhaps due to these adaptive and explorative behaviors, infants are sensitive to visual properties that varied in their amount of informational uncertainty and informational complexity, and attended to discrepancies within these properties in the current task. This would explain how greater differences in target-background deviation and entropy led infants to detect the discrepant target patch.
However, one may then ask why greater differences in other computational visual properties did not affect infants’ target detection performance. For example, alpha represents statistical aspects of naturalistic scenes, and different values of alpha have even been found to affect adults’ processing speed, recognition, and visual memory (e.g., [62, 98, 133, 134]). And indeed, adults’ detection latencies in the current study were shorter if targets differed more strongly from the background in alpha in the current search task. Nevertheless, in the current study, alpha did not predict infants’ detection performance (Table 4). Similarly, certain ranges in alpha did not predict judgments made by younger children, but did so for older children and adults [46]. It is possible that variations in alpha are much more difficult to distinguish for young children, because alpha defines proportions of spatial scales, but not the range of the spatial scales included in the structure. The discrimination of variations in alpha makes sensitivity to the full spectrum of spatial scales necessary [46] and can be seen as higher-order statistic compared to deviation.
Depth cues, but not shape predicted detection performance
We found that high rated depth of the background image hindered infants’ target detection performance. Infants start to be sensitive to stereoscopic depth and to pictorial depth within a few months after birth [11, 135, 136]. Nevertheless, the dark regions and high contrast contours which are typical for shading-defined pictorial depth might have diverted infants’ gaze and complicated the detection of the target. Adults’ detection performance was not hindered significantly by higher background depth. Instead, adults took advantage from depth-congruency, in contrast to infants whose performance was not affected by depth-congruency. Recall that this variable stated if the level of rated depth in the target image was the same as in the background image. In contrast to background depth, the visual processing of depth-congruency requires comparisons between pictorial depth cues within image regions, which may have been beyond infants’ visual abilities. The reasons why depth affected target detection may therefore be different in infants and adults. A photographic representation of complex three-dimensional arrangements might challenge an infant’s perceptual abilities, and could potentially lead to either disengagement from the task or the search for further information [3, 137]. Thus, an alternative explanation is possible: Infants did not disengage from the depth cues because spatial characteristics of scene elements or their arrangement provided opportunities for further visual exploration (e.g., [138]) and significant attentional learning processes [3, 4] that were more rewarding than searching for the target.
Interestingly, none of the other rated properties—which were all related to shape characteristics or their arrangement within the background image—affected target detection performance in infants or adults. One might argue that higher ratings of these properties were as interesting or demanding for the infants as the lower ratings. Yet, this explanation is unlikely, because (i) curved shape is preferred to angular shape very early in life [29], (ii) symmetry is processed in a basic way by 1-year-olds and may attract attention because it provides learning opportunities [139], and (ii) repeated elements which represent regularity are reliably used as backgrounds in infant search paradigms (e.g., [89]). As with pictorial depth, these findings were obtained with graphical stimuli and therefore may not fully describe the effect of these properties on infants’ visual abilities if they are present in complex naturalistic scenes.
The current results show that depth cues can play an important role in scene perception and segmentation. However, it remains unclear why the particular shape properties defining complex naturalistic structures we assessed in this study did not affect detection performance in either of the participant groups.
Infants’ visual search relates to preschoolers’, but not adults’ similarity judgments
Remarkably, judgments of image dissimilarity made by preschoolers, but not adults, in a recent study [93] predicted infants’ search performance. This raises the question of which aspects of structure perception are shared between infants and preschoolers, but are less relevant for adults.
The perception of complex naturalistic structures is a hierarchical organizational process and relies on several perceptual mechanisms: smaller elements can be grouped, compared, segregated, or perceived in their configural relations, leading to more global elements with which the operations can be repeated [140, 141]. Some visual abilities that are important for structure perception—such as spatial acuity, contrast sensitivity, and particularly the higher-order ability of perceptual integration—are only adult-like after the preschool years [6, 8, 15].
In the current search task, infants’ detection performance was less affected by variables that relied on higher-order processing, such as alpha or depth congruency, whereas these properties enhanced performance in the adult participants. Likewise, during the sorting task from which the dissimilarity judgments were derived, preschoolers more reliably attended to properties that required little hierarchical processing, while adults integrated properties of several hierarchical processing levels into their judgments [93]. Neuroimaging studies have found an increase of horizontal intra- and interhemispheric connectivity at least until 13 years of age, as well as increasing feedback connectivity from extrastriate visual areas to V1—changes which are thought to be involved in the detection, grouping and spatial integration of distributed visual elements [15, 142, 143]. It is therefore likely that immaturities in perceptual integration that contribute to difficulties processing complex visual images are an important overlap between infants and preschoolers.
Another overlap could be that preschoolers’ perceptual judgments relied on their experiences with the structures which are likely related to different aspects of the entities than those of adults. For example, young children’s relation to their environment is to a large extent guided by explorative actions. These are described as the primary behavioral mechanism for generating perceptual information [144–146]. In contrast, adults actions more strongly refer to the context or usage of the entities (e.g., [79]). Moreover, perception of infants and young children before school age is less affected by cultural norms than perception in older children and adults [147]. Such fundamental differences in the gathering of experiences very likely affect perception in young children differently than that of adults. Next to related immaturities in visual abilities, overlaps in a learning- and exploration-driven relation to the environment between preschoolers and infants might therefore have caused the predictive value of preschoolers’, but not adults’ judgments on infants’ performance. With this, the finding provides an example for age-dependent variations in the significance of visual aspects of the environment.
Did categorical information affect infants’ detection performance?
One of our main interests was whether certain superordinate categories that have significance for humans—vegetation, artifacts, and natural elements—affect scene-segmentation in infancy. However, we did not find clear evidence that differences in these categories between target and background images affected target detection or segmentation ability in infants or adults. Infants’ search performance was only affected by category information that was supported by differences in luminance. Detection probability of a target was higher if it belonged to a category which was congruent to the background category and also differed more strongly from the background in luminance. In contrast, incongruent category combinations were less affected by differences in luminance, but led to better detection success than congruent category combinations if the full range of luminance differences is considered, Fig 4C. In adults, luminance differences did not interact with categorical information, and there were no differences in detection latency between congruent and incongruent category combinations.
It is difficult to separate visual properties from category information. Accordingly, we do not see categories as independent or opposite from visual properties—categories must necessarily be defined by a set of properties. In order to rule out biases in the stimuli that might have led to a facilitated detection of congruent or incongruent categories in infants, we analyzed if any of the selected visual properties differed between category-congruent and category-incongruent stimuli. Table 5 compares the visual properties of congruent and incongruent category combinations, respectively, within the stimuli of all trials that were presented to the participants in the eight versions of the experiment. The comparison shows that none of the visual properties included in the current analysis differed between the congruent and incongruent category combinations.
Table 5. Visual properties as function of category-congruency.
Variable | M congruent a | M incongruent a | d | t b | p b |
---|---|---|---|---|---|
Diff_luminance | 0.70 | 0.61 | -0.11 | 0.89 | .373 |
Diff_alpha | 0.50 | 0.42 | -0.90 | 0.74 | .458 |
Diff_deviation | 0.47 | 0.50 | 0.03 | -0.25 | .802 |
Diff_entropy | 0.48 | 0.35 | -0.15 | 1.16 | .248 |
Diff_skew | 0.09 | 0.18 | 0.09 | -0.77 | .444 |
Curvature | -0.01 | -0.08 | -0.08 | 0.60 | .546 |
Depth | 0.27 | 0.31 | 0.04 | -0.30 | .767 |
Regularity | -0.19 | -0.27 | -0.09 | 0.69 | .492 |
Symmetry | -0.05 | -0.15 | -0.10 | 0.83 | .408 |
a Stimuli with congruent (N = 96) or incongruent (N = 192) target-background category combinations.
b T-test between category-congruent and -incongruent stimuli.
Note. The analysis refers to N = 288 data points distributed over 36 test-trials in the respective 8 versions of the experiment.
A similar supporting role of luminance differences for infants’ detection of categorical information was found in previous studies investigating the effect of low-level salience on the detection of faces. In those studies, stronger luminance contrasts within the face target facilitated its detection in infants younger than 1 year of age ([39], but see: [37]) and stronger luminance contrasts between competing stimuli hindered the detection of the faces in 4-month-old infants [41]. Fig 4C shows that—compared to category-incongruent stimuli—category-congruent stimuli profited much more from high luminance contrasts, whereas low luminance contrasts only resulted in a minor disadvantage. Thus, the current findings do not provide evidence for a better distinction of structures belonging to a different general category, as we had expected based on the existing literature on categorization in infants (e.g., [39, 68, 69, 72]).
One reason for this may be because evidence for sensitivity to particular content-related visual information in photographic scenes comes from studies with iconic and object-like targets. In these studies, targets were commonly the only exemplar of their general category in the context of a different general category (e.g., landscapes including an animal, interiors surrounding a person’s face or body; [39, 64]). Visual abilities underlying the detection of a particular structure patch (as in the current study) may be somewhat distinct from those underlying the detection of delineated objects. For example, hierarchical perceptual integration and grouping mechanisms play a stronger role in the perceptual organization of structures compared to bounded objects (e.g., [141]).
Limitations and future questions
It cannot be ruled out that infants used cues to detect the targets beyond those analyzed. For example, they might have learned to associate rewards with round areas of a certain size—despite the efforts to hide the contours of the targets. This might have led them to preferably fixate round salient patches, leading to faster detection if the round patch actually included the target. However, we think that such cues did not strongly alter search performance, since otherwise, rated background curvature would have affected detection significantly, and it did not.
The current visual properties only represent a small selection of many possible visual properties that could play a role in the perception and segmentation of naturalistic structure. It cannot be ruled out that some properties that were not assessed in the current study affected detection performance. Still, we think that the current selection of visual properties highlighted several important aspects of visual development (e.g., action or exploration related attention, differences in processing effort between lower and higher order statistics), and included important characteristics of the environment that affected search performance (e.g., their distribution of spatial scales and contrasts). Future investigations should build on some of the current findings. In particular the effect of an increase in pictorial depth of the background image on infants’ detection performance needs to be evaluated more deeply. Teasing apart the roles of attention or explorative reactions to functional spatial information on the one hand, and distractions due to shading and contrast on the other hand, would deepen insight into the impact of functional significance on early vision development.
The monochromatic stimuli we used allowed us to focus on structure-related visual properties. If the stimuli would have maintained their original colors, target detection would have been dominated by color differences. However, the effect of category information on search performance might have been affected by this decision: Color cues provide important information supporting the discrimination of significant categories [56, 148].
With regard to the ambiguous influence of category information on the segmentation of visual scenes, it would be interesting conduct additional investigations with different age groups using a similar search task. Target-background combinations of different superordinate categories could be compared to combinations of sub-groups within particular superordinate categories. By also examining the interplay of luminance contrasts and category contrasts this search task would extend developmental research on visual salience (e.g., [37, 48, 149]).
One key question concerning infants’ visual processing of naturalistic stimuli is whether basic visual abilities, which are typically assessed with graphic stimuli, transfer to naturalistic scenes. In adults, visual responses to naturalistic stimuli exceed the performance that can be expected from responses to graphic stimuli [150]. In future studies, it might therefore be useful to add psychophysical search stimuli to the experiment that vary in contrast and fine detail. An assessment of the individual infant’s low-level visual abilities would allow to better uncover if sensitivity to content- or structure-related properties during the segmentation of naturalistic scenes correlates with the ability to react to fine detail in abstract stimuli.
Conclusion
The current study revealed that 8-months-olds were able to search for discontinuities in photographs depicting naturalistic surfaces or assemblies of elements. The task to search for a discrepant target image patch was solely defined by a gaze-contingent reward. Infants detection performance was primarily affected by luminance differences and by structure-related visual properties such as scaling invariance (i.e., deviation) and entropy. Differences in category membership did not affect search performance independent of luminance differences.
The pattern of results suggests that infants’ gaze was largely guided by statistical properties. Higher-order visual information such as pictorial depth cues hindered detection. Possibly, opportunities to explore this significant property increased infants’ distraction by high pictorial depth. Visual properties affecting infants’ detection performance differed from the properties affecting performance of an adult comparison group. For example, in contrast to the adult participants, infants’ gaze was not affected by the statistical property alpha or by the level of rated depth between target and background image. Additionally, infants’ detection performance was predicted by preschool children’s but not adults’ perceptual judgments of the same image combinations, whereas adults’ detection latency was only predicted by the dissimilarity judgments of adults. Thus, maturing visual abilities necessary for naturalistic structure perception such as the perceptual integration of distributed elements seems to affect the perception of these naturalistic structures in infants.
Infants were sensitive to variations in the amount of grey scales defining a structure, and to variations in spatial frequency distributions indicating scaling invariance (assessed by the property deviation). These properties typically differ between artifact categories and significant domains of natural categories (i.e., vegetation, natural elements [59, 62, 98]). Variations in these visual properties include visual information that exceeds the infants’ low-level visual abilities [6, 8, 46]. The sensitivity to variability within these properties we found in infants may be related to the importance of these properties in distinguishing the significant category domains. However, given the many reasons why infants might attend to one visual cue but not another—including exploration and visual learning—the significance of visual information applies to more than particular entities or category domains. The current study showed that infants’ visual abilities allow them to perceptually organize complex structures within their environment by reacting to visual information, even if it was uncertain or incomplete. Thus, visual aspects or physical qualities that are part of an infant’s developmental tasks in supporting his or her interaction with the environment may be as important as particular entities or categories.
In congruence with other studies using naturalistic images in controlled laboratory settings (e.g., [36, 38, 39]) we argue that the inclusion of naturalistic images in infant vision research is important and might lead to different results than research with artificial objects or graphic stimuli.
Supporting information
A. The effect of movement on detection performance. B. Did changes between the differently colored stimuli affect infants’ detection performance? S1 Table: Number of trials per factor level, in the original experiment and in the infant and adult data.
(PDF)
Acknowledgments
We thank our participants and their parents, Janek Stahlberg and the members of the Max Planck Research Group Naturalistic Social Cognition for their assistance. We also thank the team of the Department of Developmental Psychology: Infancy and Childhood at the University of Zurich for the supportive discussions of this project. Over the course of the project, Karola Schlegelmilch was external Fellow of the International Max Planck Research School on the Life Course (LIFE).
Data Availability
All files are available from the osf database in the project: "Data_Visual_segmentation_of_naturalistic_structures_in_infant_eye-tracking_search_task" under the link: https://osf.io/uyg76/?view_only=14e8e992abfe46e992e5a963776fc70b.
Funding Statement
This work was funded by the Max Planck Society. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Bronson GW. Infants’ transitions toward adult‐like scanning. Child Dev. 1994;65(5):1243–61. doi: 10.1111/j.1467-8624.1994.tb00815.x [DOI] [PubMed] [Google Scholar]
- 2.Colombo J. The development of visual attention in infancy. Annu Rev Psychol. 2001;52(1):337–67. doi: 10.1146/annurev.psych.52.1.337 [DOI] [PubMed] [Google Scholar]
- 3.Courage ML, Reynolds GD, Richards JE. Infants’ attention to patterned stimuli: Developmental change from 3 to 12 months of age. Child Dev. 2006;77(3):680–95. doi: 10.1111/j.1467-8624.2006.00897.x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Colombo J, Cheatham CL. The emergence and basis of endogenous attention in infancy and early childhood. In: Kail RV, editor. Advances in Child Development and Behavior [Internet]. JAI; 2006. [cited 2021 Jan 12]. p. 283–322. http://www.sciencedirect.com/science/article/pii/S0065240706800108 doi: 10.1016/s0065-2407(06)80010-8 [DOI] [PubMed] [Google Scholar]
- 5.Atkinson J, Braddick O. Visual Development. Oxf Handb Dev Psychol; Vol 1 [Internet]. 2013. Mar 21 [cited 2019 Aug 6]; https://www.oxfordhandbooks.com/view/10.1093/oxfordhb/9780199958450.001.0001/oxfordhb-9780199958450-e-10 [Google Scholar]
- 6.Ellemberg D, Lewis TL, Liu CH, Maurer D. Development of spatial and temporal vision during childhood. Vision Res. 1999;39(14):2325–33. doi: 10.1016/s0042-6989(98)00280-6 [DOI] [PubMed] [Google Scholar]
- 7.Kovács I, Kozma P, Fehér Á, Benedek G. Late maturation of visual spatial integration in humans. Proc Natl Acad Sci. 1999;96(21):12204–9. doi: 10.1073/pnas.96.21.12204 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Siu C, Murphy K. The development of human visual cortex and clinical implications. Eye Brain. 2018. Apr;Volume 10:25–36. doi: 10.2147/EB.S130893 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Aslin RN, Smith LB. Perceptual Development. Annu Rev Psychol. 1988;39(1):435–73. doi: 10.1146/annurev.ps.39.020188.002251 [DOI] [PubMed] [Google Scholar]
- 10.Braddick O, Atkinson J. Development of human visual function. Vision Res. 2011. Jul;51(13):1588–609. doi: 10.1016/j.visres.2011.02.018 [DOI] [PubMed] [Google Scholar]
- 11.Kellman PJ, Arterberry ME. Infant Visual Perception. In: Handbook of Child Psychology [Internet]. John Wiley & Sons, Inc.; 2007. [cited 2017 Feb 9]. http://onlinelibrary.wiley.com/doi/10.1002/9780470147658.chpsy0203/abstract [Google Scholar]
- 12.Lewis TL, Maurer D. Multiple sensitive periods in human visual development: evidence from visually deprived children. Dev Psychobiol J Int Soc Dev Psychobiol. 2005;46(3):163–83. doi: 10.1002/dev.20055 [DOI] [PubMed] [Google Scholar]
- 13.Brown AM, Lindsey DT. Contrast Insensitivity: The Critical Immaturity in Infant Visual Performance: Optom Vis Sci. 2009. Jun;86(6):572–6. doi: 10.1097/OPX.0b013e3181a72980 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Pirchio M, Spinelli D, Fiorentini A, Maffei L. Infant contrast sensitivity evaluated by evoked potentials. Brain Res. 1978;141(1):179. doi: 10.1016/0006-8993(78)90628-5 [DOI] [PubMed] [Google Scholar]
- 15.van den Boomen C, van der Smagt MJ, Kemner C. Keep your eyes on development: the behavioral and neurophysiological development of visual mechanisms underlying form processing. Front Psychiatry. 2012;3:16. doi: 10.3389/fpsyt.2012.00016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Morrone MC, Burr DC. Evidence for the existence and development of visual inhibition in humans. Nature. 1986;321(6067):235–7. doi: 10.1038/321235a0 [DOI] [PubMed] [Google Scholar]
- 17.Slater A, Morison V, Somers M. Orientation discrimination and cortical function in the human newborn. Perception. 1988;17(5):597–602. doi: 10.1068/p170597 [DOI] [PubMed] [Google Scholar]
- 18.Almoqbel FM, Irving EL, Leat SJ. Visual acuity and contrast sensitivity development in children: Sweep visually evoked potential and psychophysics. Optom Vis Sci. 2017;94(8):830–7. doi: 10.1097/OPX.0000000000001101 [DOI] [PubMed] [Google Scholar]
- 19.Leat SJ, Yadav NK, Irving EL. Development of Visual Acuity and Contrast Sensitivity in Children. J Optom. 2009;2(1):19–26. [Google Scholar]
- 20.Lewis TL, Kingdon A, Ellemberg D, Maurer D. Orientation discrimination in 5-year-olds and adults tested with luminance-modulated and contrast-modulated gratings. J Vis. 2007;7(4):9–9. doi: 10.1167/7.4.9 [DOI] [PubMed] [Google Scholar]
- 21.Putzar L, Hötting K, Rösler F, Röder B. The development of visual feature binding processes after visual deprivation in early infancy. Vision Res. 2007;47(20):2616–26. doi: 10.1016/j.visres.2007.07.002 [DOI] [PubMed] [Google Scholar]
- 22.Maurer D, Lewis TL. Sensitive Periods in Visual Development. Oxf Handb Dev Psychol; Vol 1 [Internet]. 2013. Mar 21 [cited 2019 Aug 6]; https://www.oxfordhandbooks.com/view/10.1093/oxfordhb/9780199958450.001.0001/oxfordhb-9780199958450-e-8 [Google Scholar]
- 23.Mandler JM, McDonough L. Studies in inductive inference in infancy. Cognit Psychol. 1998;37(1):60–96. doi: 10.1006/cogp.1998.0691 [DOI] [PubMed] [Google Scholar]
- 24.Quinn PC. Born to categorize. In: Goswami UC, editor. The Wiley-Blackwell handbook of childhood cognitive development. 2nd ed. Blackwell Publishers Ltd; 2011. p. 129–52. [Google Scholar]
- 25.Rakison DH, Yermolayeva Y. Infant categorization. WIREs Cogn Sci. 2010;1(6):894–905. doi: 10.1002/wcs.81 [DOI] [PubMed] [Google Scholar]
- 26.Hoehl S. The development of category specificity in infancy—What can we learn from electrophysiology? Neuropsychologia. 2016. Mar 1;83:114–22. doi: 10.1016/j.neuropsychologia.2015.08.021 [DOI] [PubMed] [Google Scholar]
- 27.Quinn PC, Eimas PD, Rosenkrantz SL. Evidence for representations of perceptually similar natural categories by 3-month-old and 4-month-old infants. Perception. 1993;22(4):463–75. doi: 10.1068/p220463 [DOI] [PubMed] [Google Scholar]
- 28.Pauen S. Evidence for knowledge–based category discrimination in infancy. Child Dev. 2002;73(4):1016–33. doi: 10.1111/1467-8624.00454 [DOI] [PubMed] [Google Scholar]
- 29.Fantz RL, Nevis S. Pattern preferences and perceptual-cognitive development in early infancy. Merrill-Palmer Q Behav Dev. 1967;13(1):77–108. [Google Scholar]
- 30.Mondloch CJ, Lewis TL, Budreau DR, Maurer D, Dannemiller JL, Stephens BR, et al. Face perception during early infancy. Psychol Sci. 1999;10(5):419–22. [Google Scholar]
- 31.LoBue V, Adolph KE. Fear in infancy: Lessons from snakes, spiders, heights, and strangers. Dev Psychol. 2019. Sep;55(9):1889–907. doi: 10.1037/dev0000675 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Rakison DH, Derringer J. Do infants possess an evolved spider-detection mechanism? Cognition. 2008. Apr;107(1):381–93. doi: 10.1016/j.cognition.2007.07.022 [DOI] [PubMed] [Google Scholar]
- 33.Włodarczyk A, Elsner C, Schmitterer A, Wertz AE. Every rose has its thorn: Infants’ responses to pointed shapes in naturalistic contexts. Evol Hum Behav. 2018;39(6):583–93. [Google Scholar]
- 34.Kellman PJ. Separating processes in object perception. J Exp Child Psychol. 2001;78(1):84–97. doi: 10.1006/jecp.2000.2604 [DOI] [PubMed] [Google Scholar]
- 35.Balas B, Woods R. Infant Preference for Natural Texture Statistics is Modulated by Contrast Polarity. Infancy. 2014. May;19(3):262–80. doi: 10.1111/infa.12050 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Balas B, Saville A, Schmidt J. Neural sensitivity to natural texture statistics in infancy. Dev Psychobiol. 2018;60(7):765–74. doi: 10.1002/dev.21764 [DOI] [PubMed] [Google Scholar]
- 37.Amso D, Haas S, Markant J. An eye tracking investigation of developmental change in bottom-up attention orienting to faces in cluttered natural scenes. PloS One. 2014;9(1):e85701. doi: 10.1371/journal.pone.0085701 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Frank MC, Amso D, Johnson SP. Visual search and attention to faces during early infancy. J Exp Child Psychol. 2014;118:13–26. doi: 10.1016/j.jecp.2013.08.012 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Kelly DJ, Duarte S, Meary D, Bindemann M, Pascalis O. Infants rapidly detect human faces in complex naturalistic visual scenes. Dev Sci [Internet]. 2019. Nov [cited 2020 Jul 30];22(6). Available from: https://onlinelibrary.wiley.com/doi/abs/10.1111/desc.12829 doi: 10.1111/desc.12829 [DOI] [PubMed] [Google Scholar]
- 40.Itti L, Koch C. Computational modelling of visual attention. Nat Rev Neurosci. 2001;2(3):194–203. doi: 10.1038/35058500 [DOI] [PubMed] [Google Scholar]
- 41.Kwon M-K, Setoodehnia M, Baek J, Luck SJ, Oakes LM. The development of visual search in infancy: Attention to faces versus salience. Dev Psychol. 2016. Apr;52(4):537–55. doi: 10.1037/dev0000080 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Elder JH, Goldberg RM. Ecological statistics of Gestalt laws for the perceptual organization of contours. J Vis. 2002;2(4):5–5. doi: 10.1167/2.4.5 [DOI] [PubMed] [Google Scholar]
- 43.Landy M, Graham N. Visual perception of texture. In: The visual neurociences. MIT Press; 2004. p. 1106–18. [Google Scholar]
- 44.Marr D. Early processing of visual information. Philos Trans R Soc Lond B Biol Sci. 1976;275(942):483–519. doi: 10.1098/rstb.1976.0090 [DOI] [PubMed] [Google Scholar]
- 45.Panis S, De Winter J, Vandekerckhove J, Wagemans J. Identification of everyday objects on the basis of fragmented outline versions. Perception. 2008;37(2):271–89. doi: 10.1068/p5516 [DOI] [PubMed] [Google Scholar]
- 46.Ellemberg D, Hansen BC, Johnson A. The developing visual system is not optimally sensitive to the spatial statistics of natural images. Vision Res. 2012. Aug 15;67:1–7. doi: 10.1016/j.visres.2012.06.018 [DOI] [PubMed] [Google Scholar]
- 47.Sireteanu R, Rieth C. Texture segregation in infants and children. Behav Brain Res. 1992;49(1):133–9. doi: 10.1016/s0166-4328(05)80203-7 [DOI] [PubMed] [Google Scholar]
- 48.Pomaranski KI, Hayes TR, Kwon M-K, Henderson JM, Oakes LM. Developmental changes in natural scene viewing in infancy. Dev Psychol. 2021;57(7):1025. doi: 10.1037/dev0001020 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Bhatt RS, Quinn PC. How Does Learning Impact Development in Infancy? The Case of Perceptual Organization: PERCEPTUAL LEARNING IN INFANCY. Infancy. 2011. Jan;16(1):2–38. doi: 10.1111/j.1532-7078.2010.00048.x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Fiser J, Aslin RN. Statistical learning of new visual feature combinations by infants. Proc Natl Acad Sci. 2002. Nov 26;99(24):15822–6. doi: 10.1073/pnas.232472899 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Kirkham NZ, Slemmer JA, Johnson SP. Visual statistical learning in infancy: Evidence for a domain general learning mechanism. Cognition. 2002;83(2):B35–42. doi: 10.1016/s0010-0277(02)00004-5 [DOI] [PubMed] [Google Scholar]
- 52.Kavšek MJ. Infants’ perception of directional alignment of texture elements on a spherical surface. Infant Child Dev. 2003. Sep 1;12(3):279–92. [Google Scholar]
- 53.Yang J, Otsuka Y, Kanazawa S, Yamaguchi MK, Motoyoshi I. Perception of surface glossiness by infants aged 5 to 8 months. Perception. 2011;40(12):1491–502. doi: 10.1068/p6893 [DOI] [PubMed] [Google Scholar]
- 54.Yonas A, Granrud CE. Infants’ perception of depth from cast shadows. Percept Psychophys. 2006;68(1):154–60. doi: 10.3758/bf03193665 [DOI] [PubMed] [Google Scholar]
- 55.Walther DB, Shen D. Nonaccidental Properties Underlie Human Categorization of Complex Natural Scenes. Psychol Sci. 2014. Apr 1;25(4):851–60. doi: 10.1177/0956797613512662 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Geisler WS. Visual perception and the statistical properties of natural scenes. Annu Rev Psychol. 2008;59:167–92. doi: 10.1146/annurev.psych.58.110405.085632 [DOI] [PubMed] [Google Scholar]
- 57.Torralba A, Oliva A. Statistics of natural image categories. Netw Comput Neural Syst. 2003;14(3):391–412. [PubMed] [Google Scholar]
- 58.Tenenbaum JM, Barrow HG. Experiments in interpretation-guided segmentation. Artif Intell. 1977;8(3):241–74. [Google Scholar]
- 59.Frazor RA, Geisler WS. Local luminance and contrast in natural images. Vision Res. 2006;46(10):1585–98. doi: 10.1016/j.visres.2005.06.038 [DOI] [PubMed] [Google Scholar]
- 60.Isherwood ZJ, Schira MM, Spehar B. The tuning of human visual cortex to variations in the 1/f α amplitude spectra and fractal properties of synthetic noise images. NeuroImage. 2017. Feb;146:642–57. doi: 10.1016/j.neuroimage.2016.10.013 [DOI] [PubMed] [Google Scholar]
- 61.Olshausen BA, Field DJ. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature. 1996;381(6583):607–9. doi: 10.1038/381607a0 [DOI] [PubMed] [Google Scholar]
- 62.Hansen BC, Hess RF. Discrimination of amplitude spectrum slope in the fovea and parafovea and the local amplitude distributions of natural scene imagery. J Vis. 2006. Jun 19;6(7):3. doi: 10.1167/6.7.3 [DOI] [PubMed] [Google Scholar]
- 63.Geisler WS, Diehl RL. Bayesian natural selection and the evolution of perceptual systems. Philos Trans R Soc Lond B Biol Sci. 2002;357(1420):419–48. doi: 10.1098/rstb.2001.1055 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Crouzet SM, Joubert OR, Thorpe SJ, Fabre-Thorpe M. Animal Detection Precedes Access to Scene Category. PLOS ONE. 2012. Oct 12;7(12):e51471. doi: 10.1371/journal.pone.0051471 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Thorpe S, Fize D, Marlot C. Speed of processing in the human visual system. nature. 1996;381(6582):520–2. doi: 10.1038/381520a0 [DOI] [PubMed] [Google Scholar]
- 66.Crouzet SM. Fast saccades toward faces: Face detection in just 100 ms. J Vis. 2010;10(4):1–17. doi: 10.1167/10.4.16 [DOI] [PubMed] [Google Scholar]
- 67.LoBue V, DeLoache JS. Detecting the Snake in the Grass: Attention to Fear-Relevant Stimuli by Adults and Young Children. Psychol Sci. 2008. Mar 1;19(3):284–9. doi: 10.1111/j.1467-9280.2008.02081.x [DOI] [PubMed] [Google Scholar]
- 68.Elsner B, Jeschonek S, Pauen S. Event-related potentials for 7-month-olds’ processing of animals and furniture items. Dev Cogn Neurosci. 2013. Jan 1;3:53–60. doi: 10.1016/j.dcn.2012.09.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Opfer JE, Gelman SA. Development of the animate-inanimate distinction. In: Goswami UC, editor. The Wiley-Blackwell handbook of childhood cognitive development. Blackwell Publishers Ltd; 2011. p. 213–38. [Google Scholar]
- 70.Rakison DH, Poulin-Dubois D. Developmental origin of the animate–inanimate distinction. Psychol Bull. 2001;127(2):209. doi: 10.1037/0033-2909.127.2.209 [DOI] [PubMed] [Google Scholar]
- 71.LoBue V, DeLoache JS. Superior detection of threat-relevant stimuli in infancy. Dev Sci. 2010. Jan 1;13(1):221–8. doi: 10.1111/j.1467-7687.2009.00872.x [DOI] [PubMed] [Google Scholar]
- 72.Elsner C, Wertz AE. The seeds of social learning: Infants exhibit more social looking for plants than other object types. Cognition. 2019. Feb 1;183:244–55. doi: 10.1016/j.cognition.2018.09.016 [DOI] [PubMed] [Google Scholar]
- 73.Wertz AE, Wynn K. Can I eat that too? 18-month-olds generalize social information about edibility to similar looking plants. Appetite. 2019. Jul;138:127–35. doi: 10.1016/j.appet.2019.02.013 [DOI] [PubMed] [Google Scholar]
- 74.Wertz AE. How plants shape the mind. Trends Cogn Sci. 2019;23(7):528–31. doi: 10.1016/j.tics.2019.04.009 [DOI] [PubMed] [Google Scholar]
- 75.Şerban P, Wilson JRU, Vamosi JC, Richardson DM. Plant Diversity in the Human Diet: Weak Phylogenetic Signal Indicates Breadth. BioScience. 2008. Feb 1;58(2):151–9. [Google Scholar]
- 76.Hardy K. Plant use in the Lower and Middle Palaeolithic: Food, medicine and raw materials. Quat Sci Rev. 2018;191:393–405. [Google Scholar]
- 77.Pauen S, Hoehl S. Preparedness to Learn About the World: Evidence from Infant Research. In: Breyer T, editor. Epistemological Dimensions of Evolutionary Psychology [Internet]. New York, NY: Springer New York; 2015. p. 159–73. doi: 10.1080/09297049.2014.899571 [DOI] [Google Scholar]
- 78.Read D, Van Der Leeuw S. Biology is only part of the story…. Philos Trans R Soc B Biol Sci. 2008;363(1499):1959–68. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Carrara M, Mingardo D. Artifact Categorization. Trends and Problems. Rev Philos Psychol. 2013. Sep;4(3):351–73. [Google Scholar]
- 80.Oña L, Oña LS, Wertz AE. The evolution of plant social learning through error minimization. Evol Hum Behav. 2019;40(5):447–56. [Google Scholar]
- 81.Maguire EA, Burgess N, O’Keefe J. Human spatial navigation: cognitive maps, sexual dimorphism, and neural substrates. Curr Opin Neurobiol. 1999;9(2):171–7. doi: 10.1016/s0959-4388(99)80023-3 [DOI] [PubMed] [Google Scholar]
- 82.Tversky B. Navigating by mind and by body. In: International Conference on Spatial Cognition. Springer; 2002. p. 1–10.
- 83.Gelman SA. The development of induction within natural kind and artifact categories. Cognit Psychol. 1988;20(1):65–95. doi: 10.1016/0010-0285(88)90025-4 [DOI] [PubMed] [Google Scholar]
- 84.Schlegelmilch K, Hanussek C. Architectures of Particularities. In: Pinther K, Förster L, Hanussek C, editors. Afropolis: city media art. English edition. Auckland Park: Jacana; 2012. p. 6–73. [Google Scholar]
- 85.Adelson EH. On seeing stuff: the perception of materials by humans and machines. In 2001 [cited 2016 Nov 20]. p. 1–12. 10.1117/12.429489 [DOI]
- 86.Schuppli C, Graber SM, Isler K, van Schaik CP. Life history, cognition and the evolution of complex foraging niches. J Hum Evol. 2016. Mar;92:91–100. doi: 10.1016/j.jhevol.2015.11.007 [DOI] [PubMed] [Google Scholar]
- 87.Smuda M. Landschaft. Vol. 2069. Frankfurt am Main: Suhrkamp; 1986.
- 88.Wang Q, Bolhuis J, Rothkopf CA, Kolling T, Knopf M, Triesch J. Infants in Control: Rapid Anticipation of Action Outcomes in a Gaze-Contingent Paradigm. Sporns O, editor. PLoS ONE. 2012. Feb 17;7(2):e30884. doi: 10.1371/journal.pone.0030884 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 89.Hessels RS, Hooge ITC, Kemner C. An in-depth look at saccadic search in infancy. J Vis. 2016. Jun 1;16(8):10–10. doi: 10.1167/16.8.10 [DOI] [PubMed] [Google Scholar]
- 90.Jones PR, Kalwarowsky S, Atkinson J, Braddick OJ, Nardini M. Automated measurement of resolution acuity in infants using remote eye-tracking. Invest Ophthalmol Vis Sci. 2014;55(12):8102–10. doi: 10.1167/iovs.14-15108 [DOI] [PubMed] [Google Scholar]
- 91.Heaps C, Handel S. Similarity and features of natural textures. J Exp Psychol Hum Percept Perform. 1999;25(2):299–320. [Google Scholar]
- 92.Schmidt F, Hegele M, Fleming RW. Perceiving animacy from shape. J Vis. 2017. Sep 1;17(11):10–10. doi: 10.1167/17.11.10 [DOI] [PubMed] [Google Scholar]
- 93.Schlegelmilch K, Wertz AE. Grass and Gravel: Investigating visual properties preschool children and adults use when distinguishing naturalistic images. PsyArXiv [Internet]. 2020; https://psyarxiv.com/tgmd3
- 94.Contini EW, Wardle SG, Carlson TA. Decoding the time-course of object recognition in the human brain: From visual features to categorical decisions. Neuropsychologia. 2017. Oct 1;105:165–76. doi: 10.1016/j.neuropsychologia.2017.02.013 [DOI] [PubMed] [Google Scholar]
- 95.Baumgartner E, Gegenfurtner KR. Image Statistics and the Representation of Material Properties in the Visual Cortex. Front Psychol [Internet]. 2016. [cited 2019 Jan 29];7. Available from: https://www.frontiersin.org/articles/10.3389/fpsyg.2016.01185/full doi: 10.3389/fpsyg.2016.01185 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 96.Gerhard HE, Wichmann FA, Bethge M. How Sensitive Is the Human Visual System to the Local Statistics of Natural Images? PLOS Comput Biol. 2013. Jan 24;9(1):e1002873. doi: 10.1371/journal.pcbi.1002873 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 97.Hiramatsu C, Goda N, Komatsu H. Transformation from image-based to perceptual representation of materials along the human ventral visual pathway. NeuroImage. 2011. Jul;57(2):482–94. doi: 10.1016/j.neuroimage.2011.04.056 [DOI] [PubMed] [Google Scholar]
- 98.Redies C, Hasenstein J, Denzler J. Fractal-like image statistics in visual art: similarity to natural scenes. Spat Vis. 2007;21(1–2):137–48. doi: 10.1163/156856807782753921 [DOI] [PubMed] [Google Scholar]
- 99.Burton GJ, Moorhead IR. Color and spatial structure in natural scenes. Appl Opt. 1987;26(1):157–70. doi: 10.1364/AO.26.000157 [DOI] [PubMed] [Google Scholar]
- 100.Shannon CE. A mathematical theory of communication. Bell Syst Tech J. 1948;27(3):379–423. [Google Scholar]
- 101.Graham D, Schwarz B, Chatterjee A, Leder H. Preference for luminance histogram regularities in natural scenes. Vision Res. 2016. Mar;120:11–21. doi: 10.1016/j.visres.2015.03.018 [DOI] [PubMed] [Google Scholar]
- 102.Motoyoshi I, Nishida S, Sharan L, Adelson EH. Image statistics and the perception of surface qualities. Nature. 2007. May;447(7141):206–9. doi: 10.1038/nature05724 [DOI] [PubMed] [Google Scholar]
- 103.Long B, Störmer VS, Alvarez GA. Mid-level perceptual features contain early cues to animacy. J Vis. 2017. Jun 1;17(6):20–20. doi: 10.1167/17.6.20 [DOI] [PubMed] [Google Scholar]
- 104.Rao AR, Lohse GL. Towards a texture naming system: identifying relevant dimensions of texture. Vision Res. 1996;36(11):1649–69. doi: 10.1016/0042-6989(95)00202-2 [DOI] [PubMed] [Google Scholar]
- 105.Açık A, Onat S, Schumann F, Einhäuser W, König P. Effects of luminance contrast and its modifications on fixation behavior during free viewing of images from different categories. Vision Res. 2009. Jun 1;49(12):1541–53. doi: 10.1016/j.visres.2009.03.011 [DOI] [PubMed] [Google Scholar]
- 106.Gonzalez RC, Woods RE. Digital image processing. New York, NY: Pearson; 2018. 1168 p. [Google Scholar]
- 107.Ruff HA, Rothbart MK. Attention in early development: Themes and variations. Oxford: Oxford University Press; 1996. [Google Scholar]
- 108.Reynolds GD, Courage ML, Richards JE. The Development of Attention. Oxf Handb Cogn Psychol [Internet]. 2013. Mar 11 [cited 2019 Aug 6]; https://www.oxfordhandbooks.com/view/10.1093/oxfordhb/9780195376746.001.0001/oxfordhb-9780195376746-e-63 [Google Scholar]
- 109.Emberson LL, Rubinstein DY. Statistical learning is constrained to less abstract patterns in complex sensory input (but not the least). Cognition. 2016;153:63–78. doi: 10.1016/j.cognition.2016.04.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 110.Zhang F, Jaffe‐Dax S, Wilson RC, Emberson LL. Prediction in infants and adults: A pupillometry study. Dev Sci. 2019;22(4):e12780. doi: 10.1111/desc.12780 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 111.Adler SA, Gallego P. Search asymmetry and eye movements in infants and adults. Atten Percept Psychophys. 2014. Aug;76(6):1590–608. doi: 10.3758/s13414-014-0667-6 [DOI] [PubMed] [Google Scholar]
- 112.Forssman L, Ashorn P, Ashorn U, Maleta K, Matchado A, Kortekangas E, et al. Eye-tracking-based assessment of cognitive function in low-resource settings. Arch Dis Child. 2017. Apr 1;102(4):301–2. doi: 10.1136/archdischild-2016-310525 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 113.Harel J, Koch C, Perona P. Graph-based visual saliency. In: Advances in neural information processing systems. 2007. p. 545–52. [Google Scholar]
- 114.Schlegelmilch K, Wertz AE. The effects of calibration target, screen location, and movement type on infant eye‐tracking data quality. Infancy. 2019;24(4):636–62. doi: 10.1111/infa.12294 [DOI] [PubMed] [Google Scholar]
- 115.EyeLink 1000 Plus User Manual. Version 1.0.6. Mississauga, Ontario, Canada: SR Research Ltd.; 2015.
- 116.Holmqvist K, Nyström M, Mulvey F. Eye Tracker Data Quality: What It is and How to Measure It. In: Proceedings of the Symposium on Eye Tracking Research and Applications [Internet]. New York, NY, USA: ACM; 2012. [cited 2017 Jan 23]. p. 45–52. (ETRA ‘12). [Google Scholar]
- 117.Helo A, Rämä P, Pannasch S, Meary D. Eye movement patterns and visual attention during scene viewing in 3- to 12-month-olds. Vis Neurosci. 2016;33:E014. doi: 10.1017/S0952523816000110 [DOI] [PubMed] [Google Scholar]
- 118.Kliegl R, Wei P, Dambacher M, Yan M, Zhou X. Experimental effects and individual differences in linear mixed models: Estimating the relationship between spatial, object, and attraction effects in visual attention. Front Psychol. 2011;1:238. doi: 10.3389/fpsyg.2010.00238 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 119.Valuch C, Pflüger LS, Wallner B, Laeng B, Ansorge U. Using eye tracking to test for individual differences in attention to attractive faces. Front Psychol [Internet]. 2015. Feb 2 [cited 2020 Aug 18];6. Available from: http://journal.frontiersin.org/Article/10.3389/fpsyg.2015.00042/abstract doi: 10.3389/fpsyg.2015.00042 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 120.Bates D, Mächler M, Bolker B, Walker S. Fitting Linear Mixed-Effects Models Using lme4. J Stat Softw. 2015;67(i01). [Google Scholar]
- 121.Hartig F. DHARMa: Residual Diagnostics for Hierarchical (Multi-Level / Mixed) Regression Models [Internet]. 2020. https://CRAN.R-project.org/package=DHARMa
- 122.Fox J, Weisberg S. An R companion to applied regression [Internet]. 3rd ed. Thousand Oaks Ca: Sage publications; 2019. https://socialsciences.mcmaster.ca/jfox/Books/Companion/ [Google Scholar]
- 123.Graham MH. Confronting multicollinearity in ecological multiple regression. Ecology. 2003;84(11):2809–15. [Google Scholar]
- 124.O’Brien RM. A caution regarding rules of thumb for variance inflation factors. Qual Quant. 2007;41(5):673–90. [Google Scholar]
- 125.Niehorster DC, Cornelissen THW, Holmqvist K, Hooge ITC, Hessels RS. What to expect from your remote eye-tracker when participants are unrestrained. Behav Res Methods. 2017. Feb 15;1–15. doi: 10.3758/s13428-015-0685-x [DOI] [PMC free article] [PubMed] [Google Scholar]
- 126.R Core Team. R: A language and environment for statistical computing [Internet]. R Foundation for Statistical Computing; 2019. https://www.R-project.org/
- 127.Aslin RN. Perceptual organization of visual structure requires a flexible learning mechanism. Infancy. 2011;16(1):39–44. doi: 10.1111/j.1532-7078.2010.00053.x [DOI] [PubMed] [Google Scholar]
- 128.Sireteanu R, Rettenbach R, Wagner M. Transient preferences for repetitive visual stimuli in human infancy. Vision Res. 2009. Sep;49(19):2344–52. doi: 10.1016/j.visres.2008.08.006 [DOI] [PubMed] [Google Scholar]
- 129.Bartoń K. MuMIn: Multi-model inference. [Internet]. 2019. https://CRAN.R-project.org/package=MuMIn [Google Scholar]
- 130.Köster M, Kayhan E, Langeloh M, Hoehl S. Making Sense of the World: Infant Learning From a Predictive Processing Perspective. Perspect Psychol Sci. 2020. May;15(3):562–71. doi: 10.1177/1745691619895071 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 131.Oudeyer P-Y, Smith LB. How Evolution May Work Through Curiosity-Driven Developmental Process. Top Cogn Sci. 2016. Apr;8(2):492–502. doi: 10.1111/tops.12196 [DOI] [PubMed] [Google Scholar]
- 132.Vygotsky L. Interaction between learning and development. Read Dev Child. 1978;23(3):34–41. [Google Scholar]
- 133.Ruderman DL. Origins of scaling in natural images. Vision Res. 1997;37(23):3385–98. doi: 10.1016/s0042-6989(97)00008-4 [DOI] [PubMed] [Google Scholar]
- 134.White BJ, Stritzke M, Gegenfurtner KR. Saccadic Facilitation in Natural Backgrounds. Curr Biol. 2008. Jan 22;18(2):124–8. doi: 10.1016/j.cub.2007.12.027 [DOI] [PubMed] [Google Scholar]
- 135.Kavšek MJ, Yonas A, Granrud CE. Infants’ sensitivity to pictorial depth cues: A review and meta-analysis of looking studies. Infant Behav Dev. 2012. Feb 1;35(1):109–28. doi: 10.1016/j.infbeh.2011.08.003 [DOI] [PubMed] [Google Scholar]
- 136.Nardini M, Bedford R, Mareschal D. Fusion of visual cues is not mandatory in children. Proc Natl Acad Sci. 2010;107(39):17041–6. doi: 10.1073/pnas.1001699107 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 137.Ruff HA, Rothbart MK. Attention in Early Development [Internet]. Oxford University Press; 2001. [cited 2020 Jul 17]. http://www.oxfordscholarship.com/view/10.1093/acprof:oso/9780195136326.001.0001/acprof-9780195136326 doi: 10.1242/dev.128.23.4815 [DOI] [Google Scholar]
- 138.Bertenthal BI. Origins and Early Development of Perception, Action, and Representation. Annu Rev Psychol. 1996;47(1):431–59. [DOI] [PubMed] [Google Scholar]
- 139.Bornstein MH, Ferdinandsen K, Gross CG. Perception of symmetry in infancy. Dev Psychol. 1981;17(1):82. [Google Scholar]
- 140.Kimchi R. The perception of hierarchical structure. Oxf Handb Percept Organ. 2015;129–49. [Google Scholar]
- 141.Wagemans J, Elder JH, Kubovy M, Palmer SE, Peterson MA, Singh M, et al. A century of Gestalt psychology in visual perception: I. Perceptual grouping and figure–ground organization. Psychol Bull. 2012;138(6):1172. doi: 10.1037/a0029333 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 142.Fornari E, Rytsar R, Knyazeva MG. Development of spatial integration depends on top-down and interhemispheric connections that can be perturbed in migraine: a DCM analysis. Neurol Sci. 2014. May 1;35(1):215–24. doi: 10.1007/s10072-014-1777-6 [DOI] [PubMed] [Google Scholar]
- 143.Knyazeva MG. Splenium of corpus callosum: patterns of interhemispheric interaction in children and adults. Neural Plast. 2013;2013. doi: 10.1155/2013/639430 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 144.Gibson EJ. Perceptual learning in development: Some basic concepts. Ecol Psychol. 2000;12(4):295–302. [Google Scholar]
- 145.Adolph KE, Kretch KS. Gibson’s theory of perceptual learning. Int Encycl Soc Behav Sci. 2015;10:127–34. [Google Scholar]
- 146.Cole WG, Robinson SR, Adolph KE. Bouts of steps: The organization of infant exploration. Dev Psychobiol. 2016;58(3):341–54. doi: 10.1002/dev.21374 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 147.Köster M, Castel J, Gruber T, Kärtner J. Visual cortical networks align with behavioral measures of context-sensitivity in early childhood. NeuroImage. 2017. Dec 1;163:413–8. doi: 10.1016/j.neuroimage.2017.08.008 [DOI] [PubMed] [Google Scholar]
- 148.Gegenfurtner KR, Rieger J. Sensory and cognitive contributions of color to the recognition of natural scenes. Curr Biol. 2000. Jun 1;10(13):805–8. doi: 10.1016/s0960-9822(00)00563-7 [DOI] [PubMed] [Google Scholar]
- 149.van Renswoude DR, Visser I, Raijmakers MEJ, Tsang T, Johnson SP. Real-world scene perception in infants: What factors guide attention allocation? Infancy. 2019;24(5):693–717. doi: 10.1111/infa.12308 [DOI] [PubMed] [Google Scholar]
- 150.Kayser C, Körding KP, König P. Processing of complex stimuli and natural scenes in the visual cortex. Curr Opin Neurobiol. 2004;14(4):468–73. doi: 10.1016/j.conb.2004.06.002 [DOI] [PubMed] [Google Scholar]