Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2020 Nov 10;15(11):e0242009. doi: 10.1371/journal.pone.0242009

Head and body structure infants’ visual experiences during mobile, naturalistic play

Chuan Luo 1, John M Franchak 1,*
Editor: Thomas A Stoffregen2
PMCID: PMC7654772  PMID: 33170881

Abstract

Infants’ visual experiences are important for learning, and may depend on how information is structured in the visual field. This study examined how objects are distributed in 12-month-old infants’ field of view in a mobile play setting. Infants wore a mobile eye tracker that recorded their field of view and eye movements while they freely played with toys and a caregiver. We measured how centered and spread object locations were in infants’ field of view, and investigated how infant posture, object looking, and object distance affected the centering and spread. We found that far toys were less centered in infants’ field of view while infants were prone compared to when sitting or upright. Overall, toys became more centered in view and less spread in location when infants were looking at toys regardless of posture and toy distance. In sum, this study showed that infants’ visual experiences are shaped by the physical relation between infants’ bodies and the locations of objects in the world. However, infants are able to compensate for postural and environmental constraints by actively moving their head and eyes when choosing to look at an object.

Introduction

Infants are immersed in a world full of objects to explore. Visual attention is one of the channels through which objects are selected for further processing while filtering out other potential distractions. Infants develop the ability to selectively attend to visual stimuli early in the first year of life [1, 2], which is important for cognitive development [35]. However, most of our knowledge about infant visual selection comes from studies using screen eye trackers (SETs). In a typical SET study setup, infants are seated in a chair or on a caregiver’s lap while they look at images or videos on a screen. Although it is a useful technique to examine some aspects of infants’ selective attention [6, 7], it precludes measuring the gross motor aspects of attention and how they shape infants’ visual experiences of objects in daily life [8].

First, motor tasks shape where infants need to look. Screen-viewing studies do not place demands on the visual system for guiding gross motor action. In real-world adult studies of vision and action, visual attention is needed to guide the hand to objects that are relevant to an ongoing task [911]. When making a sandwich, people looked at bread before reaching for it and then fixated peanut butter jar while putting down the bread slice on the plate [12]. When building block models, adults moved their eyes in sequence with hand movements to pick up and position each block [13, 14]. Like adults, infants move their eyes to look at objects while reaching [15, 16]. Infants may selectively attend to objects to engage in object manipulation and explore object properties [15, 1720]. Second, body posture (e.g., prone, sitting, and upright) shapes infants’ visual attention [2124]. Whereas in a screen-viewing study participants are seated with information easily viewed in front of the face, real-world visual attention is shaped by the orientation of the body. While prone on all fours, infants struggle to see faces because holding up the head to look at caregivers’ faces is effortful [21, 25]. Looking at caregivers’ faces is more frequent while sitting or upright when it is motorically easier to bring faces into view. Also, compared to walking infants, crawling infants look more often at the floor and have less visual access to distant and elevated objects [25].

Thus, the broad goal of the current research is to examine infants’ visual experiences of objects when infants can move their heads and bodies and interact with objects in a 3D environment. To do so, we employ mobile eye trackers (METs, Fig 1), which have been developed to investigate mobile infants’ visual experiences in more naturalistic settings [15, 17, 21]. Two cameras on a headband simultaneously record infants’ eye movements and their approximate head-centered field of view, revealing how movements of body in space, and the head within the body, shape what is in view from moment to moment. Most prior MET work asked how the motor system influences visual experiences by comparing rates of face/object looking between postures [21] or between observers of different sizes [26]. Little is known about the effect of head movements on visual attention within different postures.

Fig 1. Mobile eye tracker (MET).

Fig 1

The MET contains a scene camera and an eye camera. The scene camera points straight ahead to record the infant’s approximate head-centered field of view. The eye camera points in to the infant’s eye to record eye movements.

Prior work with adults shows the importance of not only measuring the frequency of gazing at targets but also measuring how the head centers targets in view while looking. Previous adult studies show that head movements function to localize gaze targets and stabilize the field of view, which then allows eye movements to make fine-grained movements [2729]. By using the head to center objects in view, extreme eye movements can be avoided, which facilitates visual processing [30] and makes visual exploration more efficient [31]. For example, inhibiting head movement—and thus removing the ability to center objects in view—decreases the speed and accuracy of object identification [32]. Despite the importance of using the head to center objects in view, little is known about how infants coordinate head and eye movements to orient their visual attention.

Recent work with infants has begun to look at head orientation within a sitting posture to determine how it changes over development [33]. In the study, 9- to 24-month-olds sat in a chair and played with toys at a table while wearing a MET. Infants’ first-person field of view was measured using the MET’s scene camera, which was worn on infants’ foreheads and faced out to capture their view (Fig 1). The locations of toys in the head-centered field of view were measured to reveal how infants’ head movements shaped visual experiences of objects. The findings showed that infants moved their head to keep toys centered in view, and indicated that the ability to center information in view improved with age. Furthermore, toys were more centered in view when infants looked at toys versus when they did not look at toys, and centered toys were fixated longer than non-centered toys. This suggests that centering objects in view contributes to the ability to sustain attention, which is a fundamental component of cognitive development [1, 3436].

However, that study only looked at head movements within a single posture, sitting, while the infants were stationary. In everyday life, infants are free to move and switch postures (e.g., prone, sitting, upright). How do mobile infants coordinate head and eyes to structure their field of view in different postures? Body posture imposes constraints on infants’ ability to access visual information [21, 2325]. Specifically, infants’ view of their surroundings is constrained when prone on all fours compared to when sitting or upright (i.e., standing on two feet). How do infants use the head to center objects in view in different postures that place different motor constraints on looking? These questions remain unanswered.

Another factor that is neglected in most prior work is how the spatial configuration of targets in naturalistic settings impacts visual experiences. In one notable exception, Yamamoto and colleagues [37] showed that the distance between infant and caregiver affects infant-caregiver mutual gaze: Infants and caregivers are less likely to make eye contact if they are too close to one another or too far apart. Similarly, object distance may affect infants’ visual attention to objects, but this has yet to be tested since most work employs SETs. Although experimenters can display images of objects on a screen that convey distance information through pictorial cues, distance information in an image lacks behavioral relevance—infants cannot reach to an apparently close object nor to an apparently far object on a computer monitor. And in past work that did use METs during table-top play with real toys, toys were always near infants [18, 19, 33]. Because distance could not vary, it is not clear whether infants can center close and far toys equally well. First, the head and body movements needed to center close and far toys might differ, and may differ within postures. Far toys are located higher in infants’ field of view while close toys are in relatively lower locations in the view (Fig 2A). Thus, centering of close and far toys may depend on how infants choose to orient their heads in different body postures, which, in turn, depends on biomechanical constraints on head movements in each posture. Second, close and far toys have different relevance for motor action. Close toys can be reached to and manipulated whereas far toys cannot be immediately acted on. It is possible that infants tend to look at and center close toys in view for manual exploration, but it may be less important to center far toys in view.

Fig 2. Schematic drawings of infants’ field of view (FOV) in sitting and prone positions.

Fig 2

An infant (A) sitting or (B) prone on the ground with a close toy (blue oval) and a far toy (orange oval). The rectangles on the right of the figure depict infants’ approximate first-person FOV in sitting and prone positions (not drawn to scale). (A) In the sitting infant’s FOV, close toys are in relatively lower locations, whereas far toys are in relatively higher locations. (B) In the prone infant’s FOV, close toys appear large, whereas far toys are out of view.

Finally, prior work that studied infants’ visual experiences focused only on the frequency of looking—how often infants’ eye gaze was fixated on toys during a period of time. However, measuring only looking frequency neglects differences in how visual experiences are structured—what does the field of view contain and in what spatial configuration—and how that structure changes over time. The structure and complexity of object locations in the field of view in real time are important to know because they may affect infant visual attention, and ultimately, learning. Some prior studies suggested that infants’ selective field of view may relate to infants’ ability to learn novel words [19, 3841]. Results showed that infants’ field of view is often occupied with only a few objects. With fewer objects in view, infants have less difficulty of matching objects with the words they hear, which facilitates word learning. Thus, the structure and complexity of infants’ visual experiences—not just the frequency of looking—should be studied to understand how infants’ visual experiences may be relevant to learning.

Current study

The purpose of the study was to investigate how infant posture, toy looking, and toy distance affect the spatial distribution of toys in infants’ field of view. This project presents a secondary analysis of [21], in which 12-month-olds played with caregivers in a room with six toys. Infants could freely move around (e.g. crawl, cruise, walk) or remain stationary (e.g., sit, stand), which allowed examining infants’ visual experiences in different postures in a naturalistic context. We focused on toys rather than faces in this study for three reasons: 1) the distance of toys changes their relevance to behavior (e.g., close toys can be manipulated, far toys can not), 2) centering objects facilitates sustained attention [33], which is an important component of learning words for novel objects [4144], and 3) infants rarely look at faces but frequently look at toys in naturalistic play [15, 17, 18, 21], yielding sufficient data to study toy distributions but not face distributions.

The first aim of the study was to test how the relative position between infants and toys (i.e. infant posture and toy distance from infant) may affect infants’ visual experiences. Like [33], the current study measured the centering and spread of toys in infants’ field of view as they reflect the role of infants’ head orientation in structuring visual experiences. Infants wore a MET which recorded their first-person field of view and eye movements (Fig 1). Human coders identified the location of toys in each frame of the field of view videos to determine the distance of each toy to the center of the field of view. The mean and standard deviation of each toy’s distance to the center were defined as centering and spread, respectively. We hypothesized that the centering and spread of close/far toys in infants’ field of view would vary across postures. Centering and spread of visual information in the field of view may relate to sustained attention [33] and visual selection [19], which are pivotal to infant learning. By measuring the centering and spread of close and far toys in the field of view, we will determine how infants’ visual experiences of objects at different distances are shaped by head orientation in different postures.

The second aim of the study was to test whether infants, as active explorers, can compensate for the constraints imposed by posture and toy distance to center targets of interest in view when looking at toys. In the current study, eye movement data captured by the MET determined looking episodes—when infants foveated each toy by pointing the eyes directly at the toy. As visual experiences are shaped by where infants choose to look, we hypothesized that infants’ decisions to look at a toy would alter toy locations in the field of view. Past work showed that toys become more centered and less spread in the field of view during looking episodes compared with non-looking episodes [33]. However, the extent to which infants’ looking can modify their view of toys may depend on posture and toy distance. For instance, it is more difficult to lift the head to look at far toys in prone position, which may constrain whether infants are able to bring far toys to the center of the view while prone even when they choose to look at toys.

It is reasonable to ask why we would focus on infants’ visual experiences of objects at moments that the object is not fixated. First, it should be pointed out that what appears in infants’ field of view (even though infants are not looking) is not random. Infants’ head and body movements shape the content and properties of the field of view, even if these changes are unintentional byproducts. For instance, infants’ field of view is substantially different while crawling versus walking: Crawling infants have much less access to distant and elevated objects compared to walking infants [25]. Those differences in the field of view while infants are not looking reflect the impact of infants’ head and body movements on their visual experiences. Second, the content and properties of infants’ field of view are important because they affect the opportunities infants have to look at a particular stimulus. Just as wearable audio recorders (e.g., LENA devices) measure the number of words infants are exposed to without regard for whether infants are attending speech [4548], measuring the properties of objects in the field of view while infants may not be looking provides insights into the variability in what infants are visually exposed to. Thus, the degree to which infants’ head and body movements structure the content of the field of view reflects the processes that shape infants’ opportunities for visual learning about objects.

Method

This study used a previously published MET video dataset [21] to score new behaviors to address the questions above. The data [49], including procedural videos and coding materials, were downloaded from Databrary (Databrary.org), an online digital data library that stores and shares videos for developmental researchers. First, we briefly describe aspects of the original study relevant to the current investigation. Afterwards, we explain the new coding and analyses conducted in the current study, which are shared on OSF (DOI: 10.17605/OSF.IO/3M8JB).

Description of original study

Ethics statement

All experimental protocols and consent materials were approved by the New York University Institutional Review Board. Parents of all participating infants provided written informed consent. The legal guardians of the individuals whose pictures appear in this manuscript have given written informed consent (as outlined in PLOS consent form) to publish these case details.

Participants

The final sample of the original study included 17 caregiver-infant dyads recruited from the New York City metropolitan area. Most participants were White and middle class. The age range of the infants was 11.8-12.4 months (M age = 12.10, SD = 0.18). Eight infants (5 male, 3 female) could sit, crawl, and stand but had not yet begun to walk; of the 8 non-walkers, 7 could walk with external support (e.g. cruising on furniture or with support from their caregiver). The remaining 9 infants (4 male, 5 female) could sit, crawl, stand, and walk both independently and with support. Thus, all infants in the sample were capable of being in sitting, prone (i.e., crawling), and upright (i.e., standing, supported walking, and independent walking) positions.

Apparatus

Infants wore a Positive Science head-mounted eye tracker attached to a hat. The eye tracker has two cameras (Fig 1): a scene camera and an eye camera. The scene camera was positioned above infants’ right eye and faced out to capture infants’ first-person field of view (54.4° × 42.2° field of view). Given that the distance between the right eye and head midline is very small in infants, the field of view captured by the scene can be treated as an approximate head-centered field of view. The eye camera pointed to the infants’ right eye to record eye movements. The field of view videos were 640 pixels horizontal by 480 pixels vertical recorded at 30 frames/sec. Infants also wore a harness which was held by an experimenter who followed infants throughout sessions to keep infants from falling.

The experiment took place in a laboratory space (4.5 m × 6 m). Six common toys (jingling apple, xylophone, ball, plush dog, saxophone, and plastic car) were distributed all over the room for infants to play with during the session. An overhead camera recorded the whole room and a hand-held digital camcorder operated by an assistant captured a closer third-person view of the participants.

Procedure

The experimenter put on the eye tracker for infants. To calibrate the eye tracker, the experiment displayed a gridded board about 1 m in front of infants and rattled a small toy to draw infants’ attention to specific locations on the board [15, 21, 25]. The experimenter adjusted the position of the scene camera so that the entire calibration board was visible in the scene video [25] to calibrate the scene camera, approximating a head-centered field of view. Although the angle of the scene camera could not be perfectly equated across infants, we expect the differences between infants would be small because the scene camera has a limited range of motion and the scene camera must be placed in a consistent location above the eye due to the configuration of the eye-tracking headgear. The experimenter also ensured the eye camera captured the right eye’s pupil and corneal reflection across the full range of eye movements. By measuring the position of the eye while looking at different locations on the calibration board, the eye tracker calculated gaze location. After the calibration, the play session started. Infants and caregivers played with six toys in a room for 5-15 minutes. Caregivers were told to play with their children as they usually did at home. Both caregivers and children could move around freely in the room. The same procedure of calibration was conducted again at the end of the session to ensure eye tracking accuracy and consistency. The spatial accuracy of gaze location estimates for the eye-tracking data was M = 1.55° of error [21].

Infant posture coding

Infant posture was coded in the original study for each frame as prone, sitting, or upright. Infants in prone posture could be on their hands and knees or lying on their belly. Infants in sitting posture could be tripod sitting, independent sitting, or kneeling with legs tucked under the bottom. Infants in upright posture could be standing or walking with/without support. The moments when infants were in other postures, such as being held by caregivers, were rare and excluded from analyses (<1% of the data).

Secondary data coding and processing in current study

In the current study, we coded the videos to determine: 1) toy locations in infants’ field of view, 2) looking episodes of each toy, and 3) toy distance from infants. Two outcome measures, centering and spread, were calculated based on toy locations to describe infants’ visual experiences.

Toy location in infants’ field of view

Human coders went through the first 5 min of infants’ field of view videos frame by frame and identified any toy that appeared in the view by drawing regions of interest (ROIs) around them using the Dynamic ROI Coder for Matlab v0.2 (www.github.com/JohnFranchak/roi_coding) (Fig 3). The locations of ROIs in the screen were measured in pixel coordinates. The X and Y coordinates of each ROI’s center represented toy location in infants’ field of view. Despite the relatively brief play session, the total dataset across the 17 infants comprised 323,217 hand-coded video frames.

Fig 3. Dynamic ROI coder.

Fig 3

Coders used the Matlab program Dynamic ROI Coder to draw a box around the toy in each frame of infants’ field of view video.

Because of the large volume of data and the slow, manual coding procedure, 8% of each infant’s data was randomly selected to be coded by a second independent coder to evaluate inter-rater reliability of ROIs. The proportion of frames on which the coders agreed whether a toy was in view was close to ceiling (M across toys = .99). The average correlation between coders for six toys’ X coordinates was r = .97 and r = .93 for Y coordinates. The average location differences between coders were 8.44 pixels for X coordinates and 6.84 pixels for Y coordinates. Thus, coders were able to identify the presence and location of toys with excellent reliability.

Looking episodes

Looking episodes for each toy were defined as frames during which infants looked at that toy. Eye movement data determined the gaze location (green cross in Fig 4) in infants’ field of view in pixel coordinates. It was previously reported that the spatial accuracy of gaze location for these gaze data was M = 1.55° of error [21]. To account for the uncertainty in gaze position, ROIs were enlarged to the bounding ellipse (yellow ellipse in Fig 4) to avoid underestimating looking episodes. If eye gaze fell within or on the edge of the ellipse, the frame was defined as a looking episode of the toy. Otherwise, it was a non-looking episode of that toy. Because it was possible for elliptical ROIs to overlap, infants could possibly be counted as looking at more than one toy in a given video frame. On average, M = 29.57% (SD = 14.22) of video frames were defined as looking episodes, and M = 70.43% (SD = 14.22) were non-looking episodes in the data analyses.

Fig 4. Looking episode and centering.

Fig 4

A frame was defined as a looking episode when eye gaze (green cross) fell on or within ROI (yellow ellipse). Centering (red line) was the distance of toy location to the center of the field of view (white dot).

Toy distance

Toy distance from infants was coded using Datavyu (Datavyu.org). Coders went through third-person view videos and scored the distance of each toy from the infant for every frame. Toy distance was scored dichotomously as close or far according to a reference distance: the distance between the infant and the experimenter who held the falling protection harness for the infant (Fig 5). This reference distance was chosen because as the infants moved around in the room, the experimenter kept a relatively consistent distance (∼0.5 m) with the infants while holding the harness. Having the experimenter as a fixed reference point allowed coding of distance as the infant/experimenter moved in depth within the fixed camera used for coding. If a toy’s distance to the infant was greater than the reference distance from the infant, the toy was counted as far. Otherwise, the toy was scored as close. All frames in all videos (100%) were coded by two independent coders. Inter-rater agreement for distance codes was reached M = 97.02%.

Fig 5. Distance coding.

Fig 5

Coders went through third-person view videos to score toy distance. Toys within the reference distance (yellow line) were coded as close to the infant, whereas toys beyond the reference distance were coded as far.

Centering and spread

On every video frame, we calculated the distance between the each toy’s location and the center of the infant’s field of view (red line in Fig 4). Then, for each infant, we aggregated all toys that were in view for every combination of independent variable levels (infant posture, toy looking, and toy distance), and calculated two outcome variables. The first outcome variable was centering: the mean distance of a toy to the center of the field of view. The second outcome variable was spread: the SD of each toy’s distance to the center of the field of view. Thus, centering reflected the average centering of toys in the field of view and spread described the variability of centering for a given combination of conditions (e.g., looking at close toys while prone).

Statistical approach

Linear mixed models (LMMs) were applied to test the effects of posture, looking, and toy distance on both centering and spread. For each outcome variable, a separate LMM was calculated using the lme4 [50] package in R [51] with posture, looking, and toy distance (and their interactions) as fixed factors and with random intercepts for participant (random slope models failed to converge). Posture, looking, and toy distance were treated as categorical factors with three levels of posture (prone, sitting, and upright), two levels of looking (looking, non-looking), and two levels of toy distance (close, far). ANOVAs were used to test significance of main effects and interactions using the lmerTest package in R [52]. Degrees of freedom were determined by the Satterthwaite approximation [53, 54]. Follow-up pairwise comparisons used the Holm-Bonferroni correction to adjust for multiple comparisons.

Results

Centering of toy locations in infants’ field of view

As Fig 6A shows, posture and distance moderated the degree to which toys were centered in infants’ field of view. While prone, far toys were more poorly centered in the field of view compared to when sitting and upright. In addition, toys were more centered in the field of view during looking episodes compared to non-looking episodes regardless of posture and toy distance. Table 1 lists descriptive statistics for centering according to posture, looking, and toy distance.

Fig 6. Centering and spread by posture, toy looking, and toy distance.

Fig 6

(A) Centering and (B) spread of close toys (blue symbols) and far toys (orange symbols) during looking and non-looking episodes across postures. For centering, smaller values on the y-axis indicate better centering. For spread, smaller values on the y-axis indicate less variability of toy locations in the field of view.

Table 1. Descriptive statistics (M and SD in pixels) for centering by posture, looking, and toy distance.

Non-looking Looking
Prone
 Close 223.01 (43.89) 143.62 (28.70)
 Far 262.68 (49.00) 176.10 (68.01)
Sitting
 Close 230.11 (29.99) 153.25 (34.21)
 Far 218.15 (22.96) 139.43 (51.59)
Upright
 Close 226.31 (21.89) 153.28 (22.59)
 Far 210.78 (18.28) 148.48 (44.15)

A LMM predicting centering from posture, looking, and toy distance confirmed the main effects of posture and looking and a significant posture × toy distance interaction (Table 2). To follow up on the significant posture × toy distance interaction, pairwise comparisons were calculated separately for far toys and close toys by posture while collapsing across looking and non-looking episodes. Far toys were less centered while prone compared to sitting (t(164) = 4.20, p < .001) and upright (t(166) = 4.14, p < .001). Centering for far toys did not differ between sitting and upright (t(162) = .08, p = .935). In contrast, the centering of close toys was similar across postures (ps >.05). Thus, posture influenced infants’ visual experiences of far toys but not close toys.

Table 2. Summary of LMM results predicting centering from posture, looking, and toy distance.

Degrees of freedom and resulting p values obtained using the Satterthwaite approximation.

df F p
Posture 2, 163.27 3.86 .023 *
Looking 1, 161.16 207.04 <.001 ***
Distance 1, 161.53 .75 .387 n.s.
Posture×Looking 2, 160.99 .68 .508 n.s.
Posture×Distance 2, 161.41 8.77 <.001 ***
Looking×Distance 1, 161.22 .006 .941 n.s.
Posture×Looking×Distance 2, 160.97 0.23 .795 n.s.

* p < .05,

** p < .01,

*** p < .001.

Spread of toy locations in infants’ field of view

The spread of toy locations in infants’ field of view was related to posture, looking, and toy distance (Fig 6B). Overall, the spread was smaller during looking episodes compared to non-looking episodes, suggesting toy locations were more concentrated in the field of view at the moment of looking. During non-looking episodes (Fig 6B left panel) specifically, toy locations were more spread out in sitting and upright positions compared to prone. Far toys’ locations were more variable than close toys’ locations. In contrast, during looking episodes (Fig 6B right panel), close toys were more spread out than far toys in prone position, whereas the spread of close and far toys were similar in sitting and upright positions. Table 3 shows the means and standard deviations of spread for each combination of posture, looking, and toy distance.

Table 3. Descriptive statistics (M and SD in pixels) for spread by posture, looking, and toy distance.

Non-looking Looking
Prone
 Close 64.48 (20.90) 63.88 (11.86)
 Far 71.93 (23.45) 39.20 (21.30)
Sitting
 Close 70.57 (11.82) 53.65 (10.39)
 Far 81.53 (10.76) 53.52 (7.75)
Upright
 Close 69.23 (9.45) 52.92 (8.45)
 Far 83.34 (6.86) 48.31 (18.74)

A LMM predicting spread from posture, looking, and toy distance confirmed a significant looking effect, a significant looking × toy distance interaction, and a significant posture × toy distance interaction (Table 4). Overall, the spread of toy locations decreased by 21.17 pixels from non-looking to looking episodes (Mnon-look = 73.66, Mlook = 52.49, t(162) = 10.68, p < .001). To follow up on the looking × toy distance and posture × toy distance interactions, a LMM predicting the spread from posture and toy distance was conducted separately for looking episodes and for non-looking episodes.

Table 4. Summary of LMM results predicting spread from posture, looking, and toy distance.

Degrees of freedom and resulting p values obtained using the Satterthwaite approximation.

df F p
Posture 2, 163.35 1.97 .143 n.s.
Looking 1, 160.91 114.24 <.001 ***
Distance 1, 161.37 .05 .825 n.s.
Posture×Looking 2, 160.69 1.50 .226 n.s.
Posture×Distance 2, 161.21 4.78 .010 **
Looking×Distance 1, 160.98 26.32 <.001 ***
Posture×Looking×Distance 2, 160.66 2.23 .111 n.s.

* p < .05,

** p < .01,

*** p < .001.

Within non-looking episodes, the results of LMM confirmed the main effects of posture (F(2,77.54) = 3.23, p = .045) and toy distance (F(1,76.80) = 14.45, p < .001). Visual inspection of Fig 6B (left panel) suggested that the main effect of posture (across toy distance) is due to increasing spread from prone to sitting and from sitting to upright. However, pairwise comparisons between postures were not statistically significant. The main effect of toy distance indicated far toys had larger spread than close toys (p < .001).

Posture and toy distance influenced spread differently for looking episodes. The results of LMM for looking episodes showed a main effect of toy distance (F(1,85) = 11.76, p < .001), and a posture × toy distance interaction (F(2,85) = 6.43, p = .003). Pairwise comparisons followed up the posture × toy distance interaction. Results suggested that during looking episodes, the spread was larger for close toys than far toys in prone position (t(73.3) = 4.59, p < .001). But there was no difference between close and far toys in sitting (t(71) = .03, p = .977) and upright (t(70.2) = .98, p = .330).

Discussion

This study investigated how infants’ visual experiences of toys during play depend on the physical relation between infant and toy, which is created by infants’ eye, head, and body movements. Specifically, we found that the centering and spread of toy locations in infants’ field of view were affected by infant posture, looking, and toy distance. Overall, infants’ decision to look at toys led to better centering and decreased spread of toy locations, regardless of toy distance and posture, demonstrating how attention serves to structure visual experiences amidst variations in the spatial relation between infant and object.

Beyond the overall effect of looking, complex interactions between posture, distance, and looking show the value of studying attention in mobile infants in a naturalistic play task as opposed to in a stationary screen task. First, far toys were less centered in the field of view while infants were prone compared to when they were sitting and upright. However, posture did not influence the centering of close toys. It implies that infants while prone have a disadvantaged access to far toys but not close toys. This moderating effect of distance may be due to infants while prone pointing their head down—restricting how far the upper bound of the field of view extends in the distance [25]. Consequently, far toys are usually located near the top edge of the field of view. Second, toy locations in the field of view became more spread from prone to sitting and from sitting to upright during non-looking episodes, regardless of toy distance. Most likely, this results from a difference in elevation of the head—and thus viewpoint—between upright, sitting, and prone positions. As the head is elevated higher off the ground, the field of view encompasses a larger area of the environment which allows for a larger spatial distribution of object locations in the field of view. In contrast, during looking episodes the spread of toys in the field of view varied only between close and far toys while prone. This shows that through the act of looking, infants focus on a single object and actively maintain it in the center of view through head movements, resulting in a narrow spread. The reason why the spread of close toys while prone is larger compared to in sitting and upright positions during looking episodes may be because infants have more difficulty stabilizing the field of view while prone. Tilting the head up and down is effortful while prone, which may hinder head movements to re-center objects in view. Thus, infants are less likely to keep the toys at consistent locations in the field of view when prone, but do so more easily when the range of head movement while sitting and upright is relatively unrestricted.

Documenting visual experiences and how they are shaped by gross motor behavior is important because infants learn from what they see. Most previous studies investigated infants’ visual experiences by reporting only the frequency of looking at different targets, such as objects and faces [18, 21]. Although looking frequency tells us what opportunities infants have to learn about different categories of visual stimuli in different situations, it overlooks the spatial aspect of infants’ visual experiences—where and how objects are distributed in the field of view at each moment. The spatial aspects of visual experiences are important in that infants select what to attend in the field of view in real time actively through eye, head, and body movements. How centered and complex the distribution is affects infants’ visual attention and, potentially, learning. Studies with adults suggest that centering objects in the field of view by aligning eyes and head facilitates visual processing [31, 55]. Moreover, past developmental work indicates that infants’ field of view is typically dominated by one object, which decreases the difficulty of visual selection for infants and facilitates learning [19, 39]. Another study showed that infants tend to center visual information in the field of view, which may support sustained attention [33]. However, those studies were based on settings where infants were sitting at a table and toys were placed very close to them.

But infants are mobile in everyday life, and the current study goes a step farther than past work to reveal how infants shape visual experiences in the context of mobility. This study is novel in investigating the object distribution in infants’ field of view when infants can freely move their body in different postures and move closer and farther from objects. Our results extend past work showing that infants center objects in their field of view when looking [33] by showing that they do so in prone and upright postures, not just sitting, and they do so for far objects, not just close objects. Also, the current study found that infants are overall exposed to increasingly variable visual experiences from prone to sitting and from sitting to upright (among non-looking episodes, which comprised most visual experiences of objects). This implies the acquisition of new motor abilities, such as during the transition from crawling to walking, may increase the complexity of infants’ visual experiences. Whether this helps or hinders learning has yet to be tested. Complex and variable visual experiences may further encourage infants to move and explore the environment, promoting both motor and cognitive development (e.g. spatial memory) [56, 57]. Alternatively, the more complex and cluttered experiences in some postures, such as when upright, could be a hindrance to learning. For example, pairing object labels to the correct visual referent may be more difficult when upright compared to when sitting if upright infants experience a more cluttered visual scene. The downstream effects of changing visual experiences in development should be studied in the future to discern between these possibilities [58].

Differences in visual experiences in different postures are amplified when object distance is taken into account. The current study differentiated close and far objects in the environment—another benefit to testing visual experiences in a mobile task as opposed to a seated task. We found that infants while prone have difficulty centering far objects in view compared to when sitting and upright, perhaps because lifting the head up to look at far objects is effortful while prone. As a result, far objects are more likely to reside at the top of their field of view and be less centered. But in spite of the constraints imposed by posture and environment, infants are by no means passive observers. In contrast, they actively structure their visual experiences by coordinating head and eyes to look. Regardless of the differences across postures, within every posture infants actively centered and decreased the spread of toy locations when choosing to look at toys. Although infants may not be able to overcome the constraints of a posture completely, they do compensate for the postural and environmental constraints to some degree when they choose to look, using the body to facilitate sustained object attention.

We acknowledge several limitations in the study. First, we only considered whether infants’ posture affected visual experiences, but not whether infants were stationary or mobile. When upright, infants can be standing still, cruising along furniture, or walking independently from one place to another. Likewise, while prone infants could have been either stationary or crawling. Future work should compare infants’ head orientation when stationary versus moving. It is possible that infants look more at close toys when stationary while more at far toys when locomoting. The biomechanics of maintaining balance and need for visual guidance while moving versus stationary may also impact how infants orient the head. A second variable that was not considered was the elevation of toys off the ground. Elevation, not just distance, determines how the head must be oriented to center a toy in view. A previous study suggested that infants in different postures have different access when toy height was manipulated [25]. Infants while prone have more difficulty lifting their head to look at objects at high elevations, especially when toys are close. In the current study, toys could be placed on the ground, but could also be picked up by infants and caregivers and placed on low platforms in the room. Future studies should test how toy elevation may affect infants’ centering of information in the field of view. Last, this study did not investigate the relation between the spatial distribution of objects and infants’ hand action. It is unknown whether the properties of visual experiences affect the probability of infants’ reaching toward objects, and how visual experiences change as visual system guides reaching. Future work should look at the association between infants’ visual experiences and hand actions.

In conclusion, the findings suggest that infants have different visual experiences of close and far objects in different postures. Infants are constrained by the body and environment. Head orientation is constrained by the body posture: Tilting the head up and down is more effortful while prone, resulting in distinct visual experiences. Also, infants’ visual experiences are affected by their spatial relationship with the objects in the environment—whether toys are located close versus far from infants. Nevertheless, infants actively coordinate the head and eyes to choose what in view and structure the location of visual information in view. The resulting spatial characteristics in the field of view may affect infants’ visual attention in real time and learning over developmental time.

Acknowledgments

We thank Karen Adolph and Kari Kretch, co-authors of the original study, for sharing this dataset on Databrary to allow reanalysis. We also thank the members of UCR Perception, Action, and Development Lab for their assistance in coding the data.

Data Availability

The original data from this study are available through the Databrary video data repository (DOI: 10.17910/B7.135), and the new coding files, aggregate data, and analysis script from the current study are available on OSF (DOI: 10.17605/OSF.IO/3M8JB).

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1. Colombo J. The development of visual attention in infancy. Annu Rev of Psychol. 2001;52:337–367. 10.1146/annurev.psych.52.1.337 [DOI] [PubMed] [Google Scholar]
  • 2. Aslin RN, Salapatek P. Saccadic localization of visual targets by the very young human infant. Percept Psychophys. 1975;17(3):293–302. 10.3758/BF03203214 [DOI] [Google Scholar]
  • 3. Colombo J, Mitchell DW, Dodd J, Coldren JT, Horowitz FD. Longitudinal correlates of infant attention in the paired-comparison paradigm. Intelligence. 1989;13(1):33–42. [Google Scholar]
  • 4. Ruff HA, Rothbart MK. Attention in early development: themes and variations. New York, NY: Oxford University Press; 1996. [Google Scholar]
  • 5. Choudhury N, Gorman KS. The relationship between sustained attention and cognitive performance in 17–24-month-old toddlers. Infant Child Dev. 2000;9(3):127–146. [DOI] [Google Scholar]
  • 6. Oakes LM. Advances in eye tracking in infancy research. Infancy. 2012;17(1):1–8. 10.1111/j.1532-7078.2011.00101.x [DOI] [PubMed] [Google Scholar]
  • 7. Richmond J, Nelson CA. Relational memory during infancy: evidence from eye tracking. Dev Sci. 2009;12(4):549–556. 10.1111/j.1467-7687.2009.00795.x [DOI] [PubMed] [Google Scholar]
  • 8. Franchak JM. Visual exploratory behavior and its development In: Federmeier K, Schotter E, editors. The psychology of learning and motivation: Gazing toward the future: Advances in eye movement theory and applications. vol. 73 Elsevier; 2020. p. 59–94. [Google Scholar]
  • 9. Land MF, Hayhoe MM. In what ways do eye movements contribute to everyday activities? Vis Res. 2001;41:3559–3565. [DOI] [PubMed] [Google Scholar]
  • 10. Hayhoe MM. Vision using routines: a functional account of vision. Vis Cogn. 2000;7(1-3):43–64. 10.1080/135062800394676 [DOI] [Google Scholar]
  • 11. Land MF. Eye movements and the control of actions in everyday life. Prog Retin Eye Res. 2006;25(3):296–324. 10.1016/j.preteyeres.2006.01.002 [DOI] [PubMed] [Google Scholar]
  • 12. Hayhoe MM, Shrivastava A, Mruczek R, Pelz JB. Visual memory and motor planning in a natural task. J Vis. 2003;3(1):6 10.1167/3.1.6 [DOI] [PubMed] [Google Scholar]
  • 13. Smeets JB, Hayhoe MM, Ballard DH. Goal-directed arm movements change eye-head coordination. Exp Brain Res. 1996;109(3):434–440. [DOI] [PubMed] [Google Scholar]
  • 14. Pelz J, Hayhoe MM, Loeber R. The coordination of eye, head, and hand movements in a natural task. Exp Brain Res. 2001;139(3):266–277. 10.1007/s002210100745 [DOI] [PubMed] [Google Scholar]
  • 15. Franchak JM, Kretch KS, Soska KC, Adolph KE. Head-mounted eye tracking: a new method to describe infant looking. Child Dev. 2011;82(6):1738–1750. 10.1111/j.1467-8624.2011.01670.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Franchak JM, Yu C. Visual-motor coordination in natural reaching of young children and adults. In: Proceedings of the Annual Meeting of the Cognitive Science Society; 2015. [PMC free article] [PubMed]
  • 17. Slone LK, Abney DH, Borjon JI, Chi-Hsin C, Franchak JM, Pearcy D, et al. Gaze in action: head-mounted eye tracking of children’s dynamic visual attention during naturalistic behavior. J Vis Exp. 2018;141:e58496 10.3791/58496 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Yu C, Smith LB. Joint attention without gaze following: human infants and their parents coordinate visual attention to objects through eye-hand coordination. PLoS One. 2013;8(11):e79659 10.1371/journal.pone.0079659 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Yu C, Smith LB, Shen H, Pereira AF, Smith T. Active information selection: visual attention through the hands. IEEE T Auton Ment De. 2009;1(2):141–151. 10.1109/TAMD.2009.2031513 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Soska KC, Adolph KE, Johnson SP. Systems in development: motor skill acquisition facilitates three-dimensional object completion. Dev Psychol. 2010;46(1):129 10.1037/a0014618 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Franchak JM, Kretch KS, Adolph KE. See and be seen: infant-caregiver social looking during locomotor free play. Dev Sci. 2018;21(4):e12626 10.1111/desc.12626 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Yamamoto H, Sato A, Itakura S. Transition from crawling to walking changes gaze communication space in everyday infant-parent interaction. Front Psychol. 2020;10:2987 10.3389/fpsyg.2019.02987 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Frank MC, Simmons K, Yurovsky D, Pusiol G. Developmental and postural changes in children’s visual access to faces. In: Proceedings of the Annual Meeting of the Cognitive Science Society; 2013.
  • 24.Sanchez A, Long B, Kraus AM, Frank MC. Postural developments modulate children’s visual access to social information. In: Proceedings of the Annual Meeting of the Cognitive Science Society; 2018.
  • 25. Kretch KS, Franchak JM, Adolph KE. Crawling and walking infants see the world differently. Child Dev. 2014;85(4):1503–1518. 10.1111/cdev.12206 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Franchak JM, Adolph KE. Visually guided navigation: head-mounted eye-tracking of natural locomotion in children and adults. Vis Res. 2010;50(24):2766–2774. 10.1016/j.visres.2010.09.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Solman GJ, Foulsham T, Kingstone A. Eye and head movements are complementary in visual selection. R Soc Open Sci. 2017;4(1):160569 10.1098/rsos.160569 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Einhauser W, Schumann F, Vockeroth J, Bartl K, Cerf M, Harel J, et al. Distinct roles for eye and head movements in selecting salient image parts during natural exploration. Ann NY Acad Sci. 2009;1164(1):188–193. 10.1111/j.1749-6632.2008.03714.x [DOI] [PubMed] [Google Scholar]
  • 29. Foulsham T, Walker E, Kingstone A. The where, what and when of gaze allocation in the lab and the natural environment. Vis Res. 2011;51(17):1920–1931. 10.1016/j.visres.2011.07.002 [DOI] [PubMed] [Google Scholar]
  • 30. Einhäuser W, Schumann F, Bardins S, Bartl K, Böning G, Schneider E, et al. Human eye-head coordination in natural exploration. Network Comp Neural. 2007;18(3):267–297. 10.1080/09548980701671094 [DOI] [PubMed] [Google Scholar]
  • 31. Nakashima R, Shioiri S. Why do we move our head to look at an object in our peripheral region? Lateral viewing interferes with attentive search. PloS One. 2014;9(3):e92284 10.1371/journal.pone.0092284 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Biguer B, Prablanc C, Jeannerod M. The contribution of coordinated eye and head movements in hand pointing accuracy. Exp Brain Res. 1984;55(3):462–469. [DOI] [PubMed] [Google Scholar]
  • 33.Franchak JM, Yu C. Infants structure their visual experiences to favor toys over faces during play. Poster presented at the International Congress on Infant Studies, New Orleans, LA. 2016 May;.
  • 34. Lawson K, Ruff H. Early attention and negative emotionality predict later cognitive and behavioural function. Int J Behav Dev. 2004;28(2):157–165. 10.1080/01650250344000361 [DOI] [Google Scholar]
  • 35. Wass S, Porayska-Pomsta K, Johnson MH. Training attentional control in infancy. Curr Biol. 2011;21(18):1543–1547. 10.1016/j.cub.2011.08.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Mundy P, Newell L. Attention, joint attention, and social cognition. Curr Dir Psychol Sci. 2007;16(5):269–274. 10.1111/j.1467-8721.2007.00518.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Yamamoto H, Sato A, Itakura S. Eye tracking in an everyday environment reveals the interpersonal distance that affords infant-parent gaze communication. Sci Rep. 2019;9(1):1–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38. Clerkin EM, Hart E, Rehg JM, Yu C, Smith LB. Real-world visual statistics and infants’ first-learned object names. Philos T R Soc B: Biol Sci. 2017;372 (1711). 10.1098/rstb.2016.0055 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39. Smith LB, Yu C, Pereira AF. Not your mother’s view: the dynamics of toddler visual experience. Dev Sci. 2011;14(1):9–17. 10.1111/j.1467-7687.2009.00947.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40. Yoshida H, Smith LB. What’s in view for toddlers? Using a head camera to study visual experience. Infancy. 2008;13(3):229–248. 10.1080/15250000802004437 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41. Yu C, Smith LB. Embodied attention and word learning by toddlers. Cognition. 2012;125(2):244–262. 10.1016/j.cognition.2012.06.016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Tomasello M, Farrar MJ. Joint attention and early language. Child Dev. 1986;57(6):1454–1463. 10.2307/1130423 [DOI] [PubMed] [Google Scholar]
  • 43. de Villiers Rader N, Zukow-Goldring P. Caregivers’ gestures direct infant attention during early word learning: the importance of dynamic synchrony. Lang Sci. 2012;34(5):559–568. 10.1016/j.langsci.2012.03.011 [DOI] [Google Scholar]
  • 44. Estigarribia B, Clark EV. Getting and maintaining attention in talk to young children. J Child Lang. 2007;34(4):799–814. 10.1017/S0305000907008161 [DOI] [PubMed] [Google Scholar]
  • 45. Ganek H, Eriks-Brophy A. Language environment analysis (LENA) system investigation of day long recordings in children: a literature review. J Commun Disord. 2018;72:77–85. 10.1016/j.jcomdis.2017.12.005 [DOI] [PubMed] [Google Scholar]
  • 46. Weisleder A, Fernald A. Talking to children matters: early language experience strengthens processing and builds vocabulary. Psychol Sci. 2013;24(11):2143–2152. 10.1177/0956797613488145 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Walle E, Warlaumont AS. Infant locomotion, the language environment, and language development: a home observation study. In: Proceedings of the Annual Meeting of the Cognitive Science Society; 2015.
  • 48. Ramírez-Esparza N, García-Sierra A, Kuhl PK. The impact of early social interactions on later language development in Spanish–English bilingual infants. Child Dev. 2017;88(4):1216–1234. 10.1111/cdev.12648 [DOI] [PubMed] [Google Scholar]
  • 49. Adolph KE, Franchak JM. See and be seen: infant-caregiver social looking during locomotor free play (Databrary volume); 2015. Available from: 10.17910/B7.135. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. Bates D, Mächler M, Bolker B, Walker S. Fitting linear mixed-effects models using lme4. J Stat Softw. 2014;67(1):1–48. [Google Scholar]
  • 51.R Core Team. R: a language and environment for statistical computing; 2019. Available from: https://www.R-project.org/.
  • 52. Kuznetsova A, Brockhoff PB, Christensen RHB. lmerTest package: tests in linear mixed effects models. J Stat Softw. 2017;82(13). 10.18637/jss.v082.i13 [DOI] [Google Scholar]
  • 53. Satterthwaite FE. Synthesis of variance. Psychometrika. 1941;6(5):309–316. 10.1007/BF02288586 [DOI] [Google Scholar]
  • 54. Luke SG. Evaluating significance in linear mixed-effects models in R. Behav Res Methods. 2017;49(4):1494–1502. 10.3758/s13428-016-0809-y [DOI] [PubMed] [Google Scholar]
  • 55. Dunham DN. Cognitive difficulty of a peripherally presented visual task affects head movements during gaze displacement. Int J Psychophysiol. 1997;27(3):171–182. 10.1016/S0167-8760(97)00054-8 [DOI] [PubMed] [Google Scholar]
  • 56. Adolph KE, Robinson SR. Motor development In: Lerner RM, editor. Handbook of child psychology and developmental science. Hoboken, NJ: Wiley Online Library; 2015. p. 114–157. [Google Scholar]
  • 57. Gibson EJ. Exploratory behavior in the development of perceiving, acting, and the acquiring of knowledge. Ann Rev Psychology. 1988;39(1):1–42. 10.1146/annurev.ps.39.020188.000245 [DOI] [Google Scholar]
  • 58. Franchak JM. The ecology of infants’ perceptual-motor exploration. Curr Opin Psychol. 2020;32:110–114. 10.1016/j.copsyc.2019.06.035 [DOI] [PubMed] [Google Scholar]

Decision Letter 0

Thomas A Stoffregen

28 Aug 2020

PONE-D-20-22503

Head and body structure infants’ visual experiences during mobile, naturalistic play

PLOS ONE

Dear Dr. Franchak,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please see my comments, below.

Please submit your revised manuscript by Oct 12 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

We look forward to receiving your revised manuscript.

Kind regards,

Thomas A Stoffregen, PhD

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please provide additional details regarding participant consent. In the ethics statement in the Methods and online submission information, please ensure that you have specified whether consent was informed.

3. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.

4.We note that Figure [1] includes an image of a patient / participant in the study. 

As per the PLOS ONE policy (http://journals.plos.org/plosone/s/submission-guidelines#loc-human-subjects-research) on papers that include identifying, or potentially identifying, information, the individual(s) or parent(s)/guardian(s) must be informed of the terms of the PLOS open-access (CC-BY) license and provide specific permission for publication of these details under the terms of this license. Please download the Consent Form for Publication in a PLOS Journal (http://journals.plos.org/plosone/s/file?id=8ce6/plos-consent-form-english.pdf). The signed consent form should not be submitted with the manuscript, but should be securely filed in the individual's case notes. Please amend the methods section and ethics statement of the manuscript to explicitly state that the patient/participant has provided consent for publication: “The individual in this manuscript has given written informed consent (as outlined in PLOS consent form) to publish these case details”.

If you are unable to obtain consent from the subject of the photograph, you will need to remove the figure and any other textual identifying information or case descriptions for this individual.

Additional Editor Comments (if provided):

I have received two reviews from experts in the field. One requests quite substantial changes, while the other suggests relatively minor changes. I have categorized things as "major revision", though I'd prefer a category intermediate between major and minor. In any event, I trust you will carefully consider both reviews, and I hope (and expect) you will be able to submit a suitable revision.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This manuscript describes the re-analysis of the eye-tracking data of 12-month-old infants in a mobile play setting. The authors examined how centered and variable toy locations in infants’ field of view, and investigated how infant posture, toy looking, and toy distance affected the centering and variability. Beyond the overall effect of toy looking, systematic interactions were found between posture, distance, and toy looking. Among others, the results showed that far toys were less centered in the field of view while infants were prone compared to when they were sitting and upright, and that close toys were more variable in the field of view compared to far toys while infants were in prone. The authors concluded that infants have different visual experiences of close and far objects in different postures, and that infants actively coordinate the head and eyes to look at an object. For the most part, the study described in this paper was well-motivated and competently conducted. However, I have a couple of concerns about the method and the interpretation of the results that need to be addressed before I could recommend publication.

1. Head-centered field of view

How was the center of the scene camera image aligned to the center of the field of view of the infant’s right eye (i.e., how was the head-centered field of view defined)? I believe that calibration in the eye tracker is done so as to match the position of the pupil in the eye-camera image and the location of the object in the scene camera image. The eye tracker can track the gaze accurately even if the orientation of scene camera is not perfectly aligned to the center of the actual field of view of the right eye. Slight changes in the orientation of the scene camera could result in different “center” in the scene camera image. In other words, although the eye tracker is able to pick up the location of gaze (green cross in Fig. 4) accurately, the center of the scene camera image and resulting distance between the image center and the green cross may not be so accurate.

2. Non-looking episodes

By definition, the toys in non-looking episodes were what randomly got in infant’s field of view. They were not something that infant’s visual system was oriented to. As such, I’m not sure whether and how the analysis of the data of “non-looking” episode could deepen our understanding the nature of infants’ visual experiences during naturalistic tasks. For instance, what does the result imply that far toys are less centered compared to close toys not only in looking episodes but also in “non-looking” episodes when infants are in prone (Fig. 6A)? Why did toy locations in the field of view become more spread from prone to sitting and from sitting to upright during “non-looking” episodes, but not during looking episodes (Fig. 6B)? What underlies the increased spread in the gaze toward closely located toys while infants are prone in looking episodes (Fig 6B), which was the exact opposite to the tendency found in “non-looking” episode just mentioned above? I have difficulty in interpreting the results of the analysis of what infants were not looking at, and the comparison between looking and non-looking episodes.

3. The effect of distance

The authors used the distance in the 2D image of the scene camera to measure how centered and variable toy locations in infants’ field of view. Close toy occupies the greater area of the scene camera image (Fig 3) than far toy does (Fig 4). I wonder if there are greater chances for eye gaze (green cross) to fall near the ROI when a toy is closely located and takes up the scene camera image compared to the situation where the toy is located far. If the actual distance between the object and the infant is available, visual angle instead of distance on 2D image might be a better measure, as it may compensate for such an effect.

Reviewer #2: Abstract

- No comments

Introduction

- The authors state, “Although it is a useful technique to examine some aspects of infants' selective attention, it precludes measuring the motor aspects of attention and how they shape infants' visual experiences of objects in daily life.” Researchers who use SET to examine infants’ visual attention may classify eye movement as “motor movements.” It would be useful for the authors to specify that they are referring to gross motor movements. This would strengthen their argument about the role of body position on visual attention in infancy.

Method

- Under procedure, is there a publication that can be cited to elaborate on the calibration procedure?

Results

- Please include units for Tables 1 and 3.

Discussion

- I think that an additional limitation or perhaps future direction could be pairing visual attention with the probability of reaching toward particular objects. This would offer information about visually guided reaching and prospective control with upper extremity tasks in infancy.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Decision Letter 1

Thomas A Stoffregen

26 Oct 2020

Head and body structure infants’ visual experiences during mobile, naturalistic play

PONE-D-20-22503R1

Dear Dr. Franchak,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Thomas A Stoffregen, PhD

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: (No Response)

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors have addressed my concerns as well as those of the other reviewers. I believe that the manuscript can now be accepted for publication.

Reviewer #2: The authors have satisfactorily responded to the reviewers' concerns. I think that the paper is now ready to be accepted.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Acceptance letter

Thomas A Stoffregen

30 Oct 2020

PONE-D-20-22503R1

Head and body structure infants’ visual experiences during mobile, naturalistic play

Dear Dr. Franchak:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Thomas A Stoffregen

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Response to Reviewers.pdf

    Data Availability Statement

    The original data from this study are available through the Databrary video data repository (DOI: 10.17910/B7.135), and the new coding files, aggregate data, and analysis script from the current study are available on OSF (DOI: 10.17605/OSF.IO/3M8JB).


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES