Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2019 Nov 7;14(11):e0224402. doi: 10.1371/journal.pone.0224402

Characterizing the Randot Preschool stereotest: Testability, norms, reliability, specificity and sensitivity in children aged 2-11 years

Jenny C A Read 1,*, Sheima Rafiq 1, Jess Hugill 1, Therese Casanova 1, Carla Black 1, Adam O’Neill 1, Vicente Puyat 1, Helen Haggerty 2, Kathryn Smart 2, Christine Powell 2, Kate Taylor 2, Michael P Clarke 1,2, Kathleen Vancleef 1
Editor: Guido Maiello3
PMCID: PMC6837395  PMID: 31697704

Abstract

Purpose

To comprehensively assess the Randot Preschool stereo test in young children, including testability, normative values, test/retest reliability and sensitivity and specificity for detecting binocular vision disorders.

Methods

We tested 1005 children aged 2–11 years with the Randot Preschool stereo test, plus a cover/uncover test to detect heterotropia. Monocular visual acuity was assessed in both eyes using Keeler Crowded LogMAR visual acuity test for children aged 4 and over.

Results

Testability was very high: 65% in two-year-olds, 92% in three-year-olds and ~100% in older children. Normative values: In 389 children aged 2–5 with apparently normal vision, 6% of children scored nil (stereoblind). In those who obtained a threshold, the mean log threshold was 2.06 log10 arcsec, corresponding to 114 arcsec, and the median threshold was 100 arcsec. Most older children score 40 arcsec, the best available score. We found a small sex difference, with girls scoring slightly but significantly better. Test/retest reliability: ~99% for obtaining any score vs nil. Agreement between stereo thresholds is poor in children aged 2–5; 95% limit of agreement = 0.7 log10 arcsec: five-fold change in stereo threshold may occur without any change in vision. In children over 5, the test essentially acts only as a binary classifier since almost all non-stereoblind children score 40 arcsec. Specificity (true negative rate): >95%. Sensitivity (true positive rate): poor, <50%, i.e. around half of children with a demonstrable binocular vision abnormality score well on the Randot Preschool.

Conclusions

The Randot Preschool is extremely accessible for even very young children, and is very reliable at classifying children into those who have any stereo vision vs those who are stereoblind. However, its ability to quantify stereo vision is limited by poor repeatability in children aged 5 and under, and a very limited range of scores relevant to children aged over 5.

Introduction

Stereotests assess binocular visual function by measuring the smallest depth difference between two adjacent surfaces which a person can detect purely by using their stereoscopic vision. This stereo threshold is largely independent of viewing distance when expressed in angular terms [1]. It is usually expressed in seconds of arc (1 arcsec = 1/3600 deg), and is often referred to as stereoacuity.

Stereoacuity is clinically important because stereoscopic vision is considered the “gold standard” of binocular vision [2], requiring good vision in both eyes, good oculomotor control as well as cortical neurons to combine the two eyes’ inputs and extract disparity [3]. Accordingly, stereoacuity is a primary or secondary outcome measure in interventions for strabismus and amblyopia [2,47], and is routinely measured when children are referred to eye clinics with these conditions.

Several stereo tests are commonly used in the clinic. These include the near Frisby, Frisby-Davis Distance, Lang, TNO, Randot and Preschool Randot stereotests, each with their own properties and (dis)advantages[3,810]. Of these, the Randot family of stereotests produced by the Stereo Optical Company (stereooptical.com) are the mostly widely used. The Randot stereotest is the clinical stereotest most often used in the USA and Canada [11], while a PubMed search indicates that Randot tests are also one of the most commonly used for research (the search "testname"[Title/Abstract] AND ((stereo*) OR stereopsis OR amblyopia OR binocular OR strabismus” on 26th March 2019 returned 165 results for testname =“Randot”, as opposed to 252 for “Titmus”, 143 for “TNO”, 88 for “Frisby”, 83 for “Lang”).

In using any stereotest, it is important to know

  1. testability, i.e. how many children at each age have the cognitive and other capacity to obtain a meaningful measure on the test.

  2. the normative data, i.e. the distribution of values expected at different age-ranges for visually normal individuals.

  3. the test/retest reliability, effectively a measure of the “error” on the test. This is particularly important where one wishes to track changes in stereoacuity over time, e.g. as a result of treatment. One needs to know when a given difference in scores reflects real change, and when it is consistent with the measurement error.

  4. the sensitivity and specificity with which the test detects binocular visual problems. This is important since stereotests are sometimes included in screening programmes. A child who fails a stereotest will usually be referred for further investigation for a binocular visual problem such as strabismus or amblyopia, while a child who passes the stereotest (and other tests) may be assumed to have no need of referral. Thus it is important to know how many children can be expected to be referred unnecessarily on this basis (the false positive rate, or 1 –specificity) and how many children with binocular visual problems will be missed (the false negative rate, or 1 –sensitivity).

Various studies, summarised in Table 1, have assessed these for the Randot Preschool Stereotest[8,1219]. However, no study has assessed all of them together, and estimates of reliability, sensitivity and specificity are not available for all age ranges. In this paper, we report these values for the Randot Preschool Stereotest, from a cohort of around a thousand children tested in 2016 in North East England. Comparing them to results from previously published studies, we find generally good agreement, confirming the consistency of the Randot Preschool in different populations. A specific contribution of our paper is a set of equations describing the probability of obtaining a given Randot Preschool stereoacuity as a function of age in years, rather than a simple mean and standard deviation, for visually normal children.

Table 1. Summary of results from this and previous studies.

The study is identified in the “Reference” column; “.” indicates this study (values in red). Numbers of children for our study refer to testability and are smaller for other columns; see relevant sections for details. Values in italics were inferred from the data provided in the paper, e.g. by reading off values from figures and/or using the formulae in Table 2. Note that Yang et al[15] claim to provide sensitivity and specificity for the Randot Preschool, but this appears to be incorrect. First, their sample is described as “100 normal children without ocular disease” and specifically excluded children with strabismus or amblyopia, so it is not clear who the “true positives” would be. Second, the denominators used in the calculation of sensitivity and specificity are the number of children in each age-group, not the number of children passing/failing the test. Third, the sensitivity values are over 90% in all age-groups, far higher than any other report. Thus, we have excluded their data.

Age in years Reference Number of children Testability (% who understand & cooperate) Normative data (i.e. in children with apparently normal vision) Reliability Validity
% of testable scoring nil Stereo thresholds in arcsec Stereo log-thresholds in log10 arcsec 95% coefficient of repeatability (1.96 SD of differences[20] in log10arcsec) Pearson correlation Sensitivity \ True Positive % Specificitiy \ True Negative %
Mean Median SD Mean SD Estimate 95% CI
2 . 49 65 14 382c 400 287 2.44 0.38
[14] 33
[15] 19 47 332
[16] 411 31
[17] 130 32
[13] 400h 2 216 2.17 0.37
3 . 150 92 7 214c 100 234 2.13 0.40 0.71 0.42–1.0 0.72 14j 91 j
[14] 73
[15] 34 85 135
[16] 366 67
[17] 287 75
[12] 138 100a 150 1.74 0.47
[13] 1606h 3 114 1.92 0.35 27h 99h
4 . 161 99 2 134c 100 148 1.97 0.34 0.82 0.60–1.03 0.05
[14] 96
[15] 25 96 71
[16] 365 88
[17] 297 96
[12] 217 100a 50 1.95 0.21
[13] 400h 1 94 1.80 0.38
[18] 100e 0.64e 0.97e
5 . 101 99 5 87c 60 76 1.84 0.27 0.64 0.42–0.87 0.59 60 j 98 j
[14] 98
[15] 22 95 51
[16] 373 95
[17] 300 98
[12] 104 60a 70 1.59 0.40
[13] 400h 0 71 1.69 0.38
6–7 . 256 100 1 91c 40 120 1.80 0.31 0.56 0.42–0.70 0.72 31 j 99 j
[12] 46 60ab 20b 1.76b 0.14b
10–11 . 195 100 1 63c 40 93 1.69 0.23 0.28 0.20–0.36 0.86 26 j 99 j
[12] 56i 40ai 40 10i 24g
[8] 19 0.60d 0.35–0.85
[19] 47f 0.23f 0.12–0.35

a: Study [12] states means are “rounded to the next larger disparity level available in the Randot Preschool”, but this is not consistent with the fact that in some age-groups the mean is given as 40, the best available score, and yet the SD is non-zero, meaning that some children must have scored worse than 40. In this case the pre-rounding mean must have been >40 and so “rounding to the next larger disparity level” would have given a rounded mean of 60. Means may have been rounded to the closest available level.

b. Data from 46 children aged 6 years.

c: We calculated means for non-stereoblind children only, which would reduce our estimates compared to those including all children.

d. For 19 children aged 7–18.

e. For 100 children aged 2–12, but mainly aged around 4. This cohort included 75 clinical patients and 31 out of the 100 were stereoblind, thus ensuring perfect agreement (fail both times).

f. For 47 participants with microtropia, heterophoria or orthophoria, aged 3–80 years, around half children.

g: For 242 patients with amblyogenic conditions, aged 3–18 years, counting 800 arcsec as a fail (57/242 failed).

h: For 1606 children aged 2–5 years, screening for strabismus, counting 800 arcsec as a fail. Sensitivity was 24% screening for amblyopia, and 9% screening for anisometropia; specificity was similarly high for all conditions.

i: Data from 56 children aged 9 and 10 years.

j: Our sensitivity/specificity analysis used larger age-groups, see relevant section.

Methods

Comparisons with previous studies

To compare our results with previous studies, we conducted a Pubmed search for “Preschool Randot”, on 6th March 2019. This returned 85 citations, which we reviewed manually to find those which contained relevant data. We excluded any studies which contained data solely for ages over 10 years. For normative values, sensitivity and specificity, we also excluded any studies reporting values solely in a clinical population. We did include reliability measures assessed in clinical populations of the relevant ages. The results of our analysis are summarised in Table 1.

Participants

A total of 1005 children (488 boys, 517 girls) participated in the study. They were aged between 2 and 12 years old (numbers in each age-group are provided in the Results sections below). The children were recruited through local primary schools, preschools, nurseries, personal contacts, and at local science centres. Testing took place in schools, nurseries, at Newcastle University and at local science centres in the city of Newcastle upon Tyne in North East England. Children were targeted in four UK school year-groups: Nursery (2- and 3-year-olds), Reception (4- and 5-year-olds), Year 2 (6- and 7-year-olds), Year 6 (10- and 11-year-olds). UK school years include children born from September to August of the following year. The study included one 9-year-old who was grouped with the 10- and 11-year-olds. For the Reception year group, our study was combined with the routine Orthoptic School Vision Screening programme, and only children who participated in the screening were eligible to participate. Participation in the Orthoptic School Vision Screening assured that all children were screened for visual problems and the recommended referral pathway was followed. For the other age groups, all children within the targeted ages were eligible to participate. For the three studies reported below (reliability, validity, and normative data), subsamples were used. The criteria and characteristics of the samples are detailed below in the relevant sections.

Ethics

The study protocol was compliant with the Declaration of Helsinki and was approved by the Ethics Committee of the Newcastle University Faculty of Medical Sciences (approval number 01078). All parents received an information leaflet about the study. For most testing at schools and nurseries, we used opt-out consent, where parents could return a form withdrawing their child from participation. Opt-out consent was approved by our ethics committee in order to ensure a representative sample [21,22]. If requested by the school or nursery and for testing sessions at Newcastle University and local science centres an opt-in consent procedure was used. Children were always asked for oral or non-verbal assent at the time of testing. Parents and children were informed about the results on standard vision screening test (Visual Acuity and Cover Test) and referred to an optometrist or orthoptists when failing either of these tests.

Data analysis

Data analysis and statistics were carried out using R (version 3.5.2, "Eggshell Igloo") in Rstudio (version 1.1.463), https://CRAN.R-project.org/. The R data files along with R markdown code to carry out all analysis and figures for this paper are available at https://doi.org/10.25405/data.ncl.9755045.

Study design and procedures

To evaluate validity and collect a normative data sample, we assessed visual impairments via a questionnaire, visual acuity with a Crowded logMAR test, performed a Cover Test, and a Preschool Randot stereoacuity test. To evaluate reliability a subsample of children was requested to participate in a second session. The second session only included the Randot Preschool stereotest.

Vision questionnaire

Parents of participating children were asked to provide information about their child’s eye sight: whether they needed glasses for near and / or far vision tasks, whether they were receiving patching of atropine treatment for amblyopia, and to report any other vision problems. In the case of schools and nurseries, this questionnaire was sent home with the child and we requested its return completed. Questionnaires were returned for around half of children (numbers specified in Results sections). To avoid sample bias, we did not exclude children for whom questionnaires were not returned [21,22]. If indicated on the questionnaire, we asked children to wear their glasses during testing. The full questionnaire is available in the Supplementary Material.

Visual acuity

Visual acuity was measured in participants tested in non-nursery settings (thus in almost all participants aged 4 years and over) with the Keeler Crowded LogMAR visual acuity test (Keeler Ltd, UK), which is the standard visual acuity test used across the UK [23]. In this test, participants identify or match letters of various sizes presented at a distance of 3 meters with one eye covered. First a screening card is presented to the child and the size of the letters is reduced until an error is made. Once the child answers incorrectly, the examiner starts with the test card two sizes above the last correctly identified letter. If the child is able to identify 2 or more letters on a line, then the next test card is presented. The examiner proceeds until the child is unable to correctly identify 2 letters or more out of the 4 on a line, then the examiner returns to the size above and completes that line. If the 0.800 letter was not seen at 3 meters, our protocol specified that the examiner would walk closer, adding log units to correct for the change in distance. However, all children examined scored at least 0.75 logMAR in both eyes, so this protocol was not used.

Virtually all Reception-year children (4 and 5-year-olds) in our study had their acuity measured as part of the regional Orthoptic School Vision Screening programme. In their protocol, no threshold visual acuity was obtained and the best possible acuity achievable was 0.2 logMAR. Children with a visual acuity above this value were referred to an optician or orthoptist.

Visual acuity measurement was not attempted in nursery settings, as the clinical co-authors advised that in their experience it was not feasible to obtain reliable results with children of this age (2 or 3 years) in such a setting in the available time.

Cover test

A cover test was performed to detect manifest strabismus [24]. In this test one eye in turn is covered with an occluder for a short moment while the participant fixates on a near object or penlight, or a distant object. The cover is then briefly removed and the eyes are observed to see if they move as the occluded eye acquires fixation of the test object. Movement of the unoccluded eye indicates heterotropia. A cover test was performed at near and at distance. The near cover test was always performed at 33 centimetres. The distance for the distance cover test depended on the size of the room that was available. For 42 children the distance for the distance cover test was not recorded. For the other children, the target for the distance cover test was shown at 3 to 6.9 meters away (mean = 4.9, SD = 1.2). Both visual acuity and cover tests were carried out by a qualified orthoptist, either from the study team (SKR) or by an orthoptist from the Newcastle upon Tyne Orthoptic School Vision Screening programme that runs for Reception year group children (4–5 years old).

Randot Preschool stereotest

Researchers administered the Randot Preschool Stereoacuity Test. The test consists of 3 pages. At the left-hand side of each page black-and-white silhouettes of everyday objects are presented. The right-hand side shows random dot patterns. In each set of four random-dot patterns one contains no object (is flat), while the remaining contain disparity-defined objects matching one of the silhouette objects presented on the left. The objects are only visible when wearing 3D polarized glasses. The participant has to identify the object in each random dot pattern or point to the matching object on the left page. The available levels are 800, 400, 200, 100, 60 and 40 arcsec. The test distance for Randot Preschool is not specified in the test’s own manual, nor in the test protocol described on the Pediatric Eye Disease Investigation Group (www.pedig.net), but we followed previous authors[8,12,13] in performing it at 40cm. Stereopsis was tested at 800 arcsec, 400 arcsec, 200 arcsec, 100 arcsec, 60 arcsec, and 40 arcsec, in that order, following a non-stereo pre-test to check understanding and cooperation. Note that we followed the protocol used in the PEDIG studies rather than that supplied with the test (ATS Miscellaneous Testing Procedures Manual downloaded from www.pedig.net). A lower disparity was shown only if the child could identify at least 2 out of 4 shapes correctly at the previous level. The final score was calculated as the lowest level measured at which 2 of more shapes were correctly identified.

Converting between threshold and log threshold

We prefer to work in terms of log-threshold (specifically, the common or decadic logarithm of the threshold in arcsec, measured in log10 arcsec), since the distribution of log-threshold is closer to normal than the distribution of threshold itself [2528]. Other workers in the field have reported statistics on the threshold itself.

This raises problems when comparing results, since the logarithm of the mean threshold is not the same as the mean of the log-threshold, as illustrated in Fig 1. However, if we assume that log-threshold is indeed distributed normally, then it is possible to derive formulae for converting between statistics on the thresholds themselves and on the log-thresholds. These formulae are provided in Table 2. Note that they require us to know two statistics about the distribution, e.g. both the mean and the SD. In this way, we were able to estimate mean and SD of the log-threshold for previous studies which reported the mean and SD of the threshold in arcsec (Table 1). This was not possible for studies which reported only the mean threshold.

Fig 1.

Fig 1

Theoretical distribution of stereo thresholds in arcsec (A), assuming that the distribution of log-thresholds is normal (B).The mean, median and mode of the log-thresholds are all equal (1.7 log10 arcsec corresponding to 50 arcsec), but the mean of the thresholds is higher (58 arcsec) and the mode is lower (38 arcsec). The conversion relating the mean, medians etc of these two distributions uses the formulae given in Table 2. The values chosen for the mean and SD of the log-distribution do not matter for the present purpose of illustration, but were taken from the values we obtained in 10 and 11-year-olds.

Table 2. How to convert between the mean, median, mode and standard deviation of log-thresholds and thresholds, assuming that log-thresholds are distributed normally (Fig 1).
Quantity Units Symbol Formulae to convert between these
Mean of log-thresholds log10 arcsec M M=log10μ12log10[1+(σμ)2]
Standard deviation of log-thresholds log10 arcsec S S=log10[1+(σμ)2]ln10
Median of log-thresholds log10 arcsec = M
Mode of log-thresholds log10 arcsec = M
Mean of thresholds arcsec μ μ = 10Mexp(0.5(S ln 10)2)
Standard deviation of thresholds arcsec σ σ=μexp((Sln10)2)1
Median of thresholds arcsec =10M=μ[1+(σμ)2]0.5
Mode of thresholds arcsec =10MS2ln10=μ[1+(σμ)2]1.5

Results and discussion

Selection bias with questionnaire

Most of our children were recruited via opt-out consent. Consent forms and study questionnaires were sent home with the children, but if they were not returned, children were still included in the study. We justified this procedure by the non-invasive nature of the study and the importance of avoiding selection bias. Our data-set permitted us to examine the likely effect of any such bias, by comparing children for whom questionnaires were or were not returned.

Sample

In this section, we used our whole cohort of 1005 children.

Results

Overall, questionnaires were returned for 54% of our children. Questionnaires were much more likely to be returned for younger children (e.g. 83% of two-year-olds vs 44% of 11-year-olds; p = 0.0001, logistic regression on age). There was no significant difference by sex. In every age-group, the mean visual acuity and stereoacuity were lower (i.e. better) in children for whom questionnaires were returned (Fig 2). This difference was significant in both cases (visual acuity: regression of visual acuity on log age with questionnaire as a categorical factor, p = 0.0008; stereoacuity: ordinal logistic regression (see below) with log age and questionnaire as factors, p = 0.04). For stereoacuity, the advantage of having parents who returned a questionnaire was equivalent to being 0.6 years older.

Fig 2.

Fig 2

(A) Visual acuity (average of left and right eyes) and (B) stereoacuity, by age-group, for children for whom a questionnaire was (blue, right) or was not (red, left) returned. Size of symbols show proportion obtaining the given score; Tukey box and whisker plots show interquartile ranges. Note that most children aged under 6 only underwent a screening acuity test, on which the best score was 0.2 logMAR.

Discussion

We had imagined that parents might be more likely to return questionnaires for children who had a diagnosed eye condition such as amblyopia. With such an effect, vision would have tended to be worse in children with questionnaires. In fact, we find the opposite effect. The possible cause is beyond the scope of our study; nevertheless these results suggest that our sample would have been biased towards children with better vision if we had used opt-in consent.

Testability with Randot Preschool stereotest

Sample

For the testability analysis, we included all 1005 children. The data is provided in data-file RandotPreschoolTestability.RData and the analysis in this section can be recreated by running R markdown file AnalyseTestabilityData.Rmd (see Supplementary Material).

Results

Fig 3 and Table 3 shows the percentage of children who were testable in each age group. Testability rose from 65% in two-year-olds and 92% in three-year olds, to virtually 100% in older children. Slightly more girls than boys were testable in the youngest age-group, but this difference was not significant. Reasons for non-testability included failing the non-stereo pre-test (which requires naming simple luminance-defined black-on-white shapes), not being willing to wear the 3D glasses (including one two-year-old who burst into tears after putting them on!), and not understanding what they were being asked to do.

Fig 3. Testability on the Randot Preschool, by age-group and sex.

Fig 3

Table 3. Testability on the Randot Preschool, by age-group and sex.
Age Group N children % testable N girls % testable N boys % testable
2yo 78 65 35 74 43 58
3yo 167 92 86 91 81 94
4yo 171 99 96 100 75 99
5yo 106 99 53 100 53 98
6 & 7yo 267 100 139 100 128 99
9–11yo 216 100 108 100 108 100

Discussion

Our results agree with previous studies in finding virtually 100% testability in children aged 4 and up [1417,29], but we find substantially higher testability in two-year-olds than previous studies: 65% compared to 31% [16], 32% [17], 33% [14], or 47% [15].

Test/retest reliability of Randot Preschool stereotest

Sample

A subset of children aged three and older were retested on Preschool Randot in a separate session within three weeks of the first test (max 21, mean 16 days apart). We compared results on the two sessions to calculate the test/retest reliability. For this analysis, we did not exclude any children because of visual conditions, but we did exclude children who were recorded as having worn optical correction on one session but not on the other, since we did not want to confound poor test reliability with changes in optical correction. We also excluded children who did not understand the test the first time, as we wanted to understand repeatability of results independent of changes in understanding. In total, this sample consisted of 182 children from 3 to 11 years old.

Results

We initially looked at the reliability of a binary pass/fail classification. We define “failing” as not passing the 800 arcsec level. In the ideal case where the test is 100% reliable and independent of cognitive effects or motivation, children should either pass the test on both sessions or fail on both. P

Fig 4 shows the proportion of children in each situation, classified by their age-group on the first session. Overall 96% of children passed Preschool Randot both times. Only 1% (2 children) failed on the second session having previously passed, and one of these, an 11-year-old, had obtained 800 arcsec on the first session. Thus, the Randot Preschool is extremely consistent at classifying children into those who do versus those who do not have any demonstrable stereo vision.

Fig 4. 187 children were tested in two sessions (22 three-year-olds, 50 four-year-olds,27 five-year-olds, 48 aged 6 or 7 and 40 aged 10 and 11) and on each session we classify them as passing/failing the stereotest.

Fig 4

The light-colored bars (green & beige) show the high level of agreement.

To assess agreement in more detail, we first examined the correlation between scores on the two tests. The Spearman correlation coefficient, which examines the ranking of scores, was 0.59. To compute the Pearson correlation, following previous workers[18], we first replaced a “nil” score with a notional level of 1600 arcsec, i.e. one log-level up from the highest available score of 800 arcsec. The Pearson correlation coefficient between the log-thresholds on the two sessions was 0.62. Both correlations were extremely significant (p<10−10).

We then carried out a Bland-Altman analysis, shown in Fig 5. Following previous workers [8,18,19], we analysed log-thresholds rather than thresholds, since these are closer to normally distributed. The mean difference between results on the two sessions was -0.114 log-arcsec or a factor of 0.77, and this was significantly different from zero (p<10−5). Thus, children tended to obtain a better score on the second session, presumably due to practice effects.

Fig 5. Difference between log-thresholds in the two sessions (second minus first) plotted against means, for 187 children aged 3–11 years.

Fig 5

The size of the symbol indicates the number of children (see legend). Positive differences indicate worse performance on re-test, negative differences indicate better; no change is indicated by the solid black line. Dashed line shows the mean of the differences, dotted lines show the 95% limits of reliability about 0. The minor grid lines indicate the possible mean and differences given the available values on the Randot Preschool.

Since this improvement is relatively small compared with the variability between the two sessions (Fig 5), we will neglect it in computing the reliability. We quantify reliability using the Bland-Altman 95% limit of agreement, where a value of L means that one can be 95% confident that the result of a second test will lie within ± L of the first. For a normal distribution, this corresponds to 1.96 times the standard deviation of the differences. In our Preschool Randot data, this gives L = 0.63 log10 arcsec, corresponding to a factor of 4.3 in thresholds. For example, if a children scores 200 arcsec in the first session, their score in the second session could be between 50 arcsec and 800 arcsec, without any change in their binocular vision.

One can also compute the 95% confidence interval on the estimate L. We follow Bland & Altman’s (1986) recipe for this, estimating the standard error on the limit of agreement as √(3s2/n), where s is the standard deviation of the differences and n is the sample size. We then estimate the 95% confidence interval as the original estimate ± t times the standard error, where t is the t-statistic corresponding to the 95% confidence interval (1.96 for an infinite sample). In this way we estimate a 95% confidence interval of 0.55 to 0.71 log10 arcsec (factors of 3.5 to 5.2).

Fig 6 shows the reliability as a function of age. Values are given in Table 4. Reliability is similar between ages 3–7, but better in the oldest age-group, 10 & 11 year old. A linear regression of absolute difference in log-thresholds against age, with sex as a covariate, revealed a highly significant decrease in absolute difference with age (p<0.058). Absolute differences were slightly higher for boys, though this was not significant (offset = 0.074, p = 0.058).

Fig 6. Differences in Preschool Randot thresholds (second session minus first session) are plotted by age-group for 182 children.

Fig 6

The thick horizontal lines show the 95% limits of reliability for the 5 age-groups; the vertical lines show the estimated 95% confidence interval on this limit.

Table 4. 95% limits of reliability and 95% confidence interval by age-group.

The low correlations obtained for the 4-year-olds are correct; they are partly because only one child in this age-group scored 1600 arcsec on either test (and this child then scored 40 arcsec on retest), so there was a lower range of values (note the high confidence intervals).

Age-group N children 95% limits of reliability and 95% confidence interval on these Correlation coefficients
In log10 arcsec As a factor Spearman Pearson
3 year olds 20 0.71 (0.42 to 1.01) 5.15 (2.61 to 10.13) 0.47 (0.036–0.76) 0.72 (0.4–0.88)
4 year olds 47 0.82 (0.60 to 1.03) 6.54 (4.02 to 10.65) 0.13 (-0.17–0.4) 0.045 (-0.25–0.33)
5 year olds 27 0.64 (0.42 to 0.87) 4.38 (2.61 to 7.35) 0.79 (0.58–0.9) 0.59 (0.28–0.79)
6 & 7 year olds 48 0.56 (0.42 to 0.70) 3.63 (2.61 to 5.06) 0.65 (0.46–0.79) 0.72 (0.54–0.83)
10 & 11 year olds 40 0.28 (0.20 to 0.36) 1.91 (1.59 to 2.30) 0.47 (0.19–0.68) 0.86 (0.74–0.92)
All 182 0.63 (0.55 to 0.71) 4.27 (3.54 to 5.15) 0.59 (0.49–0.68) 0.62 (0.52–0.7)

Discussion

Our estimate of the reliability of Preschool Randot is very similar to previous estimates, with the 95% limits of repeatability being ±0.64 log10 arcsec or a factor of 4.3 in threshold. Fawcett & Birch [18], in 102 children aged 2–12 years, obtained exactly the same value as us: 0.64 log10 arcsec. Adams et al[8] report a similar value of ±0.60 log10 arcsec (a factor of 4) with 19 children aged 7–18, while Smith et al obtained a slightly lower value, 0.46 log10 arcsec (a factor of 3), in 47 people aged 3 to 80 years. The differences likely reflect an improvement in reliability with age.

A previous study [18] concluded there was no change in test/retest reliability over the range 3–12 years. However, this conclusion was based on a linear regression and t-test on the differences themselves, rather than the absolute difference. This is in fact testing for a change in the bias (i.e. the mean of the differences), rather than in the reliability. An increase in reliability with age would be expected to decrease the variance of the difference in the scores obtained on two sessions, without changing its mean. This is why we did our linear regression on the absolute values of the differences. An F-test confirms that there is a highly significant decrease in the variance of differences in log-thresholds with age. In our data, this variance is 2.38 times higher in children under 5 than in those aged 5 and over, and in study [18] the figure is actually 3.00 times higher (data read off from their Fig 3 [18]); p<10−3 for both.

However, this apparent improvement in reliability with age may in fact reflect improved stereo thresholds combined with a floor effect in the scores available. As reported in the section on Normative values, almost all older children obtain the best possible threshold of 40 arcsec on the Randot Preschool. Suppose that the reliability is in fact a factor of 4 at all ages. A young child whose true threshold is 200 arcsec may obtain 400 arcsec on one session and 100 on the next. Yet an older child whose true threshold is 20 arcsec will obtain 40 arcsec on both sessions, simply because a score of 10 arcsec is not available. Reliability will appear higher for the older child but this will be a side-effect of their improved stereo, combined with the available scores. The Randot Preschool was not designed to assess genuine changes in the repeatability of stereo thresholds in general over this age range.

Normative values of Randot Preschool stereotest

Sample

We next investigated the distribution of stereo thresholds obtained with the Randot Preschool in children who, as far as we could tell, had normal vision. To this end, we excluded participants who failed a cover test, or in whom cover test data were not available. We also excluded participants whose parents reported that they were diagnosed or treated for amblyopia or strabismus, or under assessment in an eye clinic for suspected vision problems, or who when tested were not wearing glasses when their parents reported that they needed glasses; but we did not exclude children for whom parental questionnaires were not available. For children aged 4 and over, we also excluded participants for whom visual acuity data were not available, or whose visual acuity was worse than 0.2 logMAR in either eye, or whose interocular acuity difference exceeded 0.2 logMAR. Table 5 summarises the visual acuity of our normative sample. In the 2 and 3 year-olds, who were tested in nurseries, visual acuity data was not available. Our “normative” sample may therefore include some children in this age-range with undiagnosed poor vision. We also excluded any children who were not testable with Randot because they refused to wear the glasses or did not cooperate with the test in another way (see section on Testability for age breakdown). The remaining 826 participants were 402 boys and 424 girls aged between 2.00 and 11.6 years (Fig 7 and Table 5). Analysis code is provided in AnalyseNormativeData.Rmd using data-file RandotPreschoolNormative.RData (Supplementary Material).

Table 5. Visual acuity in the two older age-groups of our normative sample.

All visual acuities are in logMAR, mean ± SD. Children with visual acuity worse than 0.2 logMAR in either eye, or an interocular visual acuity difference >0.2 logMAR, were excluded from the normative sample.

Age 6 & 7 years old 10 & 11 years old
Number of children 247 190
Monocular acuity in better eye 0.047 +/- 0.067 -0.012 +/- 0.084
Monocular acuity averaged across eyes 0.07 +/- 0.064 0.01 +/- 0.081
Interocular acuity difference 0.0033 +/- 0.062 0.011 +/- 0.067
Acuity in left eye 0.068 +/- 0.073 0.0049 +/- 0.087
Acuity in right eye 0.071 +/- 0.069 0.016 +/- 0.088
Fig 7. Age distribution of normative data sample, showing a similar number of boys (mauve, total 465) and girls (green, 487) for each age.

Fig 7

Results

Fig 8 and Table 6 report the distribution of Randot Preschool scores in each age-group. There is a clear improvement in stereoacuity with age. Additionally, scores are consistently slightly better for girls. To quantify this, we carried out an ordinal logistic regression, using function polr from R package MASS [30]. The main effect of age and sex were both significant (p<10−10 for age and p = 0.002 for sex). The better performance of girls was most pronounced in the 2 and 3 year-olds, where visual acuity was not measured, but remained significant even if only children in the two older age-groups (over 5 years) are considered.

Fig 8. Distribution of normative Randot Preschool stereo thresholds in different age-groups, for children with normal vision, separated by sex.

Fig 8

In children aged 4 and up, the most common outcome is 40 arcsec, the best score available. Dots show the fit of a descriptive model fitted to this data (Eq 1). For each child in the study, given their age and sex, the model estimated the probability of each possible score. We then averaged these across all children of a given sex and age-group to obtain the points shown. With 8 parameters fitted to the data, the model gives a reasonable account of the data in these 84 bins.

Table 6. Randot Preschool stereo thresholds by age-group, for children with normal vision who were judged as being able to understand and cooperate with the test.

Mean and SDs are reported for those children who could complete at least the 800 arcsec test level. Percentiles (type 1 quantile from Hyndman and Fan (1996)) are for all children, including those who scored “nil”. Stereothresholds in the normal range for each age group are marked in green (up to the 75% percentile). “Nil” means unable to perform 800 arcsec plate of Randot Preschool despite passing non-stereo pre-test.

Age-group Number tested % scoring nil Mean, SD computed on those who could complete at least 800 arcsec Percentiles computed on all tested
Threshold in arcsec Log-threshold in log10 arcsec Percentile in arcsec
Mean SD Mean SD 25% 50% 75% 90% 95%
2yo 49 14. 382. 287. 2.44 0.380 200 400 800 Nil Nil
3yo 150 6.7 214. 234. 2.13 0.395 60 100 400 800 Nil
4yo 98 0.0 134. 148. 1.97 0.338 40 100 200 200 400
5yo 92 5.4 87.1 76.3 1.84 0.265 40 60 100 200 Nil
6 & 7yo 247 0.81 90.6 120. 1.80 0.306 40 40 100 200 400
10 & 11yo 188 1.1 63.0 92.6 1.69 0.225 40 40 40 100 200
All 824 3.2 126. 173. 1.89 0.368 40 60 100 400 800

In the ordinal logistic regression model, the probability of obtaining a stereo threshold equal to or better than a particular Randot Preschool level L is modelled as

P(RL)=[1+exp(αL+βA+γM)]1

where R is score on Randot Preschool stereotest, A is log age in log10(years), M is a categorical variable specifying sex (M = 1 for males and M = 0 for females), β is a fitted parameter describing the effect of age A, γ is a fitted parameter describing the effect of sex, and the six fitted parameters αL depend on the level L (Table 7). For our data, the 8 fitted parameters (the regression slope for age and gender plus 6 coefficients describing the transitions between the 7 different Randot levels) were as specified in Table 7. The predicted probability of obtaining each Randot Preschool score as a function of age A in years is then:

P(R=40arcsec)=[1+exp(α40|60+βA+γM)]1P(R=60arcsec)=[1+exp(α60|100+βA+γM)]1[1+exp(α40|60+βA+γM)]1P(R=100arcsec)=[1+exp(α100|200+βA+γM)]1[1+exp(α60|100+βA+γM)]1P(R=200arcsec)=[1+exp(α200|400+βA+γM)]1[1+exp(α100|200+βA+γM)]1P(R=400arcsec)=[1+exp(α400|800+βA+γM)]1[1+exp(α200|400+βA+γM)]1P(R=800arcsec)=[1+exp(α800|Nil+βA+γM)]1[1+exp(α400|800+βA+γM)]1P(R=Nil)=1[1+exp(α800|Nil+βA+γM)]1 Eq 1
Table 7. Fitted parameters for the ordinal logistic model (see Eq 1 for how to use).

The negative regression slope β represents the improvement with age; the positive γ represents the poorer performance of boys. The model residual deviance is 2305 compared to 2635 for a model ignoring age and sex (thus with two fewer free parameters).

Parameter Estimated value Significance
Age regression slope β
(per log10 year)
-3.079 <10−10
Effect of being male γ +0.4152 0.002
α40|60 -5.513 <10−10
α60|100 -4.623 <10−10
α100|200 -3.639 <10−10
α200|400 -2.679 <10−10
α400|800 -2.007 <10−10
α800|Nil -0.9797 0.003

Discussion

As it was designed to, the Randot Preschool contains a range of test values suitable for assessing threshold stereoacuity in pre-school children (aged 2–5 years). For older children, most visually normal children will obtain the best available score, and so the test is more appropriate as a screening tool.

Normative values from ours and other studies are reported in Table 1. Our values are a little higher than in previous studies [12,13,15]. A potential reason could be that, as described in the Methods, we followed the PEDIG protocol for the Randot Preschool rather than the manufacturer’s instructions. This means that we started the stereo part of the test at 800 arcsec and proceeded to smaller disparities, while the manufacturer instructs testers to begin at 200 arcsec and move up or down depending on results. The latter increases the probability of obtaining a good score by chance, and similarly means that a child who ceases responding after a few trials will obtain a better score (if poor motivation is mistaken for inability to do the task). However, previous studies using the PEDIG [13] or manufacturer’s [12] protocols have obtained similar results, suggesting that this is unlikely to be responsible. Perhaps the most plausible reason could be poor acuity in our “normative” sample. Visual acuity was not measured in our 2- and 3-year-olds, so these groups likely included some children with undiagnosed poor vision. However, we also obtain higher scores in the older age-groups, where all children had visual acuity better than 0.2 logMAR in both eyes. Additionally, an unknown fraction of our sample may have not been wearing the correct refractive correction. The previous studies performed more thorough screening for normal vision, e.g. performing cycloplegic refraction and excluding children with anisometropia >1D [13]. However, anisometropia would be expected to affect stereoacuity via an effect on visual acuity, and we still obtain slightly higher scores even with more stringent limits on visual acuity and interocular acuity differences (S1 Table in S1 File).

Surprisingly, we have found a small but significant sex difference in stereoacuity, with girls scoring slightly better (i.e. lower) than boys, especially at younger ages. This has not previously been reported. In adults, female interpupillary distance is smaller than males’, meaning that a given depth creates a smaller angular disparity at the retina. Thus, females might be expected to develop sensitivity to smaller disparities, i.e. better stereoacuity. To our knowledge, only one publication has reported an effect of interpupillary distance [31], and this was in the wrong direction (the study used the Frisby stereotest, which uses real depth, so observers with larger interpupillary distance should have experienced more disparity for a given test level; yet in fact thresholds increased with interpupillary distance). Other studies have found no sex difference in stereoacuity in adults [32,33], and also no effect of interpupillary distance [33,34]. Sex differences in interpupillary distance become more pronounced during development [35,36], whereas our sex differences in stereoacuity are more pronounced in the younger age-groups. For these reasons, we think it is unlikely that interpupillary distance accounts for the sex difference. It may be related to sex differences in binocular function reported very early in life [3739], with stereopsis emerging earlier in girl babies and their vergence responses being more responsive to disparity than boys’. It may also reflect non-visual cognitive/social/developmental sex differences, e.g. in willingness to cooperate. Boys in our cohort also showed slightly lower testability and test/retest reliability (Fig 3, Fig 6), though these differences were not significant. Further work would be needed to establish whether this sex difference in stereoacuity is reproducible and to establish its causes.

Sensitivity and specificity of Randot Preschool stereotest

Stereotests are commonly used in visual screening. The aim is to identify children with binocular vision problems by their failing the stereotest. Here, we evaluate the sensitivity and specificity of the Randot Preschool used in this way. Sensitivity or true positive rate is the proportion of children with binocular vision problems who failed the stereotest. Specificity or true negative rate is the proportion of patients without binocular vision problems who passed the stereotest (Fig 9).

Fig 9. Definition of sensitivity, specificity, and positive/negative predictive value, in the context of stereotests.

Fig 9

Sample

For this analysis, we included children in whom we were able to perform the Randot Preschool (i.e. they were testable) and the cover test; in children over 4, we also required that visual acuity was measured in both eyes. This left 892 children, for 480 of whom parental questionnaires were available. Since binocular vision problems were rare (only 37 of the 892 children had one of the binocular visual problems defined in the next paragraph), for this analysis we combined the younger age-groups so that we had at least 200 children in each group. The analysis code is available in AnalyseValidityData.Rmd using data in RandotPreschoolValidation.RData.

Results

Fig 10 shows the percentage of children scoring in each of the Randot Preschool levels, by age, color-coded to show whether their vision appeared entirely normal (green) or whether a binocular visual problem was identified. We classified children as having a problem likely to affect binocular vision if (a) their parental questionnaire (if available) clearly indicated one, e.g. “attends eye clinic for lazy eye” (purple, “Parent” in Fig 10); or (b) our cover test showed clear evidence of a problem, e.g. exotropia (red, “CTFAIL”); or (c) their visual acuity in the poorer eye exceeded 0.48 logMAR, the WHO threshold for “moderate visual impairment” (brown, “ModVI”); or (d) they had an interocular visual acuity difference in excess of 0.2 logMAR (gold, “IADonly”). Note that each class includes the ones to its right in the legend. Thus only one child appears in the graph purely because of a large interocular acuity difference (gold), but children who failed the cover test (red) may also have had a large interocular acuity difference. We included all four categories when classifying children as having a binocular vision problem.

Fig 10. Randot Preschool stereo thresholds by age-group, color-coded to indicate whether any binocular vision problems were identified.

Fig 10

Parent: parental questionnaire indicated strabismus or amblyopia; CTFAIL: cover test found heterotropia; ModVI: acuity in worse eye > 0.48 logMAR; IADonly: interocular acuity difference >0.3 logMAR (but no other abnormality). Note that each type of binocular vision problem includes the ones on the right; e.g. those colored red for ‘CTFAIL’ may also have had acuity >0.48 logMAR in one eye, but those colored brown for ‘ModVI’ must have passed the cover test.

Table 8 shows the specificity, sensitivity and positive/negative predicted value for the Randot Preschool stereo test in detecting a binocular vision problem. In all age-groups, the specificity is very high (>~ 90%) but the sensitivity is poor. The high specificity means that nearly all children without binocular vision problems pass the Randot Preschool. However, the poor sensitivity means that many children with binocular vision problems also pass. The positive predict value is fairly good, especially in older age-groups, meaning that most children over 5 who fail the Randot Preschool do have a binocular visual problem. The numbers in Table 8 reflect our fixed pass level of 800 arcsec, but as is apparent from Fig 10, there is no alternative choice of criterion–even varying it by age–which would permit higher sensitivity while retaining the high specificity. We also explored other definitions of “moderate visual impairment”, without finding a better criterion.

Table 8. Predictive value of Randot Preschool in detecting binocular vision problems, by age-group, taking “pass” as a score of 800 arcsec or lower.

N = number of children, TP = number of true positives (those who had a binocular vision problem and failed the stereotest), FP = number of false positives, TN = number of true negatives, FN = number of false negatives, PPV = positive predictive value, NPV = negative predictive value (see Fig 9).

N TP FP TN FN Sensitivity Specificity PPV NPV
2 & 3yo 203 1 17 179 6 14 91 6 97
4 & 5yo 207 3 5 197 2 60 98 38 99
6 & 7yo 266 4 2 251 9 31 99 67 97
9, 10 & 11yo 216 5 2 195 14 26 99 71 93
All 892 13 26 822 31 30 97 33 96

A possible concern is that some children for whom visual acuity or parental questionnaire were not available may have had visual problems which we did not detect. They would therefore be erroneously classed as “false positives” if they failed the Randot Preschool, rather than “true positives”. We therefore also examined the results when including only children for whom visual acuity and questionnaires were available (S1 Fig in S1 File, S2 Table in S1 File), but this did not change the conclusions. In fact the sensitivity values were even poorer.

A further potential concern is crosstalk. The Randot stereotests use linear polarisation to separate the images for the two eyes. When used correctly, this has extremely low crosstalk. It is however critical that patients do not tilt their heads or rotate the test book, as this introduces interocular crosstalk. When crosstalk is present, both eyes can see the images intended for the separate eyes. Crosstalk generally weakens stereoscopic depth perception [4042], but counter-intuitively it can help stereoblind observers pass the Randot Preschool test. This is because the images overlap differently in the disparate region of the image, leading to a visible difference in the combined image which can be used to pass the higher levels of the test with one eye when crosstalk is present (cf Fig 11). The crosstalk is minimal with the eyes parallel to the top edge of the book, and reaches 100% when the book is rotated through 45o relative to the interocular axis. Such crosstalk is a theoretical reason why a child without stereo vision could nevertheless achieve a measurable score on the Randot Preschoool. While we cannot rule out that this contributed to our results, we did attempt to ensure that children viewed the stereotest correctly.

Fig 11. Randot Preschool test card, 400 arcsec level, photographed without polarising glasses (100% crosstalk).

Fig 11

The disparate shapes are faintly but distinctly visible in this monocular image, e.g. one can see that (clockwise from top-left) the random dots depict blank, fir-tree, star, disk.

Discussion

We find that the Randot Preschool has good specificity but poor sensitivity. This is in line with previous findings about the Randot Preschool and other stereotests. Previous results for the Randot Preschool are summarised in Table 1. Afsari et al [13] found that sensitivity was 0% when counting a “fail” as nil stereoacuity; sensitivity varied between 9% and 27% when a score of 800 arcsec also counted as a fail. Birch et al [12] found a sensitivity of 24% for the same definition of fail. Some authors have suggested that random-dot stereotests are more sensitive to strabismus than to amblyopia, but neither our data nor that of Afsari et al[13] suggests a significant difference in sensitivity (S3 and S4 Tables in S1 File). Our data and previous studies indicate that around half of children with binocular vision problems can score well with Randot Preschool, so this cannot be relied upon as a screen for binocular visual problems.

Conclusions

We have carried out a comprehensive analysis of the Randot Preschool stereo test in a thousand children aged 2–11 years, and compared our results with previous studies. The Randot Preschool can be successfully completed by most children from as young as 2 years old. It contains a limited range of possible stereo thresholds, which span the range of normal values in children between 2 and 5 years. These groups have a mean score of ~100 arcsec, or 2.1±0.35 log10 arcsec (mean±SD). In older children and adults, most people will score the best possible value, 40 arcsec, making the Randot Preschool–as the name implies–not suitable for investigating individual differences in vision in these age-groups. The Randot Preschool is extremely reliable at classifying children into those with/without any stereo vision; it is very rare for a child to pass it on one occasion and fail it one another. However, the reliability of stereo thresholds themselves is poor in the youngest age-groups, with a changes of up to ±0.7 log10arcsec or a factor of 5 in stereo threshold occurring by chance. Reliability is high in clinical populations, where many patients fail every time, or in older age-groups, where many obtain the best score every time. Regarded as a screen for binocular vision abnormalities such as strabismus and/or amblyopia, the Randot Preschool has excellent specificity (true negative rate >95%), meaning that almost all people without a binocular vision abnormality pass the test. Thus, failing the Randot Preschool merits further investigation. However, its sensitivity (true positive rate) is poor, <50%, so passing the Randot Preschool, even with the best possible score, certainly does not mean that a binocular vision abnormality can be ruled out.

A strength of our study is the relatively large sample and the completeness of the analysis; few previous studies have examined testability, reliability, normative values and validity. A limitation is the limited or lacking visual acuity data in the younger age-groups, and the lack of information on refraction. Using cyclopleged refraction would have enabled us to exclude anisometropic children from the normative sample, but would have required opt-in consent, which as we have seen would have likely biased the sample towards children with better vision. Different design choices have different strengths and weaknesses, which is why it is valuable to have results from many studies with different designs.

Supporting information

S1 File. Word document containing details of analyses referred to in the main text: (1) Normative data with more stringent limits on visual acuity; (2) Sensitivity and specificity for limited data set where full data were available; (3) Sensitivity and specificity for strabismus vs amblyopia.

(DOCX)

Acknowledgments

This manuscript presents independent research commissioned by the Health Innovation Challenge Fund (HICF-R8-442 and WT102565/z/13/z), a parallel funding partnership between the Wellcome Trust and the Department of Health. The views expressed in this manuscript are those of the authors and not necessarily those of the Wellcome Trust or the Department of Health.

Disclosure

The authors have developed their own stereotest, which has recently been licensed to a company. The authors have no personal financial interest in this or any relevant product.

Data Availability

All data, along with R markdown code to analyse them and produce the figures in the paper, are available at the following link: https://doi.org/10.25405/data.ncl.9755045.

Funding Statement

This manuscript presents independent research commissioned by the Health Innovation Challenge Fund (HICF-R8-442 and WT102565/z/13/z to JCAR and MPC), a parallel funding partnership between the Wellcome Trust (wellcome.ac.uk) and the UK Department of Health (www.gov.uk/government/organisations/department-of-health-and-social-care). The views expressed in this manuscript are those of the authors and not necessarily those of the Wellcome Trust or the Department of Health. The funders played no role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Bradshaw MF, Glennerster A. Stereoscopic acuity and observation distance. Spat Vis. 2006/01/18. 2006;19: 21–36. Available: http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=16411481 [DOI] [PubMed] [Google Scholar]
  • 2.Elliott S, Shafiq A. Interventions for infantile esotropia. Cochrane Database Syst Rev. 2013;7: CD004917 10.1002/14651858.CD004917.pub3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Read JCA. Stereo vision and strabismus. Eye (Lond). 2015;29: 214–24. 10.1038/eye.2014.279 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Li T, Shotton K. Conventional occlusion versus pharmacologic penalization for amblyopia. Cochrane Database Syst Rev. 2009; CD006460 10.1002/14651858.CD006460.pub2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Rowe FJ, Noonan CP. Botulinum toxin for the treatment of strabismus. Cochrane Database Syst Rev. 2012;2: CD006499 10.1002/14651858.CD006499.pub3 [DOI] [PubMed] [Google Scholar]
  • 6.Stewart CE, Wallace MP, Stephens DA, Fielder AR, Moseley MJ, Cooperative M. The effect of amblyopia treatment on stereoacuity. J AAPOS. 2013;17: 166–173. 10.1016/j.jaapos.2012.10.021 [DOI] [PubMed] [Google Scholar]
  • 7.Wong AM. Timing of surgery for infantile esotropia: sensory and motor outcomes. Can J Ophthalmol. 2008;43: 643–651. 10.3129/i08-115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Adams WE, Leske DA, Hatt SR, Holmes JM. Defining real change in measures of stereoacuity. Ophthalmology. 2008/12/19. 2009;116: 281–285. 10.1016/j.ophtha.2008.09.012 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Broadbent H, Westall C. An evaluation of techniques for measuring stereopsis in infants and young children. Ophthalmic Physiol Opt. 1990/01/01. 1990;10: 3–7. Available: http://www.ncbi.nlm.nih.gov/pubmed/2184389 [PubMed] [Google Scholar]
  • 10.Simons K. A comparison of the Frisby, Random-Dot E, TNO, and Randot circles stereotests in screening and office use. Arch Ophthalmol. 1981/03/01. 1981;99: 446–452. Available: http://www.ncbi.nlm.nih.gov/pubmed/7213163 10.1001/archopht.1981.03930010448011 [DOI] [PubMed] [Google Scholar]
  • 11.Vancleef K, Read JCA. Which Stereotest do You Use? A Survey Research Study in the British Isles, the United States and Canada. Br Ir Orthopt J. 2019;15: 15–24. 10.22599/bioj.120 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Birch E, Williams C, Drover J, Fu V, Cheng C, Northstone K, et al. Randot Preschool Stereoacuity Test: normative data and validity. J AAPOS. 2007/08/28. 2008;12: 23–26. 10.1016/j.jaapos.2007.06.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Afsari S, Rose KA, Pai AS, Gole GA, Leone JF, Burlutsky G, et al. Diagnostic reliability and normative values of stereoacuity tests in preschool-aged children. Br J Ophthalmol. 2013;97: 308–313. 10.1136/bjophthalmol-2012-302192 [DOI] [PubMed] [Google Scholar]
  • 14.Tarczy-Hornoch K, Lin J, Deneen J, Cotter SA, Azen SP, Borchert MS, et al. Stereoacuity Testability in African-American and Hispanic Pre-School Children. Optom Vis Sci. 2008;85: 158–163. 10.1097/OPX.0b013e3181643ea7 [DOI] [PubMed] [Google Scholar]
  • 15.Yang JW, Son MH, Yun IH. A Study on the Clinical Usefullness of Digitalized Random-dot Stereoacuity Test. Korean J Ophthalmol. 2004;18: 154 10.3341/kjo.2004.18.2.154 [DOI] [PubMed] [Google Scholar]
  • 16.Pai AS, Rose KA, Samarawickrama C, Fotedar R, Burlutsky G, Varma R, et al. Testability of refraction, stereopsis, and other ocular measures in preschool children: the Sydney Paediatric Eye Disease Study. J AAPOS. 2012/04/25. 2012;16: 185–192. 10.1016/j.jaapos.2011.09.017 [DOI] [PubMed] [Google Scholar]
  • 17.Trager MJ, Dirani M, Fan Q, Gazzard G, Selvaraj P, Chia A, et al. Testability of Vision and Refraction in Preschoolers: The Strabismus, Amblyopia, and Refractive Error Study in Singaporean Children. Am J Ophthalmol. 2009;148: 235–241.e6. 10.1016/j.ajo.2009.02.037 [DOI] [PubMed] [Google Scholar]
  • 18.Fawcett SL, Birch EE. Interobserver test-retest reliability of the Randot preschool stereoacuity test. J AAPOS. 2000/12/22. 2000;4: 354–358. 10.1067/mpa.2000.110340 [DOI] [PubMed] [Google Scholar]
  • 19.Smith SJ, Leske DA, Hatt SR, Holmes JM. Stereoacuity Thresholds before and after Visual Acuity Testing. Ophthalmology. 2012;119: 164–169. 10.1016/j.ophtha.2011.06.041 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Bland JM, Altman DG. Statistical methods for assessing agreement between two methods of clinical measurement. Lancet. 1986;1: 307–310. Available: http://www.ncbi.nlm.nih.gov/pubmed/2868172 [PubMed] [Google Scholar]
  • 21.Anderman C, Cheadle A, Curry S, Diehr P, Shultz L, Wagner E. Selection Bias Related To Parental Consent in School-Based Survey Research. Eval Rev. 1995;19: 663–674. 10.1177/0193841X9501900604 [DOI] [Google Scholar]
  • 22.Severson HH, Ary D V. Sampling bias due to consent procedures with adolescents. Addict Behav. 1983;8: 433–437. 10.1016/0306-4603(83)90046-1 [DOI] [PubMed] [Google Scholar]
  • 23.Public Health England. Child vision screening: Service specification. Available: https://www.gov.uk/government/publications/child-vision-screening/service-specification
  • 24.Stidwill D, Fletcher R. Normal Binocular Vision: Theory, Investigation and Practical Aspects. Wiley-Blackwell; 2011. [Google Scholar]
  • 25.Vancleef K, Serrano-Pedraza I, Sharp C, Slack G, Black C, Casanova T, et al. ASTEROID: A New Clinical Stereotest on an Autostereo 3D Tablet. Transl Vis Sci Technol. 2019;8: 25 10.1167/tvst.8.1.25 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Serrano-Pedraza I, Herbert W, Villa-Laso L, Widdall M, Vancleef K, Read JCA. The stereoscopic anisotropy develops during childhood. Investig Ophthalmol Vis Sci. 2016;57 10.1167/iovs.15-17766 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Hess RF, Ding R, Clavagnier S, Liu C, Guo C, Viner C, et al. A Robust and Reliable Test to Measure Stereopsis in the Clinic. Investig Opthalmology Vis Sci. 2016;57: 798–804. 10.1167/iovs.15-18690 [DOI] [PubMed] [Google Scholar]
  • 28.Bosten JM, Goodbourn PT, Lawrance-Owen AJ, Bargary G, Hogg RE, Mollon JD. A population study of binocular function. Vision Res. 2015;110: 34–50. 10.1016/j.visres.2015.02.017 [DOI] [PubMed] [Google Scholar]
  • 29.Schmidt P, Maguire M, Kulp MT, Dobson V, Quinn G. Random Dot E stereotest: testability and reliability in 3- to 5-year-old children. J AAPOS. 2006/12/26. 2006;10: 507–514. 10.1016/j.jaapos.2006.08.019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Venables WN (William N., Ripley BD, Venables WN (William N). Modern applied statistics with S. Available: https://cran.r-project.org/web/packages/MASS/citation.html
  • 31.Shafiee D, Jafari AR, Shafiee AA. Correlation between Interpupillary Distance and stereo acuity. Bull Environ Pharmacol Life Sci. 2014;3: 26–33. [Google Scholar]
  • 32.Zaroff CM, Knutelska M, Frumkes TE. Variation in stereoacuity: normative description, fixation disparity, and the roles of aging and gender. Invest Ophthalmol Vis Sci. 2003;44: 891–900. Available: http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=12556426 10.1167/iovs.02-0361 [DOI] [PubMed] [Google Scholar]
  • 33.Mai MN, Schlueter MA. The Relationship Between Pupillary Distance and Depth Perception in Humans. Investigative Ophthalmology & Visual Science. [Association for Research in Vision and Ophthalmology, etc.]; 2010. pp. 4359–4359. Available: https://iovs.arvojournals.org/article.aspx?articleid=2372974 [Google Scholar]
  • 34.Eom Y, Song JS, Ahn SE, Kang SY, Suh YW, Oh J, et al. Effects of interpupillary distance on stereoacuity: the Frisby Davis distance stereotest versus a 3-dimensional distance stereotest. Jpn J Ophthalmol. 2013/07/06. 2013. 10.1007/s10384-013-0253-9 [DOI] [PubMed] [Google Scholar]
  • 35.Pryor HB. Objective measurement of interpupillary distance. Pediatrics. 1969;44: 973–7. Available: http://www.ncbi.nlm.nih.gov/pubmed/5365062 [PubMed] [Google Scholar]
  • 36.Fledelius HC, Stubgaard M. Changes in eye position during growth and adult life as based on exophthalmometry, interpupillary distance, and orbital distance measurements. Acta Ophthalmol. 1986;64: 481–6. Available: http://www.ncbi.nlm.nih.gov/pubmed/3492853 [DOI] [PubMed] [Google Scholar]
  • 37.Horwood AM, Riddell PM. Gender differences in early accommodation and vergence development. Ophthalmic Physiol Opt. 2008;28: 115–126. 10.1111/j.1475-1313.2008.00547.x [DOI] [PubMed] [Google Scholar]
  • 38.Gwiazda J, Bauer J, Held R. Binocular function in human infants: correlation of stereoptic and fusion-rivalry discriminations. J Pediatr Ophthalmol Strabismus. 1989/05/01. 1989;26: 128–132. Available: http://www.ncbi.nlm.nih.gov/pubmed/2723974 [DOI] [PubMed] [Google Scholar]
  • 39.Held R, Thorn F, Gwiazda J, Bauer J. Development of binocularity and its sexual differentiation In: Vital-Durand F, Atkinson J, Braddick OJ, editors. Infant Vision. Oxford: OUP; 1996. 10.1093/acprof:oso/9780198523161.003.0017 [DOI] [Google Scholar]
  • 40.Seuntiëns P, Meesters L, IJsselsteijn W. Perceived quality of compressed stereoscopic images: effects of symmetric and asymmetric JPEG coding and camera separation. ACM Trans Appl Percept. 2000. [Google Scholar]
  • 41.Yeh YY, Silverstein LD. Limits of fusion and depth judgment in stereoscopic color displays. Hum Factors. 1990;32: 45–60. Available: http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?cmd=Retrieve&db=PubMed&dopt=Citation&list_uids=2376407 10.1177/001872089003200104 [DOI] [PubMed] [Google Scholar]
  • 42.Tsirlin I, Wilcox LM, Allison RS. The Effect of Crosstalk on the Perceived Depth From Disparity and Monocular Occlusions. IEEE Trans Broadcast. 2011;57: 445–453. 10.1109/TBC.2011.2105630 [DOI] [Google Scholar]

Decision Letter 0

Guido Maiello

29 Aug 2019

PONE-D-19-21006

Characterizing the Randot Preschool Stereotest: testability, norms, reliability, specificity and sensitivity in children aged 2-11 years

PLOS ONE

Dear Prof Read,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Two expert reviewers have evaluated your manuscript. Reviewer 1 is very positive and provides a series of useful comments and suggestions to improve the paper. Reviewer 2 is somewhat more critical. They ask that the study be motivated better and point out a couple of potential issues regarding ceiling effects. The reviewer however also suggests how to fix these issues. I therefore expect that it should not be too difficult for the authors to address the points raised by the reviewers, and I look forward to receiving your revised work.

We would appreciate receiving your revised manuscript by Oct 13 2019 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Guido Maiello

Academic Editor

PLOS ONE

Journal Requirements:

1. When submitting your revision, we need you to address these additional requirements.

Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.

3. Thank you for stating the following in the Competing Interests section:

The authors have developed their own stereotest, which has recently been licensed to a company. The authors have no personal financial interest in this or any relevant product.

Please confirm that this does not alter your adherence to all PLOS ONE policies on sharing data and materials, by including the following statement: "This does not alter our adherence to  PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests).  If there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please include your updated Competing Interests statement in your cover letter; we will change the online submission form on your behalf.

Please know it is PLOS ONE policy for corresponding authors to declare, on behalf of all authors, all potential competing interests for the purposes of transparency. PLOS defines a competing interest as anything that interferes with, or could reasonably be perceived as interfering with, the full and objective presentation, peer review, editorial decision-making, or publication of research or non-research articles submitted to one of the journals. Competing interests can be financial or non-financial, professional, or personal. Competing interests can arise in relationship to an organization or another person. Please follow this link to our website for more details on competing interests: http://journals.plos.org/plosone/s/competing-interests

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: I Don't Know

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The study evaluates the normative values, testability, and reliability of the Randot Preschool stereotest by age and sex in children 3-11. It also assesses sensitivity and specificity of the Randot Preschool to detect binocular vision issues. The strength of the study is to have a very large sample (>1000 children) and precise procedures. The manuscript is very well written, clear, and easy to read. Each result is appropriately backed up with data and statistical tests and conclusions match the data. Each result is discussed in a very interesting way. I have only a few comments, mostly minor.

Minor comments:

pp5 line 93 = « values in italics » - there is no values in italics. Also it is not clear why the ref 13 is in red, given it is not your data, isnt't it?

Pp5 – your threshold values are systematically above the other studies. I have seen that you discuss why later. I suspect it comes from the instructions: you used PEDIG instructions while the other studies may have used manufacturer instructions. The manufacturer proposes to start at 200” rather at 800”, and then to go up or down depending on success. It certainly increases the probability to be successful only by chance which lowers the average score. The PEDIG instructions certainly lead to more accurate estimations of scores.

Pp6 – line 124 : “We also excluded any reporting values solely in a clinical population, except that we did include reliability measures in clinical populations of the relevant ages.” I do not understand what you mean here.

Pp8 – line 170 and pp12 line 278 – if I understand correctly, around half of the children did not have a questionnaire filled by the time of the screening and therefore were not asked to wear glasses during testing (even if some of them might need glasses). The fact that they are not wearing glasses could explain why their stereoacuity and visual acuity is lower. Do you know what procedure other studies used regarding the correction of refraction when testing stereopsis? Could it explain the larger stereoacuity values that you obtain (if other studies actually systematically measured refraction and used frame glasses for stereopsis testing, or excluded the children with no known prescription)?

Pp13 line 293 and 300: the main result here seems to be that testability is virtually 100% after age 4. Given you have many children, would it make sense to do a statistical test (possible showing whether a confidence interval includes 1)? I think this could show that the testability is 1 at least from age x.

Pp20 – please provide a measure of goodness of fit for your model. Would you be able to justify whether it is necessary to have 8 parameters rather than 2 or 3, using any procedure to compare models (ex: AIC)? It looks to me that the distributions in Fig 8 could be described by a gamma distribution with two parameters, provided that age and sex can at least influence the scale parameter. You could compare your model to this one.

Pp23 – line 541 – missing word? “Figure” 10.

Pp23 – line 546– why include condition (c)? It does not seem to be a binocular vision problem.

Pp24 – line 588- I do not think that specificities can be considered “high” or “excellent”’ in regards to detecting participants with no binocular problem considering that the proportion of children with binocular problems is so low (4% - the target proportion). When the target proportion is that low, it is crucial that specificity is extremely high to avoid drastically decreasing the positive predictive value normalized to the target proportion. For example, if sensitivity is 100% and specificity is 96%, positive predictive value normalized to the target proportion (4%) is only 50%. Could you report positive and negative predictive values normalized to the target population as well?

Abstract: you report in the abstract that 6% of the children are stereoblind but you do not mention it in the text. Do you mean that 6% of all testable children are stereoblind or 6% of children with otherwise normal vision?

Reviewer #2: Review of Randot paper

This paper presents compressive data on the Random Preschool Stereotest administered to over 1000 children aged 2-11 years. However, its purpose is unclear because there are published data on every one of the measures the authors provide. They state their goal as getting the measures all in the same study but it is not clear why testability from one cohort and scientist and sensitivity on another cohort measured by a different scientist are any less valuable than the data presented here in separate sections of the paper for each measure (and sometimes different cohorts). The authors need to provide a stronger rationale.

The measures are also limited by a ceiling effect (no value can be tested finer than 40 arcsec). That means there is very little variance in the data from children above age 5, severely limiting the possibility of valid measures of reliability or meaningful norms. For the tests of sensitivity and specificity, the outcome will depend on the cutoffs that are used, which likely need to differ by age up to age 6—as is clear in Figure 10. This makes Figure 9 meaningless. The small number of children with binocular problems on other measures and the variety of other measures used also make the sensitivity data difficult to interpret. A brief report on the measures that are not limited by the ceiling—i.e., for the younger children only—could be of interest if the authors explain how it differs from the existing literature.

The sex difference is interesting and reminiscent of the one identified during infancy on the onset of stereopsis, although that sex difference may be an consequence of sex differences in accommodative and vergence eye movements:

Gwiazda, J., Bauer, J., & Held, R. (1989). Binocular function in human infants: correlation of stereoptic and fusion-rivalry discriminations. J Pediatr Ophthalmol Strabismus, 26(3), 128-132. Retrieved from https://www.ncbi.nlm.nih.gov/pubmed/2723974

Horwood, A. M., & Riddell, P. M. (2008). Gender differences in early accommodation and vergence development. Ophthalmic Physiol Opt, 28(2), 115-126. doi:10.1111/j.1475-1313.2008.00547.x

Minor points:

1, The authors transform all values to log units because they are more likely to be normally distributed. That’s a sound rationale but makes it very difficult to relate the findings to the actual 6 value it is possible to get on the test.

2, The acuity chart is inappropriate for children older than 4 because there are no values smaller than 0.2 loMAR. Children of age 5+ usually achieve an acuity of logMAR .10. This limits the usefulness of the analysis of sensitivity and specificity.

3, The interocular acuity difference consider indicative of abnormal binocular vision was greater than logMAR .20 (which is more than 2 lines). Amblyopia is normally defined as a difference of 2 lines or more.

4, Taking 800 arcsecs (Table 8) is an inappropriate cutoff because by age 4 almost all children score better than that.

5, Because there are only 6 discrete values and the test is not designed to determine a threshold, it’s not surprising that reliability is poor. The authors should make this point and not spend so much space analyzing the reliability data.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

Decision Letter 1

Guido Maiello

27 Sep 2019

PONE-D-19-21006R1

Characterizing the Randot Preschool Stereotest: testability, norms, reliability, specificity and sensitivity in children aged 2-11 years

PLOS ONE

Dear Prof Read,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Reviewer 1 was satisfied with your revisions. Reviewer 2 was not available to review your revised manuscript before next month. Therefore, in order to expedite the process, I decided to assess your revisions myself. Overall, I think you did a fine job at addressing the reviewer comments, and that the manuscript has improved because of this. I’ve spotted a few remaining minor issues, which however should be very easy to address.

We would appreciate receiving your revised manuscript by Nov 11 2019 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Guido Maiello

Academic Editor

PLOS ONE

Additional Editor Comments (if provided):

Minor Comments:

1) I believe the x axis in figures 8 and 10 should read “Randot stereoacuity (arcsec)” instead of “Age-group”.

2) I know you have correctly uploaded your data to a public repository, but I’ve also noted that in a few places you reference analyses without reporting them. The PLOS guidelines state:

“Please note that PLOS does not permit references to “data not shown.” Authors should provide the relevant data within the manuscript, the Supporting Information files, or in a public repository. If the data are not a core part of the research study being presented, we ask that authors remove any references to these data.”

Therefore, I ask you to add these analyses as brief supporting information. Specifically, I am referring to:

Page 22 line 515: “data not shown”

Page 24 line 580: “not shown”

Page 25 Line 615 “Some authors have suggested that random-dot stereotests are more sensitive to strabismus than to amblyopia, but neither our data nor that of Afsari et al(13) suggests a significant difference in sensitivity.”

3) On page 24 line 585 you say that crosstalk “makes it possible to pass the higher levels of the test with one eye”. I think you should explain briefly how this might work, as it is not very intuitive. Additionally, you should mention that crosstalk might also worsen your results, since crosstalk decreases perceived depth:

Tsirlin, I., Wilcox, L. M., & Allison, R. S. (2012). Effect of crosstalk on depth magnitude in thin structures. Journal of Electronic Imaging, 21(1), 011003.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

Decision Letter 2

Guido Maiello

14 Oct 2019

Characterizing the Randot Preschool Stereotest: testability, norms, reliability, specificity and sensitivity in children aged 2-11 years

PONE-D-19-21006R2

Dear Dr. Read,

We are pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it complies with all outstanding technical requirements.

Within one week, you will receive an e-mail containing information on the amendments required prior to publication. When all required modifications have been addressed, you will receive a formal acceptance letter and your manuscript will proceed to our production department and be scheduled for publication.

Shortly after the formal acceptance letter is sent, an invoice for payment will follow. To ensure an efficient production and billing process, please log into Editorial Manager at https://www.editorialmanager.com/pone/, click the "Update My Information" link at the top of the page, and update your user information. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, you must inform our press team as soon as possible and no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

With kind regards,

Guido Maiello

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

A few final, extremely minor notes:

page 24 line 602: the caption of Figure 11, if the order is meant to be clockwise, I believe should read:

"the random dots depict blank, fir-tree,disk, star."

And perhaps the panels on the right side of figure 11 could be removed, they are slightly confusing as they don't match the stereograms on the right.

page 25, line 621: please reference the supplementary analyses, e.g. by amending the text to read

"Some authors have suggested that random-dot stereotests are more sensitive to strabismus than to amblyopia (Supporting Information, section 3)

Typo in the legend of table 4 of the supporting information: "amblyopiat".

Reviewers' comments:

Acceptance letter

Guido Maiello

23 Oct 2019

PONE-D-19-21006R2

Characterizing the Randot Preschool Stereotest: testability, norms, reliability, specificity and sensitivity in children aged 2-11 years

Dear Dr. Read:

I am pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

For any other questions or concerns, please email plosone@plos.org.

Thank you for submitting your work to PLOS ONE.

With kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Guido Maiello

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. Word document containing details of analyses referred to in the main text: (1) Normative data with more stringent limits on visual acuity; (2) Sensitivity and specificity for limited data set where full data were available; (3) Sensitivity and specificity for strabismus vs amblyopia.

    (DOCX)

    Attachment

    Submitted filename: ResponsetoReviewers.docx

    Attachment

    Submitted filename: RevComms.docx

    Data Availability Statement

    All data, along with R markdown code to analyse them and produce the figures in the paper, are available at the following link: https://doi.org/10.25405/data.ncl.9755045.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES