Abstract
Individuals with an autism spectrum disorder (ASD) diagnosis are often described as having an eye for detail. But it remains to be shown that a detail-focused processing bias is a ubiquitous property of vision in individuals with ASD. To address this question, we investigated whether a greater number of autistic traits in neurotypical subjects is associated with an increased reliance on image details during a natural image recognition task. To this end, we use a novel reverse correlation-based method (feature diagnosticity mapping) for measuring the relative importance of low-level image features for object recognition. The main finding of this study is that image recognition in participants with an above-median number of autistic traits benefited more from the presence of high-spatial frequency image features. Furthermore, we found that this reliance-on-detail effect was best predicted by the presence of the most clinically relevant autistic traits. Therefore, our findings suggest that a greater number of autistic traits in neurotypical individuals is associated with a more detail-oriented visual information processing strategy and that this effect might generalize to a clinical ASD population.
Subject terms: Cognitive neuroscience, Perception, Object vision, Autism spectrum disorders
Introduction
Autism is a developmental disorder, now referred to as autism spectrum disorder (ASD), that manifests itself in a variety of forms. Diagnostic criteria for ASD include persistent deficits in social interactions and communication and repetitive patterns of behavior1. An intriguing aspect of ASD is that it is also associated with superior performance for tasks that involve the processing of visual detail. For example, ASD has been associated with superior performance for the embedded figures test2–4, which involves searching for a simple shape contained by a complex figure, and faster identification of the odd-man-out in cluttered displays5–7. A well-known, albeit unrepresentative, case of ASD-related enhanced processing of visual detail is the savant ability of Stephen Wiltshire, who is able to draw highly detailed urban landscapes after having seen his subject only briefly (https://www.stephenwiltshire.com).
In short, ASD has been associated with having an eye for detail. This, however, has been argued to come at the cost of a reduced ability to ‘see the big picture’ according to the influential weak central coherence (WCC) theory8,9. The original WCC formulation10 proposed that a bias towards processing details might underlie deficits in social functioning central to ASD11: a focus on details could cause individuals with ASD to miss socially meaningful cues that are global in nature, like facial expressions12. However, a string of studies finding no evidence for a relationship between perceptual measures of weak central coherence and measures of theory of mind and social skills13–17 led to a revised version of WCC9 suggesting that a bias towards processing details and social deficits might be two distinct aspects of ASD. Consequently, Happé and Frith9 pointed out that the future veracity of WCC critically depends on establishing a relationship between a detail-focused processing bias and real-life abilities and difficulties in ASD individuals. One way of supporting the feasibility of such a relationship would be to demonstrate that this bias is a ubiquitous property of vision in individuals with ASD.
ASD-related detail-focused processing has been initially supported by superior detail-focused abstract visual tasks, including the embedded figure task3 and the Navon task17. These early findings, however, are highly controversial given the fact that more recent studies have frequently been unable to replicate them18. An important endeavour of ASD research has been to determine if this detail-focused perceptual style is grounded in deviant low-level sensory processing in early sensory brain areas19. Despite the fact that several basic measures of visual sensitivity appear to be unaffected by ASD16,20–23, detail-focused processing has been associated with ASD in the context of face and object perception20,22,24. In addition, mixed evidence has been provided for ASD being related to enhanced sensitivity to high vs. low spatial frequency grating stimuli, as earlier studies found no such relationship25,26 while a more recent study by Kéïta et al.20 did find a relationship between ASD and enhanced sensitivity for high-spatial frequency grating stimuli.
Given the restricted and mixed nature of the current literature on ASD-related detail-focused processing, further research is needed to resolve whether such a bias represents a ubiquitous property of natural day-to-day image perception in individuals with ASD. Therefore, we here use a new reverse-correlation based ‘feature diagnosticity mapping’ paradigm to determine how the presence of autistic traits affects the relative contribution of low-level visual features to natural image recognition. Specifically, we used the 50-item Autism Spectrum Quotient (AQ) questionnaire for adults24 to split our neurotypical participants into a high- and low AQ group and assessed if image recognition depended more on high-spatial frequency stimulus features in the high AQ group. This approach relies on the notion that the comparison of individuals with many vs. few autistic traits can effectively model differences between individuals with and without a clinical ASD diagnosis4,25–27. In addition to testing for a relationship between ASD and detail-based image recognition, this approach is also used to test if previous reports of reduced gaze duration towards eyes28–30 and increased gaze durations for the central area of images31 can be related to image recognition, with individuals with ASD relying more on these aspects of images.
Results
We developed a new experimental paradigm to measure the relative contribution of low-level image features to image recognition, using a technique similar to reverse correlation32,33. During the experiment, 52 participants were presented with partial reconstructions of five cat and five dog images. To create these stimuli, we first selected 1,000 Gabor wavelets (with varying position, spatial frequency and orientation) which, when summed, provided a good estimate of pixel intensity values of the original cat and dog images which one can easily recognize as a dog or cat image (sums of all features are displayed in the Supplementary figure). Partial reconstructions contained a random selection of 90 of the 1,000 features (Fig. 1). Via button presses, participants indicated whether they recognized a dog, a cat or whether they were not sure.
AQ-scores of the participants were all within the neurotypical range (< 32) and ranged from 5 to 30 (M = 14.3, SD = 5.9). This range of AQ scores is somewhat lower than the typically reported AQ score range34 which can be explained by the fact that our participants were predominantly female (41 female, 11 male). In this context, it is worth noting the AQ test has been primarily been validated for ASD patient groups that were predominantly male24. Participants were assigned to the high AQ group (n = 25) if they had an autistic-spectrum quotient24 (AQ) higher than the median across all participants (AQ > 14) while the others were assigned to the low AQ group (n = 27). We opted for this AQ cut-off approach because it enabled us to include all participants in the analysis and because the outcome of this analysis type, in contrast to a correlational analysis, depends less on hard-to-interpret effects of small AQ-score difference. The AQ-score cut-off used during this study (AQ > 14) was comparable to the cut-offs used by previous studies using the same median-split approach35–38.
On average, participants successfully recognized 49.7% (SD = 15.6%) of the partial reconstructions. During these trials they either reported having recognized a cat when a cat was shown or reported having recognized a dog when a dog was shown. Note that participants, in addition to reporting that they recognized the image as a cat/dog, could also indicate that they were ‘unsure’, which they did during 29.3% of the trials (SD = 22.9%). During the remaining 21.0% of trials participants reported seeing a different animal than displayed. As a result, recognition performance reported here should not be confused with recognition performance for two-alternative forced choice paradigms. A repeated-measure ANOVA revealed that there was no difference in recognition performance between cat and dog images (50.9% and 48.4% respectively, F(1, 100) = 0.50, p = 0.48), no effect of AQ group on recognition performance (high AQ group: 47.4%; low AQ group: 51.9%; F(1, 100) = 2.49, p = 0.12), nor an interaction between these two factors (F(1, 100) = 1.23, p = 0.27).
To quantify the relative importance of visual features we computed a feature diagnosticity index (FDi) for all 10,000 features (10 images with 1,000 features each) based on the average recognition accuracy for trials containing each feature. This value, however, is affected by variance due to differences in the recognizability between images and by overall performance differences between participants (across all images and image features). This variance, however, is undesirable because we are only interested in the relative contribution of features to image recognition within each image and within each participant. Therefore, we z-scored FDi values within each participant and image (across the 1,000 features of each image). If FDi values truly reflect feature importance for image recognition, images resulting from the summation of features with the highest FDi values should be most recognizable. Visual inspection of images reconstructed from the 200 features with the highest FDi values revealed that these images were easier to recognize than images reconstructed from the 200 features with the lowest FDi values (shown for one exemplary image in Fig. 1c and for all images in the Supplementary figure). This observation, however, does not quantitatively validate the efficacy of our method. Therefore, we have, in addition, validated our method by assessing the replicability of the pattern of FDi values (across all 10,000 features) between participants. This analysis revealed significant replicability of FDi patterns across participants (Pearson r = 0.081, p < 0.0001, permutation-based test, Fig. 1c).
After having established that FDi values measure the importance of features for image recognition, we tested if FDi values are elevated for high-spatial frequency features in high-AQ individuals. To this end, we grouped features into five equally-sized bins (containing 2000 features each) with ascending feature spatial frequencies. Spatial frequency ranges for bin 1–5 in cycles per degree visual angle were resp.: 0.24–0.60, 0.60–0.89, 0.89–1.21, 1.21–1.52 and 1.52–2.07. We tested for an interaction between spatial frequency and AQ group using a repeated-measure ANOVA with average bin FDi values as the dependent variable. This analysis revealed a main effect of spatial frequency (F (4, 250) = 3.97, p < 0.005, Fig. 2), and an interaction between spatial frequency and AQ group (F (4, 250) = 4.12, p < 0.005, Fig. 2). The interaction was mostly driven by the high-AQ group having elevated FDi values for the highest spatial frequency bin (0.005 and − 0.013 resp.; t(50) = 3.80, p < 0.001). Therefore, our results are consistent with high-AQ individuals relying more on local details for image recognition.
Previous behavioural studies have linked ASD to reduced gaze durations for the eye-region in human faces28–30, and increased gaze durations for the central area of images31. To assess this, we repeated the analysis while grouping features into five ascending bins (containing 2,000 features each) according to feature distance from the nearest eye (bin 1–5 resp. ranged from 0° to 2.9°, 2.9°–5.3°, 5.3°–7.3°, 7.3°–9.8° and 9.8°–18.4° visual angle) or distance from the image centre (bin 1–5 resp. ranged from 0.1° to 4.6°, 4.6°–6.7°, 6.7°–8.6°, 8.6°–10.6° and 10.6°–15.7° visual angle). This analysis revealed two main effects which indicate that FDi decreases as a function of the distance from the nearest eye (F (4, 250) = 27.4, p < 0.00001, Fig. 2) and distance from the image centre (F (4, 250) = 14.3, p < 0.00001, Fig. 2). Importantly, neither of these effects was modulated by AQ group. Therefore, our data does not replicate previously reported ASD effects on image centre and eye-region processing. It is important to note that in contrast to these previous studies, our results are not based on eye movements nor on data from clinically diagnosed ASD patients.
Does our finding of an increased reliance-on detail for visual recognition in high AQ individuals generalize to individuals with an ASD diagnosis? Our study does not provide direct evidence for this as we measured a neurotypical student population. However, we are able to provide indirect evidence by testing if visual-detail-reliance depends most on the presence of clinically diagnostic AQ traits (with ‘trait’ we refer to a positive score on one of the 50 items of the AQ questionnaire). To this end, we quantified the clinical diagnosticity the 50 autism traits as the (natural) log of the trait prevalence ratio between clinically diagnosed ASD individuals and neurotypical students, making use of previously published prevalence data24. For example, the highest log odds ratio (2.04) was assigned to the trait measured with the item “I enjoy social occasions”, which ASD individuals disagree with 7.7 times more often than neurotypical students. In addition, we quantified the reliance-on-detail for each participant as the linear regression coefficient between their average FDi values for each spatial frequency bin and the ascending bin numbers (1–5, see Fig. 2). Subsequently, we performed a robust regression analysis that confirmed our hypothesis by revealing that our reliance-on-detail measure was best predicted by the presence of the most clinical diagnostic autistic traits (slope = 0.0023, t(48) = 3.40, p < 0.005, Fig. 3).
In addition, we assessed how trait-related increases of reliance-on-detail depended on the five different trait types (social skill, attention switching, attention to detail, communication and imagination) that the AQ questionnaire was designed to measure (with ten items for each type of ASD trait)24. Note here that the trait types labelled as ‘social skill’, ‘attention switching’, ‘communication’ and ‘imagination’ all refer to difficulties in the respective domains while ‘attention to detail’ refers to its domain positively. A one-way ANOVA revealed an effect of trait type (F(4, 45) = 5.89, p < 0.001). Five post-hoc t-tests (Bonferroni corrected) testing for a difference between each trait type versus all others revealed that the presence of ‘social skill’ traits increase reliance-on-detail more than all other trait types (t(48) = 3.72, p < 0.005) while ‘attention to detail’ traits give rise to a below-average effect (t(48) = − 3.48, p < 0.01). Therefore, our findings suggest that enhanced reliance on detail for image recognition is most predictive for autistic traits related to self-reported social difficulties.
Finally, we analysed reaction times (M = 895 ms, SD = 313 ms). We found that high-AQ participants took longer to respond than low-AQ participants (993 ms vs 804 ms, t(50) = 2.25, p < 0.05). Unfortunately, reaction times were found to carry no information about the relative importance of visual features for image recognition as reaction time based FDi values did not replicate across participants (Pearson r = 0.0077, p = 0.14, permutation-based test).
Discussion
In summary, we developed a method that measures the relative contribution of low-level visual features to image recognition. With this method, we obtained evidence for natural image recognition depending more on high spatial frequency features in individuals with an above-median number of autistic traits. Therefore, the presence of a greater number of autistic traits appears to predict enhanced reliance on fine details for natural image recognition. This effect was found to be driven most by the presence of autistic traits with the highest clinical relevance, which increases the likelihood that our main finding generalizes to clinically diagnosed ASD individuals.
Our study sheds new light on the ongoing debate regarding whether a bias towards processing details underlies ASD-related social deficits4,9,10,13–16,39 as we find that reliance on details is best predicted by positive scores for AQ items from the AQ subscale measuring self-reported social difficulties. Therefore, in contrast to findings from a fairly large body of research10,13–16,39, our results suggest that social deficits central to ASD might be related to having an eye for detail. Our findings, however do converge with a recent study demonstrating that performance for the embedded figure task is positively related to scores for items of the same social skill AQ-subscale4. In this context, it is important to note that previous research indicates that all AQ-subscales, except for the ‘attention to detail’ subscale, reflect a single general ASD attribute, thought to be related to difficulties in social interactions40, rather than specific ASD symptoms41. Consistent with this, we found that scores for all subscales, except for ‘attention to detail’, were positively associated with our measure of detail-focused visual information processing. Furthermore, we would like to point out that the ‘attention to detail’ AQ subscale contains items measuring attention towards both perceptual and more abstractly defined details (e.g. item 32: “I notice patterns in things all the time.”), which could explain why we observed no relationship between scores for this subscale and detail-focused visual information processing.
In sum, our findings suggest that original proposal of detail-focused processing underlying the social deficits associated with ASD8 appears to be worth revisiting. An important limitation of our study, however, is that our study is restricted to comparing groups of neurotypical participants with above- and a below median AQ scores. Despite the fact that this comparison has proven to be extremely useful for modelling ASD42, an ultimate verification of a relationship between detail-focused visual information processing during natural image recognition and ASD-related social difficulties would require replication of our findings for a clinically diagnosed ASD population.
If ASD is associated with image recognition being driven more by details, what could be the underlying neural mechanisms? One possibility is that individuals with ASD process visual information differently already at an early stage of cortical visual information processing, e.g. already within the primary visual cortex19. This would be compatible with the finding that ASD individuals appear to have enhanced sensitivity for high-spatial frequency gratings20 and with the finding of early (125 ms post stimulus) enhanced EEG responses to high spatial frequency information in 4-year old children with ASD43. A second possibility—that is compatible with reports suggesting that ASD does not affect elementary vision22,44—would be that the reliance-on-detail effect reported here results from higher level brain areas responsible for object recognition, e.g. the inferior temporal cortex, being more reliant on lower-level representations of high-spatial frequency features. Differentiating between these two possible underlying neural mechanisms will require neuroimaging studies aimed at revealing whether ASD enhances the extent to which high-spatial frequency information is encoded in the primary visual cortex and/or if ASD is associated with enhanced transmission of high-spatial frequency visual information to higher level brain areas.
In contrast to previous eye-tracking studies29–31, we found no evidence for ASD-related reduced processing of the eye-region, nor an increased processing of the central area of images. One possible explanation for this discrepancy might be the fact that we investigated ASD by comparing individuals with above- and below-median AQ scores instead of comparing clinically diagnosed ASD individuals with healthy controls. Another intriguing possible explanation could be that ASD leads to elevated gaze durations to the image centre and eye-region, but that this does not translate into object recognition relying more on the features inside these areas. To test the latter, future studies will need to combine our novel psychophysical paradigm with eye-tracking.
In conclusion, our results show that natural image recognition is driven more by visual details in neurotypical individuals with an elevated number of autistic traits. Given that the most clinically relevant autistic traits best predict this effect, this finding is suggestive for a detail-focused information processing bias being a ubiquitous property of vision in ASD individuals. In addition, this detail-focused processing bias was found to predict increased self-reported social difficulties. Therefore, we propose that the original proposal of WCC8 that enhanced local information processing underlies the social deficits associated with ASD might be worth revisiting.
Methods
Participants
52 healthy student volunteers (11 male, 41 female: average age = 19.4, SD = 1.05) with normal or corrected-to-normal vision took part in this experiment. All participants gave their informed consent after being introduced to the experimental procedure in accordance with the Declaration of Helsinki. The experimental procedure was approved by the ethics committee of the University of Birmingham (ethics reference ERN_15-1374P). Participants all filled out the 50-item Autism Spectrum Quotient (AQ) questionnaire for adults24 and were assigned to the high- and low AQ groups depending on whether their AQ scores exceeded the median AQ score for all participants (AQ > 14).
Stimuli
Portrait images of five cats and five dogs were converted to 250 × 250 grey-scale images. We then used a custom-made algorithm, implemented in Matlab 2016a, to reduce images to 1,000 Gabor wavelets. The aim of this algorithm was to find a set of wavelets that is able to describe most of the coarse and fine image details (see Fig. 1a for a visualization of the features selected for one of the images). Wavelets considered had 29 (n = [1, 2, …, 29]) exponentially increasing spatial frequencies (sf) between 0.24 and 2.07 cycles/visual degree angle:
Wavelets were considered with 18 equidistant orientations between 0° and 180° (0°–170° in steps of 10°). Features were selected iteratively from the lowest to the highest spatial frequency. During the first iteration, wavelets were selected based on the original grey-scale image. For consecutive iterations, the input image was a residual image resulting from least-square regression of the input to the previous iteration while using previously selected wavelets as regressors. During each iteration, eighteen Gabor wavelet filters (one per orientation) were applied to the input image. The output of this analysis enabled us to find the optimally-fitting orientation and phase for Gabor wavelets cantered on each pixel of the input image for the current iterations’ spatial frequency. From these 2502 wavelets, we considered only 25% with the highest amplitude. From these wavelets, we then randomly selected a number of wavelets (nw) increasing as a function of spatial frequency:
The exact parameters of this formula were determined exploratively based on how well the sum of all selected wavelets captured all coarse and fine visual details of the original images (based on visual inspection). This resulted in 21 features being selected for 0.24 cycles/visual degree angle wavelets and 1,008 features for 2.07 cycles/visual degree angle wavelets. During each iteration, we computed the covariance of each feature with the input image and discarded wavelets with covariances smaller than a fifth of the maximum observed covariance value. Finally, we selected 1,000 wavelets from all spatial frequencies (from all iterations) having the highest covariance with the original grey-scale image. Amplitudes where set to an equal value for all wavelets. Partial reconstructions of the images were created by randomly selecting 90 wavelets from the set of 1,000 and summating them (see Fig. 1a). The pixel intensity range of the resulting images was kept constant, covering the full 0–255 range.
Experimental procedure
Participants viewed partial reconstructions on an LCD from 70 cm distance [visual degree angle (°), horizontal and vertical dimensions of the screen: 51.6° × 30.4°]. Each trial started with a grey screen and a central fixation cross which participants were instructed to fixate, which we presented for a duration between 250 and 400 ms. Afterwards, a partial reconstruction was presented on a central area of the screen (covering 22.5° × 22.5°) and remained on the screen until a button press was made. Participants used their right hand to press one of the three available buttons with which they indicated having recognized a cat, a dog or that they weren’t sure about the type of animal they were shown. The next trial started as soon as a button was pressed. Due to the self-paced nature of the paradigm, participants completed a variable number of sessions (6, 5, 4, 3 and 2 sessions were completed by 2, 14, 25, 10 and 1 participants resp.), which each consisted of 50 trials per image. Stimuli were presented and behavioural data was recorded using Matlab 2016a and the PsychToolbox45 (version 3).
Data analysis
First, we assessed for each trial whether the image depicted was recognized. Then, we computed the average recognition performance for each participant and image feature by computing the average performance for trials containing the respective feature. This provided us with a three-dimensional matrix of recognition performances with the dimension number of participants (52), number of images (10) and number of features (1,000). Next, we obtained our feature diagnosticity index (FDi) values by z-transforming the performance values within each participant and image. This final step is important because it precludes FDi values being higher for features of images that are easier to recognize. Furthermore, it ensures that participants’ relative contributions to the following analyses do not depend on their average recognition performance nor the variability of their responses.
To evaluate the replicability of the observed FDi values, we randomly split the data into two halves (two times 26 participants) 100 times and computed the Pearson correlation between the average FDi values across splits. Replicability was then measured as the average Pearson correlation value across these 100 splits. The probability of observing this value by chance was determined by computing a null-distribution by re-computing this value 10,000 times while permuting the relative feature labels across splits.
To assess effects of spatial frequency, distance from the nearest eye and distance from image centre on FDi values, we created five equally sized ascending bins based on each of these parameters. Thereafter, we computed the average FDi value within each of these bins separately for each participant. In addition, we assigned each participant to the high AQ group and low AQ group depending on whether their AQ was or was not higher than the median AQ across all participants. This enabled us to perform three 5 × 2 repeated measure ANOVAs with average within-bin FDi values as the dependent variable. Each ANOVAs second factor was AQ group while the first factor was the binning feature: spatial frequency, distance from the nearest eye or distance from image centre.
Supplementary information
Acknowledgements
This work was supported by a Marie Curie fellowship (753441) to A.A. and an ERC starter Grant (ERC-StG-759432) to I.C. We thank Anisha Rungapillay, Fiona Boxford, and Noor Hassan for their help in collecting the behavioural data.
Author contributions
A.A. and I.C. designed the experiments. A.A. and I.C. wrote the paper; A.A. analyzed the data; I.C. conducted the experiments.
Funding
Open access funding provided by Projekt DEAL.
Data availability
Behavioural data and Matlab code for our data analyses can be downloaded from the online public GitHub repository: https://github.com/arjenalink/AQ_EyeForDetail_NSR.
Competing interests
The authors declare no competing interests.
Footnotes
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
is available for this paper at 10.1038/s41598-020-70953-8.
References
- 1.American Psychiatric Association. Diagnostic and statistical manual of mental disorders (DSM-5®). (American Psychiatric Pub, 2013). https://www.appi.org/Diagnostic_and_Statistical_Manual_of_Mental_Disorders_DSM-5_Fifth_Edition
- 2.Bölte S, Holtmann M, Poustka F, Scheurich A, Schmidt L. Gestalt perception and local-global processing in high-functioning autism. J. Autism Dev. Disord. 2007;37:1493–1504. doi: 10.1007/s10803-006-0231-x. [DOI] [PubMed] [Google Scholar]
- 3.Shah A, Frith U. An islet of ability in autistic children: A research note. J. Child Psychol. Psychiatry. 1983;24:613–620. doi: 10.1111/j.1469-7610.1983.tb00137.x. [DOI] [PubMed] [Google Scholar]
- 4.Russell-Smith SN, Maybery MT, Bayliss DM, Sng AA. Support for a link between the local processing bias and social deficits in autism: An investigation of embedded figures test performance in non-clinical individuals. J. Autism Dev. Disord. 2012;42:2420–2430. doi: 10.1007/s10803-012-1506-z. [DOI] [PubMed] [Google Scholar]
- 5.O’riordan MA. Superior visual search in adults with autism. Autism. 2004;8:229–248. doi: 10.1177/1362361304045219. [DOI] [PubMed] [Google Scholar]
- 6.O’riordan MA, Plaisted KC, Driver J, Baron-Cohen S. Superior visual search in autism. J. Exp. Psychol. Hum. Percept. Perform. 2001;27:719. doi: 10.1037//0096-1523.27.3.719. [DOI] [PubMed] [Google Scholar]
- 7.Plaisted K, O’Riordan M, Baron-Cohen S. Enhanced visual search for a conjunctive target in autism: A research note. J. Child Psychol. Psychiatry. 1998;39:777–783. [PubMed] [Google Scholar]
- 8.Frith U, Happé F. Autism: Beyond “theory of mind.”. Cognition. 1995;50(1–3):115–132. doi: 10.1016/0010-0277(94)90024-8. [DOI] [PubMed] [Google Scholar]
- 9.Happé F, Frith U. The weak coherence account: detail-focused cognitive style in autism spectrum disorders. J. Autism Dev. Disord. 2006;36:5–25. doi: 10.1007/s10803-005-0039-0. [DOI] [PubMed] [Google Scholar]
- 10.Happé FG. Central coherence and theory of mind in autism: Reading homographs in context. Br. J. Dev. Psychol. 1997;15:1–12. [Google Scholar]
- 11.Baron-Cohen S. Mindblindness: An essay on autism and theory of mind. Cambridge: MIT Press; 1997. [Google Scholar]
- 12.Gross TF. Global-local precedence in the perception of facial age and emotional expression by children with autism and other developmental disabilities. J. Autism Dev. Disord. 2005;35:773. doi: 10.1007/s10803-005-0023-8. [DOI] [PubMed] [Google Scholar]
- 13.Burnette CP, et al. Weak central coherence and its relations to theory of mind and anxiety in autism. J. Autism Dev. Disord. 2005;35:63–73. doi: 10.1007/s10803-004-1035-5. [DOI] [PubMed] [Google Scholar]
- 14.Morgan B, Maybery M, Durkin K. Weak central coherence, poor joint attention, and low verbal ability: Independent deficits in early autism. Dev. Psychol. 2003;39:646. doi: 10.1037/0012-1649.39.4.646. [DOI] [PubMed] [Google Scholar]
- 15.Pellicano E, Maybery M, Durkin K, Maley A. Multiple cognitive capabilities/deficits in children with an autism spectrum disorder: “Weak” central coherence and its relationship to theory of mind and executive control. Dev. Psychopathol. 2006;18:77–98. doi: 10.1017/S0954579406060056. [DOI] [PubMed] [Google Scholar]
- 16.Pellicano E, Gibson L, Maybery M, Durkin K, Badcock DR. Abnormal global processing along the dorsal visual pathway in autism: A possible mechanism for weak visuospatial coherence? Neuropsychologia. 2005;43:1044–1053. doi: 10.1016/j.neuropsychologia.2004.10.003. [DOI] [PubMed] [Google Scholar]
- 17.Plaisted K, Swettenham J, Rees L. Children with autism show local precedence in a divided attention task and global precedence in a selective attention task. J. Child Psychol. Psychiatry. 1999;40:733–742. [PubMed] [Google Scholar]
- 18.White SJ, Saldaña D. Performance of children with autism on the embedded figures test: A closer look at a popular task. J. Autism Dev. Disord. 2011;41:1565–1572. doi: 10.1007/s10803-011-1182-4. [DOI] [PubMed] [Google Scholar]
- 19.Robertson CE, Baron-Cohen S. Sensory perception in autism. Nat. Rev. Neurosci. 2017;18:671–684. doi: 10.1038/nrn.2017.112. [DOI] [PubMed] [Google Scholar]
- 20.Kéïta L, Guy J, Berthiaume C, Mottron L, Bertone A. An early origin for detailed perception in Autism Spectrum Disorder: Biased sensitivity for high-spatial frequency information. Sci. Rep. 2014;4:5475. doi: 10.1038/srep05475. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Tavassoli T, Latham K, Bach M, Dakin SC, Baron-Cohen S. Psychophysical measures of visual acuity in autism spectrum conditions. Vision Res. 2011;51:1778–1780. doi: 10.1016/j.visres.2011.06.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.De Jonge MV, et al. Visual information processing in high-functioning individuals with autism spectrum disorders and their parents. Neuropsychology. 2007;21:65. doi: 10.1037/0894-4105.21.1.65. [DOI] [PubMed] [Google Scholar]
- 23.Bertone A, Mottron L, Jelenic P, Faubert J. Enhanced and diminished visuo-spatial information processing in autism depends on stimulus complexity. Brain. 2005;128:2430–2441. doi: 10.1093/brain/awh561. [DOI] [PubMed] [Google Scholar]
- 24.Baron-Cohen S, Wheelwright S, Skinner R, Martin J, Clubley E. The autism-spectrum quotient (AQ): Evidence from asperger syndrome/high-functioning autism, malesand females, scientists and mathematicians. J. Autism Dev. Disord. 2001;31:5–17. doi: 10.1023/a:1005653411471. [DOI] [PubMed] [Google Scholar]
- 25.Almeida RA, Dickinson JE, Maybery MT, Badcock JC, Badcock DR. A new step towards understanding Embedded Figures Test performance in the autism spectrum: The radial frequency search task. Neuropsychologia. 2010;48:374–381. doi: 10.1016/j.neuropsychologia.2009.09.024. [DOI] [PubMed] [Google Scholar]
- 26.Grinter EJ, et al. Global visual processing and self-rated autistic-like traits. J. Autism Dev. Disord. 2009;39:1278–1290. doi: 10.1007/s10803-009-0740-5. [DOI] [PubMed] [Google Scholar]
- 27.Russell-Smith SN, Maybery MT, Bayliss DM. Are the autism and positive schizotypy spectra diametrically opposed in local versus global processing? J. Autism Dev. Disord. 2010;40:968–977. doi: 10.1007/s10803-010-0945-7. [DOI] [PubMed] [Google Scholar]
- 28.Pelphrey KA, et al. Visual scanning of faces in autism. J. Autism Dev. Disord. 2002;32:249–261. doi: 10.1023/a:1016374617369. [DOI] [PubMed] [Google Scholar]
- 29.Spezio ML, Adolphs R, Hurley RSE, Piven J. Analysis of face gaze in autism using “Bubbles”. Neuropsychologia. 2007;45:144–151. doi: 10.1016/j.neuropsychologia.2006.04.027. [DOI] [PubMed] [Google Scholar]
- 30.Dalton KM, et al. Gaze fixation and the neural circuitry of face processing in autism. Nat. Neurosci. 2005;8:519–526. doi: 10.1038/nn1421. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Wang S, et al. Atypical visual saliency in autism spectrum disorder quantified through model-based eye tracking. Neuron. 2015;88:604–616. doi: 10.1016/j.neuron.2015.09.042. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Gosselin F, Schyns PG. Bubbles: A technique to reveal the use of information in recognition tasks. Vision Res. 2001;41:2261–2271. doi: 10.1016/s0042-6989(01)00097-9. [DOI] [PubMed] [Google Scholar]
- 33.Eggermont JJ, Johannesma PIM, Aertsen AMHJ. Reverse-correlation methods in auditory research. Q. Rev. Biophys. 1983;16:341–414. doi: 10.1017/s0033583500005126. [DOI] [PubMed] [Google Scholar]
- 34.Ruzich E, et al. Measuring autistic traits in the general population: a systematic review of the Autism-Spectrum Quotient (AQ) in a nonclinical population sample of 6,900 typical adult males and females. Mol. Autism. 2015;6:2. doi: 10.1186/2040-2392-6-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Bayliss AP, Kritikos A. Brief report: Perceptual load and the autism spectrum in typically developed individuals. J. Autism Dev. Disord. 2011;41:1573–1578. doi: 10.1007/s10803-010-1159-8. [DOI] [PubMed] [Google Scholar]
- 36.Hudson M, Nijboer TC, Jellema T. Implicit social learning in relation to autistic-like traits. J. Autism Dev. Disord. 2012;42:2534–2545. doi: 10.1007/s10803-012-1510-3. [DOI] [PubMed] [Google Scholar]
- 37.Lindell AK, Notice K, Withers K. Reduced language processing asymmetry in non-autistic individuals with high levels of autism traits. Laterality. 2009;14:457–472. doi: 10.1080/13576500802507752. [DOI] [PubMed] [Google Scholar]
- 38.O’Keefe N, Lindell AK. Reduced interhemispheric interaction in non-autistic individuals with normal but high levels of autism traits. Brain Cogn. 2013;83:183–189. doi: 10.1016/j.bandc.2013.08.005. [DOI] [PubMed] [Google Scholar]
- 39.Walter E, Dassonville P, Bochsler TM. A specific autistic trait that modulates visuospatial illusion susceptibility. J. Autism Dev. Disord. 2009;39:339–349. doi: 10.1007/s10803-008-0630-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Hoekstra RA, Bartels M, Cath DC, Boomsma DI. Factor structure, reliability and criterion validity of the autism-spectrum quotient (AQ): A study in dutch population and patient groups. J. Autism Dev. Disord. 2008;38:1555–1566. doi: 10.1007/s10803-008-0538-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Murray AL, McKenzie K, Kuenssberg R, Booth T. Do the autism spectrum quotient (AQ) and autism spectrum quotient short form (AQ-S) primarily reflect general ASD traits or specific ASD traits? A bi-factor analysis. Assessment. 2017;24:444–457. doi: 10.1177/1073191115611230. [DOI] [PubMed] [Google Scholar]
- 42.Landry O, Chouinard PA. Why we should study the broader autism phenotype in typically developing populations. J. Cogn. Dev. 2016;17:584–595. [Google Scholar]
- 43.Vlamings PHJM, Jonkman LM, van Daalen E, van der Gaag RJ, Kemner C. Basic abnormalities in visual processing affect face processing at an early age in autism spectrum disorder. Biol. Psychiatry. 2010;68:1107–1113. doi: 10.1016/j.biopsych.2010.06.024. [DOI] [PubMed] [Google Scholar]
- 44.Koh HC, Milne E, Dobkins K. Spatial contrast sensitivity in adolescents with autism spectrum disorders. J. Autism Dev. Disord. 2010;40:978–987. doi: 10.1007/s10803-010-0953-7. [DOI] [PubMed] [Google Scholar]
- 45.Brainard DH. The psychophysics toolbox. Spat. Vis. 1997;10:433–436. [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Behavioural data and Matlab code for our data analyses can be downloaded from the online public GitHub repository: https://github.com/arjenalink/AQ_EyeForDetail_NSR.