Abstract
Autism spectrum disorder (ASD) is associated with significant social, communication, and behavioral challenges. The insufficient number of trained clinicians coupled with limited accessibility to quick and accurate diagnostic tools resulted in overlooking early symptoms of ASD in children around the world. Several studies have utilized behavioral data in developing and evaluating the performance of machine learning (ML) models toward quick and intelligent ASD assessment systems. However, despite the good evaluation metrics achieved by the ML models, there is not enough evidence on the readiness of the models for clinical use. Specifically, none of the existing studies reported the real-life application of the ML-based models. This might be related to numerous challenges associated with the data-centric techniques utilized and their misalignment with the conceptual basis upon which professionals diagnose ASD. The present work systematically reviewed recent articles on the application of ML in the behavioral assessment of ASD, and highlighted common challenges in the studies, and proposed vital considerations for real-life implementation of ML-based ASD screening and diagnostic systems. This review will serve as a guide for researchers, neuropsychiatrists, psychologists, and relevant stakeholders on the advances in ASD screening and diagnosis using ML.
Keywords: autism spectrum disorder, screening, diagnosis, artificial intelligence, machine learning
1. Introduction
Autism spectrum disorder (ASD) is a lifelong neurodevelopmental disorder associated with communication impairment, restrictive and compulsive behavior. According to the fifth edition of the diagnostic and statistical manual of mental disorders (DSM-5), the primary indicators for diagnosing ASD are deficits in social communication and the manifestation of repetitive and restricted patterns of activities, behavior, or interests [1]. The rising prevalence of ASD necessitates the need for early and cost-effective diagnosis to set the path for efficient, and appropriate treatment [2,3]. Moreover, early diagnosis of ASD leads to improved outcomes in communication and social interaction and guides parents to the right interventions in school, home, and clinic [4,5,6]. However, apart from the cost-ineffectiveness of the current diagnostic instruments, studies have indicated the delay of the clinical processes of diagnosing ASD [7,8,9,10]. Addressing these challenges lead to several suggestions, including the so-called quick and accurate Machine Learning (ML)-enabled ASD assessment systems [11,12,13,14]. The promising results realized with ML algorithms across various research fields motivated these suggestions and made it a vital step toward quick and cost-effective assessment of ASD symptoms.
The gap in the existing literature is the absence of a definitive explanation on the sufficiency and readiness of the ML models toward real-life implementation. Recently, there is an increasing number of studies on the development of ML models for diagnosing ASD based on either genetic [15,16], brain imaging [17,18,19], physical biomarkers [20,21,22,23,24], or behavioral data. However, despite the high evaluation metrics reported in the ML-based behavioral studies, there is little evidence on the clinical use of the resulting ML models [11]. Generally, apart from improving the accuracy metrics of the ML models, previous studies focused on improving diagnostic speed by reducing the model parameters using various dimensionality reduction techniques. Worthy of note, both the ML algorithms and the dimensionality reduction techniques are data-centric; they are independent of the conceptual basis upon which professionals build and utilize ASD assessment instruments [25]. Thus, the clinical validity of the resulting ML models could be explained based on the alignment of the data-centric techniques with the conceptual basis of diagnosing ASD. Nonetheless, other factors that might limit the clinical validity and real-life implementation of the models include the reported discrepancies within the data repositories [26,27].
The present review explores the advances in the application of machine learning in the behavioral assessment of ASD. Accordingly, recent articles were systematically reviewed on the application of machine learning models toward quick and accurate assessment of ASD. Based on the reviewed literature, we sought the answer on whether the recent findings could sufficiently translate to real-life implementation of ML-based ASD screening and diagnostic models. Nonetheless, previous literature reviews assessed the performance of ML models in ASD screening and diagnosis based on the common evaluation metrics of sensitivity, specificity, and accuracy, among others [25,28]. However, none of the existing literature reviews systematically analyzed the subject area and provided enough evidence on the readiness and sufficiency of the models toward real-life implementation of the ML-based systems. For instance, Song et al. [28] reviewed 13 relevant studies that utilized varying data types and discussed the possibility of achieving effective classification of ASD based on the study findings. Similarly, Thabtah [25] identified some limitations within the commonly employed research methodologies and proposed intuitive stages toward appending the ML models into ASD screening apps. In this work, key challenges were highlighted alongside the commonly utilized assessment tools, datasets, and data intelligence techniques, and solutions were suggested toward valid implementations of real-life ML-based ASD screening and diagnostic systems.
2. Methodology
2.1. Search Strategy
The present review involved a systematic search, which is conducted in October 2020. To identify the most relevant studies, the authors ensured careful planning and allocation of tasks at every stage of the systematic literature review. The search strategy was tailored to the four most popular scientific databases of the study field, namely, Web of Science, PubMed, IEEEXplore, and Scopus. Furthermore, the search query utilized includes the following terms “Autism Spectrum Disorder” OR “Autistic Disorder” OR “Autism” AND “Screening” OR “Assessment” OR “Identification” OR “Test” OR “Detection” AND “Machine Learning” OR “Artificial Intelligence”. The search filters covered a period of ten years from 2011 to 2020 and were limited to journal articles published in the English language. Beyond the above-mentioned databases, relevant publications were accessed from other databases on the advances in ASD assessment.
2.2. Selection Criteria
The article selection process was based on the PRISMA statement [29]. Relevant studies have utilized PRISMA in providing critical appraisal on the advances in the assessment of autism and other neuropsychiatric disorders [19,24,28,30,31,32,33]. The determining factor in the inclusion criteria involves any published full-text journal article on the use of ML in ASD screening or diagnosis. At the initial screening stage, after duplicates removal, the authors assessed the records against the inclusion criteria to decide on worthy articles for the systematic literature review. The decisions for inclusion/exclusion on the records were recorded in a separate column within the combined excel sheet imported from the databases. Thus, for records whose titles and corresponding abstracts aligned with the preset inclusion criteria, full-text articles of the studies were retrieved for the subsequent screening stage. In the next PRISMA screening stage, all the authors reviewed the downloaded papers, independently, to ascertain their relevance with the search query used, as well as the set research question. The authors utilized the WhatsApp discussion group in resolving disagreements in the selection process.
Specifically, three hundred and sixty-seven records were carefully assessed for eligibility. One hundred and eighty studies out of the 367 records were discarded, due to the following reasons: Book chapters (n = 17), conference papers (n = 138), editorial materials (n = 11), literature reviews (n = 15), not written in English (n = 9). The remaining one hundred and seventy-seven studies were further assessed; one hundred and forty-four records were eliminated because they are either based on brain imaging data (n = 57), genetic data (n = 35), or physical/metabolic biomarkers (n = 32), while others are intervention studies (n = 20). Consequently, thirty-three full-text articles were retrieved, read, and qualitatively assessed. Nonetheless, additional articles were excluded because ML is not the main method employed (n = 7), and ASD is not the main neuropsychiatric disorder assessed (n = 4). Finally, 22 studies met the inclusion criteria. The PRISMA flow diagram (Figure 1) summarized the above-mentioned systematic literature review process, and Table 1 itemized the key items of the inclusion and exclusion criteria of the study.
Table 1.
Inclusion Criteria |
---|
Journal articles published in the English language |
Documents published within the last ten years from 2011 to date |
Full-text papers that are accessible and downloadable |
Studies that utilized behavioral data |
Studies that employed machine learning as the main technique |
Studies that considered autism as the main disorder assessed |
Exclusion criteria |
Papers that are written in other languages |
Duplicated papers |
Full-text of the document is not accessible on the internet |
The study aim is not clearly defined |
Studies that are not relevant to the stated research question |
Relevant studies, but machine learning is not the main method |
Relevant studies, but autism is not the main disorder assessed |
Conferences papers, editorial materials, and literature reviews |
Studies that utilized data from either brain imaging, genetic, or physical/metabolic biomarkers. |
Intervention studies |
2.3. Quality Assessment
The authors carefully adhered to the planned, systematic literature review process to maintain the study’s quality. Particularly, at every phase of the systematic literature review, the authors ensured careful planning and allocation of tasks. The first author created an online Mendeley repository and monitored the progress of the review based on preset milestones to ensure that all tasks complied with the scheduled deadlines. The Mendeley repository was also used in keeping track of the data extraction stages, noting essential observations and sharing vital contents related to the study. The authors further upheld peer-reviewing at each phase of the study to enhance the systematic literature review. Nevertheless, unbiased and constructive assessments on the systematic approach used in this study were sought from external professionals on ASD diagnostic procedures with expertise in systematic literature reviews.
2.4. Data Extraction
As the final stage of the study’s PRISMA, the data extraction stage, 22 articles were appraised critically, and the following information was extracted from the studies:
Author(s) (year),
Number of citations,
Source(s) of the research data,
Data collection/assessment instrument,
ML model(s)developed,
Best performing model(s),
The key finding(s).
3. Results
3.1. Descriptive Analysis on Trends and Status of the Study on ML in ASD Assessment
Based on the exported data, the trend of studies on the use of ML in the behavioral assessment of ASD showed the most cited references, the most cited journals, as well as citation and publication frequencies across the years.
With the increasing application of ML in healthcare studies, as shown in Figure 2, there are more publications on ML and ASD assessment. From 2012 to 2018, not so many studies cared about the application of ML in ASD assessment. However, with the recently increased patronage of ML techniques across various fields, there is an increasing demand for intelligent tools for accurate assessment of ASD. From Figure 3, most of the articles contributing to the area were published in Translational Psychiatry (n = 5), followed by the Health Informatics Journal (n = 3). The remaining fifteen journals depicted published one article, each.
Based on the citation data exported, as shown in Table 2, we can see that the most cited references are Wall et al. [34] (n = 160), Wall et al. [35] (n = 106), Duda et al. [36] (n = 89), Kosmicki et al. [37] (n = 84), and Bone et al. [38] (n = 77). Most of the significant references; with the highest number of citations, were published in Translational Psychiatry [34,36,37] (Figure 4, n = 408) in the years 2012 (Figure 5, n = 266), 2015 (Figure 5, n = 84), and 2016 (Figure 5, n = 166). Figure 4 highlighted the citation data of the eight most cited journals involved in the study; Translational Psychiatry (n = 408), PLoS One (n = 106), Journal of Children Psychological Psychiatry (n = 77), and so on.
Table 2.
Article/ Citations |
Aim | Tool | Data Source | FS/FT | FS/FT Method | Modeling Algorithms | Key Findings |
---|---|---|---|---|---|---|---|
Goel et al. [51] C = 10 |
Proposed Optimization Algorithm for improved performance over common ML | AQ-10 (child, adolescent, adult) | ASDTest | - | - | GOA, BACO, LR, NB, KNN, RF-CART + ID3, * MGOA | The proposed MGOA (GOA with Random Forest classifier) predicted ASD cases with approximate accuracy, specificity, and sensitivity of 100%. |
Shahamiri and Thabtah [11] C = 0 |
Implementation and evaluation of CNN-based ASD scoring system | Q-CHAT-10, AQ-10 | ASDTest | - | - | C4.5, Bayes Net, RIDOR, * CNN | The performance evaluation showed the superior performance of CNN over other algorithms; indicating the robustness of the implemented system. |
Thabtah and Peebles [52] C = 28 |
Demonstrate the superiority of Rules-based ML over other models | Q-CHAT-10, AQ-10 (child, Adolescent, adult) | ASDTest | - | - | RIPPER, RIDOR, Nnge, Bagging, CART, C4.5, and PRISM, * RML | Empirically evaluated rule induction, Bagging, Boosting, and decision trees algorithms on different ASD datasets. The superiority of the RML model was reported in not only classifying ASD but also offer rules that can be utilized in understanding the reasons behind the classification. |
Wall et al. [35] C = 106 |
Streamlining ADR-I and evaluate ML performance | ADI-R | AGRE, SSC, AC | FS | Trial-error | * ADTree, BFTree, ConjunctiveRule, DecisionStump, FilteredClassifier, J48, J48graft, JRip, LADTree, Nnge, OneR, OrdinalClassClassifier, PART, Ridor, and SimpleCart | The best model utilized 7 of the 93 items contained in the ADI-R in classifying ASD with 99.9% accuracy. |
Duda et al. [39] C = 50 |
Streamlining ADOS and demonstrate the superior performance of ADTree over common hand-crafted methods | ADOS | AC, AGRE, SSC, NDAR, SVIP | FS | Trial-error | ADTree | 72% reduction in the items from ADOS-G with >97% accuracy. |
Küpper et al. [40] C = 2 |
Streamlining ADOS and demonstrate the performance of SVM | ADOS | ASD outpatient clinics in Germany | FS | Recursive Feature Selection | SVM | SVM achieved good sensitivity and specificity with fewer ADOS items pointing to 5 behavioral features. |
Wall et al. [34] C = 160 |
Streamlining ADOS and evaluate ML performance | ADOS | AC, AGRE, SSC | FS | Trial-error | * ADTree, BFTree, Decision Stump, Functional Tree, J48, J48graft, Jrip, LADTree, LMT, Nnge, OneR, PART, Random Tree, REPTree, Ridor, Simple Cart | The ADTree model utilized 8 of the 29 items in Module 1 of the ADOS and classified ASD with 100% accuracy. |
Levy et al. [50] C = 21 |
Streamlining ADOS and evaluate ML performance | ADOS | AC, AGRE, SSC, SVIP | FS | Sparsity/parsimony enforcing regularization techniques | LR, Lasso, Ridge, Elastic net, Relaxed Lasso, Nearest shrunken centroids, LDA, * LR, * SVM, ADTree, RF, Gradient boosting, AdaBoost | With at most 10 features from ADOS′s Module 3 and Module 2, AUC of 0.95 and 0.93 was achieved, respectively. |
Kosmicki et al. [37] C = 84 |
Streamlining ADOS and evaluate ML performance | ADOS | AC, AGRE, SSC, NDAR, SVIP | FS | Stepwise Backward Feature Selection | ADTree, * SVM, Logistic Model Tree, * LR, NB, NBTree, RF | The best performing models have utilized 9 of the 28 items from module 2, and 12 of the 28 items from module 3 in classifying ASD with 98.27% and 97.66% accuracy, respectively. |
Thabtah [13] C = 31 |
Propose ASDTest; AQ-based mobile screening app, streamline AQ-10 items, and evaluate the performance of 2 ML models | AQ-10 (child, adolescent, adult) | ASDTest | FS | Trial-error | NB, * LR | Feature and predictive analyses demonstrate small groups of autistic traits improving the efficiency and accuracy of screening processes. |
Thabtah et al. [46] C = 47 |
Demonstrate the superiority of Va over other FS methods based on the performance of ML models on the streamlined datasets | Q-CHAT-10, and AQ-10 (child, adolescent, adult) | ASDTest | FS | Va, IG, Correlation, CFS, and CHI | Repeated Incremental Pruning to Produce Error Reduction (RIPPER), C4.5 (Decision Tree) | Va derived fewer features from adults, adolescents, and child datasets with optimal model performance. Demonstrate the efficacy of Va over IG, Correlation, CFS, and CHI in reducing AQ-10 items |
Thabtah et al. [48] C = 13 |
Streamlining AQ-10 and demonstrate the superior performance of LR over common hand-crafted methods | AQ-10 (adolescent, adult) | ASDTest | FS | IG, CHI | LR | LR showed acceptable performance in terms of sensitivity, specificity, and accuracy among others. |
Suresh Kumar and Renugadevi [49] C = 0 |
Algorithm Optimization (improvement in accuracy compared to common ML) | AQ-10 (child, adolescent, adult) | ASDTest | FS | SFS | SVM, ANN, * DE SVM, DE ANN | DE optimized SVM outperformed ANN and DE optimized ANN in classifying ASD. DE is effective. |
Pratama et al. [47] C = 0 |
Input Optimization using Va | AQ-10 (child, adolescent, adult) | ASDTest | FS | Va | SVM, * RF, ANN | RF succeeded in producing higher adult AQ sensitivity (87.89%), and a rise in the specificity level of AQ-Adolescents was better produced using SVM (86.33%). |
Usta et al. [45] C = 9 |
ML Performance Evaluation | Autism Behavior Checklist, Aberrant Behavior Checklist, Clinical Global Impression | Ondokuz Mayis University Samsun | FS | Trial-error | NB, LR, * ADTree | The ML modeling revealed the significant influence of other demographic parameters in ASD classification. |
Wingfield et al. [12] C = 3 |
Propose PASS; a culturally sensitive app embedded with ML model | PASS | VPASS app | FS | CFS, mRMR | * RF, NB, Adaboost, Multilayer Perceptron, J48, PART, SMO | PASS app overcomes the cultural variation in interpreting ASD symptoms, and the study demonstrated the possibility of removing feature redundancy. |
Duda et al. [36] C = 89 |
ML Performance Evaluation in classifying ASD from ADHD | SRS | AC, AGRE, SSC | FS | Forward Feature Selection | ADTree, RF, SVM, LR, Categorical lasso, LDA | All the models could classify ASD from ADHD by utilizing 5 of the 65 items of SRS with high average accuracy (AUC = 0.965). |
Duda et al. [53] C = 25 |
Improve models’ reliability using expanded datasets for classifying ASD from ADHD | SRS | AC, AGRE, SSC, and crowdsourced data | FS | - | SVM, LR, * LDA | LDA model achieved an AUC of 0.89 with 15 items. |
Bone et al. [38] C = 77 |
Demonstrate the improved accuracy of SVM over common hand-crafted rules | ADI-R, SRS | Balanced Independent Dataset | FT | Tuned parameters across multiple levels of cross-validation | SVM | The SVM model utilized five of the fused ADI-R and SRS items and classified ASD sufficiently with below (above) 89.2% (86.7%) sensitivity and 59.0% (53.4%) specificity. |
Puerto et al. [42] C = 17 |
Propose MFCM-ASD and evaluate its performance against other ML models | ADOS, ADI-R | APADA | FT | Inputs fuzzification | * MFCM-ASD, SVM, Random forest, NB | The superior performance of MFCM characterized by its robustness makes it an effective ASD diagnostic technique. |
Akter et al. [44] C = 6 |
Compare FT methods and evaluate the performance of ML models on the transformed datasets | Q-CHAT-10, and AQ-10 (child, adolescent, adult) | ASDTest | FT | Log, Z-score, and Sine FT | Adaboost, FDA, C5.0, LDA, MDA, PDA, SVM, and CART | Varying superior performances of the ML models and FT approaches were achieved across the datasets. |
Baadel et al. [43] C = 2 |
Input Optimization using a clustering approach | AQ-10 (child, adolescent, adult) | ASDTest | FT | CATC | OMCOKE, RIPPER, PART, * RF, RT, ANN | CATC showed significant improvement in screening ASD based on traits′ similarity as opposed to scoring functions. The improvement was more pronounced with RF classifier. |
ASD, autism spectrum disorder; FS, feature selection; FT, feature transformation; ML, machine learning; ANN, artificial neural network; SVM, support vector machine; CNN, convolutional neural network; RF, random forest; LR, logistic regression; ADTree, alternative decision tree; LDA, linear discriminant analysis; MGOA, modified grasshopper optimization algorithm; BACO, binary ant colony optimization; NB, naïve Bayes; KNN, K-nearest neighbor; RIPPER, repeated incremental pruning to produce error reduction; ADOS, autism diagnostic observation schedule; ADI-R, autism diagnostic interview-revised; Q-CHAT, quantitative checklist for autism toddlers; AQ, autism quotient; SRS, social responsiveness scale; PASS, pictorial autism assessment schedule; AC, boston autism consortium; AGRE, autism genetic resource exchange; SSC, Simons Simplex Collection; NDAR, National Database for Autism Research; SVIP, Simons Variation In Individuals Project; APADA, Association of Parents and Friends for the Support and Defense of the Rights of People with Autism; MFCM, multilayer fuzzy cognitive maps; CATC, clustering-based autistic trait classification. * Best performing models.
3.2. Dimensionality Reduction Techniques
Most of the studies primarily aimed at streamlining the data collection instruments, followed by evaluating the performance of various ML algorithms on the streamlined datasets [35,37,39,40,41]. While various feature selection methods were applied in streamlining the most influential features of the data collection instruments from the datasets, other studies utilized various feature transformation techniques in reducing the input parameters. For instance, in the work of Puerto et al. [42], the inputs were fuzzified into membership values before applying the classification algorithms. Similarly, before implementing the classification models, Baadel et al. [43] and Akter et al. [44] transformed the inputs using clustering and feature transformation functions, respectively. Nonetheless, other studies employed a trial-error approach in selecting the most influential features. The trial-error approach involves repetitive evaluation of the ML models using a varying combination of the features; the most influential combination achieves superior results with fewer input parameters. Specifically, the studies utilized various feature selection techniques, including trial-error [13,34,35,39,45], Variable Analysis (Va) [46,47], information gain (IG) and chi-square testing (CHI) [48], sequential feature selection (SFS) [49], correlation-based feature selection (CFS) and minimum redundancy maximum relevance (mRMR) [12]. Additionally, ML-based feature selection techniques employed include recursive feature selection [40], sparsity/parsimony enforcing regularization techniques [50], stepwise backward feature selection [37], and forward feature selection [36].
3.3. Models Implementation
As shown in Table 2, the commonly implemented ML algorithms are Random Forest (RF) [12,43,47,51], Support Vector Machines (SVM) [37,38,40,49,50], Alternative Decision Tree (ADTree) [34,35,39,45], and Logistic Regression (LR) [13,37,48]. To achieve comparative results, most of the studies employed several algorithms, such as Adaboost, Artificial Neural Network (ANN), Linear Discriminant Analysis (LDA), Naïve Bayes, and K-Nearest Neighbor (KNN).
3.4. Data Collection/Assessment Instruments
The most utilized data collection instruments are AQ-10 [11,13,43,44,46,47,48,49,51,52], Q-CHAT-10 [11,44,46,52], ADOS [34,37,39,40,42,50], ADI-R [35,38,42], and Social Responsiveness Scale (SRS) [36,38,53]. Others include Autism Behavior Checklist, Aberrant Behavior Checklist, Clinical Global Impression [45], and MCHAT-based Pictorial Autism Assessment Schedule (PASS) [12]. Thus, the need for improving the reliability of these assessment instruments and ascertaining their relevance in ML modelling remains.
3.5. Sources of Data
The most prominent sources of data utilized in the studies include Boston Autism Consortium (AC), Autism Genetic Resource Exchange (AGRE), Simons Simplex Collection (SSC) [34,35,36,37,39,50,53], National Database for Autism Research (NDAR) [37,39], and Simons Variation In Individuals Project (SVIP) [37,39,50]. Other studies utilized data sets from ASDTest: Kaggle and UCI ML repository [11,13,43,44,46,47,48,49,51,52], Association of Parents and Friends for the Support and Defense of the rights of people with Autism (APADA) [42], PASS app [12], Ondokuz Mayis University Samsun [45], and ASD outpatient clinics in Germany [40]. To achieve standardized comparative results, there is a need for standardized ASD data repositories for machine learning studies [25].
3.6. Research Procedures
Apart from the common aim of streamlining the various data collection instruments followed by model evaluation, other studies focused on either optimizing the machine-learning algorithms [49,51], proposing input optimization techniques [43,44,46,47], or implementing ML-based screening apps [11,12]. For instance, Goel et al. [51] proposed Modified Grasshopper Optimization Algorithm (MGOA) for improved performance over common ML algorithms. The proposed MGOA (GOA with Random Forest classifier) outperformed other basic models and predicted ASD with approximate accuracy, specificity, and sensitivity of 100%. Similarly, Suresh et al. [49] proposed Differential Evaluation (DE) Algorithm to find the optimal solution of SVM parameters. The proposed DE tuned SVM achieved better performance over SVM, ANN and DE optimized ANN in classifying ASD. As stated earlier, apart from trial-error, studies employed either feature selection or transformation techniques for dimensionality reduction. For instance, Thabtah et al. [46] demonstrated the superiority of Va over IG, Correlation, CFS, and CHI in reducing AQ-10 items. Va derived fewer features, while maintaining competitive predictive accuracy, sensitivity, and specificity rates. A replicated study by Pratama et al. [47] produced a higher sensitivity of 87.89% in Adults AQ with RF and an increased specificity level of 86.33% in Adolescents AQ with SVM. Despite the good performance of the above-mentioned techniques in automating feature selection processes across various applications [54,55], none of the previous studies justified the conformity of the feature selection methods with the conceptual basis upon which professionals built and utilize ASD diagnostic instruments. Furthermore, unlike other medical diagnoses, the absence of definitive measures and medical tests for diagnosing ASD makes it difficult to numerical quantify the disorder based on few parameters. Notably, accurate assessment of ASD relied on the precise application of the commonly used behavioral scales built based on the knowledge and expertise of the professionals. Thus, applying human knowledge is imperative to reliable ASD diagnosis. Based on that, there is a need for quantifying the trade-offs of dimensionality reduction (ensuring fewer items for quick assessment) and validity (preservation of the human knowledge for correct diagnosis). Specifically, a machine-learning model built based on fewer behavioral features that do not sufficiently capture the human knowledge of the assessment instrument, will not be valid for clinical use. Thus, there is a need for applying dimensionality reduction techniques that professionals could track their ability to preserve the validity of the assessment instruments.
Nonetheless, various feature transformation techniques were equally utilized in the dimensionality reduction processes. For instance, Akter et al. [44] utilized three feature transformation techniques; Log, Z-score, and Sine functions, and evaluated the performance of nine different ML models on the transformed datasets. Log, Z-score, and Sine functions normalize data by converting excessively skewed entities into a normal distribution, converting features into −1 to 1 value range, and transforming instances to the Sine 0–2π value intervals, respectively. Akter et al. [44] recorded varying superior performances of the ML models, and the feature transformation approaches across the datasets. The feature transformations resulting in the best classifications were Z-score and Sine function on children, adolescents, and toddlers’ datasets, respectively. However, despite the reported improved performances of the ML models on the transformed datasets and the theoretical understanding of the capabilities of the transformation functions, studies have demonstrated how these transformations compromise the relevance of the original data to the transformed data [56,57,58,59]. Researchers ought to be mindful of the limitations in using these transformations in terms of the relevance of the original to the transformed data during results interpretation. For instance, Feng [59] demonstrated such irrelevancies between the statistical findings of standard tests performed on original and log-transformed data. Similarly, several studies have highlighted some of the pitfalls and inconsistencies in the application of Z-scores and its concepts that overlooked the meaning of the original data, its standard deviations, and confusing applications [56,57,58].
Recent studies further demonstrated how ML-enable ASD screening and diagnostic models could be developed, evaluated, and implemented. Recently, Baadel et al. [43] proposed Clustering-based Autistic Trait Classification (CATC), which identifies ASD-based traits’ similarity, unlike the commonly used scoring functions. CATC showed significant improvement in the ASD classification based on clustered inputs. Comparative evaluation of various classification algorithms showed better improvement with the Random Forest classifier. On the implementation of mobile apps for ASD screening, Wingfield et al. [12], and Shahamiri and Thabtah [11] embedded RF and CNN-based scoring models, respectively, while Thabtah [13] employed ML to validate ASDTest; a mobile screening app embedded with non-ML functions. In all the foregoing studies, the commonly used evaluation metrics are classification accuracy, sensitivity, and specificity. Specificity is the ratio of non-ASD cases that are correctly classified (i.e., true negatives rate) and sensitivity is the ratio of true ASD cases that are correctly classified (i.e., true positives rate), while classification accuracy is derived from sensitivity and specificity—as the measure of precisely classified cases from the total number of the cases.
4. Discussion
The search for cost-effective ASD assessment coupled with the global rise in ASD cases attracted the implementation of quick and accurate assessment measures based on data intelligence techniques, including machine-learning algorithms. Despite the various attempts in ML-based ASD assessment using functional magnetic resonance imaging (MRI), eye tracking, and genetic data, among others, the promising results based on behavioral data call for further research. For instance, Plitt et al. [60] found that ASD classification via behavioral measures consistently surpassed rs-fMRI classifiers. Accordingly, in line with the common research aim of the behavioral studies, various dimensionality reduction techniques were employed to improve the diagnostic speed of the resulting ML models. However, unlike the reduced dimensions, there is enough evidence on the good reliability, high internal consistency, and convergent validity between the common assessment instruments within large samples [61,62,63,64,65]. Furthermore, studies have ascertained the robustness of the common assessment instruments in the quantitative measurement of the various dimensions of communication, interpersonal behavior, and stereotypic/repetitive behavior associated with ASD. Therefore, it will be difficult to sufficiently measure the key dimensions of the instruments using the fewer items generated by the common dimensionality reduction techniques. For instance, while professionals interpret SRS scores based on the sum of its 65 items, Bone et al. [38], Duda et al. [36], and Duda et al. [53] implemented SRS-enabled machine-learning models with at most 5, 5, and 15 items, respectively. Specifically, Duda et al. [36] and Duda et al. [53] focused on classifying ASD from ADHD using the SRS data from AC, AGRE, SSC. Duda et al. [36] implemented ADTree, RF, SVM, LR, Categorical lasso, and LDA models and achieved the highest area under the curve (AUC) of 0.965 in classifying ASD from ADHD by utilizing five of the 65 items of SRS identified using forward feature selection. Duda et al. [53] validated the findings of Duda et al. [36] with crowdsourced data to improve the model’s capability on ‘real-world’ data, and the findings revealed that LDA outperformed LR and SVM by achieving an AUC of 0.89 with 15 items. Despite the high metrics reported by the studies, based on the standard clinical procedures for ASD diagnosis, the ML models are neither clinically sufficient nor readily implementable for real-life use.
Similarly, Wall et al. [35] compared the performance of 15 different ML algorithms on AGRE, SSC, and AC datasets and found ADTree to outperformed other models by utilizing 7 of the 93 items contained in the ADI-R in classifying ASD with 99.9% accuracy. In a similar study by Wall et al. [34], ADTree outperformed 17 comparative models by achieving 100% accuracy with 8 of the 29 items in Module 1 of ADOS. Moreover, Duda et al. [39] demonstrated the superior performance of ADTree in achieving 97% classification accuracy with a 72% reduction in ADOS-G items. Nonetheless, Levy et al. [50] and Kosmicki et al. [37] reduced the items of ADOS using sparsity/parsimony enforcing regularization and stepwise backward feature selection techniques, respectively, and reported the superior performance of LR and SVM over other ML algorithms. Specifically, in the study by Levy et al. [50], with at most 10 features from ADOS’s Module 3 and Module 2, AUC of 0.95 and 0.93 was achieved, respectively. While Kosmicki et al. [37] recorded an accuracy of 98.27% and 97.66% with 9 of the 28 items from module 2, and 12 of the 28 items from module 3, respectively. Recently, Küpper et al. [40] utilized ADOS data from a clinical sample of adolescents and adults with ASD and reported good performance of SVM on fewer items reduced using the recursive feature selection technique. The foregoing studies have demonstrated how ML-enable ASD screening and diagnostic models could be developed and evaluated. However, numerous challenges associated with the behavioral assessment instruments, data repositories, and applied data intelligence algorithms need to be understood and addressed.
Although ML-based approaches are data-centric and are expected to improve objectivity and automation [66], with the global rise in ASD cases, the capacity to quickly and accurately assess ASD requires a careful understanding of the conceptual basis of the assessment instruments, as well as their relevance to the logical concepts of the ML algorithms. Nonetheless, discrepancies within the data repositories, such as data imbalance, limit the clinical relevance of the high evaluation metrics reported in the studies [26,27]. For instance, Torres et al. [67] studied the statistical properties of ADOS scores from 1324 records and identified various factors that could undermine the scientific viability of the scores. Particularly, the empirical distributions in the generated scores break the theoretical conditions of normality and homogeneous variance, which are critical for independence between bias and sensitivity. Thus, Torres et al. [67] suggested readjusting the scientific use of ADOS, due to the variation in the distribution of the scores, lack of appropriate metrics for characterizing changes, and the impact of both on sensitivity-bias codependencies and longitudinal tracking of ASD. In essence, the applied data intelligence algorithms, and the resulting models, missed the human knowledge upon which the assessment instruments were built and applied by the professionals [25]. Additionally, most of the studies overlooked the inherent limitations associated with the dimensionality reduction techniques, and the assessment instruments [7,8,9]. Thus, the need for ascertaining the clinical relevance of the data-centric approaches and readjusting the scientific use of the assessment instruments remains. Obviously, in the future, it can be said that the trend in the application of ML in the behavioral assessment of ASD will go on. On the other hand, the pressing demands for cost-effective assessment of ASD remain. Thus, future studies need to revisit the relevance of the data collection instruments to ML algorithms.
5. Conclusions and Recommendations
Machine learning has been broadly applied in the behavioral assessment of ASD based on a variety of data types as input to data-intelligence algorithms. Commonly utilized inputs include the items of screening tools, such as ADI-R and ADOS-G. Popular ML algorithms used are SVMs, variants of the decision trees, random forests, and neural networks. However, the multitudes of challenges in accurate ASD assessments are yet to be addressed by the suggested machine learning approaches. Specifically, the high metrics achieved with the data-intelligence techniques have not guaranteed the clinical relevance of the ML models. Additionally, the commonly used evaluation measures of classification accuracy, specificity, and sensitivity, among others cannot sufficiently reflect the human knowledge applied by professionals in assessing behavioral symptoms of ASD. Consequently, understanding the clinical basis of the assessment tools and the logical concepts of the data-intelligence techniques will lead to promising studies on the real-life implementation of cost-effective ASD assessment systems. The novelty in the present review is that while previous literature reviews focused on the performance of various data intelligent techniques on different data sets, this work systematically reviewed the literature and provide a definitive explanation on the relevance of the reported findings toward the real-life implementation of the ML-based assessment systems. The authors hope that the findings of this systematic literature review will guide researchers, caregivers, and relevant stakeholders on the advances in ASD assessment with ML.
Nonetheless, a few of the limitations associated with the present work include overlooking other non-English documents. Thus, possible excellent studies reported in other languages might have been missed. Secondly, the search filters spanned ten years and were limited to the four scientific databases mentioned. Furthermore, the records retrieved relied on the few search terms utilized in the search query. Therefore, relaxing the search filters across additional databases could yield additional relevant studies. Lastly, the present review considered only full-text online journal articles. Consequently, the findings are limited to the studies included. The future research agenda will be based on relaxing the search criteria to incorporate other scholastic databases for further comparative results. In addition, future studies could relax the search filters to include books, conference papers, and so on. Noteworthy, to build on or replicate the reviewed studies, future research should explore data-intelligence techniques that will achieve not only excellent evaluation metrics, but also adhere to the conceptual basis upon which professionals diagnose ASD.
Author Contributions
Conceptualization, A.A.L., N.C. and Z.I.; methodology, A.A.L., N.C., Z.I., S.T., U.I.A., A.D. and A.H.; resources, A.A.L., N.C., Z.I. and A.D.; data curation, A.A.L.; writing—original draft preparation, A.A.L. and N.C.; writing—review and editing, A.A.L., N.C., A.D., S.T., U.I.A., A.H. and Z.I.; visualization, A.A.L.; supervision, N.C., S.T., U.I.A. and A.D. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Data sharing is not applicable to this article.
Conflicts of Interest
The authors declare no conflict of interest.
Footnotes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.American Psychiatric Association . Diagnostic and Statistical Manual of Mental Disorders (DSM-5) 5th ed. American Psychiatric Publishing; Philadelphia, PA, USA: 2013. [Google Scholar]
- 2.Chauhan A., Sahu J., Jaiswal N., Kumar K., Agarwal A., Kaur J., Singh S., Singh M. Prevalence of autism spectrum disorder in Indian children: A systematic review and meta-analysis. Neurol. India. 2019;67:100. doi: 10.4103/0028-3886.253970. [DOI] [PubMed] [Google Scholar]
- 3.Baio J., Wiggins L., Christensen D.L., Maenner M.J., Daniels J., Warren Z., Kurzius-Spencer M., Zahorodny W., Rosenberg C.R., White T., et al. Prevalence of Autism Spectrum Disorder Among Children Aged 8 Years—Autism and Developmental Disabilities Monitoring Network, 11 Sites, United States, 2014. MMWR Surveill. Summ. 2018;67:1–23. doi: 10.15585/mmwr.ss6706a1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Durkin M.S., Elsabbagh M., Barbaro J., Gladstone M., Happe F., Hoekstra R.A., Lee L.C., Rattazzi A., Stapel-Wax J., Stone W.L., et al. Autism screening and diagnosis in low resource settings: Challenges and opportunities to enhance research and services worldwide. Autism Res. 2015;8:473–476. doi: 10.1002/aur.1575. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Matson J.L., Konst M.J. Early intervention for autism: Who provides treatment and in what settings. Res. Autism Spectr. Disord. 2014;8:1585–1590. doi: 10.1016/j.rasd.2014.08.007. [DOI] [Google Scholar]
- 6.Case-Smith J., Weaver L.L., Fristad M.A. A systematic review of sensory processing interventions for children with autism spectrum disorders. Autism. 2015;19:133–148. doi: 10.1177/1362361313517762. [DOI] [PubMed] [Google Scholar]
- 7.Guthrie W., Wallis K., Bennett A., Brooks E., Dudley J., Gerdes M., Pandey J., Levy S.E., Schultz R.T., Miller J.S. Accuracy of autism screening in a large pediatric network. Pediatrics. 2019;144:e20183963. doi: 10.1542/peds.2018-3963. [DOI] [PubMed] [Google Scholar]
- 8.Øien R.A., Candpsych S.S., Volkmar F.R., Shic F., Cicchetti D.V., Nordahl-Hansen A., Stenberg N., Hornig M., Havdahl A., Øyen A.S., et al. Clinical features of children with autism who passed 18-month screening. Pediatrics. 2018;141:e20173596. doi: 10.1542/peds.2017-3596. [DOI] [PubMed] [Google Scholar]
- 9.Surén P., Saasen-Havdahl A., Bresnahan M., Hirtz D., Hornig M., Lord C., Reichborn-Kjennerud T., Schjølberg S., Øyen A.-S., Magnus P., et al. Sensitivity and specificity of early screening for autism. BJPsych Open. 2019;5:1–8. doi: 10.1192/bjo.2019.34. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Yuen T., Carter M.T., Szatmari P., Ungar W.J. Cost-Effectiveness of Universal or High-Risk Screening Compared to Surveillance Monitoring in Autism Spectrum Disorder. J. Autism Dev. Disord. 2018;48:2968–2979. doi: 10.1007/s10803-018-3571-4. [DOI] [PubMed] [Google Scholar]
- 11.Shahamiri S.R., Thabtah F. Autism AI: A New Autism Screening System Based on Artificial Intelligence. Cognit. Comput. 2020;12:766–777. doi: 10.1007/s12559-020-09743-3. [DOI] [Google Scholar]
- 12.Wingfield B., Miller S., Yogarajah P., Kerr D., Gardiner B., Seneviratne S., Samarasinghe P., Coleman S. A predictive model for paediatric autism screening. Health Inform. J. 2020;26:2538–2553. doi: 10.1177/1460458219887823. [DOI] [PubMed] [Google Scholar]
- 13.Thabtah F. An accessible and efficient autism screening method for behavioural data and predictive analyses. Health Inform. J. 2019;25:1739–1755. doi: 10.1177/1460458218796636. [DOI] [PubMed] [Google Scholar]
- 14.Campbell K., Carpenter K.L.H., Espinosa S., Hashemi J., Qiu Q., Tepper M., Calderbank R., Sapiro G., Egger H.L., Baker J.P., et al. Use of a Digital Modified Checklist for Autism in Toddlers–Revised with Follow-up to Improve Quality of Screening for Autism. J. Pediatr. 2017;183:133–139.e1. doi: 10.1016/j.jpeds.2017.01.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Ghafouri-Fard S., Taheri M., Omrani M.D., Daaee A., Mohammad-Rahimi H., Kazazi H. Application of Single-Nucleotide Polymorphisms in the Diagnosis of Autism Spectrum Disorders: A Preliminary Study with Artificial Neural Networks. J. Mol. Neurosci. 2019;68:515–521. doi: 10.1007/s12031-019-01311-1. [DOI] [PubMed] [Google Scholar]
- 16.Sekaran K., Sudha M. Predicting autism spectrum disorder from associative genetic markers of phenotypic groups using machine learning. J. Ambient Intell. Humaniz. Comput. 2020;12:1–14. doi: 10.1007/s12652-020-02155-z. [DOI] [Google Scholar]
- 17.Jack A. Neuroimaging in neurodevelopmental disorders: Focus on resting-state fMRI analysis of intrinsic functional brain connectivity. Curr. Opin. Neurol. 2018;31:140–148. doi: 10.1097/WCO.0000000000000536. [DOI] [PubMed] [Google Scholar]
- 18.Fu C.H.Y., Costafreda S.G. Neuroimaging-based biomarkers in psychiatry: Clinical opportunities of a paradigm shift. Can. J. Psychiatry. 2013;58:499–508. doi: 10.1177/070674371305800904. [DOI] [PubMed] [Google Scholar]
- 19.Moon S.J., Hwang J., Kana R., Torous J., Kim J.W. Accuracy of machine learning algorithms for the diagnosis of autism spectrum disorder: Systematic review and meta-analysis of brain magnetic resonance imaging studies. J. Med. Internet Res. 2019;6:e14108. doi: 10.2196/14108. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Sarabadani S., Schudlo L.C., Samadani A.A., Kushski A. Physiological Detection of Affective States in Children with Autism Spectrum Disorder. IEEE Trans. Affect. Comput. 2020;11:588–600. doi: 10.1109/TAFFC.2018.2820049. [DOI] [Google Scholar]
- 21.Liu W., Li M., Yi L. Identifying children with autism spectrum disorder based on their face processing abnormality: A machine learning framework. Autism Res. 2016;9:888–898. doi: 10.1002/aur.1615. [DOI] [PubMed] [Google Scholar]
- 22.Alcañiz Raya M., Chicchi Giglioli I.A., Marín-Morales J., Higuera-Trujillo J.L., Olmos E., Minissi M.E., Teruel Garcia G., Sirera M., Abad L. Application of Supervised Machine Learning for Behavioral Biomarkers of Autism Spectrum Disorder Based on Electrodermal Activity and Virtual Reality. Front. Hum. Neurosci. 2020;14:90. doi: 10.3389/fnhum.2020.00090. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Hashemi J., Dawson G., Carpenter K.L.H., Campbell K., Qiu Q., Espinosa S., Marsan S., Baker J.P., Egger H.L., Sapiro G. Computer Vision Analysis for Quantification of Autism Risk Behaviors. IEEE Trans. Affect. Comput. 2018;3045:1–12. doi: 10.1109/TAFFC.2018.2868196. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Dahiya A.V., McDonnell C., DeLucia E., Scarpa A. A systematic review of remote telehealth assessments for early signs of autism spectrum disorder: Video and mobile applications. Pract. Innov. 2020;5:150–164. doi: 10.1037/pri0000121. [DOI] [Google Scholar]
- 25.Thabtah F. Machine learning in autistic spectrum disorder behavioral research: A review and ways forward. Inform. Health Soc. Care. 2019;44:278–297. doi: 10.1080/17538157.2017.1399132. [DOI] [PubMed] [Google Scholar]
- 26.Alahmari F. A Comparison of Resampling Techniques for Medical Data Using Machine Learning. J. Inf. Knowl. Manag. 2020;19:1–13. doi: 10.1142/S021964922040016X. [DOI] [Google Scholar]
- 27.Abdelhamid N., Padmavathy A., Peebles D., Thabtah F., Goulder-Horobin D. Data Imbalance in Autism Pre-Diagnosis Classification Systems: An Experimental Study. J. Inf. Knowl. Manag. 2020;19:1–16. doi: 10.1142/S0219649220400146. [DOI] [Google Scholar]
- 28.Song D.-Y., Kim S.Y., Bong G., Kim J.M., Yoo H.J. The Use of Artificial Intelligence in Screening and Diagnosis of Autism Spectrum Disorder: A Literature Review. J. Korean Acad. Child Adolesc. Psychiatry. 2019;30:145–152. doi: 10.5765/jkacap.190027. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Moher D., Liberati A., Tetzlaff J., Altman D.G., Group T.P. Preferred Reporting Items for Systematic Reviews and Meta-Analyses: The PRISMA Statement. PLoS Med. 2009;6:e1000097. doi: 10.1371/journal.pmed.1000097. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Low D.M., Bentley K.H., Ghosh S.S. Automated assessment of psychiatric disorders using speech: A systematic review. Laryngoscope Investig. Otolaryngol. 2020;5:96–116. doi: 10.1002/lio2.354. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Lebersfeld J.B., Swanson M., Clesi C.D., Kelley S.E.O. Systematic Review and Meta-Analysis of the Clinical Utility of the ADOS-2 and the ADI-R in Diagnosing Autism Spectrum Disorders in Children. J. Autism Dev. Disord. 2021;51:1–14. doi: 10.1007/s10803-020-04839-z. [DOI] [PubMed] [Google Scholar]
- 32.Kulage K.M., Goldberg J., Usseglio J., Romero D., Bain J.M., Smaldone A.M. How has DSM-5 Affected Autism Diagnosis? A 5-Year Follow-Up Systematic Literature Review and Meta-analysis. J. Autism Dev. Disord. 2020;50:2102–2127. doi: 10.1007/s10803-019-03967-5. [DOI] [PubMed] [Google Scholar]
- 33.Smith I.C., Reichow B., Volkmar F.R. The Effects of DSM-5 Criteria on Number of Individuals Diagnosed with Autism Spectrum Disorder: A Systematic Review. J. Autism Dev. Disord. 2015;45:2541–2552. doi: 10.1007/s10803-015-2423-8. [DOI] [PubMed] [Google Scholar]
- 34.Wall D., Kosmicki J., Deluca T., Harstad E., Fusaro V. Use of machine learning to shorten observation-based screening and diagnosis of autism. Transl. Psychiatry. 2012;2:e100-8. doi: 10.1038/tp.2012.10. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Wall D., Dally R., Luyster R., Jung J.Y., DeLuca T. Use of artificial intelligence to shorten the behavioral diagnosis of autism. PLoS ONE. 2012;7:e43855. doi: 10.1371/journal.pone.0043855. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Duda M., Ma R., Haber N., Wall D. Use of machine learning for behavioral distinction of autism and ADHD. Transl. Psychiatry. 2016;6:1–5. doi: 10.1038/tp.2015.221. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Kosmicki J., Sochat V., Duda M., Wall D. Searching for a minimal set of behaviors for autism detection through feature selection-based machine learning. Transl. Psychiatry. 2015;5:1–7. doi: 10.1038/tp.2015.7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Bone D., Bishop S.L., Black M.P., Goodwin M.S., Lord C., Narayanan S.S. Use of machine learning to improve autism screening and diagnostic instruments: Effectiveness, efficiency, and multi-instrument fusion. J. Child Psychol. Psychiatry Allied Discip. 2016;57:927–937. doi: 10.1111/jcpp.12559. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Duda M., Kosmicki J., Wall D. Testing the accuracy of an observation-based classifier for rapid detection of autism risk. Transl. Psychiatry. 2015;5:e556. doi: 10.1038/tp.2015.51. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Küpper C., Stroth S., Wolff N., Hauck F., Kliewer N., Schad-Hansjosten T., Kamp-Becker I., Poustka L., Roessner V., Schultebraucks K., et al. Identifying predictive features of autism spectrum disorders in a clinical sample of adolescents and adults using machine learning. Sci. Rep. 2020;10:1–11. doi: 10.1038/s41598-020-61607-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Bellesheim K.R., Cole L., Coury D.L., Yin L., Levy S.E., Guinnee M.A., Klatka K., Malow B.A., Katz T., Taylor J., et al. Family-driven goals to improve care for children with autism spectrum disorder. Pediatrics. 2018;142:e20173225. doi: 10.1542/peds.2017-3225. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Puerto E., Aguilar J., López C., Chávez D. Using Multilayer Fuzzy Cognitive Maps to diagnose Autism Spectrum Disorder. Appl. Soft Comput. J. 2019;75:58–71. doi: 10.1016/j.asoc.2018.10.034. [DOI] [Google Scholar]
- 43.Baadel S., Thabtah F., Lu J. A clustering approach for autistic trait classification. Inform. Health Soc. Care. 2020;45:309–326. doi: 10.1080/17538157.2019.1687482. [DOI] [PubMed] [Google Scholar]
- 44.Akter T., Shahriare Satu M., Khan M.I., Ali M.H., Uddin S., Lio P., Quinn J.M.W., Moni M.A. Machine Learning-Based Models for Early Stage Detection of Autism Spectrum Disorders. IEEE Access. 2019;7:166509–166527. doi: 10.1109/ACCESS.2019.2952609. [DOI] [Google Scholar]
- 45.Usta M.B., Karabekiroglu K., Sahin B., Aydin M., Bozkurt A., Karaosman T., Aral A., Cobanoglu C., Kurt A.D., Kesim N., et al. Use of machine learning methods in prediction of short-term outcome in autism spectrum disorders. Psychiatry Clin. Psychopharmacol. 2019;29:320–325. doi: 10.1080/24750573.2018.1545334. [DOI] [Google Scholar]
- 46.Thabtah F., Kamalov F., Rajab K. A new computational intelligence approach to detect autistic features for autism screening. Int. J. Med. Inform. 2018;117:112–124. doi: 10.1016/j.ijmedinf.2018.06.009. [DOI] [PubMed] [Google Scholar]
- 47.Pratama T.G., Hartanto R., Setiawan N.A. Machine learning algorithm for improving performance on 3 AQ-screening classification. Commun. Sci. Technol. 2019;4:44–49. doi: 10.21924/cst.4.2.2019.118. [DOI] [Google Scholar]
- 48.Thabtah F., Abdelhamid N., Peebles D. A machine learning autism classification based on logistic regression analysis. Health Inf. Sci. Syst. 2019;7:1–11. doi: 10.1007/s13755-019-0073-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Suresh Kumar R., Renugadevi M. Differential evolution tuned support vector machine for autistic spectrum disorder diagnosis. Int. J. Recent Technol. Eng. 2019;8:3861–3870. doi: 10.35940/ijrte.B3063.078219. [DOI] [Google Scholar]
- 50.Levy S., Duda M., Haber N., Wall D. Sparsifying machine learning models identify stable subsets of predictive features for behavioral detection of autism. Mol. Autism. 2017;8:1–17. doi: 10.1186/s13229-017-0180-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Goel N., Grover B., Gupta D., Khanna A., Sharma M. Modified Grasshopper Optimization Algorithm for detection of Autism Spectrum Disorder. Phys. Commun. 2020;41:101115. doi: 10.1016/j.phycom.2020.101115. [DOI] [Google Scholar]
- 52.Thabtah F., Peebles D. A new machine learning model based on induction of rules for autism detection. Health Inform. J. 2020;26:264–286. doi: 10.1177/1460458218824711. [DOI] [PubMed] [Google Scholar]
- 53.Duda M., Haber N., Daniels J., Wall D. Crowdsourced validation of a machine-learning classification system for autism and ADHD. Transl. Psychiatry. 2017;7:2–8. doi: 10.1038/tp.2017.86. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Alhaj T.A., Siraj M.M., Zainal A., Elshoush H.T., Elhaj F. Feature Selection Using Information Gain for Improved Structural-Based Alert Correlation. PLoS ONE. 2016;11:e0166017. doi: 10.1371/journal.pone.0166017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55.Roobaert D., Karakoulas G., Chawla N.V. Feature Extraction. Vol. 207. Springer; Berlin/Heidelberg, Germany: 2006. Information Gain, Correlation and Support Vector Machines; pp. 463–470. [Google Scholar]
- 56.Wiesen J. Benefits, Drawbacks, and Pitfalls of z-Score Weighting; Proceedings of the 30th Annual IPMAAC Conference; Las Vegas, NV, USA. 27 June 2006; pp. 1–41. [Google Scholar]
- 57.Lapteacru I. On the Consistency of the Z-Score to Measure the Bank Risk. SSRN Electron. J. 2016;4:1–33. doi: 10.2139/ssrn.2787567. [DOI] [Google Scholar]
- 58.Curtis A., Smith T., Ziganshin B., Elefteriades J. The Mystery of the Z-Score. AORTA. 2016;4:124–130. doi: 10.12945/j.aorta.2016.16.014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Feng C., Wang H., Lu N., Chen T., He H., Lu Y., Tu X.M. Log-transformation and its implications for data analysis. Shanghai Arch. Psychiatry. 2014;26:105–109. doi: 10.3969/j.issn.1002-0829.2014.02. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Plitt M., Barnes K.A., Martin A. Functional connectivity classification of autism identifies highly predictive brain features but falls short of biomarker standards. NeuroImage Clin. 2015;7:359–366. doi: 10.1016/j.nicl.2014.12.013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 61.Chan W., Smith L.E., Hong J., Greenberg J.S., Mailick M.R. Validating the social responsiveness scale for adults with autism. Autism Res. 2017;10:1663–1671. doi: 10.1002/aur.1813. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 62.Becker M.M., Wagner M.B., Bosa C.A., Schmidt C., Longo D., Papaleo C., Riesgo R.S. Translation and validation of Autism Diagnostic Interview-Revised (ADI-R) for autism diagnosis in Brazil. Arq. Neuropsiquiatr. 2012;70:185–190. doi: 10.1590/S0004-282X2012000300006. [DOI] [PubMed] [Google Scholar]
- 63.Falkmer T., Anderson K., Falkmer M., Horlin C. Diagnostic procedures in autism spectrum disorders: A systematic literature review. Eur. Child Adolesc. Psychiatry. 2013;22:329–340. doi: 10.1007/s00787-013-0375-0. [DOI] [PubMed] [Google Scholar]
- 64.Medda J.E., Cholemkery H., Freitag C.M. Sensitivity and Specificity of the ADOS-2 Algorithm in a Large German Sample. J. Autism Dev. Disord. 2019;49:750–761. doi: 10.1007/s10803-018-3750-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Chojnicka I., Pisula E. Adaptation and Validation of the ADOS-2, Polish Version. Front. Psychol. 2017;8:1–14. doi: 10.3389/fpsyg.2017.01916. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Achenie L.E.K., Scarpa A., Factor R.S., Wang T., Robins D.L., McCrickard D.S. A Machine Learning Strategy for Autism Screening in Toddlers. J. Dev. Behav. Pediatr. 2019;40:369–376. doi: 10.1097/DBP.0000000000000668. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Torres E.B., Rai R., Mistry S., Gupta B. Hidden aspects of the research ADOS are bound to affect autism science. Neural Comput. 2020;32:515–561. doi: 10.1162/neco_a_01263. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Data sharing is not applicable to this article.