Skip to main content
PLOS One logoLink to PLOS One
. 2024 May 14;19(5):e0302236. doi: 10.1371/journal.pone.0302236

A hybrid CNN-SVM model for enhanced autism diagnosis

Linjie Qiu 1,*,#, Jian Zhai 1,#
Editor: Umer Asgher2
PMCID: PMC11093301  PMID: 38743688

Abstract

Autism is a representative disorder of pervasive developmental disorder. It exerts influence upon an individual’s behavior and performance, potentially co-occurring with other mental illnesses. Consequently, an effective diagnostic approach proves to be invaluable in both therapeutic interventions and the timely provision of medical support. Currently, most scholars’ research primarily relies on neuroimaging techniques for auxiliary diagnosis and does not take into account the distinctive features of autism’s social impediments. In order to address this deficiency, this paper introduces a novel convolutional neural network-support vector machine model that integrates resting state functional magnetic resonance imaging data with the social responsiveness scale metrics for the diagnostic assessment of autism. We selected 821 subjects containing the social responsiveness scale measure from the publicly available Autism Brain Imaging Data Exchange dataset, including 379 subjects with autism spectrum disorder and 442 typical controls. After preprocessing of fMRI data, we compute the static and dynamic functional connectivity for each subject. Subsequently, convolutional neural networks and attention mechanisms are utilized to extracts their respective features. The extracted features, combined with the social responsiveness scale features, are then employed as novel inputs for the support vector machine to categorize autistic patients and typical controls. The proposed model identifies salient features within the static and dynamic functional connectivity, offering a possible biological foundation for clinical diagnosis. By incorporating the behavioral assessments, the model achieves a remarkable classification accuracy of 94.30%, providing a more reliable support for auxiliary diagnosis.

Introduction

Autism spectrum disorder (ASD) is a neurodevelopmental disease marked by language impairment, social impairment, and stereotyped behavior. According to statistical data, approximately one in every 100 children is diagnosed with autism [1]. Despite its characteristics often being identifiable in early childhood, definitive diagnosis typically requires an extended period of time. Given that individuals with ASD generally need specialized medical care to mitigate the risk of co-occurring mental health conditions, timely and accurate diagnosis is imperative.

Nowadays, machine learning is widely employed in auxiliary diagnosis across various medical conditions, such as brain tumors [2, 3], autism [4, 5], depression [6] and more. They learn features from data and help doctors screen and diagnose diseases. These methods include traditional machine learning techniques, such as support vector machine (SVM) [7], random forest (RF), and deep learning models, such as convolutional neural network (CNN) [8] and vision transformer (ViT) [9]. However, compared to doctors who make comprehensive judgments based on all aspects of information, most models tend to rely on a certain aspect of subjects, such as patients’ medical imaging data. When dealing with conditions characterized by lesions, such as brain tumors, a reliance on brain imaging data proves effective [10]. However, for mental disorders like autism, a singular data source may be insufficient. Therefore, a more rational approach involves the integration of multiple facets through an ensemble learning model.

In the context of autism, numerous studies have proposed various models from different perspectives, offering diverse strategies for diagnosis. There are classification methods based on extracting features from facial image data [11], classification based on extracted features from audio and video data of subjects’ behavior [12], classification based on subjects’ eye-tracking data [13], and models based on functional magnetic resonance imaging (fMRI) data [5], etc. Among these, fMRI technology has gained widespread usage in research due to its high resolution and non-invasive nature, both in task-based and resting-state scenarios [14, 15]. Based on fMRI technology, there are models that directly utilize MRI data for deep learning construction [16] and others that employ various preprocessing tools to extract features for subsequent analysis. Functional connectivity (FC) is a commonly used analytical tool. The intricate physiological activities within the human body orchestrated through the coordinated interplay among various brain regions, and the relationship between brain regions evolve over time. Leveraging these connections holds promise for diagnosis of autism and finding areas that cause brain disorders. Resting-state functional magnetic resonance imaging (rs-fMRI), with its advantages of not requiring additional complex tasks, has been widely used in previous study. Constructing functional connectivity based rs-fMRI data serves as an efficient method for characterizing relationships among brain areas. Early scholarly articles often focused solely on static functional connectivity, which presupposes that brain interrelations remain constant throughout the whole scanning process. However, recent literature reveals that neural connections fluctuate during the scan, thus giving rise to the broader application of dynamic functional connectivity in brain activity research [1719]. To furnish a more comprehensive evaluation of the brain activities, this paper employs both static and dynamic functional connectivity to analyze differences between the subjects with ASD and typical controls (TCs), trying to identify areas causing ASD. Nevertheless, only the analysis of brain activity may prove.

As mentioned earlier, relying solely on the analysis of brain activity connections may not suffice for the effective diagnosis of psychiatric disorders. Furthermore, since autism mainly manifests impairments in social behavior, FC may not adequately reflect this aspect. To address this limitation, more information is required. Current psychiatric screening or diagnostic procedures predominantly rely on behavioral observations rooted in symptomatology. For example, the Social Responsiveness Scale–Second Edition (SRS-2) serves as an early screening tool for ASD [20], quantifying social functionality in five aspects: awareness, cognition, communication, motivation, and mannerisms. Substantial evidence supports its effectiveness and sensitivity in identifying autism symptoms in school-age children. However, the questionnaire mainly focuses on social communication, and only a few items involve stereotyped behavior. Researchers are still gathering more data on the stability and effectiveness of the SRS-2 [2123]. While other diagnostic tools such as the Autism Diagnostic Observation Schedule, the Childhood Autism Rating Scale, and the Autism Diagnostic Interview-Revised are commonly employed in clinical practice, these methods are more time-consuming. However, these methods are inevitably affected by clinical training, tools, and cultural background, thereby interfering with clinicians’ subjective observations [24, 25]. To overcome this limitation, it is not enough to only use the scale results to assist diagnosis. Consequently, we have integrated analyses of functional connectivity to better assist diagnosis.

To harness the full potential of early screening questionnaires and brain functional connectivity for autism diagnosis, this paper introduces a hybrid CNN-SVM model. This model employs a CNN architecture to extract deep features from both static and dynamic functional connectivity. In the learning process, the feature extraction based on the frequency band and the convolution kernel based on the functional connectivity matrix are used, and the attention mechanism is introduced to weight the learned features. Finally, the learned features combined with the features from the SRS are sent to SVM for classification. After studying 379 subjects with ASD and 442 typical controls, our model achieves good classification accuracy and provides the most discriminating brain regions and scanning bands. The contributions of this work are summarized as:

  1. A novel ensemble learning model is proposed, which integrates information from FC in fMRI data with the SRS scores. SRS scores reflect the degree of social impairment in subjects.

  2. Leveraging fMRI data, we construct both static and dynamic FC. Based on the CNN structure of extracted features, the brain areas and frequency bands that play a greater role in classification are found. And the attention mechanism is introduced to enhance the extracted features from FC.

  3. To assess performance, we conduct comparative evaluations with various machine learning models, RF, logistic regression (LR), multilayer perceptron (MLP) and SVM. We also show the model’s performance on data across genders, age groups and sites. Besides, we compare our model with some previous works.

The structure of this paper is as follows. Section 2 introduces data preprocessing and network architecture. Section 3 presents our results. Section 4 discusses the limitations of our study and the future work. Section 5 provides the conclusions.

Materials and methods

The whole process of this experiment is illustrated in Fig 1. More details of Data preprocessing and CNN for feature extraction are shown in the following sections.

Fig 1. The entire process of this experiment.

Fig 1

Details of the CNN for extracting features from static and dynamic FC parts are illustrated in Figs 3 and 4 separately. The left figure in the raw data part is republished from [26] under a CC BY license, with permission from Pixabay, original copyright 2017. The right figure in the raw data part is republished from [27] under a CC BY license, with permission from Pixabay, original copyright 2015.

To enhance the comprehension of Fig 1, we shall describe the flowchart in algorithmic form as follows:

Input:

  • Training dataset D = {(x1, y1), (x2, y2), …, (xn, yn)}, where xi denotes features, and yi represents the corresponding labels.

  • The number of folds for cross-validation, denoted as k = 10.

Output: A well-trained model.

Algorithm Description:

  1. Preprocess the fMRI data xi from the training dataset D, following the steps outlined in the subsection 0.1. This yields a new dataset D1 = {(a1, t1, y1), (a2, t2, y2), …, (an, tn, yn)}, where ai represents SRS features, ti corresponds to the ROI time series, and yi denotes labels.

  2. Based on ti, calculate the required static FC bi and dynamic FC ci, with detailed computational formulas provided in subsections 0.2 and 0.3. This results in a new dataset D2 = {(a1, b1, c1, y1), (a2, b2, c2, y2), …, (an, bn, cn, yn)}.

  3. Partition the training dataset D2 into k non-overlapping subsets, typically employing a K-fold cross-validation approach. For each fold i, designate it as the validation set while using the remaining k − 1 folds as the training set.

  4. Feed the static FC bi and dynamic FC ci of the training set into a CNN for deep feature extraction. Utilize the backpropagation algorithm and Adam optimizer to update model parameters. Employ the cross-entropy loss function, monitor its convergence during training, and set the maximum number of epochs to 50. The details of neural network architecture can be found in subsection 0.4. Assuming the extracted deep features are denoted as di and ei, combine them with ai as (ai, di, ei) and feed them into a linear kernel SVM for training.

  5. Utilize the well-trained model to make predictions on the validation set and compute performance metrics.

  6. Repeat steps 4 and 5 iteratively, using each fold as the validation set, to obtain a set of performance metrics.

  7. Calculate the average performance metrics from the k folds to assess the model’s performance. Concurrently, save the model parameters from the training process with the best performance metrics as the final model for deployment.

0.1 Data preprocessing

Firstly, we select resting-state fMRI data that includes SRS metrics from the publicly available Autism Brain Imaging Data Exchange (ABIDE) dataset [28]. The distribution and age of the subjects selected are shown in the Table 1 and Fig 2. Upon acquiring the data from the ABIDE dataset, we perform the following preprocessing steps:

Table 1. Distribution of the data from rs-fMRI ABIDE database used in this study.

Site ASD TC Total AgeRange
LEUVEN 29 35 64 12-32
BNI 28 29 57 18-64
ETH 11 22 33 13-31
GU 51 55 106 8-14
IU 15 11 26 17-54
KKI 52 148 200 8-13
NYU 75 28 103 5-34
OHSU 33 46 79 8-15
SDSU 32 25 57 7-18
TCD 21 21 42 10-20
UCD 18 14 32 12-18
USM 14 8 22 9-39
total 379 442 821

Fig 2. The distribution of subjects in the dataset.

Fig 2

  1. Extract the evaluation scores of the five aspects of the SRS, convert them into a one-dimensional vector and subsequently send it to the final SVM for classification, as depicted in the SRS feature in Fig 1.

  2. Remove the first ten time points of each subject to mitigate errors caused by the instability of the gradient magnetic field at the beginning of the scan; correct slice scan time to ensure that the resampled data can be seen as scanned at the same time point; carry out head motion correction to mitigate the impact of head movements to some extent; and perform nuisance regression to eliminate the influence of extraneous factors.

  3. Register the structural image data and then map it to the standard space-MNI space, which aligns the images of each subject to ensure that the anatomical structures of different subjects correspond to the same voxels, and perform spatial smoothing to enhance registration effects.

  4. Employ band–pass filtering (0.01–0.08 Hz). Most nuclear magnetic resonance signals are low–frequency signals. Given that magnetic resonance signals are primarily low–frequency signals, this frequency range is chosen to minimize the impact of physiological noise from cardiac (∼ 0.15) and respiratory (∼0.3 Hz) activities [29].

  5. Utilize the Automated Anatomical Labeling (AAL) atlas, which partitions the brain into 116 regions. These regions serve as our regions of interest (ROIs), and the corresponding time series are extracted.

This study uses SPM12 and DPARSF [30] to preprocess fMRI data as mentioned above, and the extracted time series is used for subsequent processing.

0.2 Static functional connectivity

The construction of a static functional connectivity matrix is predicated on the Pearson correlation coefficients between various regions of interest in the brain, as shown in Eq 1.

R(X,Y)=(xi-x¯i)(yi-y¯i)(xi-x¯i)2(yi-y¯i)2, (1)

where X and Y are two distinct brain regions, while xi and yi denote the corresponding time series. x¯i and y¯i are mean values of xi and yi separately. The underlying assumption of static functional connectivity is that the interconnections between different brain areas remain invariant throughout the entire scanning procedure, and the Pearson correlation coefficient characterizes this relationship [18]. The resulting static functional connectivity matrix has dimensions of 116 × 116, as illustrated in the input part of Fig 3.

Fig 3. The entire process of processing static functional connectivity.

Fig 3

0.3 Dynamic functional connectivity

Throughout the scanning process, the functional connectivity between different brain regions exhibits temporal fluctuations, indicating that a representation solely based on static functional connectivity would be overly simplistic. This has led to the introduction of dynamic functional connectivity, which also serves to enrich the data set. Sliding-window analysis is a method often considered for dynamic functional connectivity. It describes changes in the brain’s connection patterns by using a fixed window size and moving it at a specific stride. However, there is no consensus on the optimal size for the fixed temporal window. Additionally, variations have been observed in connectivity matrices at different frequencies [31, 32]. These issues cannot be solved by sliding-window method. We therefore perform a time-frequency analysis of the signal. We choose to use wavelet analysis to characterize dynamic functional connectivity, which effectively extracts information from signals through operations such as scaling and translation, thus enabling a multi-scale, fine-grained examination.

Initially, we employ continuous wavelet transforms to process the time series for each brain area, as delineated by the Eq 2.

W(s,τ)=1s-x(t)ϕ*(t-τs)dt, (2)

where s is the wavelet scale, τ denotes the translation value, x(t) is the signal, ϕ represents a mother wavelet and * denotes the complex conjugate [33]. Scale s controls the expansion and contraction of the wavelet function, which is inversely proportional to frequency. The translation τ controls the translation of the wavelet function, which corresponds to time points. When we traverse the required scales and translations, we can obtain the needed spectrogram from the unstable signal for subsequent analysis. We employ the Morlet wavelet as the mother wavelet, which offers an optimal ratio between frequency bands and wavelet scales, facilitating the understanding of data within specific frequency ranges [34]. In our analysis, we partition the frequency range of 0.01-0.08 Hz into 40 segments, and the translation τ corresponds to the sampled time points.

After obtaining the time-frequency spectra for each region of interest, we proceed to compute their dynamic functional connectivity. Here, in order to describe the time-frequency spatial relationship between two signals, we initially introduce the concept of cross-wavelet power [33], which identifies regions of the common high power between the two signals, as shown in Eq 3.

CWxy(s,τ)=Wx(s,τ)Wy*(s,τ), (3)

where Wx and Wy are the continuous wavelet of x and y separately. To further describe the coherence of these two cross wavelets in time-frequency space as the functional connectivity between two brain areas, we introduce the concept of wavelet coherence [33], calculated as Eq 4.

Rxy=|S(CWxy(s,τ))|2S(|Wx(s,τ)|2)·S(|Wy(s,τ)|2), (4)

where S(⋅) denotes a smoothing operator related to scale and time, for which we employ the most commonly used moving average technique.

Finally, after processing different times and scales, we convert the signals of any two regions of interest into a matrix WCRm×n, here m corresponds to the number of chosen scales, specifically m = 40, and n corresponds to the number of time points. In this way, for the same individual, we can get 6612 completely different matrices. Obviously, the amount of data is too large. At the same time, the data obtained by different institutions have different time points n, and so they cannot be easily passed to the same neural network for learning. To this end, we consider employing Principal Component Analysis (PCA) on the matrices [35]. In order to facilitate the unification of data from different institutions, we opt to perform dimensionality reduction on the temporal dimension, as detailed in the following procedures:

  1. Calculate wavelet coherence for two regions of interest to obtain a matrix WC = [wc1, …, wcn]

  2. For the matrix WC, remove the average value to get WC^=[w1,,wn], where wi=wci-j=1nwci/n,i=1,2,,n.

  3. Construct matrix 1nWC^·WC^T. Use the eigenvalue decomposition method to calculate its eigenvalues and the corresponding unit eigenvectors. Suppose there are l eigenvalues and unit eigenvectors, and eigenvalues λ1, λ2, …, λl are arranged from large to small. Then 1nWC^WC^TXi=λiXi,i=1,2,,l.

  4. In order to retain 99% of the information, it is necessary to find the smallest k that satisfies i=1kλii=1lλi0.99. These k unit eigenvectors can form an eigenvector matrix P.

  5. Finally, the matrix is mapped to the space composed of the corresponding feature vectors, that is, XPRm×k.

After performing principal component analysis on all matrices WCRm×n, we find that k = 1 can satisfy the condition in step 4. Thus we finally obtain a matrix of 116 × 116 × 40 for each subject as the dynamic functional connectivity.

0.4 CNN for feature extraction

As illustrated in Fig 1, the input to our model is based on data from SRS tables, static FC and dynamic FC. Here, we employ CNNs to extract features from static FC and dynamic FC. For ordinary pictures, feature extraction is proficiently achieved through the utilization of 3 × 3 or 5 × 5 convolution kernels [36, 37]. However, when it comes to the FC matrix for a subject, each row or column of the matrix corresponding to the correlation between two brain areas. The aforementioned convolution kernel cannot summarize such information, and there is no reasonable explanation for the local features. Therefore, we believe that a more reasonable convolution kernel should be 1 × n or n × 1, where n corresponds to the number of regions of interest, that is, 116. The convolution kernel proposed here is more capable of extracting brain areas corresponding to functional connectivity that make greater contributions to classification.

For static FC, each subject corresponds to a 116 × 116 × 1 tensor, as shown in Fig 3. We send it into the CNN. The first layer comprises 32 filters, and the corresponding convolution kernel is 1 × 116. After passing through the batch normalization (BN) layer and rectified linear unit (ReLu) layer, it is sent to the second convolution layer. The second layer comprises 64 filters, and the corresponding convolution kernel is 116 × 1. After passing through the BN layer and ReLu layer, it is sent to the third convolution layer. The third layer comprises 16 filters, and the corresponding convolution kernel is 1 × 1. Finally, it goes through the BN layer and ReLu layer. The third layer of convolutional layer is to prevent the extracted feature dimension from being too large and affecting subsequent classification. In addition to this step, we also use a dropout layer to prevent overfitting, where the dropout rate is set to 50%. All activation functions employed are of the LeakyReLU type (α = 0.01).

The most discriminative deep features need to be distinguished through the attention mechanism [38], which follows the dropout layer. This attention network is bifurcated into two fully connected layers: the first boasting 8 neurons, followed by a ReLU layer, and the second containing 16 neurons, which is the same number of the input neurons. Then, it passes through the Sigmoid function to obtain the normalized weight of each input feature. These normalized weights are utilized for weighted summation to extract the requisite static functional connectivity features. To facilitate backpropagation for parameter tuning, an additional layer is added, the size of which corresponds to the number of classification categories, which is 2. The loss between the actual labels and the predicted outcomes is minimized via the adaptive moment estimation optimizer, with a learning rate set at 0.0001. The employed loss function is cross-entropy. However, the focus is not on the network’s predictive capacity. We just use it to train the network parameters and so the final hidden layer’s parameters serve as the learned static functional connectivity features, as shown in the staticFC feature in Fig 1.

For dynamic FC, we can know that each object corresponds to a matrix of 116 × 116 × 40 after PCA processing, as shown in Fig 4. The first step is channel compression. As mentioned before, each 116 × 116 layer corresponds to a distinct frequency. The compression of channels can be interpreted as the neural network looking for the most discriminative frequency bands for classification, which may have certain reference value for the analysis of fMRI data. Hence, in comparison to the extraction of features from static FC, the processing of dynamic FC necessitates an additional step of channel compression. The first layer comprises 1 filter with a corresponding 1 × 1 convolutional kernel. Then it is sent to the BN layer and the ReLu layer. The subsequent operations is similar to the operations that performed on the static FC. Finally the dynamicFC features required in the Fig 1 are obtained.

Fig 4. The entire process of processing dynamic functional connectivity.

Fig 4

Subsequently, SRS features are concatenated with those gleaned from both static FC and dynamic FC. This composite feature set is then submitted to a SVM with a linear kernel for further classification. This experiment uses a 10-fold cross-validation of the data. The reason why this study uses SVMs instead of the multi-layer perceptron (MLP) for classification is because in experiments we find that the MLP model has poor generalization capability and performs poorly on the test set. In the support vector machine part, we use the more common linear kernel for classification, because the extracted feature dimension is small enough compared to the data set. In order to illustrate the advantages of SVM, we also conduct experiments using MLP and RF for classification. The architecture of the MLP consists of two hidden layers, featuring 64 and 16 neurons respectively. The number of neurons in the input layer is the dimension after concatenating three features, and the output layer is 2 neurons. As for RF, 100 trees are selected for classification. And for LR, we choose a threshold of 0.5 to distinguish subjects with ASD and TCs. These experiments all adopt the 10-fold cross-validation, and the results are averaged after several experiments.

Results

0.5 Classification comparison

According to the results of cross-validation, sensitivity (SEN), specificity (SPE), accuracy (ACC), false positive rate (FPR), false negative rate (FNR) and F1 Score (F1) are used as evaluation indicators for classification results. Here, true positives (TPs) are considered to be correctly diagnosed ASD patients, true negatives (TNs) are considered to be truly diagnosed TCs, false positives (FPs) and false negatives (FNs) are respectively incorrectly diagnosed ASD patients and TCs. SEN, SPE and ACC are calculated as follows,

ACC=TP+TNTP+FP+TN+FN (5)
SEN=TPFN+TP (6)
SPE=TNFP+TN (7)
FPR=FPFP+TN (8)
FNR=FNFN+TP (9)
F1=2Precision*SENPrecision+SEN, (10)

where Precison = TP/(TP + FP). Precision focuses on the model’s accuracy in predicting positive examples. In order to balance the precision and SEN, the F1 score are used to comprehensively evaluate model performance. The ACC shows the model’s ability to predict correctly, while SEN and SPE tell the model’s ability to identify subjects and negative examples respectively. The FPR refers to the rate at which the model incorrectly classifies actual negative examples as positive examples. The FNR refers to the rate at which the model incorrectly classifies actual positive examples as negative examples. After ten-fold cross-validation, our framework achieves 94.30% ACC, 92.87% SEN, 95.47% SPE, 7.12%FPR, 4.52% FNR and 94.73% F1. Fig 5 illustrates the performance of different classifiers.

Fig 5. The performance of different classifiers.

Fig 5

Since the classification performance of the model can be affected by gender and age, we also analyze these two variables according to the results, as shown in Figs 6 and 7.

Fig 6. The performance of the CNN-SVM model on different gender.

Fig 6

Fig 7. The performance of the CNN-SVM model on different age range.

Fig 7

From Fig 6, we can see that the overall performance of women is better than men when the model makes predictions. The dataset used in this article involves 175 women and 646 men. This is because the prevalence rate in men is four times that of women, which may lead to an imbalance in the collected data. Thus, further confirmation of whether the model is better at predicting female subjects than men requires collecting more data. Overall, the performace of our model does not differ significantly across gender.

From Fig 7, we can see that the model’s prediction results are pretty good in multiple age groups, and its performance is only slightly worse for subjects aged 20-30. This means that our model is less affected by age.

At the same time, we also compare the prediction results of the model on data from different sites. As shown in Fig 8, we find that the model perform relatively well at most sites, but the performance is not very satisfactory at only two sites, ETH and IU. It shows that our model has a certain degree of robustness when processing data from different sites.

Fig 8. The performance of the CNN-SVM model on data from different sites.

Fig 8

To measure the performance of our model, we compare it with some models proposed in previous papers, as shown in Table 2.

Table 2. Performance comparison of multiple papers based on ABIDE dataset.

DNN: Deep neural network; DBN: Deep Belief Network; CNNG:convolutional neural network and gate recurrent unit; BNC-DGHL:a brain network classification method based on deep graph hashing learning; C-GAN:Conditional Generative Adversarial Network; A-GCL: an adversarial self-supervised graph neural network based on graph contrastive learning; AWSO: the Adam war strategy optimization.

Paper Feature Model ACC SEN SPE
[39] static FC DNN 93.17% 91.21% 94.93%
[40] static FC DBN 76.40% 76.20% 73.10%
[41] fMRI images CNN 71.81% 81.25% 68.75%
[16] fMRI images CNNG 72.46% 71.35% 79.25%
[42] static FC BNC-DGHL 77.1% 77.8% 75.8%
[43] static FC MLP 74.82% 67.33% 80.75%
[44] texture feature, dynamic FC C-GAN 74.00% 79.00% 74.00%
[45] ALFF, static FC A-GCL 80.65% - 82.28%
[46] fMRI images AWSO-DBN 92.40% 93.00% 93.50%
this paper SRS score, static FC, dynamic FC CNN-SVM 94.30% 92.87% 95.47%

0.6 Analysis

Previous research has revealed differences in the complex relationships between TCs and subjects with ASD, and that these differences in FC exist in multiple brain regions. To identify which brain regions may have undergone alterations, we leverage the convolutional layer weights in our selected model. This allows us to ascertain which interactions between brain areas have more pronounced impact on classification, thereby suggesting that these areas have important differences between autistic and normal individuals. Examining the network weight of the first layer of the static FC part and the network weight of the second layer of the dynamic FC part, we find that the corresponding weight is a matrix of 32 × 116. By taking the absolute value of the weight of the channel and summing it, we can obtain a matrix of 1 × 116. we then select the brain areas corresponding to the top 10 features with the largest absolute values as the most discriminative brain areas, as shown in Table 3 and Fig 9 with the help of BrainNet Viewer [47].

Table 3. The discriminating brain areas of static FC and dynamic FC.

Order Static FC Dynamic FC
1 SupraMarginal_L Cerebellum_Crus2_R
2 Cingulum_Post_L Thalamus_R
3 Cerebelum_Crus1_R Cerebellum_3_R
4 Heschl_R Heschl_R
5 Occipital_Inf_R Temporal_Mid_L
6 Cerebelum_7b_L Frontal_Inf_Oper_R
7 Cerebelum_9_R Precentral_R
8 Cerebelum_3_R Vermis_6
9 Frontal_Sup_Orb_R Frontal_Sup_Orb_R
10 Temporal_Inf_R Angular_R

Fig 9. The most discriminating brain areas related to ASD.

Fig 9

As shown in Table 3, we list the top ten regions that contribute most to the classification in static FC and dynamic FC respectively. Among them, the heschl and superior frontal gyrus areas both play an important role in the classification based on two types of FC. We also find that both have a certain proportion in the cerebellum region [48, 49]. Previous classification models or physiological studies also have some similar conclusions that differences exist in these brain areas [5053]. At the same time, based on the weights coming from the first convolution layer of dynamic FC part, we can also get the smaller frequency bands in which there may be greater differences in the brain activity between subjects with ASD and TCs. After analyzing the weights, we find that the frequency bands between 0.04231Hz-0.04769Hz and 0.06385Hz-0.067436Hz contribute more to classification. Perhaps there are more differences in brain activity between subjects with ASD and TCs in these frequency bands, which requires further physiological experimental confirmation.

Furthermore, in order to elucidate the contributions of SRS scores, static FC and dynamic FC to the model’s predictions, we employ the concept of Shapley value [54]. The Shapley value, derived from game theory, is utilized in allocating gains. In the realm of machine learning, Shapley values serve to expound the extent of influence each feature exerts on prediction outcomes within the model. For a given feature set S, the formula for calculating Shapley values is expressed as

ϕS(x)=TS\i|T|!(|S|-|T|-1)!|S|![f(xTi)-f(xT)], (11)

where i represents the feature to be calculated the Shapley value, T denotes the subset of S except i, xT signifies the input containing only the features from T, and xTi refers to the input containing features from both T and i. The meaning of this formula is that for any feature i, when it is added to a feature set S, its contribution to the final prediction result is calculated by combining i with other features in S.

In our specific case, for the purpose of comparing the contributions of the three types of inputs, we take the absolute values of the Shapley values computed for each subcategory of features and sum them. We have randomly selected several subjects, and the impact of SRS scores, static FC and dynamic FC on the output is illustrated in the Fig 10. As shown in Fig 10, the prediction of our model is not heavily dependent on one type of inputs.

Fig 10. The Shapley value corresponds to SRS scores, static FC and dynamic FC from random selected subjects.

Fig 10

Discussion

In pursuit of delineating the physiological disparities between individuals with ASD and TCs, we deploy a synergistic approach that melds both static FC and dynamic FC, defining specialized convolutional kernels for feature extraction from FC matrices. This method facilitates the identification of critical brain regions that are particularly contributory to distinguishing between the two groups–some of which have been corroborated by previous studies. Moreover, we discern two brain regions, the heschl and superior frontal gyrus, that contribute substantially to both static FC and dynamic FC for classifications. Additionally, we pinpoint narrower frequency bands where dynamic FC differences may exist. To further augment the diagnostic capabilities for autism, we incorporate SRS, combining them with FC features to auxiliary diagnose ASD effectively. In this way, our model achieves better classification results compared to some previous papers.

0.7 Limitation

Meanwhile, our model also has certain limitations. When applying this model to a new dataset, it can be difficult to obtain social measures. When applying this model to other mental illnesses, such as depression, there are different perspectives and understandings when using related measures, which requires corresponding adjustments when applying the model.

Besides, autism has different severity levels, and these different levels of severity may exert an influence on categorization. Regrettably, our model does not reflect this well. Furthermore, our proposed model cannot infer the severity of the participants. In clinical practice, a comprehensive assessment, incorporating tools such as the diagnostic and statistical manual of mental disorders and clinical observations, is employed to definitively ascertain the severity of autism. Regrettably, our model does not have the ability to do this.

0.8 Future prospects

For future research, the integration of more diverse data sources for classification could be explored–for instance, by amalgamating various brain templates or utilizing time-frequency analysis methods to obtain additional FC information. Facial information of subjects can also be integrated for autism diagnosis [11]. Another way is merging the text information obtained from communicating with patients and performing natural language processing to assist in screening. The introduction of these information not only advances the prospects of AI in medical diagnostics but also minimizes the likelihood of false positives or identifying those who pretend to be patients. In addition, the proposed model holds potential for aiding the diagnosis of other psychological conditions, such as depression.

Besides, although we use Shapley value as a visual display of model interpretability, there is no similar paradigm or formula that tells us why the model makes such a decision. It does not tell us how the model relates deep features to the inputs. The clinical interpretability of the model can be figured out. In our research, we find that certain brain regions make a greater contribution to the diagnosis of autism. We also find that some previous articles also find differences in these regions. In clinical practice, further analysis may be possible, such as calculating the FC of these brain areas, and integrating the FC of multiple brain areas to obtain a threshold for distinguishing subjects with ASD from TCs. This requires more experimental subjects to obtain a more generalizable result.

Finally, since ASD is a developmental disorder characterized by potential fluctuations in behavior and brain function over time, longitudinal data could potentially offer additional insights into how ASD evolves over time. Subsequent scholars may consider seeking participants for months or even years of follow-up studies, meticulously documenting changes in their social behaviors and MRI data, thereby facilitating further study.

Conclusion

In summary, this article offers a valuable insights into autism prediction. The proposed model transcends the confines of a singular data type, presenting a hybrid CNN-SVM network model that combines ASD early screening tools with resting-state brain fMRI data. This model demonstrates good performance in classifying data across different genders, age groups and sites. Moreover, using FC, the model not only identifies brain regions with significant influence on classification outcomes but also elucidates the frequency bands that impact classification. These findings may offer clues to the etiological mechanisms and the determination of biological markers for autism. This research extends the comprehension of autism diagnosis models, encouraging comprehensive predictive analysis through the integration of multifaceted information.

In conclusion, this study’s significance lies in the introduction of a hybrid CNN-SVM model that integrates social metrics and fMRI data. It charts a promising method for future research, which is integrating diverse sources of information for prediction.

Data Availability

The datasets we used are common public datasets from URL: http://fcon_1000.projects.nitrc.org/indi/abide/. The codes associated with this study available at: https://github.com/jackykyu/A-hybrid-CNN-SVM-model-for-enhanced-autism-diagnosis.git.

Funding Statement

This work was supported by the National Natural Science Foundation of China under Grant numbers 11671354. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Zeidan J, Fombonne E, Scorah J, Ibrahim A, Durkin M, Saxena S, et al. Global prevalence of autism: A systematic review update. Autism Research. 2022;15(5):778–790. doi: 10.1002/aur.2696 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Saleh A, Sukaik R, Abu-Naser SS. Brain Tumor Classification Using Deep Learning. In: 2020 International Conference on Assistive and Rehabilitation Technologies (iCareTech); 2020. p. 131–136.
  • 3. Qureshi SA, Hussain L, Ibrar U, Alabdulkreem E, Nour MK, Alqahtani MS, et al. Radiogenomic classification for MGMT promoter methylation status using multi-omics fused feature space for least invasive diagnosis through mpMRI scans. Scientific reports. 2023;13(1):3291. doi: 10.1038/s41598-023-30309-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Hu Z, Wang J, Zhang C, Luo Z, Luo X, Xiao L, et al. Uncertainty Modeling for Multicenter Autism Spectrum Disorder Classification Using Takagi–Sugeno–Kang Fuzzy Systems. IEEE Transactions on Cognitive and Developmental Systems. 2022;14(2):730–739. doi: 10.1109/TCDS.2021.3073368 [DOI] [Google Scholar]
  • 5. Chola Raja K, Kannimuthu S. Deep learning-based feature selection and prediction system for autism spectrum disorder using a hybrid meta-heuristics approach. Journal of Intelligent & Fuzzy Systems. 2023; p. 797–807. doi: 10.3233/JIFS-223694 [DOI] [Google Scholar]
  • 6. Niu M, Tao J, Liu B, Huang J, Lian Z. Multimodal spatiotemporal representation for automatic depression level detection. IEEE transactions on affective computing. 2020;. [Google Scholar]
  • 7. Maqsood S, Damaševičius R, Maskeliūnas R. Multi-modal brain tumor detection using deep neural network and multiclass SVM. Medicina. 2022;58(8):1090. doi: 10.3390/medicina58081090 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Grampurohit S, Shalavadi V, Dhotargavi VR, Kudari M, Jolad S. Brain tumor detection using deep learning models. In: 2020 IEEE India Council International Subsections Conference (INDISCON). IEEE; 2020. p. 129–134.
  • 9. Li Z, et al. Vision transformer-based weakly supervised histopathological image analysis of primary brain tumors. iScience 26, 1, 105872 (2023). doi: 10.1016/j.isci.2022.105872 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Qureshi SA, Raza SEA, Hussain L, Malibari AA, Nour MK, Rehman Au, et al. Intelligent ultra-light deep learning model for multi-class brain tumor detection. Applied Sciences. 2022;12(8):3715. doi: 10.3390/app12083715 [DOI] [Google Scholar]
  • 11. Elshoky BRG, Younis EM, Ali AA, Ibrahim OAS. Comparing automated and non-automated machine learning for autism spectrum disorders classification using facial images. ETRI Journal. 2022;44(4):613–623. doi: 10.4218/etrij.2021-0097 [DOI] [Google Scholar]
  • 12.Rajagopalan SS. Computational behaviour modelling for autism diagnosis. In: Proceedings of the 15th ACM on International conference on multimodal interaction; 2013. p. 361–364.
  • 13. Wei Q, Cao H, Shi Y, Xu X, Li T. Machine learning based on eye-tracking data to identify Autism Spectrum Disorder: A systematic review and meta-analysis. Journal of Biomedical Informatics. 2022; p. 104254. [DOI] [PubMed] [Google Scholar]
  • 14. Mier W, Mier D. Advantages in functional imaging of the brain. Frontiers in Human Neuroscience. 2015;9. doi: 10.3389/fnhum.2015.00249 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Daliri mr, Behroozi M. Advantages and Disadvantages of Resting State Functional Connectivity Magnetic Resonance Imaging for Clinical Applications. OMICS Journal of Radiology. 2014;3. doi: 10.4172/2167-7964.1000e123 [DOI] [Google Scholar]
  • 16. Jiang X, Yan J, Zhao Y, Jiang M, Chen Y, Zhou J, et al. Characterizing functional brain networks via spatio-temporal attention 4D convolutional neural networks (STA-4DCNNs). Neural Networks. 2023;158:99–110. doi: 10.1016/j.neunet.2022.11.004 [DOI] [PubMed] [Google Scholar]
  • 17. Hutchison RM, Womelsdorf T, Allen EA, Bandettini PA, Calhoun VD, Corbetta M, et al. Dynamic functional connectivity: promise, issues, and interpretations. Neuroimage. 2013;80:360–378. doi: 10.1016/j.neuroimage.2013.05.079 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Menon SS, Krishnamurthy K. A comparison of static and dynamic functional connectivities for identifying subjects and biological sex using intrinsic individual brain connectivity. Scientific reports. 2019;9(1):5729. doi: 10.1038/s41598-019-42090-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Patil AU, Ghate S, Madathil D, Tzeng OJ, Huang HW, Huang CM. Static and dynamic functional connectivity supports the configuration of brain networks associated with creative cognition. Scientific reports. 2021;11(1):165. doi: 10.1038/s41598-020-80293-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20. Volkmar FR. Encyclopedia of autism spectrum disorders. Springer; 2021. [Google Scholar]
  • 21. Lyall K, Rando J, Toroni B, Ezeh T, Constantino JN, Croen LA, et al. Examining shortened versions of the Social Responsiveness Scale for use in autism spectrum disorder prediction and as a quantitative trait measure: Results from a validation study of 3–5 year old children. JCPP advances. 2022;2(4):e12106. doi: 10.1002/jcv2.12106 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Borges L, Otoni F, Lima THd, Schelini PW. Social Responsibility Scale (SRS-2): Validity Evidence Based on Internal Structure. Psicologia: Teoria e Pesquisa. 2023;39:11. [Google Scholar]
  • 23. Kovacs Balint Z, Raper J, Michopoulos V, Howell LH, Gunter C, Bachevalier J, et al. Validation of the Social Responsiveness Scale (SRS) to screen for atypical social behaviors in juvenile macaques. PLOS ONE. 2021;16(5):1–19. doi: 10.1371/journal.pone.0235946 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Eslami T, Almuqhim F, Raiker JS, Saeed F. Machine learning methods for diagnosing autism spectrum disorder and attention-deficit/hyperactivity disorder using functional and structural MRI: A survey. Frontiers in neuroinformatics. 2021; p. 62. doi: 10.3389/fninf.2020.575999 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Bahathiq RA, Banjar H, Bamaga AK, Jarraya SK. Machine learning for autism spectrum disorder diagnosis using structural magnetic resonance imaging: Promising but challenging. Frontiers in Neuroinformatics. 2022;16:949926. doi: 10.3389/fninf.2022.949926 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.MartinD23. Pixabay Creatures https://pixabay.com/illustrations/clipboard-checklist-business-list-2537569/. Content License: https://pixabay.com/service/terms/
  • 27.toubibe. Pixabay Creatures https://pixabay.com/illustrations/mri-magnetic-resonance-roentgen-782457/. Content License: https://pixabay.com/service/terms/
  • 28. Craddock C, Benhajali Y, Chu C, Chouinard F, Evans A, Jakab A, et al. The neuro bureau preprocessing initiative: open sharing of preprocessed neuroimaging data and derivatives. Frontiers in Neuroinformatics. 2013;7(27):5.23658544 [Google Scholar]
  • 29. Biswal B, Zerrin Yetkin F, Haughton VM, Hyde JS. Functional connectivity in the motor cortex of resting human brain using echo-planar MRI. Magnetic resonance in medicine. 1995;34(4):537–541. doi: 10.1002/mrm.1910340409 [DOI] [PubMed] [Google Scholar]
  • 30. Yan CG, Wang XD, Zuo XN, Zang YF. DPABI: data processing & analysis for (resting-state) brain imaging. Neuroinformatics. 2016;14:339–351. doi: 10.1007/s12021-016-9299-4 [DOI] [PubMed] [Google Scholar]
  • 31. Huang W, Goldsberry L, Wymbs NF, Grafton ST, Bassett DS, Ribeiro A. Graph Frequency Analysis of Brain Signals. IEEE Journal of Selected Topics in Signal Processing. 2016;10(7):1189–1203. doi: 10.1109/JSTSP.2016.2600859 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Park YH, Cha J, Bourakova V, Lee JM. Frequency specific contribution of intrinsic connectivity networks to the integration in brain networks. Scientific reports. 2019;9(1):4072. doi: 10.1038/s41598-019-40699-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Grinsted A, Moore JC, Jevrejeva S. Application of the cross wavelet transform and wavelet coherence to geophysical time series. Nonlinear processes in geophysics. 2004;11(5/6):561–566. doi: 10.5194/npg-11-561-2004 [DOI] [Google Scholar]
  • 34. Torrence C, Compo GP. A practical guide to wavelet analysis. Bulletin of the American Meteorological society. 1998;79(1):61–78. doi: [DOI] [Google Scholar]
  • 35. Daffertshofer A, Lamoth CJ, Meijer OG, Beek PJ. PCA in studying coordination and variability: a tutorial. Clinical biomechanics. 2004;19(4):415–428. doi: 10.1016/j.clinbiomech.2004.01.005 [DOI] [PubMed] [Google Scholar]
  • 36.Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:14091556. 2014;.
  • 37.Howard A, Sandler M, Chu G, Chen LC, Chen B, Tan M, et al. Searching for mobilenetv3. In: Proceedings of the IEEE/CVF international conference on computer vision; 2019. p. 1314–1324.
  • 38. Niu Z, Zhong G, Yu H. A review on the attention mechanism of deep learning. Neurocomputing. 2021;452:48–62. doi: 10.1016/j.neucom.2021.03.091 [DOI] [Google Scholar]
  • 39. Wang C, Xiao Z, Wang B, Wu J. Identification of Autism Based on SVM-RFE and Stacked Sparse Auto-Encoder. IEEE Access. 2019;7:118030–118036. doi: 10.1109/ACCESS.2019.2936639 [DOI] [Google Scholar]
  • 40. Huang ZA, Zhu Z, Yau CH, Tan KC. Identifying autism spectrum disorder from resting-state fMRI using deep belief network. IEEE Transactions on neural networks and learning systems. 2020;32(7):2847–2861. doi: 10.1109/TNNLS.2020.3007943 [DOI] [PubMed] [Google Scholar]
  • 41. Gao J, Chen M, Li Y, Gao Y, Li Y, Cai S, et al. Multisite autism spectrum disorder classification using convolutional neural network classifier and individual morphological brain networks. Frontiers in Neuroscience. 2021;14:629630. doi: 10.3389/fnins.2020.629630 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Ji J, Zhang Y. Functional Brain Network Classification Based on Deep Graph Hashing Learning. IEEE Transactions on Medical Imaging. 2022;41(10):2891–2902. doi: 10.1109/TMI.2022.3173428 [DOI] [PubMed] [Google Scholar]
  • 43.Prasad PKC, Khare Y, Dadi K, Vinod PK, Surampudi BR. Deep Learning Approach for Classification and Interpretation of Autism Spectrum Disorder. In: 2022 International Joint Conference on Neural Networks (IJCNN); 2022. p. 1–8.
  • 44. Raja KC, Kannimuthu S. Conditional Generative Adversarial Network Approach for Autism Prediction. Computer Systems Science & Engineering. 2023;44(1). [Google Scholar]
  • 45. Zhang S, Chen X, Shen X, Ren B, Yu Z, Yang H, et al. A-GCL: Adversarial graph contrastive learning for fMRI analysis to diagnose neurodevelopmental disorders. Medical Image Analysis. 2023;90:102932. doi: 10.1016/j.media.2023.102932 [DOI] [PubMed] [Google Scholar]
  • 46. Bhandage V, K MR, Muppidi S, Maram B. Autism spectrum disorder classification using Adam war strategy optimization enabled deep belief network. Biomedical Signal Processing and Control. 2023;86:104914. doi: 10.1016/j.bspc.2023.104914 [DOI] [Google Scholar]
  • 47. Xia M, Wang J, He Y. BrainNet Viewer: A Network Visualization Tool for Human Brain Connectomics. PLOS ONE. 2013;8(7):1–15. doi: 10.1371/journal.pone.0068910 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48. Sydnor LM, Aldinger KA. Structure, function, and genetics of the cerebellum in autism. Journal of psychiatry and brain science. 2022;7. doi: 10.20900/jpbs.20220008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49. Laidi C, Neu N, Watilliaux A, Martinez-Teruel A, Razafinimanana M, Boisgontier J, et al. Preserved navigation abilities and spatio-temporal memory in individuals with autism spectrum disorder. Autism Research. 2023;16(2):280–293. doi: 10.1002/aur.2865 [DOI] [PubMed] [Google Scholar]
  • 50. Prigge MD, Bigler ED, Fletcher PT, Zielinski BA, Ravichandran C, Anderson J, et al. Longitudinal Heschl’s Gyrus Growth During Childhood and Adolescence in Typical Development and Autism. Autism Research. 2013;6(2):78–90. doi: 10.1002/aur.1265 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51. Patriquin MA, DeRamus T, Libero LE, Laird A, Kana RK. Neuroanatomical and neurofunctional markers of social cognition in autism spectrum disorder. Human brain mapping. 2016;37(11):3957–3978. doi: 10.1002/hbm.23288 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52. Arutiunian V, Gomozova M, Minnigulova A, Davydova E, Pereverzeva D, Sorokin A, et al. Structural brain abnormalities and their association with language impairment in school-aged children with Autism Spectrum Disorder. Scientific Reports. 2023;13(1):1172. doi: 10.1038/s41598-023-28463-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53. Banks MI, Krause BM, Berger DG, Campbell DI, Boes AD, Bruss JE, et al. Functional geometry of auditory cortical resting state networks derived from intracranial electrophysiology. PLOS Biology. 2023;21(8):1–44. doi: 10.1371/journal.pbio.3002239 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Shapley LS, et al. A value for n-person games. 1953;.

Decision Letter 0

Umer Asgher

23 Oct 2023

PONE-D-23-32535A hybrid CNN-SVM model for enhanced autism diagnosisPLOS ONE

Dear Dr. Qiu,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Major rework and revision is needed based on the reviewers' feedback and comments appended below.

Please submit your revised manuscript by Dec 07 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Umer Asgher, PhD

Academic Editor

PLOS ONE

Journal requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. Thank you for stating the following financial disclosure:

“This work was supported by the National Natural Science Foundation of China under Grant numbers 11671354.”

Please state what role the funders took in the study. If the funders had no role, please state: "The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript."

If this statement is not correct you must amend it as needed.

Please include this amended Role of Funder statement in your cover letter; we will change the online submission form on your behalf.

4. Thank you for stating the following in the Acknowledgments Section of your manuscript:

“This work was supported by the National Science Foundation of China under Grant numbers 11671354.”

We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form.

Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows:

“This work was supported by the National Natural Science Foundation of China under Grant numbers 11671354.”

Please include your amended statements within your cover letter; we will change the online submission form on your behalf.

5. We note that Figures 1 and 6 in your submission contain copyrighted images. All PLOS content is published under the Creative Commons Attribution License (CC BY 4.0), which means that the manuscript, images, and Supporting Information files will be freely available online, and any third party is permitted to access, download, copy, distribute, and use these materials in any way, even commercially, with proper attribution. For more information, see our copyright guidelines: http://journals.plos.org/plosone/s/licenses-and-copyright.

We require you to either (1) present written permission from the copyright holder to publish these figures specifically under the CC BY 4.0 license, or (2) remove the figures from your submission:

1. You may seek permission from the original copyright holder of Figures 1 and 6 to publish the content specifically under the CC BY 4.0 license.

We recommend that you contact the original copyright holder with the Content Permission Form (http://journals.plos.org/plosone/s/file?id=7c09/content-permission-form.pdf) and the following text:

“I request permission for the open-access journal PLOS ONE to publish XXX under the Creative Commons Attribution License (CCAL) CC BY 4.0 (http://creativecommons.org/licenses/by/4.0/). Please be aware that this license allows unrestricted use and distribution, even commercially, by third parties. Please reply and provide explicit written permission to publish XXX under a CC BY license and complete the attached form.”

Please upload the completed Content Permission Form or other proof of granted permissions as an "Other" file with your submission.

In the figure caption of the copyrighted figure, please include the following text: “Reprinted from [ref] under a CC BY license, with permission from [name of publisher], original copyright [original copyright year].”

2. If you are unable to obtain permission from the original copyright holder to publish these figures under the CC BY 4.0 license or if the copyright holder’s requirements are incompatible with the CC BY 4.0 license, please either i) remove the figure or ii) supply a replacement figure that complies with the CC BY 4.0 license. Please check copyright information on all replacement figures and update the figure caption with source information. If applicable, please specify in the figure caption text when a figure is similar but not identical to the original image and is therefore for illustrative purposes only.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Partly

Reviewer #5: Yes

Reviewer #6: No

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: N/A

Reviewer #2: N/A

Reviewer #3: Yes

Reviewer #4: No

Reviewer #5: Yes

Reviewer #6: No

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: Yes

Reviewer #4: Yes

Reviewer #5: Yes

Reviewer #6: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

Reviewer #3: No

Reviewer #4: No

Reviewer #5: Yes

Reviewer #6: No

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: I think the authors need to do some efforts for comparing his approach with recent research studies such as in:

https://onlinelibrary.wiley.com/doi/full/10.4218/etrij.2021-0097

You will find that this paper compares only one AutoML package from all available packages with Transfer Learning and Traditional Machine Learning. In addition, Linear Regression (LR) produce linear models which are similar to SVM and I am wondering why the authors did not compare the performance of the hybrid CNN-LR with CNN-SVM.

Reviewer #2: While the paper presents a novel and promising approach for diagnosing Autism Spectrum Disorder (ASD), it's important to acknowledge the limitations of the work. Some potential limitations of this research could include:

1. Keywords are missing in the abstract.

2. The sample size of 379 subjects with ASD and 442 typical controls, while substantial, may not be representative of the entire ASD population. The diversity of individuals with ASD in terms of age, gender, and severity of symptoms could impact the generalizability of the findings.

3. The accuracy and reliability of the resting-state fMRI data can greatly influence the results. Issues such as motion artifacts, data preprocessing, and the consistency of data collection across different sites could affect the quality of the data.

4. Deep learning models, including CNNs, have a risk of overfitting, especially when working with relatively small datasets. It's crucial to address how the model avoids overfitting and whether any techniques like cross-validation or regularization were used.

5. The model's performance might be optimized for the specific dataset and experimental setup used in the study. The paper should discuss the potential challenges in applying this model to new or different datasets.

6. The paper emphasizes the use of both static and dynamic functional connectivity features alongside SRS metrics. However, the paper should address whether the model's performance is heavily reliant on one type of feature and how generalizable these features are to other ASD populations.

7. While achieving a high classification accuracy is important, it's also crucial to consider the clinical interpretability of the model. Understanding the biological or clinical significance of the identified features and regions is essential for making practical use of the findings.

8. The paper should discuss the interpretability of the CNN-SVM model. Deep learning models are often seen as "black boxes," making it challenging to explain the model's decision-making process.

9. Any potential bias in the data collection, preprocessing, or in the model itself should be addressed. Additionally, the ethical considerations regarding the use of machine learning in healthcare, including issues related to privacy and consent, should be discussed.

10. ASD is a developmental disorder, and it often involves changes in behavior and brain function over time. This study seems to focus on cross-sectional data. Longitudinal data might provide additional insights into how ASD evolves over time.

11. The study should discuss the validation methods employed, such as cross-validation, and the reproducibility of the results to ensure the robustness of the model's performance.

12. The authors used some evaluation metrics but they do not give us a clear understanding of the level of false positives and false negatives directly. Therefore, recently mean Intersection over union (mIoU) has been widely used as a more reasonable and intuitive way of metrics. mIoU should be highlighted in the Abstract and conclusion.

13. The proposed work must be compared with some state-of-the-art works from 2022/2023 for better understanding the capabilities of the proposed work.

Reviewer #3: 1. Typing and grammatical errors are highlighted in the attached file

2. Unify all over the text Figure or Fig.

• 70 Figure 1.

• 175 Fig 2.

• 237 Fig 4 illustrates the performance of three different

3. Figure caption is not a sentence:

• Fig. 5 This figure compares…..

A possible caption can be:

"Performance of the proposed model compared with related work

(Wang [22]; Huang [23]; Yin [24]; Jiang [25]; and Bhandage [26])"

• Fig 6. This figure shows the most discriminating brain areas.

A possible caption can be:

"The most discriminating brain areas related to ASD"

• Punctuation ruled should be respected.

Reviewer #4: Author presented a hybrid approach for autism spectrum disorder diagnosis. They made use of CNN and SVM based model to achieve this. But, there are several drawbacks in the article that should be resolved.

1. Abstract is not clear. It should express research backround , problem statement, methods and results.

2. Introduction section failed to mention the challenges of existing system, major contribution and structure of the manuscript.

3. Literature review is poor. Several works should be studied such as Deep Learning for Autism Diagnosis and Facial Analysis in Children, Conditional Generative Adversarial Network Approach for Autism Prediction, Deep learning-based feature selection and prediction system for autism spectrum disorder using a hybrid meta-heuristics approach, Machine learning for autism spectrum disorder diagnosis using structural magnetic resonance imaging: Promising but challenging.

4. Proposed model can be expressed as an algorithm. Architecture of hybrid model can be represented.

5. What is the need for devising hybrid model?. Mention the significance of incorporating SVM with CNN model.

6. Several dataset can be used for experimentation. Hyper parameter tuning needs to be explained.

7. Discussion section can be included to state the challenges, inferences and future prospects.

8. Conclusion section can be improved.

Reviewer #5: The contributions is clear. However, authors should justify the decisions that made in the article such as the selected methods and parameters settings. The authors should state the parctical implications of the work.

Reviewer #6: The following points need to be addressed in the revision:

1. The abstract is adequate in length and structure.

2. The indication of the dataset in the abstract must be ensured.

3. Comparison with SOTA techniques should be mentioned in the abstract

4. The literature review of the study is insufficient, there are many important studies that are missing. The authors should include the following references with proper discussions:

https://doi.org/10.1109/iCareTech49914.2020.00032

https://doi.org/10.3390/app12083715

https://doi.org/10.3390/medicina58081090

https://doi.org/10.1038/s41598-023-30309-4

https://doi.org/10.1088/1742-6596/1916/1/012226

https://doi.org/10.1088/1757-899X/1055/1/012115

https://doi.org/10.1109/INDISCON50162.2020.00037

https://doi.org/10.1016/j.fcij.2017.12.001

https://doi.org/10.1016/j.cmpb.2019.05.015

5. Please add contributions of your work in bulleted form preferably in the introduction section

6. The organization of the article is missing.

7. There is irregular text in the article Lines: 62-63. Ensure a consistent heading number scheme. Correct it.

8. How you are claiming CNN static, rather the converse is true.

9. Defining static FC, you cannot change the working principle of CNN. Changing the name of the remarkable technique is no good. Please justify your approach.

10. Under the heading of Neural Network, do not use CNN. One is shallow and the other is deep.

11. Table 1 is without any whereabouts. The column headings, if possible, may be provided.

12. The distribution of instances in the dataset should be graphically represented.

13. The figures related to CNN are not clearly represented according to the norms of deep learning. No figure number is there. Figures are contradictory and do not truly represent the working principle of algorithms.

14. Bhandage has almost the same results as yours. Justify that the proposed solution would always be guaranteed through some statistical analysis.

15. Limitations of the proposed work with future recommendations should be added before the conclusion.

16. The Conclusion needs rephrasing, as it is not encircling the proposed work being asked for publication.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Reviewer #3: Yes: Said Ghoniemy; AinShams University

Reviewer #4: Yes: No

Reviewer #5: No

Reviewer #6: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: PONE-D-23-32535_reviewer_commented.pdf

pone.0302236.s001.pdf (997.8KB, pdf)
PLoS One. 2024 May 14;19(5):e0302236. doi: 10.1371/journal.pone.0302236.r002

Author response to Decision Letter 0


4 Jan 2024

Dear Editor,

Thank you for your giving us an opportunity to revise our manuscript! Now we are submitting the revised manuscript entitled "A hybrid CNN-SVM model for enhanced autism diagnosis" for consideration for publication in PLOS ONE.

In the revision, we have performed nearly all the suggested opinions, experiments and analyses, and fully addressed the comments made by the reviewers and editor. The finished new experiments and data analyses as following:

1. We have revised the Abstract and Results sections to meet the requirements required for publication. And we add more related works in the Introduction section to review the application of machine learning in medical aspects, especially in the diagnosis of autism.

2. We have added the contributions of this work and the organization of this manuscript. We also have described the flowchart in algorithmic form to enhance the comprehension of our model.

3. In order to better evaluate the proposed model, we have introduced more indicators to evaluate the performance of different classifiers in autism diagnosis, as shown in the Fig 5.

4. We have analyzed the performance of the proposed model on different genders, sites, and age groups, and the results are shown in Figs 6-8.

5. We also used the Shapley value to measure the impact of three types of inputs on model predictions, giving interpretability to a certain extent.

6. We have added Limitation and Future prospects to the discussion section for better summary.

All authors have read and approved the re-submission of the manuscript! If you have any questions, please let me know!

Thank you for your consideration of our paper and we are looking forward to hearing from you!

Sincerely yours,

Linjie, Ph.D.,

School of Mathematical Sciences, Zhejiang University, Hangzhou, Zhejiang, China

Email:11935034@zju.edu.cn

Attachment

Submitted filename: Response to Reviewers.pdf

pone.0302236.s002.pdf (201.7KB, pdf)

Decision Letter 1

Umer Asgher

12 Feb 2024

PONE-D-23-32535R1A hybrid CNN-SVM model for enhanced autism diagnosisPLOS ONE

Dear Dr. Qiu,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Specifically; Two reviewers have accepted the manuscript after re-evaluation and One reviewer has give minor reviews (Feedback),. So please improve the manuscript and re-submit as soon as possible in the light of reviewers comments. 

Please submit your revised manuscript by Mar 28 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Umer Asgher, PhD

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #3: All comments have been addressed

Reviewer #4: All comments have been addressed

Reviewer #6: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #3: Yes

Reviewer #4: Yes

Reviewer #6: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #3: Yes

Reviewer #4: Yes

Reviewer #6: No

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #3: Yes

Reviewer #4: Yes

Reviewer #6: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #3: Yes

Reviewer #4: Yes

Reviewer #6: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #3: All comments were revised and the answers are acceptable. The paper is much enhanced ans is suitable for publication

Reviewer #4: After evaluating the article, I have observed significant modifications. All the comments were addressed. Hence, this article can be considered for publication.

Reviewer #6: All the points have been thoroughly addressed.

The following points are suggested as minor revisions:

1. In line number 80 (page 37), replace “…SRS scores, which reflect the…” with “….SRS scores reflecting the…”.

2. The features that we get from CNN are called dynamic features. In your work, this seems ambiguous, and one way to remove it is to change “static feature” in figure 1 (and throughout the article) to “staticFC features”, and “dynamic feature” to “dynamicFC features”

3. Remove the text lines in the conclusions “Nevertheless, the article harbors limitations, such as its inability…… diagnosis of autism.” You have already mentioned this text in the limitations and future recommendations.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #3: Yes: Said Ghoniemy, Ain Shan\\ms University, Cairo, Egypt

Reviewer #4: Yes: KANNIMUTHU SUBRAMANIAN

Reviewer #6: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 May 14;19(5):e0302236. doi: 10.1371/journal.pone.0302236.r004

Author response to Decision Letter 1


28 Feb 2024

Dear Editor,

Thank you for your giving us an opportunity to revise our manuscript! Now we are submitting the revised manuscript entitled ”A hybrid CNN-SVM model for enhanced autism diagnosis” for consideration for publication in PLOS ONE.

In the revision, we have performed nearly all the suggested opinions and fully addressed the comments made by the reviewers and editor. We summarized the modifications as follows:

1. We have checked all references and can ensure that none of them have been withdrawn.

2. We have adjusted the images using the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool to ensure they meet PLOS requirements.

3. We have made appropriate revisions to the content based on the reviewers’ comments.

All authors have read and approved the re-submission of the manuscript! If you have any

questions, please let me know!

Thank you for your consideration of our paper and we are looking forward to hearing from

you!

Sincerely yours,

Linjie, Ph.D.,

School of Mathematical Sciences, Zhejiang University, Hangzhou, Zhejiang, China

Email:11935034@zju.edu.cn

Attachment

Submitted filename: Response to Reviewers.pdf

pone.0302236.s003.pdf (63.5KB, pdf)

Decision Letter 2

Umer Asgher

1 Apr 2024

A hybrid CNN-SVM model for enhanced autism diagnosis

PONE-D-23-32535R2

Dear Dr. Qiu,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. There are minor issues to be addressed in the light of reviewers comments.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Umer Asgher, PhD

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: PONE-D-23-32535_reviewer_commented.pdf

    pone.0302236.s001.pdf (997.8KB, pdf)
    Attachment

    Submitted filename: Response to Reviewers.pdf

    pone.0302236.s002.pdf (201.7KB, pdf)
    Attachment

    Submitted filename: Response to Reviewers.pdf

    pone.0302236.s003.pdf (63.5KB, pdf)

    Data Availability Statement

    The datasets we used are common public datasets from URL: http://fcon_1000.projects.nitrc.org/indi/abide/. The codes associated with this study available at: https://github.com/jackykyu/A-hybrid-CNN-SVM-model-for-enhanced-autism-diagnosis.git.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES