Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2024 Oct 17;19(10):e0305630. doi: 10.1371/journal.pone.0305630

Multiclass classification of Autism Spectrum Disorder, attention deficit hyperactivity disorder, and typically developed individuals using fMRI functional connectivity analysis

Caroline L Alves 1,*, Tiago Martinelli 2, Loriz Francisco Sallum 2, Francisco Aparecido Rodrigues 2, Thaise G L de O Toutain 3, Joel Augusto Moura Porto 4,5, Christiane Thielemann 6, Patrícia Maria de Carvalho Aguiar 7,8, Michael Moeckel 1
Editor: Yangsong Zhang9
PMCID: PMC11486369  PMID: 39418298

Abstract

Neurodevelopmental conditions, such as Autism Spectrum Disorder (ASD) and Attention Deficit Hyperactivity Disorder (ADHD), present unique challenges due to overlapping symptoms, making an accurate diagnosis and targeted intervention difficult. Our study employs advanced machine learning techniques to analyze functional magnetic resonance imaging (fMRI) data from individuals with ASD, ADHD, and typically developed (TD) controls, totaling 120 subjects in the study. Leveraging multiclass classification (ML) algorithms, we achieve superior accuracy in distinguishing between ASD, ADHD, and TD groups, surpassing existing benchmarks with an area under the ROC curve near 98%. Our analysis reveals distinct neural signatures associated with ASD and ADHD: individuals with ADHD exhibit altered connectivity patterns of regions involved in attention and impulse control, whereas those with ASD show disruptions in brain regions critical for social and cognitive functions. The observed connectivity patterns, on which the ML classification rests, agree with established diagnostic approaches based on clinical symptoms. Furthermore, complex network analyses highlight differences in brain network integration and segregation among the three groups. Our findings pave the way for refined, ML-enhanced diagnostics in accordance with established practices, offering a promising avenue for developing trustworthy clinical decision-support systems.

1 Introduction

1.1 Clinical background

Neurodevelopmental disorders encompass a spectrum of conditions that manifest early in life and have diverse impacts on brain development and function, often presenting with genetic and clinical heterogeneity [1]. These disorders profoundly affect neurological functioning, including cognition, communication, behavior, motor skills, and social interaction [24].

Two prominent examples of neurodevelopmental disorders are Autism Spectrum Disorder (ASD) and Attention Deficit Hyperactivity Disorder (ADHD). ASD is characterized by challenges in social interaction, communication, repetitive behaviors, and sensory sensitivities [5]. Globally, ASD affects approximately 1 in 36 children and is more prevalent in males than females [6]. ASD is a spectrum disorder, displaying a wide range of symptom severity and presentation, making diagnosis a challenging task [79]. Furthermore, ASD is marked by significant heterogeneity, with no discernible patterns consistently emerging among affected individuals [10].

ADHD, another prevalent neurodevelopmental disorder, is defined by inattention, hyperactivity, and impulsivity symptoms, which can substantially impact daily functioning [11]. It affects 5–8% of children, with a higher prevalence among boys [12]. Despite extensive research, the prevalence of ADHD remains elusive [13], and diagnosis primarily relies on the assessment of behavioral symptoms [11].

While ASD and ADHD are traditionally classified as distinct neurodevelopmental disorders, they exhibit a significant degree of symptom overlap [14]. This shared symptomatology often complicates the accurate diagnosis and treatment planning for affected individuals. Furthermore, it is noteworthy that ADHD frequently co-occurs with ASD, making it one of the most prevalent comorbidities among individuals with ASD [15]. This comorbidity adds another layer of complexity to the neurodevelopmental profile of affected individuals, contributing to the challenges in diagnosis and care. Consequently, these circumstances often lead to cases of misdiagnosis and underdiagnosis.

1.2 Previous ML approaches

Given the inherent complexity of diagnosing ASD and ADHD, many studies are using machine learning methods to improve the diagnosis [16]. In the study by [17], machine learning models were trained and tested on an imbalanced dataset from research records based on the Social Responsiveness Scale (SRS). The SRS is a parent-administered questionnaire frequently employed for measuring autism traits. Notably, the superior performance algorithms were Support Vector Machines (SVM), Ridge Regression, Least Absolute Shrinkage and Selection Operator (LASSO), and Linear Discriminant Analysis, achieving accuracies ranging from 0.962 to 0.965. These results underscore the effectiveness of machine learning techniques in accurately distinguishing individuals with ASD from those with ADHD. Moreover, in [18] it was incorporated a crowdsourced dataset comprising responses to 15 SRS-derived questions. In this subsequent analysis, Linear Discriminant Analysis (LDA) and Elastic Net (ENet) emerged as the top-performing methods, both achieving an Area Under the ROC Curve (AUC) of 0.89 when tested on survey data containing the 15 questions.

Machine methods have also been applied to neuroimaging data to develop more precise and reliable approaches for characterizing and predicting ASD and ADHD in a binary way to distinguish from TD [11, 13, 1924]. Most studies in the literature are primarily focused on distinguishing individuals with a specific condition from typically developing (TD) individuals, resulting in a binary classification problem. However, a more desirable scenario would involve the differentiation among multiple neurological conditions characterized by overlapping symptoms or co-occurrence, as seen in the case of ASD and ADHD. Preliminary efforts have been made in this direction, in which various machine learning techniques have been employed [17, 18, 25]. For instance, one of the pioneering studies to differentiate adolescents with ADHD from those with ASD or TD was published in 2013 [26]. The authors considered structural magnetic resonance imaging (MRI) data and a 3-class Gaussian Process Classification (GPC) approach to classify ADHD, ASD, and TD groups simultaneously [26]. Their model achieved a balanced accuracy of 0.682, with sensitivities of 0.759, 0.655, and 0.632 for the ADHD, TD, and ASD groups, respectively. The positive predictive values for the respective groups were also 0.629, 0.731, and 0.75.

In [27], centrality abnormalities in cortical and subcortical regions were discovered, some of them shared between ADHD and ASD. They observed increased centrality in the right striatum/pallidum for ADHD and bilateral temporolimbic areas for ASD. In [28], it was conducted a comparative analysis of the network topology patterns among ASD, ADHD, and neurotypical (NT) groups. They found substantial overlap at the global level of community structure among all groups. However, the overlap between the two clinical conditions was less than that between each condition and the control group.

Additionally, the ASD and ADHD groups exhibited a more pronounced reduction in correlation strength with increased distance compared to the NT group. Notably, the ADHD group displayed reduced wiring costs, thinner cortical regions, and lower hub degrees than the ASD group. Significant findings emerged in the study [29] regarding oscillatory patterns in children with ASD and ADHD during task conditions. Children with ASD exhibited significant hypoconnectivity in large-scale networks during these tasks, while those with ADHD showed hyperconnectivity in large-scale networks under similar conditions. In a recent study [30], an SVM algorithm with l2-regularization emerged as the top-performing method, achieving an accuracy of 0.66, an f1-score of 0.68, a precision of 0.59, and a recall of 0.82. Notably, their findings unveiled a substantial convergence in functional brain connectivity patterns between ADHD and ASD, particularly within the right ventral attention network, the salience network, and the default mode network (DMN) as observed in resting-state fMRI data.

Table 1 concisely overviews the primary research using machine learning classification methods and the ASD and ADHD groups outlined in this subsection.

Table 1. Studies with ML classification algorithms for distinguishing ASD and ADHD groups described in subsection 1.2.

Studies Dataset type ML algorithm Classification type Correlation metric AUC Accuracy Recall Precision
[17] research records survey SVM, Ridge Regression, LASSO, and LDA Binary (ASD and ADHD) - - 0.962–0.965 - -
[18] research records survey LDA and ENet Binary (ASD and ADHD) - 0.890 - - -
[25] research records survey SVM Binary (ASD and ADHD) - 0.910 - 0.680 -
[26] MRI GPC Multiclass (ASD, ADHD and TD) - - 0.682 - -
[30] fMRI SVM Multiclass (ASD, ADHD and TD) Pearson correlation - 0.66 0.82 0.59

1.3 Research gap

Previous ML models have yet to be constructed using explainable AI approaches. This limits their interpretability and, hence, the trustworthiness of their model predictions. In particular, a clear connection between established clinical symptoms and model properties has yet to be drawn, raising doubts among medical professionals and hindering the use of ML modeling in actual clinical diagnosis [31].

Upon reviewing the landscape of studies within the domain of ASD and ADHD classification, a notable trend emerges, as illustrated in the summarized literature (Table 1): a predominant focus on binary comparisons between ASD and ADHD, often reliant on survey-based symptomatic datasets, which not alllows clear separation in the cases that have overlapping of symptoms. This emphasis on a binary framework potentially oversimplifies the nuanced complexities inherent in these neurodevelopmental disorders. Moreover, the reliance on survey data in many studies raises concerns regarding introducing biases [32], particularly when compared to more direct neuroimaging modalities such as EEG and fMRI. Our previous research [33, 34] highlights the critical role of correlation metrics in constructing connectivity matrices to effectively capture brain changes associated with these conditions. While studies such as [30] use fMRI and a multiclass approach, they predominantly utilize linear measures like Pearson correlation; the robustness of alternative metrics such as normalized transfer entropy still needs to be explored. In our prior works [34, 35], we demonstrated the robustness of normalized transfer entropy in distinguishing neurodevelopmental disorders from TD individuals, underscoring the need to explore these metrics further. Furthermore, the prevailing focus on ML methodologies overlooks the exploration of complex network topology changes and needs more medical interpretation. There is a clear need for a more holistic approach that integrates diverse methodologies and prioritizes nuanced understanding over simplistic binary classifications. Such an approach holds the potential to advance our comprehension of the underlying mechanisms of ASD and ADHD and inform more effective interventions and treatments.

1.4 Objective and hypothesis

Our study aims to bridge the existing research gap by advancing beyond simplistic binary comparisons in the classification of ASD, ADHD, and TD individuals using fMRI datasets. Building upon prior work that underscores the limitations of binary frameworks and the potential bias introduced by survey-based datasets, our hypothesis posits that distinct brain activity patterns underlie these neurodevelopmental disorders and can facilitate their reliable separation. We seek to investigate whether these patterns align with existing clinical knowledge of ASD and ADHD, thereby providing deeper insights into these conditions’ underlying mechanisms and proving our methodology’s trustworthiness.

Additionally, we aim to explore the utility of complex network measures in achieving a clean separation of the groups, surpassing the conventional focus on machine learning methodologies, through the integration of diverse methodologies alongside advanced analytical techniques such as normalized transfer entropy, our study endeavors to enhance prediction accuracy while ensuring an explainable and trustworthy machine learning approach.

2 Materials and methods

2.1 Innovations in the methodology

The current paper endeavors to investigate the feasibility of automatically detecting brain changes associated with ASD and ADHD while simultaneously providing a biological rationale for these observations. We leverage the blood oxygenation level-dependent (BOLD) time series data to achieve this objective and develop a classification method to distinguish individuals with ASD, ADHD, and TD profiles. Following dataset preprocessing (described in subsection 2.2), we employed two levels of data abstraction: (A) the calculation of correlations, determined by the normalized transfer entropy between specific fMRI regions of interest (described in subsection 2.3), and (B) the extraction of complex network measures from the correlations (A) (described in subsection 2.4). It is important to note that while our methodology bears similarity to our previous work [3337], where binary classification was primarily explored, the present study aims to establish a multiclass classifier. Moreover, departing from existing literature, which often focuses on utilizing only one of these abstraction levels, our study pioneers the simultaneous use of both levels in a multiclass context, marking a novel contribution to the field.

To enhance the interpretability of our machine learning results, we incorporated cutting-edge techniques that have emerged in recent years. One such technique is the application of SHapley Additive ExPlanations (SHAP) values [38]. These values help identify the most critical features within our model, shedding light on essential brain areas and connections between regions. Notably, SHAP values have demonstrated superior performance compared to prior research efforts [33, 36, 39] in pinpointing significant brain areas and connectivity patterns and have been an integral part of our previous work.

Differently from [3337], we added three measures to analyze the segregation and integration concepts: Effective Information, determinism, and degeneracy coefficients. These measures provide a comprehensive understanding of the dynamics of brain networks in individuals with ASD, ADHD, and TD profiles. Furthermore, to the best of our knowledge, this is the first study that employed the SHAP values methodology for a multiclass classification of ASD and ADHD, enhancing the interpretability and robustness of our classification results.

The Python code with the methodology used in this work and described in this section is available at: https://github.com/Carol180619/Multiclass-ADHD-ASD.git.

2.2 Data and data preprocessing

The ADHD dataset used in this study was sourced from the Neuro Bureau ADHD-200 Preprocessed repository, as detailed in [40]. During a 6-minute resting-state fMRI scan, participants received instructions to relax, maintain closed eyes, avoid falling asleep, and refrain from engaging in specific thoughts. The resting-state fMRI data captures spontaneous fluctuations in BOLD signals, widely acknowledged as indicative of underlying brain activity. In this investigation, we utilized the ADHD-200 dataset via the Nilearn package. Nilearn is a Python package [41] tailored explicitly for analyzing neuroimaging data. Our selection of Nilearn seamlessly integrated with our existing workflow, as we had already incorporated Nilearn into our analysis pipeline. Nilearn provides a comprehensive set of tools for preprocessing, feature extraction, and statistical analysis of neuroimaging data. It was used in numerous studies [4244], making it a suitable and convenient choice for our research. Within the Nilearn package, the dataset comprises 40 child and adolescent participants aged 7 to 27, equally divided between individuals diagnosed with ADHD and healthy control subjects.

As in our previous work [34], we considered the preprocessed version of the Autism Brain Imaging Data Exchange (ABIDE) (Avaiable in https://fcon_1000.projects.nitrc.org/indi/abide/) which consists of 1112 datasets comprised of 539 ASD and 573 TD with 300s BOLD time series (7–64 years, median 14.7 years across groups). Further, it is also available for use in Nilearn’s Python package. We used the same amount of 40 subjects from the ADHD dataset. For TD groups, we randomly selected 20 subjects for the ABIDE dataset and 20 for the Neuro Bureau ADHD-200 Preprocessed repository.

In our study, rather than utilizing the entire BOLD time series obtained from each voxel in brain images, we focused on specific Brain Regions of Interest (ROIs). These ROIs are defined based on a brain atlas; only the BOLD time series from these ROIs are considered. The choice of Brain Atlas is crucial, and in our work, we employed the Bootstrap Analysis of Stable Clusters (BASC) atlas, selected for its exceptional performance in discriminating Autism Spectrum Disorder (ASD) patients using deep learning models, as reported in [45]. The BASC atlas, introduced initially in [46], is derived from group-level brain parcellation through the BASC method, an algorithm utilizing k-means clustering to identify brain networks with coherent activity in resting-state fMRI, as described in [47]. BASC map with a cluster number of 122 ROIs is used here (see Fig 1A). From our previous work [34], a manual use of Yale BioImage Suite Package web application (Avaiable in https://bioimagesuiteweb.github.io/webapp/mni2tal.html) labeled the coordinates of each ROI for the identification of their names (see Fig 1A).

Fig 1. Methodology to obtain the connectivity matrices based on [34].

Fig 1

In (A), a time series of 122 ROI is extracted from fMRI data using the BASC BOLD atlas (highlighted in blue, purple, and orange). A sliding window was performed as a data augmentation. Then, they are correlated (B) to form the connectivity matrices, where each row and column corresponds to one of the Brodmann areas for a patient with ASD, TD, and ADHD (the figure illustrates an example of a connectivity matrix with a normalized TE of a subject presenting ADHD). The same highlighted matrices represent the brain in a three-dimensional (in a top and left perspective) schematic.

After extracting the BOLD time series, a sliding time window of 20 seconds was employed for data augmentation. This duration was selected based on our previous study [34], where it demonstrated optimal performance for the ASD dataset. Additionally, to ensure comparability between the ASD and ADHD datasets, the same window size was utilized for the ADHD dataset. By employing consistent time windows across both datasets, we aimed to mitigate potential biases arising from variations in data acquisition protocols between different sites and enhance the robustness of our analyses. Through the data augmentation process, 600 matrices were randomly selected, ensuring an equal representation of each class.

Once the time series for each of the 122 regions had been extracted for each time window, they were correlated according to normalized transfer entropy (TE)(see Fig 1B). The TE is described in [48] and given by the Eq 1. The normalized TE metric was selected as our study’s primary analytical tool, building upon prior research findings [34, 35]. TE has demonstrated its efficacy in capturing changes in brain dynamics associated with neurodevelopmental disorders. This choice was motivated by several advantages it offers. Firstly, normalized TE is adept at capturing directional dependencies [49], providing a nuanced understanding of information flow within neural systems [50, 51]. Unlike linear measures, such as Spearman and Pearson correlation, TE is non-linear and model-free, making it particularly suitable for analyzing complex, non-stationary, and non-Gaussian datasets [5254]. Moreover, TE stands apart from traditional correlation measures like the Pearson correlation. While the latter is confined to assessing linear or monotonic relationships between variables, regardless of their direction of influence, TE excels in identifying and quantifying the directional dependencies between time series [55], thereby facilitating a more comprehensive assessment of neural interactions. Noteworthy, as in our previous work [34, 35], for computing TE, a Min-max normalization and then a thresholding process was performed, with a value of 0.5, since this measures deal best with binary values.

TEYX(τ)=i=1NP(xt+τi,xti,yti)logP(xt+τi|xti,yti)P(xt+τi|xti) (1)

where:

  • TEYX(τ) denotes the transfer entropy from Y to X at time lag τ.

  • xti and yti represent samples from time series X and Y respectively.

  • N is the total number of samples.

  • P(xt+τi,xti,yti) is the joint probability distribution of xt+τi, xti, and yti.

  • P(xt+τi|xti,yti) is the conditional probability distribution of xt+τi given xti and yti.

  • P(xt+τi|xti) is the conditional probability distribution of xt+τi given xti.

In addition to the aforementioned preprocessing steps, we employed the NeuroImaging Analysis Kit (NIAK) [56] to further standardize and enhance the quality of our neuroimaging data for both datasets. NIAK offers a comprehensive set of tools for preprocessing fMRI data, including motion correction, slice timing correction, spatial normalization, and nuisance signal regression [40, 57]. These preprocessing procedures are crucial for mitigating potential confounding factors introduced by differences in data acquisition protocols across multiple sites. By implementing the same preprocessing pipeline in both datasets, we aimed to minimize site-related variations and ensure the consistency and reliability of our data across different acquisition sites. This standardized approach facilitated the integration of neuroimaging data from disparate sources, enhancing the validity and generalizability of our findings.

2.3 Connectivity matrices

Similar to our prior research endeavors [3337], we leveraged Machine Learning (ML) algorithms to analyze fMRI data at different levels of abstraction the connectivity matrix (A) and the attribute matrix (B), which comprises complex network metrics derived from (A). To conduct our analysis, we employed a diverse set of ML classifiers, including the Support Vector Machine (SVM) [58], Naive Bayes (NB) [59], Multilayer Perceptron (MLP) [60], a fine-tuned Convolutional Neural Network (CNN), and Long Short-Term Memory neural networks (LSTM) [61].

Subsequently, the SHAP values method was employed for the biological interpretation, as it provides a comprehensive explanation of the predictive contribution of each feature. To ensure robustness and unbiased assessment of the machine learning models, we adopted a consistent evaluation approach: 10-fold stratified cross-validation with shuffling. This cross-validation technique partitions the dataset into ten equal stratified folds, ensuring that each fold, denoted as k, maintains a balanced distribution of samples from each class. This approach guarantees that all classes are equally represented across the folds, thus enhancing the reliability of our model evaluations. The algorithm then trains on nine of these folds and validates on the remaining fold, repeating this process ten times, each serving as the validation set once. We used k = 10, which is a common value for this method [6266]. However, to show the stability of the model’s performance, we also tested different values of k. Moreover, the shuffle strategy ensures the data is randomized before being split into folds to prevent ordering effects; this randomization helps prevent any potential bias stemming from data ordering, ensuring the robustness of the model training process.

Furthermore, as a crucial step in our experimental methodology, we performed an initial partition of 30% of the original dataset (comprising a total of 600 matrices) for final testing, constituting an exclusive reserve of 180 matrices. This separation is done before the model’s training using a 10-fold cross-validation. Adopting such a practice is commonplace in machine learning, and it serves the dual purpose of assessing model performance while mitigating overfitting and ensuring its ability to generalize to new, unseen data [67, 68].

This procedure was applied for model selection and hyperparameter optimization. It was also considered the grid search method, commonly used in the literature [6973], used for all ML algorithms except the CNN and LSMT models. In the deep learning models, we used random search optimization because it offers a more computationally efficient alternative for hyperparameter tuning compared to grid search, which is particularly advantageous, given the high computational demands of deep learning tasks. The hyper-parameter optimization values for each classifier model can be seen in more detail in [3336]. In Tables 2 and 3, we present the architectural details and hyperparameter configurations for CNN and LSTM models, respectively. Notably, dropout regularization, as indicated in Tables 2 and 3, is a widely-used technique in neural network training to combat overfitting [74]. Dropout operates by randomly deactivating neurons during training, compelling the network to learn more resilient and generalizable features [75]. Empirically, dropout has effectively enhanced the generalization capabilities of deep learning architectures and medical data [76, 77].

Table 2. Hyperparameters and layer configurations for the CNN model.

Type of Layer Tuning hyperparameter Value
Convolutional
Convolutional dropout [0.00, 0.05, 0.10, 0.15, 0.20, 0.25, 0.30, 0.35, 0.40, 0.45, 0.50]
Convolutional
Convolutional number of filters [32, 64]
Max Pooling dropout [0.00, 0.50, 0.10, 0.15, 0.20]
Flatten
Dense - units [32, 64, 96….512]
-activation [relu, tanh, sigmoid]
Dropout rate [0.00, 0.50, 0.10, 0.15, 0.20]
Adam optimization compile learning rate minvalue = 1e−4
maxvalue = 1e−2
sampling = LOG

Table 3. Hyperparameters and layer configurations for the LSTM model.

Type of Layer Tuning hyperparameter Value
LSTM
LSTM dropout [0.00, 0.05, 0.10, 0.15, 0.20, 0.25, 0.30, 0.35, 0.40, 0.45, 0.50]
LSTM
LSTM units [70, 60, 50, 40]
Dense - units 3
-activation softmax
Adam optimization compile learning rate minvalue = 1e−10
maxvalue = 1e−1
sampling = LOG

Additionally, in our training set, we applied a process known as standardization. Standardization in machine learning typically involves rescaling features to have a mean of zero and a standard deviation of one [78]. This step is pivotal as it transforms the data, facilitating more straightforward comparisons and analyses [79]. This practice is essential because it ensures that all model attributes are equally important and share a consistent scale [80]. Moreover, standardization safeguards against the undue influence of outliers and features with substantial variability on the model’s performance [81].

First and foremost, we considered the widely-used accuracy metric, which provides an overall assessment of our classification model’s correctness [8286]. We expanded our evaluation to incorporate additional standard metrics such as precision and recall [8790]. Precision, also known as a positive predictive value, measures our model’s ability to classify instances belonging to a specific class correctly. In our case, precision helps gauge our model’s accuracy in identifying the TD group. On the other hand, recall, also known as sensitivity, assesses how effectively our classifier predicts positive examples, which now encompass ASD and ADHD individuals. To visualize the performance of our classification model, we continued to utilize the Receiver Operating Characteristic (ROC) curve, a standard method for illustrating the trade-off between true and false positive rates. The Area Under the ROC Curve (AUC) remained a key evaluation metric, with values ranging from 0 to 1. An AUC of 1 signifies a flawless classification, while 0.5 suggests a random choice where the classifier cannot distinguish between classes effectively [72, 82, 91, 92].

In this three-class classification context, we calculated the micro-average AUC independently for each class (TD, ASD, or ADHD) to provide insights into individual class performance. This micro-average computation treats each class equally, allowing us to assess how well our model performs for each group. Furthermore, we introduced the concept of macro average in our evaluation, which considers the classes individually and aggregates their contributions before calculating the average. Unlike the micro average, the macro average does not treat all classes equally, providing a different perspective on the overall performance of our classification model. Subsequently, the SHAP values method was used for the biological interpretation, as it explains the predictive power of each attribute. The SHAP values method was subsequently employed for the biological interpretation, as it explains the predictive potential of each attribute. The results regarding the connectivity matrix of the first level of abstraction (A) can be found in subsection 3.1.

2.4 Complex network measures

Considering the performance and computational cost, the best ML algorithm was used to evaluate the complex network measure’s level of abstraction. To characterize the structure of the brain’s network, the following complex network measurements were computed as used in the previous work [3337]: average shortest path length (APL) [93], betweenness centrality (BC) [94], closeness centrality (CC) [95], diameter [96], assortativity coefficient [97, 98], hub score [99], eccentricity [100], eigenvector centrality (EC) [101], average degree of nearest neighbors [102] (Knn), mean degree [103], entropy of the degree distribution [104], transitivity [105, 106], second moment of the degree distribution (SMD) [107], complexity, k-core [108, 109], density [110], and efficiency [111].

In this study, we employed recently developed metrics, as comprehensively detailed in [36], to quantify the number of communities within a complex network. Our investigation also incorporated various community detection algorithms [112114]. The outcomes of community detection measures were consolidated into a single scalar value. Specifically, we focused on identifying the largest community within each network, followed by the computation of the average path length within that community, resulting in a singular metric. The suite of community detection algorithms encompassed fast greedy (FC) [115], infomap (IC) [116], leading eigenvector (LC) [117], label propagation (LPC) [118], edge betweenness (EBC) [119], spinglass (SPC) [120], and multilevel community identification (MC) [121]. For clarity and coherence, we extended the abbreviations by appending the letter ‘A’ (indicating average path length) to denote the corresponding approach, resulting in AFC, AIC, ALC, ALPC, AEBC, ASPC, and AMC.

Further, we used three measures to analyze the segregation and integration concepts: Effective Information (EI) and determinism and degeneracy coefficients. Measures of integrated information promise general applicability to questions in neuroscience, in which part-whole relations play a role, and are our interest here [122]. In this paradigm, a system can show a balance between two competing tendencies [123]:

  • integration, i.e., the system behaves as one;

  • segregation, i.e., the parts of the system behave independently.

In other words, integration in network analysis refers to how well nodes in a network are interconnected, facilitating efficient information flow, and highly integrated networks allow for smooth information exchange between nodes [124, 125]. Segregation, on the other hand, pertains to distinct subgroups or communities within a network; segregated networks have subsets of nodes that are more tightly connected within their subgroups, often forming distinct clusters or communities [126, 127].

The EI was first introduced to capture the causal influence between two subsets of neurons as a step in calculating integrated information in the brain [128]. Later, a system-wide version of EI was shown to capture fundamental causal properties such as determinism and redundancy [129, 130]. To expand the EI framework to networks, in [131], the intervention operation in the EI calculation is relaxed by assuming that Wout has modulus one and interpreted as leaving a random walker on the network. This allows us to investigate the dynamics by inspecting the graph topology. Quantitatively, the EI is based on two uncertainties: the first is the Shannon entropy of the average out-weight vector in the network, H(Wiout), which captures how distributed out-weights of the network are; the second is the average entropy of each node’s H(Wiout), giving:

EI=H(Wiout)-H(Wiout).

Further, two fundamental components of EI are the determinism and degeneracy coefficients. They are based on a network’s connectivity, specifically the degree of overlapping weight in the networks. The determinism is based on the average of how much information is not lost in a walker’s uncertainty, H(Wiout). Since log2(n) represents maximal determinism, i.e., when all walkers have the output wij = 1. Then, the determinism is given by log2(n)-H(Wiout). Meanwhile, the degeneracy describes how non-uniform the weight distribution is of the network. If all nodes lead only to one node, that network is perfectly degenerate. The degeneracy can be captured by log2(n)-H(Wiout). Together, determinism and degeneracy can be used to re-define EI:

EI=determinism-degeneracy.

However, this study considers three classes, differently from the previous ones [3337] that only consider two classes. Therefore, it was not possible to classify ADHD from TD using the complex network measures. To address this challenge and gain insights into the underlying patterns within the data, we performed a Principal Components Analysis (PCA). PCA is a dimensionality reduction technique that transforms the original high-dimensional data into a lower-dimensional representation while preserving as much of the variance in the data as possible [132, 133]. By extracting the main components, we aimed to uncover hidden structures and reduce the dimensionality of the dataset, which can be beneficial for subsequent analysis and visualization. This approach allowed us to explore the relationships between the variables and potentially reveal patterns that may not be apparent in the raw data, ultimately contributing to a deeper understanding of the complex network measures in the context of ADHD and TD classification.

After PCA, we conducted a statistical analysis using the Wilcoxon test with Bonferroni correction to compare the three classes: ASD, ADHD, and TD. The Bonferroni correction controls the family-wise error rate in multiple hypothesis testing scenarios, such as when performing multiple pairwise comparisons [134]. In our context, it helps address the issue of inflated Type I error rates that can occur when conducting multiple statistical tests simultaneously. The Wilcoxon test, also known as the Mann-Whitney U test in the case of two groups or the Kruskal-Wallis test for more than two groups [135137], is a non-parametric test used to assess whether there are statistically significant differences between groups when the assumptions of normality and equal variances are not met. In this specific analysis, the Wilcoxon test allowed us to determine if there were significant differences in some measure or variable among the ASD, ADHD, and TD groups. To apply the Bonferroni correction, the significance level (alpha) for each comparison is adjusted to reduce the overall probability of making a Type I error (Type I error, also known as a false positive, occurs when a true null hypothesis is rejected in a statistical test [138]. In the context of multiple comparisons, it refers to the increased likelihood of mistakenly concluding that there is a significant difference when there is not due to the increased number of tests being performed simultaneously. The Bonferroni correction helps to reduce this risk) across all comparisons [139]. This adjustment is achieved by dividing the original alpha level by the number of comparisons being made. The adjusted alpha becomes more stringent, making it harder to declare a result as statistically significant. Consequently, the Bonferroni correction helps to mitigate the risk of false positives when conducting multiple comparisons. These results can be seen in subsection 3.2, and the following symbols represent the statistical significance:

  • ns: 5.00e − 02 < p < = 1.00e + 00;

  • *: 1.00e − 02 < p < = 5.00e − 02;

  • **: 1.00e − 03 < p < = 1.00e − 02;

  • ***: 1.00e − 04 < p < = 1.00e − 03;

  • ****: p < = 1.00e − 04.

3 Results

3.1 Connectivity matrices

According to Table 4, the best classifiers were LSTM and SVM. LSTM performance for the test set was equal to 0.983 for the mean AUC, 0.978 for precision, 0.978 for recall, and 0.978 for accuracy. SVM performance for the test set was equal to 0.946 for the AUC, 0.928 for the precision, 0.928 for the recall, and 0.928 for the accuracy. Each classifier’s confusion matrices and ROC curves are depicted in Figs 2 and 3, respectively.

Table 4. Results from different ML algorithms.

The best ML algorithms were LSTM and SVM, whose performances are highlighted in bold.

Model Subset AUC Accuracy Recall Precision
LSTM Train 0.987±0.022 0.968±0.051 0.951±0.082 0.954±0.074
Test 0.983 0.978 0.978 0.978
SVM Train 0.951±0.020 0.936±0.026 0.936±0.026 0.940±0.025
Test 0.946 0.928 0.928 0.928
LR Train 0.948±0.022 0.931±0.029 0.931±0.029 0.934±0.028
Test 0.946 0.928 0.928 0.927
CNN Train 0.956±0.042 0.904±0.048 0.852±0.072 0.860±0.073
Test 0.938 0.920 0.913 0.913
MLP Train 0.932±0.025 0.910±0.033 0.910±0.033 0.913±0.035
Test 0.929 0.905 0.905 0.905
NB Train 0.834±0.028 0.779±0.037 0.779±0.037 0.800±0.042
Test 0.854 0.805 0.805 0.812

Fig 2. Confusion matrices depict the performance of various ML algorithms.

Fig 2

The elements in the figure labeled A to F correspond to LSTM, CNN, LR, SVM, MLP, and NB, respectively. The diagonal elements represent TP values, showcasing each algorithm’s accuracy in correctly identifying positive instances. This is noteworthy on a test sample containing 180 instances.

Fig 3. ROC curve for each ML algorithm.

Fig 3

The elements in the figure labeled A to F correspond to LSTM, CNN, LR, SVM, MLP, and NB, respectively. The dashed pink line represents the random choice classifier, the purple line the micro-average ROC curve, the gray line the macro-average ROC curve, the turquoise line the ROC curve referring to the TD class, the orange line the ROC curve referring to the ADHD class, and the green line the ROC curve referring to the ASD class.

Further, we investigated potential biases arising from variations in data acquisition protocols between different sites to enhance the robustness of our analysis; we compared the TD group for the ADHD dataset and the ASD dataset using the SVM. The results in Fig 4 show that all metric performance stands around 0.50 in a random classifier. Therefore, it is impossible to distinguish between the TD groups of the different datasets, proving that we could mitigate potential biases from variations in data acquisition protocols.

Fig 4. Investigation of potential biases arising from variations in data acquisition protocols.

Fig 4

Comparing the TD groups from different datasets results in all metric performance standing around 0.50 in a random classifier, proving that we could mitigate potential biases from variations in data acquisition protocols with our preprocessing.

Since SVM has a lower computational cost, it was chosen for the following subsequent analyses. To determine the optimal number of features required for peak performance, we conducted a Recursive Feature Elimination (RFE) analysis, as illustrated in Fig 5A. RFE, often used in the literature in prediction models in medical data [140142], iteratively removes less critical features to gauge their impact on model performance, allowing us to pinpoint the most relevant features. Fig 5A demonstrates that greater accuracy is attained while using 310 characteristics. Thus, including a complete feature set was not necessary to achieve maximum effectiveness.

Fig 5. RFE and the learning curve for the SVM model are depicted in (A) and (B), respectively.

Fig 5

The best performance is achieved with a total of 310 characteristics, as shown by (A). In (B), the learning curve is presented for the training Accuracy (blue) and test Accuracy (green) using the entire dataset (600 connectivity matrices subjects). The highest performance was achieved with 450 connectivity matrices subjects without requiring the entire dataset. Notably, the connectivity matrices were generated using the data augmentation technique sliding window, and 600 connectivity matrices were used in total in the ML approach before the sampling technique.

In addition to RFE, we generated a learning curve, as illustrated in Fig 5B, to gain insights into the influence of dataset size on our model’s performance. This curve visually represents how the number of training instances affects the model’s predictive accuracy. Therefore, RFE and the learning curve enable us to fine-tune our SVM model, ensuring it achieves the optimal equilibrium between feature selection and data volume. From Fig 5B, it can be seen that the maximum performance occurred with 450 subjects without the need for the complete dataset.

Then, we used the 310 features obtained from RFE analysis to perform the SHAP values methodology. The results can be seen in Fig 6.

Fig 6. Feature importance ranking using the SHAP values methodology for the SVM classifier with brain regions in descending order.

Fig 6

(A) Feature ranking based on the average of absolute SHAP values over all subjects considered. (B) Feature importance ranking regarding ASD class. (C) Feature importance ranking regarding ADHD class.

As in Fig 6, it can be seen that for all classes, mainly for distinguishing TD and ASD (in Fig 6A, dark blue and green, respectively), the two primary connections were Left-ParsOrbitalis-Left-PrimMotor and Left-ParsOrbitalis-Left-Thalamus. Regarding the ASD class, the primary connections found were Left-ParsOrbitalis-Left-Thalamus and Left-ParsOrbitalis-Left-PrimMotor, with a low correlation value for this class (in Fig 6B). Regarding the ADHD class, the primary connections found were Left-VisualAssoc-Outside defined BAS1 and Right-AngGyrus-Outside defined BAS1, with a low correlation value for this class (in Fig 6C). From our previous work, the Outside defined BAS1 was identified as the cerebellum. Fig 7 contains the two-dimensional schematic (ventral-axis) with the man regions found regarding ASD and ADHD.

Fig 7. The most important connections found.

Fig 7

Two-dimensional schematic (ventral-axis), where the most important connection for ADHD and ASD are highlighted in green and blue, respectively. The brain plot was developed by the Braph tool [143], and each region was plotted using a Brodmann map from the Yale BioImage Suite Package.

Furthermore, we introduced noise generated by a normal distribution, with different means (level of the noise) while keeping the standard deviation constant at 0.1. This resulted in a range of noisy matrices that we used to evaluate our SVM model’s performance. We assessed the SVM model’s performance using AUC and accuracy depicted in Fig 8, which indicates that our SVM model exhibits robustness to noise in the input data matrices. Even when noise levels vary from 0.0 to 1.5, the model maintains a relatively high AUC and accuracy, indicating its ability to accurately classify patients with ASD, ADHD, and TD individuals. The model’s performance gradually decreases as the noise level increases, which is expected. However, it is noteworthy that even at noise levels as high as 1.0 (where the data is significantly distorted), the model still achieves a reasonable AUC of 0.70 and an accuracy of 0.60 (see Fig 8). This suggests that the SVM model is resilient to noise and can provide valuable diagnostic information in real-world scenarios where data may be imperfect.

Fig 8. SVM behavior after insertion of noise.

Fig 8

The mean AUC of the test was obtained with the insertion of noise generated by a normal distribution with 0.1 standard deviation and a 0–1.5 mean range.

Further, we conducted stratified k-fold cross-validation with values beyond k = 10, namely 2, 3, 5, and 15. The resulting plot in Fig 9 reveals trends in the SVM model’s performance on the training dataset. As the value of k increases, the AUC remains relatively stable, with values ranging from approximately 0.94 to 0.95. This suggests that the SVM model consistently discriminates between the three groups—ASD, ADHD, and TD. The corresponding accuracy values remain steady, ranging from approximately 0.93 to 0.94. We observed similar stability in AUC values for the test dataset, which ranges from about 0.92 to 0.95 as k varies. This indicates that the SVM model’s ability to distinguish between groups holds when applied to unseen data. The accuracy on the test dataset also remains steady, with values ranging from approximately 0.89 to 0.93. This robustness is particularly valuable when dealing with real-world data where k can impact model stability.

Fig 9. Plot for the SVM model with performance measures.

Fig 9

The AUC and Accuracy, in the y-axis, in blue and purple, respectively, were obtained by varying the number of k in the stratified k-fold-cross-validation (x-axis))—the dashed lines corresponding to the test sample and the complete lines to the training sample. Furthermore, the shaded represents the standard deviation in the training sample.

3.2 Complex network

The performance of the test sample considering the complex network yielded the confusion matrix and the ROC curve depicted in Fig 10.

Fig 10. ML results from complex network measures.

Fig 10

(A) The confusion matrix indicates that there were a lot of incorrect predictions between the TD and ADHD groups. (B) The ROC curve, where the dashed pink line represents the random choice classifier, the purple line is the micro-average ROC curve, the gray line is the macro-average ROC curve, the turquoise line the ROC curve referring to the TD class, the orange line the ROC curve referring to the ADHD class (which can be seen the ADHD has the lowest-distinguished curve) and the green line the ROC curve referring to the ASD class (which can be seen the ASD has the best-distinguished curve).

The performance of the test sample considering the complex network yielded the confusion matrix, and the ROC curve depicted in Fig 10 indicates that the model did not perform well for the ADHD class (with an accurate positive accuracy of 0.53 and an AUC of 0.72). This suboptimal performance can be attributed to several factors. Firstly, when we performed PCA with two and three components, as shown in Fig 11, it became evident that the ADHD and TD instances formed two overlapping groups, differently from the ASD instance class. This lack of clear separation in the PCA space suggests that the initial feature set does not easily capture the inherent characteristics distinguishing ADHD from TD cases. This inherent overlap in feature distributions can significantly hinder the performance of a classifier like SVM, which relies on well-defined class boundaries.

Fig 11. PCA using the complex network measures.

Fig 11

The features for ASD, ADHD, and TD are depicted in red, green, and blue, respectively. In (A), PCA with three components, namely PC1, PC2, and PC3, is illustrated in the plot axis. In (B), PCA with two components, namely PC1 and PC2, is presented in the plot axis; further, the heatmap shows that any of the features were highly correlated with the two components.

Additionally, we observed in Fig 11 that none of the features displayed strong correlations with the principal components. This lack of feature-component solid correlations suggests that the initial feature set may need to contain clear discriminatory information, making it challenging for the SVM to distinguish between classes effectively.

Therefore, to improve the model performance, it may be necessary to consider additional domain-specific features that could better capture the nuances of ADHD and TD differentiation within the dataset.

Then, we performed a statistical t-test with Bonferroni correction. This choice was driven by our need to rigorously assess the significance of differences in the means of individual features between the ADHD and TD groups. By conducting this test, we could identify which specific features exhibit statistically significant distinctions between these two classes. The Bonferroni correction is applied to mitigate the issue of multiple comparisons, ensuring that we maintained a low family-wise error rate. In other words, it helps control the higher probability of obtaining false positives when examining numerous features simultaneously. The statistically significant ones between all groups are depicted in Fig 12, and the ones with four stars at least between one of the groups are in Fig 13. Further, the statistical test with the integrated measures can be found in Fig 14.

Fig 12. Features were statistically significant between all groups when using the t-test with Bonferroni correction.

Fig 12

The orange, pink, and purple boxplots show the features that obtained the most statistically significant differences regarding the classes ADH, ASD, and TD, respectively.

Fig 13. Features were four stars statistically significant, at least between one of the groups, when using the t-test with Bonferroni correction.

Fig 13

The orange, pink, and purple boxplots show the features that obtained the most statistically significant differences regarding the classes ADH, ASD, and TD, respectively.

Fig 14. The t-test with Bonferroni correction for the integrated measures.

Fig 14

The orange, pink, and purple boxplots show the features that obtained the most statistically significant differences regarding the classes ADH, ASD, and TD, respectively.

4 Discussion

4.1 Connectivity matrices

Overall, we obtained the best performance compared to the multiclass machine learning algorithm comparing ASD, ADHD, and TD in the literature, as described in section 1. Analysis from Table 1 reveals that our methodology outperforms existing multiclass approaches. In our prior research [33] focusing on EEG time series, we demonstrated the superior accuracy of constructing connectivity matrices compared to conventional methods employing raw EEG data. This underscores the significance of network topology in characterizing brain data.

Furthermore, in subsequent investigations [34, 35], we found that employing a distinct correlation metric yielded improved detection of brain changes associated with ASD and schizophrenia, respectively. Interestingly, TE proved effective in capturing such changes in the fMRI dataset. Thus, one of our hypotheses for achieving optimal performance revolves around selecting an appropriate correlation metric.

Furthermore, our findings, as illustrated in Fig 6, provide valuable insights into the functional roles of specific brain regions in distinguishing between individuals with ASD, ADHD, and TD based on fMRI matrices. These results shed light on the neural circuitry implicated in these neurodevelopmental conditions.

In Fig 6A, where TD and ASD are distinguished, the two primary connections of interest are Left-ParsOrbitalis-Left-PrimMotor and Left-ParsOrbitalis-Left-Thalamus. This observation suggests that these connections play a significant role in discriminating between TD and ASD individuals. The Left-ParsOrbitalis is associated with decision-making and social cognition [144], areas commonly affected in individuals with ASD [145]. The connections to the PrimMotor and Thalamus imply that motor control and sensory processing also contribute to distinguishing between these groups and are also found in our previous work [34]. Further, Fig 6B highlights the primary connections for the ASD class, with a focus on Left-ParsOrbitalis-Left-Thalamus and Left-ParsOrbitalis-Left-PrimMotor. These connections are consistent with the findings in Fig 6A, emphasizing the importance of the Left-ParsOrbitalis in distinguishing individuals with ASD. This region’s involvement in social cognition, decision-making, and language processing may reflect the cognitive and behavioral characteristics associated with ASD.

In Fig 6C, which pertains to the ADHD class, the primary connections identified are Left-VisualAssoc-Outside defined BAS1 and Right-AngGyrus-Outside defined BAS1, both with low correlation values. Our previous work has identified the “Outside defined BAS1” as the cerebellum. The cerebellum is traditionally linked to motor control [146]. However, emerging research suggests its involvement in cognitive functions, including attention and executive control [147] and in several studies associated with ADHD [146, 148, 149]. The Left-VisualAssoc connection might signify differences in visual processing [150], while the Right-AngGyrus could be related to higher-order cognitive functions [151]. Both areas play a significant role in visuospatial attention processes [152, 153]. The low correlation values may imply that these connections are less distinctive for distinguishing ADHD from the other groups, suggesting a more complex neural signature for this condition.

4.2 Complex network measures

The results in Fig 12 offer valuable insights into the network properties of three distinct groups: ASD, ADHD, and TD. We employed a range of complex network measures to assess valuable insights into the network properties of three distinct groups. Further, with these metrics, we can observe distinct patterns in integration and segregation, which are fundamental concepts in network analysis, across the three groups [154, 155].

The ASD group exhibits the lowest values across various network metrics, such as BC, Density, Eccentricity, K-core, KNN, Mean Degree, ED, and Efficiency. This suggests that individuals with ASD have a more fragmented and segregated network structure, indicating challenges in information integration within the network. In other words, their networks may have more isolated clusters of nodes that do not communicate effectively with each other, according to the literature [34, 156, 157].

On the other hand, the ADHD group demonstrates higher values in these metrics compared to ASD; however, it falls short of the performance observed in the TD group, indicating that individuals with ADHD have a network structure that is more integrated than ASD but still not as optimal as the control group. This implies that individuals with ADHD have a network structure that is more cohesive than ASD but not as cohesive as typical development.

Furthermore, as shown in Fig 13, the ASD group exhibits the lowest values for both the ASC measure, gauging the size of community networks, EC, network influence, and transitivity. In contrast, the ADHD group displays higher values in these metrics than the ASD group but still lags behind the TD group. These findings suggest that individuals with ASD may have smaller and less influential networks within their communities, while those with ADHD fall in between ASD and TD individuals regarding these network characteristics.

By computing the EI in the TD, ASD, and ADHD groups, there is a lower average value of EI in the ASD group than the others, showing a tendency to have a more segregated than integrated structure in this group (see Fig 14). Interestingly, there is an analysis of determinism and degeneracy coefficients in the groups. The greater value of determinism and degeneracy coefficients in the ASD than the others shows that the graph structure in this group resembles a star (sparse connections) instead of a complete (well-connected) network [131] strengthening the EI interpretation in the last paragraph.

In [158], functional segregation was characterized as the capacity for specialized processing within tightly interconnected brain regions. In other words, neuronal processing is distributed across functionally related regions organized into modules. These modules are described as communities exhibiting dense internal connectivity among their constituent nodes and limited communication with nodes from other communities. This network analysis can be linked to the long-observed fact by clinicians that those with ASD are impaired in their ability to generalize—that is, to relate new stimuli to past experiences. Instead, these groups are good at specializing in learned habits [159, 160].

5 Conclusion

Our study used fMRI datasets and explainable IA methods to generate an interpretable classifier for ASD, ADHD, and TD. We have found distinct brain activity patterns underlying these neurodevelopmental disorders by advancing beyond binary comparisons and integrating complex network measures alongside machine learning methodologies.

Our findings confirm the existence of unique neural signatures for ASD, ADHD, and TD groups. Notably, connections involving Left-ParsOrbitalis emerged as crucial in distinguishing between TD and ASD, possibly indicating underlying deficits in decision-making and social cognition observed in ASD. Similarly, distinct neural signatures were observed for ADHD, with connections to the cerebellum, Left-VisualAssoc, and Right-AngGyrus, highlighting potential involvement in cognitive functions and sensory processing differences. The observed connectivity patterns on which the ML classification rests agree with established diagnostic approaches based on clinical symptoms, proving the trustworthiness and efficiency of our multiclass ML approach’s interpretability technique. Moreover, we demonstrate the superior performance of our multiclass machine learning approach compared to existing literature. This heightened performance is essential for reliable discrimination between neurodevelopmental conditions, promising prospects for more precise diagnostic tools.

Furthermore, our analysis of complex network measures elucidated the network properties of each group, unveiling differences in integration and segregation patterns. The ASD group exhibited the lowest values across various network metrics, suggesting a fragmented network structure. In contrast, the ADHD group demonstrated intermediate values, indicative of a network that is more integrated than ASD but less cohesive than typical development.

Despite these significant contributions, limitations such as data quantity constrain the generalizability of our findings. Future studies should aim to overcome these limitations by incorporating larger datasets encompassing a broader range of mental health conditions. Further investigations focusing on specific brain regions could provide deeper insights into group differences in brain connectivity. Further, we propose integrating our methodology with federated learning techniques as a promising avenue for advancing diagnostics and drug trials in neurodevelopmental conditions [161163]. Federated learning offers a solution to data privacy and scalability challenges, allowing for collaborative model training across multiple datasets while preserving data decentralization [164166]. This approach holds immense potential for improving diagnostic accuracy and guiding personalized treatment strategies tailored to specific demographics or clinical settings [167].

In summary, our study represents a significant step forward in understanding the neural underpinnings of neurodevelopmental conditions. By leveraging advanced analytical techniques and machine learning methodologies, we have surpassed performance in discrimination between ASD, ADHD, and TD individuals, paving the way for refined diagnostics and promising avenues for developing trustworthy clinical decision-support systems.

Supporting information

S1 File. Connectivity matrices attached in a zip file used to generate the results.

This section contains the normalized transfer entropy connectivity matrices derived from the authors’ preprocessed BOLD time series of 122 fMRI regions. These matrices served as inputs for the algorithms employed in this study and for extracting complex network measures.

(ZIP)

S2 File

(ZIP)

pone.0305630.s002.zip (5.9MB, zip)

Data Availability

All relevant data are within the article and its Supporting information files. Neuroimaging data were obtained from two large, publicly available datasets: ADHD-200 (http://fcon_1000.projects.nitrc.org/indi/adhd200/) and ABIDE (Autism Brain Imaging Data Exchange; http://fcon_1000.projects.nitrc.org/indi/abide/).

Funding Statement

ZeWiS is funded through public research projects by the Bavarian State Ministry for Sciences and the Arts. The funder provided support in the form of salaries for authors but did not have any additional role in the study design, data collection and analysis, decision to publish, or manuscript preparation. The specific roles of these authors are articulated in the ‘author contributions’ section.

References

  • 1. Parenti I, Rabaneda LG, Schoen H, Novarino G. Neurodevelopmental disorders: from genetics to functional pathways. Trends in Neurosciences. 2020;43(8):608–621. doi: 10.1016/j.tins.2020.05.004 [DOI] [PubMed] [Google Scholar]
  • 2. Li Y, Shen M, Stockton ME, Zhao X. Hippocampal deficits in neurodevelopmental disorders. Neurobiology of learning and memory. 2019;165:106945. doi: 10.1016/j.nlm.2018.10.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Mahone EM, Warschausky S, Zabel TA. Introduction to the JINS special issue: Neurodevelopmental disorders. Journal of the International Neuropsychological Society. 2018;24(9):893–895. doi: 10.1017/S1355617718000905 [DOI] [Google Scholar]
  • 4. Thapar A, Cooper M, Rutter M. Neurodevelopmental disorders. The Lancet Psychiatry. 2017;4(4):339–346. doi: 10.1016/S2215-0366(16)30376-5 [DOI] [PubMed] [Google Scholar]
  • 5. Lord C, Elsabbagh M, Baird G, Veenstra-Vanderweele J. Autism spectrum disorder. The lancet. 2018;392(10146):508–520. doi: 10.1016/S0140-6736(18)31129-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Posar A, Visconti P. Autism Spectrum Disorder in 2023: A Challenge Still Open. Turkish Archives of Pediatrics. 2023;. doi: 10.5152/TurkArchPediatr.2023.23194 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Masi A, DeMayo MM, Glozier N, Guastella AJ. An overview of autism spectrum disorder, heterogeneity and treatment options. Neuroscience bulletin. 2017;33:183–193. doi: 10.1007/s12264-017-0100-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Constantino JN, Charman T. Diagnosis of autism spectrum disorder: reconciling the syndrome, its diverse origins, and variation in expression. The Lancet Neurology. 2016;15(3):279–291. doi: 10.1016/S1474-4422(15)00151-9 [DOI] [PubMed] [Google Scholar]
  • 9. Lord C, Charman T, Havdahl A, Carbone P, Anagnostou E, Boyd B, et al. The Lancet Commission on the future of care and clinical research in autism. The Lancet. 2022;399(10321):271–334. doi: 10.1016/S0140-6736(21)01541-5 [DOI] [PubMed] [Google Scholar]
  • 10. Lord C, Cook EH, Leventhal BL, Amaral DG. Autism spectrum disorders. Neuron. 2000;28(2):355–363. doi: 10.1016/S0896-6273(00)00115-X [DOI] [PubMed] [Google Scholar]
  • 11. Du Y, Fu Z, Calhoun VD. Classification and prediction of brain disorders using functional connectivity: promising but challenging. Frontiers in neuroscience. 2018;12:525. doi: 10.3389/fnins.2018.00525 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Salari N, Ghasemi H, Abdoli N, Rahmani A, Shiri MH, Hashemian AH, et al. The global prevalence of ADHD in children and adolescents: a systematic review and meta-analysis. Italian Journal of Pediatrics. 2023;49(1):48. doi: 10.1186/s13052-023-01456-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Arbabshirani MR, Plis S, Sui J, Calhoun VD. Single subject prediction of brain disorders in neuroimaging: Promises and pitfalls. Neuroimage. 2017;145:137–165. doi: 10.1016/j.neuroimage.2016.02.079 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Rommelse NN, Geurts HM, Franke B, Buitelaar JK, Hartman CA. A review on cognitive and brain endophenotypes that may be common in autism spectrum disorder and attention-deficit/hyperactivity disorder and facilitate the search for pleiotropic genes. Neuroscience & Biobehavioral Reviews. 2011;35(6):1363–1396. doi: 10.1016/j.neubiorev.2011.02.015 [DOI] [PubMed] [Google Scholar]
  • 15. Simonoff E, Pickles A, Charman T, Chandler S, Loucas T, Baird G. Psychiatric disorders in children with autism spectrum disorders: prevalence, comorbidity, and associated factors in a population-derived sample. Journal of the American Academy of Child & Adolescent Psychiatry. 2008;47(8):921–929. doi: 10.1097/CHI.0b013e318179964f [DOI] [PubMed] [Google Scholar]
  • 16. Uddin M, Wang Y, Woodbury-Smith M. Artificial intelligence for precision medicine in neurodevelopmental disorders. NPJ digital medicine. 2019;2(1):112. doi: 10.1038/s41746-019-0191-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17. Duda M, Ma R, Haber N, Wall D. Use of machine learning for behavioral distinction of autism and ADHD. Translational psychiatry. 2016;6(2):e732–e732. doi: 10.1038/tp.2015.221 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Duda M, Haber N, Daniels J, Wall D. Crowdsourced validation of a machine-learning classification system for autism and ADHD. Translational psychiatry. 2017;7(5):e1133–e1133. doi: 10.1038/tp.2017.86 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Nogay HS, Adeli H. Machine learning (ML) for the diagnosis of autism spectrum disorder (ASD) using brain imaging. Reviews in the Neurosciences. 2020;31(8):825–841. doi: 10.1515/revneuro-2020-0043 [DOI] [PubMed] [Google Scholar]
  • 20. Rashid B, Calhoun V. Towards a brain-based predictome of mental illness. Human brain mapping. 2020;41(12):3468–3535. doi: 10.1002/hbm.25013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21. Harikumar A, Evans DW, Dougherty CC, Carpenter KL, Michael AM. A review of the default mode network in autism spectrum disorders and attention deficit hyperactivity disorder. Brain connectivity. 2021;11(4):253–263. doi: 10.1089/brain.2020.0865 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Eslami T, Almuqhim F, Raiker JS, Saeed F. Machine learning methods for diagnosing autism spectrum disorder and attention-deficit/hyperactivity disorder using functional and structural MRI: A survey. Frontiers in neuroinformatics. 2021; p. 62. doi: 10.3389/fninf.2020.575999 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Moridian P, Ghassemi N, Jafari M, Salloum-Asfar S, Sadeghi D, Khodatars M, et al. Automatic autism spectrum disorder detection using artificial intelligence methods with MRI neuroimaging: A review. Frontiers in Molecular Neuroscience. 2022;15:999605. doi: 10.3389/fnmol.2022.999605 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24. Washington P, Wall DP. A Review of and Roadmap for Data Science and Machine Learning for the Neuropsychiatric Phenotype of Autism. Annual Review of Biomedical Data Science. 2023;6. doi: 10.1146/annurev-biodatasci-020722-125454 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Wolff N, Kohls G, Mack JT, Vahid A, Elster EM, Stroth S, et al. A data driven machine learning approach to differentiate between autism spectrum disorder and attention-deficit/hyperactivity disorder based on the best-practice diagnostic instruments for autism. Scientific Reports. 2022;12(1):18744. doi: 10.1038/s41598-022-21719-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Lim L, Marquand A, Cubillo AA, Smith AB, Chantiluke K, Simmons A, et al. Disorder-specific predictive classification of adolescents with attention deficit hyperactivity disorder (ADHD) relative to autism using structural magnetic resonance imaging. PloS one. 2013;8(5):e63660. doi: 10.1371/journal.pone.0063660 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Di Martino A, Zuo XN, Kelly C, Grzadzinski R, Mennes M, Schvarcz A, et al. Shared and distinct intrinsic functional network centrality in autism and attention-deficit/hyperactivity disorder. Biological psychiatry. 2013;74(8):623–632. doi: 10.1016/j.biopsych.2013.02.011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Bethlehem RA, Romero-Garcia R, Mak E, Bullmore E, Baron-Cohen S. Structural covariance networks in children with autism or ADHD. Cerebral Cortex. 2017;27(8):4267–4276. doi: 10.1093/cercor/bhx135 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Shephard E, Tye C, Ashwood KL, Azadi B, Johnson MH, Charman T, et al. Oscillatory neural networks underlying resting-state, attentional control and social cognition task conditions in children with ASD, ADHD and ASD+ ADHD. Cortex. 2019;117:96–110. doi: 10.1016/j.cortex.2019.03.005 [DOI] [PubMed] [Google Scholar]
  • 30.Bathelt J, Caan M, Geurts H. More similarities than differences between ADHD and ASD in functional brain connectivity. 2020;.
  • 31. Rudin C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature machine intelligence. 2019;1(5):206–215. doi: 10.1038/s42256-019-0048-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32. Althubaiti A. Information bias in health research: definition, pitfalls, and adjustment methods. Journal of multidisciplinary healthcare. 2016; p. 211–217. doi: 10.2147/JMDH.S104807 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Alves CL, Pineda AM, Roster K, Thielemann C, Rodrigues FA. EEG functional connectivity and deep learning for automatic diagnosis of brain disorders: Alzheimer’s disease and schizophrenia. Journal of Physics: Complexity. 2022;3(2):025001. [Google Scholar]
  • 34. Alves CL, Toutain TGdO, de Carvalho Aguiar P, Pineda AM, Roster K, Thielemann C, et al. Diagnosis of autism spectrum disorder based on functional brain networks and machine learning. Scientific Reports. 2023;13(1):8072. doi: 10.1038/s41598-023-34650-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Alves CL, Toutain TGdO, Porto JAM, Aguiar PMdC, Pineda A, Rodrigues FA, et al. Analysis of functional connectivity using machine learning and deep learning in different data modalities from individuals with schizophrenia. Journal of Neural Engineering. 2023;. doi: 10.1088/1741-2552/acf734 [DOI] [PubMed] [Google Scholar]
  • 36. Alves CL, Cury RG, Roster K, Pineda AM, Rodrigues FA, Thielemann C, et al. Application of machine learning and complex network measures to an EEG dataset from ayahuasca experiments. Plos one. 2022;17(12):e0277257. doi: 10.1371/journal.pone.0277257 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Alves C, Wissel L, Capetian P, Thielemann C. P 55 Functional connectivity and convolutional neural networks for automatic classification of EEG data. Clinical Neurophysiology. 2022;137:e47. doi: 10.1016/j.clinph.2022.01.086 [DOI] [Google Scholar]
  • 38.Lundberg SM, Lee SI. A unified approach to interpreting model predictions. In: Proceedings of the 31st international conference on neural information processing systems; 2017. p. 4768–4777.
  • 39. Al-Beltagi M. Autism medical comorbidities. World journal of clinical pediatrics. 2021;10(3):15. doi: 10.5409/wjcp.v10.i3.15 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40. Bellec P, Chu C, Chouinard-Decorte F, Benhajali Y, Margulies DS, Craddock RC. The neuro bureau ADHD-200 preprocessed repository. Neuroimage. 2017;144:275–286. doi: 10.1016/j.neuroimage.2016.06.034 [DOI] [PubMed] [Google Scholar]
  • 41. Abraham A, Pedregosa F, Eickenberg M, Gervais P, Mueller A, Kossaifi J, et al. Machine learning for neuroimaging with scikit-learn. Frontiers in neuroinformatics. 2014;8:14. doi: 10.3389/fninf.2014.00014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42. Lawrence RM, Bridgeford EW, Myers PE, Arvapalli GC, Ramachandran SC, Pisner DA, et al. Standardizing human brain parcellations. Scientific data. 2021;8(1):78. doi: 10.1038/s41597-021-00849-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Rubin TN, Koyejo O, Gorgolewski KJ, Jones MN, Poldrack RA, Yarkoni T. Decoding brain activity using a large-scale probabilistic functional-anatomical atlas of human cognition. PLoS computational biology. 2017;13(10):e1005649. doi: 10.1371/journal.pcbi.1005649 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44. Walsh J, Othmani A, Jain M, Dev S. Using U-Net network for efficient brain tumor segmentation in MRI images. Healthcare Analytics. 2022;2:100098. doi: 10.1016/j.health.2022.100098 [DOI] [Google Scholar]
  • 45. Subah FZ, Deb K, Dhar PK, Koshiba T. A deep learning approach to predict autism spectrum disorder using multisite resting-state fMRI. Applied Sciences. 2021;11(8):3636. doi: 10.3390/app11083636 [DOI] [Google Scholar]
  • 46. Bellec P, Rosa-Neto P, Lyttelton OC, Benali H, Evans AC. Multi-level bootstrap analysis of stable clusters in resting-state fMRI. Neuroimage. 2010;51(3):1126–1139. doi: 10.1016/j.neuroimage.2010.02.082 [DOI] [PubMed] [Google Scholar]
  • 47. Yang X, Zhang N, Schrader P. A study of brain networks for autism spectrum disorder classification using resting-state functional connectivity. Machine Learning with Applications. 2022;8:100290. doi: 10.1016/j.mlwa.2022.100290 [DOI] [Google Scholar]
  • 48. Schreiber T. Measuring information transfer. Physical review letters. 2000;85(2):461. doi: 10.1103/PhysRevLett.85.461 [DOI] [PubMed] [Google Scholar]
  • 49. Lungarella M, Pitti A, Kuniyoshi Y. Information transfer at multiple scales. Physical Review E. 2007;76(5):056117. doi: 10.1103/PhysRevE.76.056117 [DOI] [PubMed] [Google Scholar]
  • 50. Shovon MHI, Nandagopal N, Vijayalakshmi R, Du JT, Cocks B. Directed connectivity analysis of functional brain networks during cognitive activity using transfer entropy. Neural Processing Letters. 2017;45:807–824. doi: 10.1007/s11063-016-9506-1 [DOI] [Google Scholar]
  • 51. Wibral M, Rahm B, Rieder M, Lindner M, Vicente R, Kaiser J. Transfer entropy in magnetoencephalographic data: quantifying information flow in cortical and cerebellar networks. Progress in biophysics and molecular biology. 2011;105(1-2):80–97. doi: 10.1016/j.pbiomolbio.2010.11.006 [DOI] [PubMed] [Google Scholar]
  • 52. Mao X, Shang P. Transfer entropy between multivariate time series. Communications in Nonlinear Science and Numerical Simulation. 2017;47:338–347. doi: 10.1016/j.cnsns.2016.12.008 [DOI] [Google Scholar]
  • 53. Orlandi JG, Stetter O, Soriano J, Geisel T, Battaglia D. Transfer entropy reconstruction and labeling of neuronal connections from simulated calcium imaging. PloS one. 2014;9(6):e98842. doi: 10.1371/journal.pone.0098842 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Gunaratne C, Ray SK, Lourenço Alves C, Waldl M. Exogenous Shocks Lead to Increased Responsiveness and Shifts in Sentimental Resilience in Online Discussions. In: Proceedings of the 2019 International Conference of The Computational Social Science Society of the Americas. Springer; 2021. p. 57–71.
  • 55. Goetze F, Lai PY, Chan C. Identifying excitatory and inhibitory synapses in neuronal networks from dynamics using Transfer Entropy. BMC Neuroscience. 2015;16:1–1. doi: 10.1186/1471-2202-16-S1-P3025655275 [DOI] [Google Scholar]
  • 56.Bellec P, Carbonnell F, Perlbarg V, Lepage C, Lyttelton O, Fonov V, et al. A neuroimaging analyses kit for Matlab and octave. In: Human Brain Mapping HBM 2011 17th Annual Meeting of the Organization on Human Brain Mapping, Quebec City, Canada, June 26-30, 2011. Organization on Human Brain Mapping; 2011. p. 1–5.
  • 57. Khodatars M, Shoeibi A, Sadeghi D, Ghaasemi N, Jafari M, Moridian P, et al. Deep learning for neuroimaging-based diagnosis and rehabilitation of autism spectrum disorder: a review. Computers in biology and medicine. 2021;139:104949. doi: 10.1016/j.compbiomed.2021.104949 [DOI] [PubMed] [Google Scholar]
  • 58. Bottou L, Lin CJ. Support vector machine solvers. Large scale kernel machines. 2007;3(1):301–320. [Google Scholar]
  • 59. Friedman N, Geiger D, Goldszmidt M. Bayesian network classifiers. Machine learning. 1997;29(2):131–163. doi: 10.1023/A:1007465528199 [DOI] [Google Scholar]
  • 60. Hinton G, Rumelhart D, Williams R. Learning internal representations by error propagation. Parallel distributed processing. 1986;1:318–362. [Google Scholar]
  • 61. Hochreiter S, Schmidhuber J. Long short-term memory. Neural computation. 1997;9(8):1735–1780. doi: 10.1162/neco.1997.9.8.1735 [DOI] [PubMed] [Google Scholar]
  • 62.Berrar D. Cross-Validation.; 2019.
  • 63. Bengio Y, Grandvalet Y. No unbiased estimator of the variance of k-fold cross-validation. Journal of machine learning research. 2004;5(Sep):1089–1105. [Google Scholar]
  • 64. Shah AA, Khan YD. Identification of 4-carboxyglutamate residue sites based on position based statistical feature and multiple classification. Scientific Reports. 2020;10(1):1–10. doi: 10.1038/s41598-020-73107-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65. Kawamoto T, Kabashima Y. Cross-validation estimate of the number of clusters in a network. Scientific reports. 2017;7(1):1–17. doi: 10.1038/s41598-017-03623-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66. Chan J, Rea T, Gollakota S, Sunshine JE. Contactless cardiac arrest detection using smart devices. NPJ digital medicine. 2019;2(1):1–8. doi: 10.1038/s41746-019-0128-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67. Kuhn M, Johnson K, et al. Applied predictive modeling. vol. 26. Springer; 2013. [Google Scholar]
  • 68.Brownlee J. How to choose a feature selection method for machine learning. Machine Learning Mastery. 2019;10.
  • 69. Sato M, Morimoto K, Kajihara S, Tateishi R, Shiina S, Koike K, et al. Machine-learning approach for the development of a novel predictive model for the diagnosis of hepatocellular carcinoma. Scientific reports. 2019;9(1):1–7. doi: 10.1038/s41598-019-44022-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70. Zhong Z, Yuan X, Liu S, Yang Y, Liu F. Machine learning prediction models for prognosis of critically ill patients after open-heart surgery. Scientific Reports. 2021;11(1):1–10. doi: 10.1038/s41598-021-83020-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71. Arcadu F, Benmansour F, Maunz A, Willis J, Haskova Z, Prunotto M. Author Correction: Deep learning algorithm predicts diabetic retinopathy progression in individual patients. NPJ digital medicine. 2020;3(1):1–6. doi: 10.1038/s41746-020-00365-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72. Krittanawong C, Virk HUH, Kumar A, Aydar M, Wang Z, Stewart MP, et al. Machine learning and deep learning to predict mortality in patients with spontaneous coronary artery dissection. Scientific reports. 2021;11(1):1–10. doi: 10.1038/s41598-021-88172-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73. Rashidi HH, Sen S, Palmieri TL, Blackmon T, Wajda J, Tran NK. Early recognition of burn-and trauma-related acute kidney injury: a pilot comparison of machine learning techniques. Scientific reports. 2020;10(1):1–9. doi: 10.1038/s41598-019-57083-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74. Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R. Dropout: a simple way to prevent neural networks from overfitting. The journal of machine learning research. 2014;15(1):1929–1958. [Google Scholar]
  • 75. Poernomo A, Kang DK. Biased dropout and crossmap dropout: learning towards effective dropout regularization in convolutional neural network. Neural networks. 2018;104:60–67. doi: 10.1016/j.neunet.2018.03.016 [DOI] [PubMed] [Google Scholar]
  • 76. Lemay A, Hoebel K, Bridge CP, Befano B, De Sanjosé S, Egemen D, et al. Improving the repeatability of deep learning models with Monte Carlo dropout. npj Digital Medicine. 2022;5(1):174. doi: 10.1038/s41746-022-00709-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77. Li X, Dou Q, Chen H, Fu CW, Qi X, Belav`y DL, et al. 3D multi-scale FCN with random modality voxel dropout learning for intervertebral disc localization and segmentation from multi-modality MR images. Medical image analysis. 2018;45:41–54. doi: 10.1016/j.media.2018.01.004 [DOI] [PubMed] [Google Scholar]
  • 78.Bisong E, Bisong E. Introduction to Scikit-learn. Building Machine Learning and Deep Learning Models on Google Cloud Platform: A Comprehensive Guide for Beginners. 2019; p. 215–229.
  • 79. Raschka S. Python machine learning. Packt publishing ltd; 2015. [Google Scholar]
  • 80. Raschka S, Mirjalili V. Python machine learning: Machine learning and deep learning with Python, scikit-learn, and TensorFlow 2. Packt Publishing Ltd; 2019. [Google Scholar]
  • 81. Géron A. Hands-on machine learning with Scikit-Learn, Keras, and TensorFlow. “O’Reilly Media, Inc.”; 2022. [Google Scholar]
  • 82. Mincholé A, Rodriguez B. Artificial intelligence for the electrocardiogram. Nature medicine. 2019;25(1):22–23. doi: 10.1038/s41591-018-0306-1 [DOI] [PubMed] [Google Scholar]
  • 83. Tolkach Y, Dohmgörgen T, Toma M, Kristiansen G. High-accuracy prostate cancer pathology using deep learning. Nature Machine Intelligence. 2020;2(7):411–418. doi: 10.1038/s42256-020-0200-7 [DOI] [Google Scholar]
  • 84. Dukart J, Weis S, Genon S, Eickhoff SB. Towards increasing the clinical applicability of machine learning biomarkers in psychiatry. Nature Human Behaviour. 2021;5(4):431–432. doi: 10.1038/s41562-021-01085-w [DOI] [PubMed] [Google Scholar]
  • 85. Li RC, Asch SM, Shah NH. Developing a delivery science for artificial intelligence in healthcare. NPJ digital medicine. 2020;3(1):1–3. doi: 10.1038/s41746-020-00318-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86. Park Y, Kellis M. Deep learning for regulatory genomics. Nature biotechnology. 2015;33(8):825–826. doi: 10.1038/nbt.3313 [DOI] [PubMed] [Google Scholar]
  • 87. Ito Y, Unagami M, Yamabe F, Mitsui Y, Nakajima K, Nagao K, et al. A method for utilizing automated machine learning for histopathological classification of testis based on Johnsen scores. Scientific reports. 2021;11(1):1–11. doi: 10.1038/s41598-021-89369-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 88. Kim J, Lee J, Park E, Han J. A deep learning model for detecting mental illness from user content on social media. Scientific reports. 2020;10(1):1–6. doi: 10.1038/s41598-020-68764-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89. Li Y, Nowak CM, Pham U, Nguyen K, Bleris L. Cell morphology-based machine learning models for human cell state classification. NPJ systems biology and applications. 2021;7(1):1–9. doi: 10.1038/s41540-021-00180-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 90. Yu X, Pang W, Xu Q, Liang M. Mammographic image classification with deep fusion learning. Scientific Reports. 2020;10(1):1–11. doi: 10.1038/s41598-020-71431-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91. Bracher-Smith M, Crawford K, Escott-Price V. Machine learning for genetic prediction of psychiatric disorders: a systematic review. Molecular Psychiatry. 2021;26(1):70–79. doi: 10.1038/s41380-020-0825-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 92. Patel D, Kher V, Desai B, Lei X, Cen S, Nanda N, et al. Machine learning based predictors for COVID-19 disease severity. Scientific Reports. 2021;11(1):1–7. doi: 10.1038/s41598-021-83967-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 93. Albert R, Barabási AL. Statistical mechanics of complex networks. Reviews of modern physics. 2002;74(1):47. doi: 10.1103/RevModPhys.74.47 [DOI] [Google Scholar]
  • 94. Freeman LC. A set of measures of centrality based on betweenness. Sociometry. 1977; p. 35–41. doi: 10.2307/3033543 [DOI] [Google Scholar]
  • 95. Freeman LC. Centrality in social networks conceptual clarification. Social networks. 1978;1(3):215–239. doi: 10.1016/0378-8733(78)90021-7 [DOI] [Google Scholar]
  • 96. Albert R, Jeong H, Barabási AL. Diameter of the world-wide web. nature. 1999;401(6749):130–131. doi: 10.1038/43601 [DOI] [Google Scholar]
  • 97. Newman ME. The structure and function of complex networks. SIAM review. 2003;45(2):167–256. doi: 10.1137/S003614450342480 [DOI] [Google Scholar]
  • 98. Newman ME. Assortative mixing in networks. Physical review letters. 2002;89(20):208701. doi: 10.1103/PhysRevLett.89.208701 [DOI] [PubMed] [Google Scholar]
  • 99. Kleinberg JM. Hubs, authorities, and communities. ACM computing surveys (CSUR). 1999;31(4es):5–es. doi: 10.1145/345966.345982 [DOI] [Google Scholar]
  • 100. Hage P, Harary F. Eccentricity and centrality in networks. Social networks. 1995;17(1):57–63. doi: 10.1016/0378-8733(94)00248-9 [DOI] [Google Scholar]
  • 101. Bonacich P. Power and centrality: A family of measures. American journal of sociology. 1987;92(5):1170–1182. doi: 10.1086/228631 [DOI] [Google Scholar]
  • 102. Eppstein D, Paterson MS, Yao FF. On nearest-neighbor graphs. Discrete & Computational Geometry. 1997;17(3):263–282. doi: 10.1007/PL00009293 [DOI] [Google Scholar]
  • 103. Doyle J, Graver J. Mean distance in a graph. Discrete Mathematics. 1977;17(2):147–154. doi: 10.1016/0012-365X(77)90144-3 [DOI] [Google Scholar]
  • 104. Dehmer M, Mowshowitz A. A history of graph entropy measures. Information Sciences. 2011;181(1):57–78. doi: 10.1016/j.ins.2010.08.041 [DOI] [Google Scholar]
  • 105. Watts DJ, Strogatz SH. Collective dynamics of ‘small-world’ networks. Nature. 1998;393(6684):440–442. doi: 10.1038/30918 [DOI] [PubMed] [Google Scholar]
  • 106. Newman ME, Watts DJ, Strogatz SH. Random graph models of social networks. Proceedings of the National Academy of Sciences. 2002;99(suppl 1):2566–2572. doi: 10.1073/pnas.012582999 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 107. Snijders TA. The degree variance: an index of graph heterogeneity. Social networks. 1981;3(3):163–174. doi: 10.1016/0378-8733(81)90014-9 [DOI] [Google Scholar]
  • 108. Seidman SB. Network structure and minimum degree. Social networks. 1983;5(3):269–287. doi: 10.1016/0378-8733(83)90028-X [DOI] [Google Scholar]
  • 109. Newman M. Networks: an introduction. Oxford university press; 2010. [Google Scholar]
  • 110. Anderson BS, Butts C, Carley K. The interaction of size and density with graph-level indices. Social networks. 1999;21(3):239–267. doi: 10.1016/S0378-8733(99)00011-8 [DOI] [Google Scholar]
  • 111. Latora V, Marchiori M. Economic small-world behavior in weighted networks. The European Physical Journal B-Condensed Matter and Complex Systems. 2003;32(2):249–263. doi: 10.1140/epjb/e2003-00095-5 [DOI] [Google Scholar]
  • 112. Newman ME. Communities, modules and large-scale structure in networks. Nature physics. 2012;8(1):25–31. doi: 10.1038/nphys2162 [DOI] [Google Scholar]
  • 113. Kim J, Lee JG. Community detection in multi-layer graphs: A survey. ACM SIGMOD Record. 2015;44(3):37–48. doi: 10.1145/2854006.2854013 [DOI] [Google Scholar]
  • 114. Zhao X, Liang J, Wang J. A community detection algorithm based on graph compression for large-scale social networks. Information Sciences. 2021;551:358–372. doi: 10.1016/j.ins.2020.10.057 [DOI] [Google Scholar]
  • 115. Clauset A, Newman ME, Moore C. Finding community structure in very large networks. Physical review E. 2004;70(6):066111. doi: 10.1103/PhysRevE.70.066111 [DOI] [PubMed] [Google Scholar]
  • 116. Rosvall M, Axelsson D, Bergstrom CT. The map equation. The European Physical Journal Special Topics. 2009;178(1):13–23. doi: 10.1140/epjst/e2010-01179-1 [DOI] [Google Scholar]
  • 117. Newman ME. Finding community structure in networks using the eigenvectors of matrices. Physical review E. 2006;74(3):036104. doi: 10.1103/PhysRevE.74.036104 [DOI] [PubMed] [Google Scholar]
  • 118. Raghavan UN, Albert R, Kumara S. Near linear time algorithm to detect community structures in large-scale networks. Physical review E. 2007;76(3):036106. doi: 10.1103/PhysRevE.76.036106 [DOI] [PubMed] [Google Scholar]
  • 119. Girvan M, Newman ME. Community structure in social and biological networks. Proceedings of the national academy of sciences. 2002;99(12):7821–7826. doi: 10.1073/pnas.122653799 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 120. Reichardt J, Bornholdt S. Statistical mechanics of community detection. Physical review E. 2006;74(1):016110. doi: 10.1103/PhysRevE.74.016110 [DOI] [PubMed] [Google Scholar]
  • 121. Blondel VD, Guillaume JL, Lambiotte R, Lefebvre E. Fast unfolding of communities in large networks. Journal of statistical mechanics: theory and experiment. 2008;2008(10):P10008. doi: 10.1088/1742-5468/2008/10/P10008 [DOI] [Google Scholar]
  • 122. Barrett AB, Seth AK. Practical Measures of Integrated Information for Time-Series Data. PLOS Computational Biology. 2011;7(1):1–18. doi: 10.1371/journal.pcbi.1001052 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 123. Mediano PAM, Seth AK, Barrett AB. Measuring Integrated Information: Comparison of Candidate Measures in Theory and Simulation. Entropy (Basel). 2018;21(1):17. doi: 10.3390/e21010017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 124. Park HJ, Friston K. Structural and functional brain networks: from connections to cognition. Science. 2013;342(6158):1238411. doi: 10.1126/science.1238411 [DOI] [PubMed] [Google Scholar]
  • 125. Sporns O. The non-random brain: efficiency, economy, and complex dynamics. Frontiers in computational neuroscience. 2011;5:5. doi: 10.3389/fncom.2011.00005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 126. Sporns O. Structure and function of complex brain networks. Dialogues in clinical neuroscience. 2022;. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 127. Avena-Koenigsberger A, Misic B, Sporns O. Communication dynamics in complex brain networks. Nature reviews neuroscience. 2018;19(1):17–33. doi: 10.1038/nrn.2017.149 [DOI] [PubMed] [Google Scholar]
  • 128. Tononi G, Sporns O, Edelman G. Measures of degeneracy and redundancy in biological networks. Proceedings of the National Academy of Sciences of the United States of America. 1999;96:3257–62. doi: 10.1073/pnas.96.6.3257 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 129. Hoel EP, Albantakis L, Tononi G. Quantifying causal emergence shows that macro can beat micro. Proceedings of the National Academy of Sciences. 2013;110(49):19790–19795. doi: 10.1073/pnas.1314922110 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 130. Hoel EP. When the Map Is Better Than the Territory. Entropy. 2017;19(5). doi: 10.3390/e19050188 [DOI] [Google Scholar]
  • 131. Klein B, Hoel E. The Emergence of Informative Higher Scales in Complex Networks. Complexity. 2020;2020:1–12. doi: 10.1155/2020/8932526 [DOI] [Google Scholar]
  • 132. Pearson K. LIII. On lines and planes of closest fit to systems of points in space. The London, Edinburgh, and Dublin philosophical magazine and journal of science. 1901;2(11):559–572. doi: 10.1080/14786440109462720 [DOI] [Google Scholar]
  • 133. Ayesha S, Hanif MK, Talib R. Overview and comparative study of dimensionality reduction techniques for high dimensional data. Information Fusion. 2020;59:44–58. doi: 10.1016/j.inffus.2020.01.005 [DOI] [Google Scholar]
  • 134. Dunn OJ. Multiple comparisons among means. Journal of the American statistical association. 1961;56(293):52–64. doi: 10.1080/01621459.1961.10482090 [DOI] [Google Scholar]
  • 135. Kruskal WH, Wallis WA. Use of ranks in one-criterion variance analysis. Journal of the American statistical Association. 1952;47(260):583–621. doi: 10.1080/01621459.1952.10483441 [DOI] [Google Scholar]
  • 136. Mann HB, Whitney DR. On a test of whether one of two random variables is stochastically larger than the other. The annals of mathematical statistics. 1947; p. 50–60. doi: 10.1214/aoms/1177730491 [DOI] [Google Scholar]
  • 137. Wilcoxon F. Probability tables for individual comparisons by ranking methods. Biometrics. 1947;3(3):119–122. doi: 10.2307/3001946 [DOI] [PubMed] [Google Scholar]
  • 138. Andrade C. Multiple testing and protection against a type 1 (false positive) error using the Bonferroni and Hochberg corrections. Indian journal of psychological medicine. 2019;41(1):99–100. doi: 10.4103/IJPSYM.IJPSYM_499_18 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 139. Armstrong RA. When to use the Bonferroni correction. Ophthalmic and Physiological Optics. 2014;34(5):502–508. doi: 10.1111/opo.12131 [DOI] [PubMed] [Google Scholar]
  • 140. Tan L, Guo X, Ren S, Epstein JN, Lu LJ. A computational model for the automatic diagnosis of attention deficit hyperactivity disorder based on functional brain volume. Frontiers in computational neuroscience. 2017;11:75. doi: 10.3389/fncom.2017.00075 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 141. Mashrur FR, Rahman KM, Miya MTI, Vaidyanathan R, Anwar SF, Sarker F, et al. BCI-Based Consumers’ Choice Prediction From EEG Signals: An Intelligent Neuromarketing Framework. Frontiers in human neuroscience. 2022;16:861270. doi: 10.3389/fnhum.2022.861270 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 142. Richhariya B, Tanveer M, Rashid AH, Initiative ADN, et al. Diagnosis of Alzheimer’s disease using universum support vector machine based recursive feature elimination (USVM-RFE). Biomedical Signal Processing and Control. 2020;59:101903. doi: 10.1016/j.bspc.2020.101903 [DOI] [Google Scholar]
  • 143. Mijalkov M, Kakaei E, Pereira JB, Westman E, Volpe G, Initiative ADN. BRAPH: a graph theory software for the analysis of brain connectivity. PloS one. 2017;12(8):e0178798. doi: 10.1371/journal.pone.0178798 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 144. Lehne M, Rohrmeier M, Koelsch S. Tension-related activity in the orbitofrontal cortex and amygdala: an fMRI study with music. Social cognitive and affective neuroscience. 2014;9(10):1515–1523. doi: 10.1093/scan/nst141 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 145. Nickel K, Tebartz van Elst L, Manko J, Unterrainer J, Rauh R, Klein C, et al. Inferior frontal gyrus volume loss distinguishes between autism and (comorbid) attention-deficit/hyperactivity disorder—a FreeSurfer analysis in children. Frontiers in psychiatry. 2018;9:521. doi: 10.3389/fpsyt.2018.00521 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 146. Murdoch BE. The cerebellum and language: historical perspective and review. Cortex. 2010;46(7):858–868. doi: 10.1016/j.cortex.2009.07.018 [DOI] [PubMed] [Google Scholar]
  • 147. Timmann D, Daum I. Cerebellar contributions to cognitive functions: a progress report after two decades of research. The cerebellum. 2007;6:159–162. doi: 10.1080/14734220701496448 [DOI] [PubMed] [Google Scholar]
  • 148. Krain AL, Castellanos FX. Brain development and ADHD. Clinical psychology review. 2006;26(4):433–444. doi: 10.1016/j.cpr.2006.01.005 [DOI] [PubMed] [Google Scholar]
  • 149. Curatolo P, D’Agati E, Moavero R. The neurobiological basis of ADHD. Italian journal of pediatrics. 2010;36(1):1–7. doi: 10.1186/1824-7288-36-79 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 150. Ardila A, Bernal B, Rosselli M, et al. Language and visual perception associations: meta-analytic connectivity modeling of Brodmann area 37. Behavioural neurology. 2015;2015. doi: 10.1155/2015/565871 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 151. Bernard F, Lemée JM, Ter Minassian A, Menei P. Right hemisphere cognitive functions: from clinical and anatomic bases to brain mapping during awake craniotomy part I: clinical and functional anatomy. World neurosurgery. 2018;118:348–359. doi: 10.1016/j.wneu.2018.05.024 [DOI] [PubMed] [Google Scholar]
  • 152. Seghier ML. The angular gyrus: multiple functions and multiple subdivisions. The Neuroscientist. 2013;19(1):43–61. doi: 10.1177/1073858412440596 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 153. Tamm L, Menon V, Reiss AL. Parietal attentional system aberrations during target detection in adolescents with attention deficit hyperactivity disorder: event-related fMRI evidence. American Journal of Psychiatry. 2006;163(6):1033–1043. doi: 10.1176/ajp.2006.163.6.1033 [DOI] [PubMed] [Google Scholar]
  • 154. Fornito A, Zalesky A, Bullmore E. Fundamentals of brain network analysis. Academic press; 2016. [Google Scholar]
  • 155. Sporns O. The human connectome: a complex network. Annals of the new York Academy of Sciences. 2011;1224(1):109–125. doi: 10.1111/j.1749-6632.2010.05888.x [DOI] [PubMed] [Google Scholar]
  • 156. Rudie JD, Brown J, Beck-Pancer D, Hernandez L, Dennis E, Thompson P, et al. Altered functional and structural brain network organization in autism. NeuroImage: clinical. 2013;2:79–94. doi: 10.1016/j.nicl.2012.11.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 157. Keown CL, Datko MC, Chen CP, Maximo JO, Jahedi A, Müller RA. Network organization is globally atypical in autism: a graph theory study of intrinsic functional connectivity. Biological Psychiatry: Cognitive Neuroscience and Neuroimaging. 2017;2(1):66–75. doi: 10.1016/j.bpsc.2016.07.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 158. Rubinov M, Sporns O. Complex network measures of brain connectivity: Uses and interpretations. NeuroImage. 2010;52(3):1059–1069. doi: 10.1016/j.neuroimage.2009.10.003 [DOI] [PubMed] [Google Scholar]
  • 159.Rimland B. Infantile Autism. East Norwalk, CT, US: Appleton-Century-Crofts; 1964.
  • 160. de Marchena AB, Eigsti IM, Yerys BE. Brief Report: Generalization Weaknesses in Verbally Fluent Children and Adolescents with Autism Spectrum Disorder. J Autism Dev Disord. 2015;45(10):3370–3376. doi: 10.1007/s10803-015-2478-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 161. Nguyen DC, Pham QV, Pathirana PN, Ding M, Seneviratne A, Lin Z, et al. Federated learning for smart healthcare: A survey. ACM Computing Surveys (Csur). 2022;55(3):1–37. doi: 10.1145/3501296 [DOI] [Google Scholar]
  • 162. Guan H, Yap PT, Bozoki A, Liu M. Federated learning for medical image analysis: A survey. Pattern Recognition. 2024; p. 110424. doi: 10.1016/j.patcog.2024.110424 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 163. Teo ZL, Jin L, Li S, Miao D, Zhang X, Ng WY, et al. Federated machine learning in healthcare: A systematic review on clinical applications and technical architecture. Cell Reports Medicine. 2024;. doi: 10.1016/j.xcrm.2024.101481 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 164. Soltan AA, Thakur A, Yang J, Chauhan A, D’Cruz LG, Dickson P, et al. A scalable federated learning solution for secondary care using low-cost microcomputing: privacy-preserving development and evaluation of a COVID-19 screening test in UK hospitals. The Lancet Digital Health. 2024;6(2):e93–e104. doi: 10.1016/S2589-7500(23)00226-1 [DOI] [PubMed] [Google Scholar]
  • 165. Crowson MG, Moukheiber D, Arévalo AR, Lam BD, Mantena S, Rana A, et al. A systematic review of federated learning applications for biomedical data. PLOS Digital Health. 2022;1(5):e0000033. doi: 10.1371/journal.pdig.0000033 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 166. Sadilek A, Liu L, Nguyen D, Kamruzzaman M, Serghiou S, Rader B, et al. Privacy-first health research with federated learning. NPJ digital medicine. 2021;4(1):132. doi: 10.1038/s41746-021-00489-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 167. Zhang C, Meng X, Liu Q, Wu S, Wang L, Ning H. FedBrain: A robust multi-site brain network analysis framework based on federated learning for brain disease diagnosis. Neurocomputing. 2023;559:126791. doi: 10.1016/j.neucom.2023.126791 [DOI] [Google Scholar]

Decision Letter 0

Yangsong Zhang

15 Feb 2024

PONE-D-23-39121Multiclass machine learning and deep learning for analysis of fMRI functional connectivity from individuals with Autism Spectrum Disorder and Attention Deficit Hyperactivity DisorderPLOS ONE

Dear Dr. Alves,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Mar 31 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Yangsong Zhang, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

1. When submitting your revision, we need you to address these additional requirements.

Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. We suggest you thoroughly copyedit your manuscript for language usage, spelling, and grammar. If you do not know anyone who can help you do this, you may wish to consider employing a professional scientific editing service. 

Whilst you may use any professional scientific editing service of your choice, PLOS has partnered with both American Journal Experts (AJE) and Editage to provide discounted services to PLOS authors. Both organizations have experience helping authors meet PLOS guidelines and can provide language editing, translation, manuscript formatting, and figure formatting to ensure your manuscript meets our submission guidelines. To take advantage of our partnership with AJE, visit the AJE website (http://aje.com/go/plos) for a 15% discount off AJE services. To take advantage of our partnership with Editage, visit the Editage website (www.editage.com) and enter referral code PLOSEDIT for a 15% discount off Editage services. If the PLOS editorial team finds any language issues in text that either AJE or Editage has edited, the service provider will re-edit the text for free.

Upon resubmission, please provide the following:

The name of the colleague or the details of the professional service that edited your manuscript

A copy of your manuscript showing your changes by either highlighting them or using track changes (uploaded as a *supporting information* file)

A clean copy of the edited manuscript (uploaded as the new *manuscript* file)

3. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. 

When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section.

4. Thank you for stating the following financial disclosure: 

EpilabKI is funded through the Bavarian stated Ministery for Sciences and the Arts research.

Please state what role the funders took in the study.  If the funders had no role, please state: "The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript." 

If this statement is not correct you must amend it as needed. 

Please include this amended Role of Funder statement in your cover letter; we will change the online submission form on your behalf.

5. Thank you for stating the following in the Competing Interests section: 

EpilabKI is funded through the Bavarian stated Ministery for Sciences and the Arts research.

We note that one or more of the authors are employed by a commercial company.

a. Please provide an amended Funding Statement declaring this commercial affiliation, as well as a statement regarding the Role of Funders in your study. If the funding organization did not play a role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript and only provided financial support in the form of authors' salaries and/or research materials, please review your statements relating to the author contributions, and ensure you have specifically and accurately indicated the role(s) that these authors had in your study. You can update author roles in the Author Contributions section of the online submission form.

Please also include the following statement within your amended Funding Statement. 

“The funder provided support in the form of salaries for authors, but did not have any additional role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. The specific roles of these authors are articulated in the ‘author contributions’ section.”

If your commercial affiliation did play a role in your study, please state and explain this role within your updated Funding Statement. 

b. Please also provide an updated Competing Interests Statement declaring this commercial affiliation along with any other relevant declarations relating to employment, consultancy, patents, products in development, or marketed products, etc.  

Within your Competing Interests Statement, please confirm that this commercial affiliation does not alter your adherence to all PLOS ONE policies on sharing data and materials by including the following statement: "This does not alter our adherence to  PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests) . If this adherence statement is not accurate and  there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please include both an updated Funding Statement and Competing Interests Statement in your cover letter. We will change the online submission form on your behalf.

6. When completing the data availability statement of the submission form, you indicated that you will make your data available on acceptance. We strongly recommend all authors decide on a data sharing plan before acceptance, as the process can be lengthy and hold up publication timelines. Please note that, though access restrictions are acceptable now, your entire data will need to be made freely accessible if your manuscript is accepted for publication. This policy applies to all data except where public deposition would breach compliance with the protocol approved by your research ethics board. If you are unable to adhere to our open data policy, please kindly revise your statement to explain your reasoning and we will seek the editor's input on an exemption. Please be assured that, once you have provided your new statement, the assessment of your exemption will not hold up the peer review process.

7. PLOS requires an ORCID iD for the corresponding author in Editorial Manager on papers submitted after December 6th, 2016. Please ensure that you have an ORCID iD and that it is validated in Editorial Manager. To do this, go to ‘Update my Information’ (in the upper left-hand corner of the main menu), and click on the Fetch/Validate link next to the ORCID field. This will take you to the ORCID site and allow you to create a new iD or authenticate a pre-existing iD in Editorial Manager. Please see the following video for instructions on linking an ORCID iD to your Editorial Manager account: https://www.youtube.com/watch?v=_xcclfuvtxQ

Additional Editor Comments:

The authorship should address the issues from the two experts before the further consideration.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: No

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: No

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors used multiclass machine learning and deep learning algorithms to distinguish ASD, ADHD, and TD. The study has a lot of major issues, which I will outline below following the order of the manuscript:

1. The title does not accurately represent their current work. Since the present study was focused on the classification of ASD, ADHD, and TD, there should be more descriptions involving classification.

2. The Abstract of this manuscript is excessively redundant and lacks a concise summary of the article's innovation. The authors should try to rewrite this section by following a specific order and to make it more readable for the readers.

3. The Introduction needs to be roundly restructured so that it clearly reflects the hypothesis, questions, and aim of the study. Besides, in the fifth and seventh paragraphs of the Introduction, necessary contents and explanations are lacking.

4. It seems that the neuroimaging data of ASD and ADHD were obtained from two independent datasets, how did the authors eliminate the effects between different sites?

5. More explanations are needed for the choice of time window. With the current description, it is not known how their network is constructed. I suggest the authors make an effort to develop the Sections, giving more notions on the network generation and the related deep learning algorithms.

6. Likewise, in the section Connectivity matrices, some necessary descriptions are missing.

7. This submission seems to be a mere application of machine learning and deep learning algorithms to ASD and ADHD, the authors should point out clearly the degree of methodological novelty of the current submission w.r.t. the previous work.

8. A brief comparison with the state-of-the-art can be included in the discussion section to highlight the impact of the study.

9. The discussion was too sample, I think there should be more discussion about the physiological meaning of the results and why the current method achieved better classification performance.

10. Please discuss future directions regarding diagnostics and drug trials.

11. Some English typos were noticed, the authors should carefully and thoroughly check the English writing to avoid these typos.

Reviewer #2: First of all, this paper made three classifications on ASD, ADHD and TC, and carried out corresponding analysis on the two diseases. A good classification effect was achieved based on the used data, and the corresponding analysis was carried out. However, the paper still had the following shortcomings:

Disadvantages:

1. The amount of data used is mentioned in the abstract, but it is not clear whether there are 40 subjects in three categories respectively or 40 subjects in three categories altogether.

2. The amount of data used is too small, and the pre-processing part of the data should be detailed.

3. Lack of innovation, less workload, can appropriately increase the amount of data to carry out experiments.

4.LSTM has significantly better performance than SVM, so why not try LSTM for analysis instead of SVM for reducing computation?

5. In Figure 4-b, it is necessary to explain where many subjects come from and how to deal with them.

6.The paper explains the use of transfer entropy to calculate the functional network, and the specific calculation formula can be written as much as possible in this place, and there is also a question that your connectivity matrix is symmetrical in Figure 1, whether it conforms to the connectivity matrix of transfer entropy calculation.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Oct 17;19(10):e0305630. doi: 10.1371/journal.pone.0305630.r002

Author response to Decision Letter 0


14 Apr 2024

Dear all,

We would like to thank you, the reviewer, for carefully reading our manuscript. The comments have improved

the manuscript. We provide point-by-point responses to each comment below, which are attached to the author response pdf.

Best regards,

Authors

Attachment

Submitted filename: Author_response_to_review_comments_PLOS_ONE_ADHD_ASD (3).pdf

pone.0305630.s003.pdf (87.7KB, pdf)

Decision Letter 1

Yangsong Zhang

29 May 2024

PONE-D-23-39121R1Multiclass Classification of Autism Spectrum Disorder, Attention Deficit Hyperactivity Disorder, and Typically Developed Individuals Using fMRI Functional Connectivity AnalysisPLOS ONE

Dear Dr. Alves,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Jul 13 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Yangsong Zhang, Ph.D.

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments:

Some minor issues should be addressed.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: In the chapter 2.2 Data and data preprocessing, the spelling of "ABIDE" is missing "A"(page 5). You should pay attention to these spelling problems and check the words in the whole text as well as sentence problems.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Oct 17;19(10):e0305630. doi: 10.1371/journal.pone.0305630.r004

Author response to Decision Letter 1


1 Jun 2024

We would like to thank you, the reviewer, for carefully reading our manuscript. The comments have improved the manuscript. The response to the reviewers can be found in the attached file named: "Author_response_to_review-version2.pdf"

Attachment

Submitted filename: Author_response_to_review-version2.pdf

pone.0305630.s004.pdf (54.7KB, pdf)

Decision Letter 2

Yangsong Zhang

4 Jun 2024

Multiclass Classification of Autism Spectrum Disorder, Attention Deficit Hyperactivity Disorder, and Typically Developed Individuals Using fMRI Functional Connectivity Analysis

PONE-D-23-39121R2

Dear Dr. Alves,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Yangsong Zhang, Ph.D.

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

The authors addressed all the comments from the reviewers.

Reviewers' comments:

Acceptance letter

Yangsong Zhang

2 Jul 2024

PONE-D-23-39121R2

PLOS ONE

Dear Dr. Alves,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Prof. Yangsong Zhang

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. Connectivity matrices attached in a zip file used to generate the results.

    This section contains the normalized transfer entropy connectivity matrices derived from the authors’ preprocessed BOLD time series of 122 fMRI regions. These matrices served as inputs for the algorithms employed in this study and for extracting complex network measures.

    (ZIP)

    S2 File

    (ZIP)

    pone.0305630.s002.zip (5.9MB, zip)
    Attachment

    Submitted filename: Author_response_to_review_comments_PLOS_ONE_ADHD_ASD (3).pdf

    pone.0305630.s003.pdf (87.7KB, pdf)
    Attachment

    Submitted filename: Author_response_to_review-version2.pdf

    pone.0305630.s004.pdf (54.7KB, pdf)

    Data Availability Statement

    All relevant data are within the article and its Supporting information files. Neuroimaging data were obtained from two large, publicly available datasets: ADHD-200 (http://fcon_1000.projects.nitrc.org/indi/adhd200/) and ABIDE (Autism Brain Imaging Data Exchange; http://fcon_1000.projects.nitrc.org/indi/abide/).


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES