Abstract
Dysfunction of brain structural and functional connectivity is increasingly being recognized as playing an important role in many brain disorders. Diffusion tensor imaging (DTI) and functional magnetic resonance (fMR) imaging are widely used to infer structural and functional connectivity, respectively. How to combine structural and functional connectivity patterns for predictive modeling is an important, yet open, problem. We propose a new method, called Bayesian prediction based on multidimensional connectivity profiling (BMCP), to distinguish subjects at the individual level based on structural and functional connectivity patterns. BMCP combines finite mixture modeling and Bayesian network classification. We demonstrate its use in distinguishing young and elderly adults based on DTI and resting-state fMR data.
Keywords: magnetic resonance imaging, multimodality, brain structural connectivity, brain functional connectivity, classification
Introduction
Predictive modeling has great potential in clinical applications such as diagnosis and prognosis as well as monitoring functional status during and after treatment, and generating an individualized treatment plan.1 Magnetic resonance (MR) imaging is a noninvasive imaging technique to examine brain structure and function. MR-based predictive modeling centers on discriminating subjects at the individual level based on their MR imaging data.2 To the extent that diseases affect brain structure and function, MR imaging should be able to characterize these manifestations, and perhaps lead to early diagnosis and accurate prognosis, or indicate promising therapeutic avenues.
The brain is a complex network of structurally and functionally connected neural modules. The dysfunction of brain networks is recognized as playing an important role in many brain disorders.3 Diffusion tensor imaging (DTI) and functional magnetic resonance (fMR) are widely used for inferring structural and functional connectivity, respectively.4 Several studies have centered on identifying connectivity patterns predictive of clinical status.2,5–7 For example, Dosenbach et al. used support vector machine based multivariate pattern analysis to analyze functional connectivity patterns of 238 subjects, and made predictions about individuals’ brain maturity across development.2 They found that binary support vector machine classification of individuals as either children or adults was 91% accurate.
It is well recognized that combining information from several imaging modalities may improve prediction accuracy. How to combine structural and functional connectivity patterns is an open problem. We propose a new method, called Bayesian prediction based on multidimensional connectivity profiling (BMCP), to distinguish subjects at the individual level based on structural and functional connectivity patterns. In BMCP, the connection between two brain regions p and q, Ep-q, is the variable of interest. For a connection E, the joint probability distribution between its structural and functional connectivity score is modeled as a finite mixture model. We call this mixture model the multidimensional connection signature of E, and we refer to the signatures of a set of connections as their multidimensional connectivity profile, which provides an overall picture of structure-function associations. Let C denote the group-membership variable (case or control, young or elderly). Our goal is to predict C based on a multidimensional connectivity profile. We use Bayesian network (BN) classifiers for this task.
Methods
The overall approach of BMCP is depicted in Figure 1. The input to BMCP includes DTI and resting-state fMR (rs-fMR) data, and the group-membership variable C. BMCP consists of three modules. The first module performs data preprocessing, to generate structural and functional connectivity scores. The second module performs multidimensional connectivity profiling (MCP), in which, we model the joint probability density of structural and functional connectivity scores, based on finite mixture modeling. The third module generates a classifier based on the profile. We base this module on a Bayesian-network representation with inverse-tree structure, to identify a set of connections that are predictive of C; this module then generates a naïve Bayes classifier from these connection variables.
Figure 1.
The architecture of BMCP.
Data preprocessing
The data preprocessing module operates on T1-weighted, DTI, and rs-fMR image data. Algorithms in this pipeline have been widely used in the neuroimaging research community.
The T1-weighted MR image-analysis procedure includes skull stripping, segmentation, and registration. First, the brain extraction tool8 is used to exclude non-brain tissues. Then, the FAST algorithm9 is used for segmentation into gray matter, white matter, and cerebrospinal fluid. Then we normalize the image to the Montreal Neurological Institute (MNI) stereotaxic space using the nonlinear registration algorithm FNIRT in the FMRIB Software Library (FSL).9 Based on the deformation field generated by FNIRT, we parcellate each individual brain into 90 structures defined in the Automated Anatomical Labeling (AAL) template.10
We also use an FSL-based pipeline to process DTI data.11 First, we correct raw DTI data for head movement and eddy-current distortion. Then we fit a diffusion-tensor model for each voxel. For each subject, the labeled brain regions obtained by T1-weighted MR parcellation are registered to the subject’s diffusion space, using the transformation matrix from structural to diffusion space generated by FSL FDT registration; this step generates a brain parcellation defined in the diffusion space. Then we use PROBTRACTX,12 a probabilistic fiber tracking algorithm, to calculate the connectivity score between two brain regions. Based on the PROBTRACTX result, we calculate the number of samples starting with voxels in brain region p and passing through brain region q, which represents the connection strength between regions p and q; we normalize this quantity by the total number of streamline samples. Noting that the density of fiber tracts between two regions, as calculated from DTI data, is often skewed, we apply a log transformation in order to increase normality.13 This transformed value is the structural connectivity score.
Resting-state fMR data preprocessing includes: motion correction, B0 field unwarping, slice-timing correction, skull-stripping, and regression against the motion parameter time courses. We register the base fMR volume to the subject’s T1-weighted MR images. Using the deformation field generated in the registration step, we generate an fMR-space AAL parcellation. After fMR space parcellation, for each AAL structure, we compute the average time series for each subject. Then we calculate Pearson’s correlation coefficient between pairs of time series. We transform each correlation coefficient into a Z-value using Fisher's Z transformation, in order to increase normality.13 This Z-value is the functional connectivity score.
Multidimensional connectivity profiling
The goal of multidimensional connectivity profiling (MCP) is to model the joint probability density of structural and functional connectivity scores, based on finite mixture modeling.14,15 For a connection E, let Yfmr and Ydti represent the functional and structural connectivity scores, respectively. Y = (Yfmr, Ydti). f(Y) is the joint probability density of these two connectivity scores. Finite mixture modeling is an extremely flexible method of modeling unknown distribution shapes, based on a semiparametric framework. In finite mixture modeling, f(Y) is written in parametric form as
(1) |
where g is the number of components in the model, θi is parameters for the ith component, πi (∑πi = 1) are called the mixing proportions, and Ψ{π1, … , πg , θ1 , … , θg} contains all parameters. This mixture model represents the raw data set in a compact form, and can be considered a fingerprint or signature of the raw data set.
Figure 2 is a hypothetical study illustrating the conceptual framework of MCP. In this study, our analysis involves three brain regions: the right anterior cingulum (Cinglum_Ant_R), the right hippocampus (Hippocampus_R), and the left parahippocampal gyrus (ParaHippocampal_L). There are three connections: Cinglum_Ant_R - Hippocampus_R, Cinglum_Ant_R - ParaHippocampal_L, and Hippocampus_R-ParaHippocampal_L. (A-C) show the finite mixture models generated by MCP. We call the finite mixture model for a connection its signature, and we call the collection of these signatures a subject’s connectivity profile. The generated model is descriptive. For example, in (A), there are two clusters for the connection Cinglum_Ant_R - Hippocampus_R: the blue and green points represent high and low anatomical connectivity, respectively. In this example, there is significant variability in functional connectivity in both clusters.
Figure 2.
The conceptual framework of multidimensional connection profiling (MCP). In this example, our analysis involves a fully connected network (the upper portion) with three brain regions: the right anterior cingulum, the right hippocampus, and the left parahippocampal gyrus. There are three connections (thick black lines) in this network. A) The signature of the right anterior cingulum - right hippocampus. B) The signature of the right hippocampus - left parahippocampal gyrus. C) The signature of the right anterior cingulum - left parahippocampal gyrus. Each signature is represented as a finite mixture model. We call the collection of these signatures the connectivity profile.
The most widely used parametric family for fi(y; θi) is the normal family, which leads to a multivariate normal mixture. The component i is the multivariate normal density with mean μi and covariance matrix Σi. We parameterize the covariance matrix using an eigenvalue-decomposition method; Σi is represented as
(2) |
where Di is the orthogonal matrix of eigenvectors, Ai is a diagonal matrix whose elements are proportional to the eigenvalues of Σi, and λi is an eigenvalue. The shape, orientation, and volume of the components are determined by Ai, Di, and λi.
Mixture-model parameters are estimated using the expectation-maximization (EM) algorithm.16 In the E-step, the probability that an instance belongs to a cluster is estimated, conditional on the current parameter estimates. In the M-step, the parameters are updated given the current clustering result. The EM algorithm for mixture modeling is initialized with a model-based hierarchical clustering step.14 Model selection, including the specification of the covariance matrices and the number of clusters, is based on the Bayesian information criterion.17
MCP is a feature-extraction process: for each connection E, MCP generates a vector representing the group to which an instance belongs. In this manner, MCP partitions the instance space. We refer to this vector as the coloring for E. The colorings for all connections are the input to our next step.
BN classifier generation
In this step, our task is to build a predictive model of C based the colorings for all connections.
The input to our algorithm is a training data set D. D = {EC1, … , ECp, C}, where C assumes a value in {+, −}, and ECi is the coloring for connection Ei. The output is a classifier.
Our algorithm is based on BN classification; our rationale for employing BN classifiers is as follows. First, the coloring of a connection is a discrete variable; BN modeling is a powerful approach for describing interactions among discrete variables, and has a formal mathematical foundation for generating models under uncertainty.18 Second, connection-based studies contain more variables than the number of subjects, leading to undersampling. In this case, the resulting model is prone to overfitting (poor generalizability).19 BN classifiers have a variety of variable-selection mechanisms to solve this problem.
We limit our algorithm to BN classifiers with inverse-tree structure (BNCIT) for joint classification and variable selection.20–22 In BNCIT, the group-membership variable C is a leaf node, and predictive variables are parents of C. BNCIT is a computationally efficient method for detecting variables that are jointly predictive of C for high-dimensional data. In BNCIT, we use the Bayesian Dirichlet (BDe) score23,24 to measure the goodness-of-fit of a candidate BN structure to the observed data. Then we search biomarkers of C based on
(3) |
where X is the parent set of C.
However, due to undersampling, the model generated by BNCIT may not be stable to small perturbations of the data D. We employ ensemble learning to improve BNCIT stability. Ensemble learning stabilizes the model-generation process25 and improves classification performance.26–28 The procedure of BNCIT with ensemble learning is as follows. We employ bootstrap resampling to resample D K times, resulting in kth data set Dk. From Dk, we generate a BN model Bk using BNCIT. Based on Bk, we obtain the feature set, Xk(C), which is the parent set of C. We thereby obtain a feature ensemble {X1(C), … , XK(C)}. In the feature-aggregation step, we calculate the frequency of connection, freq(E). That is,
(4) |
where I[condition] = 1 if condition is true; otherwise, I[condition] = 0. We rank connections based on freq(E), and choose the top-ranked features as the aggregated feature set. The number of highly ranked features to be included in the classifier can be determined by internal cross-validation or a threshold. We denote this ensemble-derived aggregated feature set by Xe.
Given Xe, we can construct a naïve Bayes classifier to classify C. We use a naïve Bayes classifier because, given a set of predictors that are informative of C, naïve Bayes manifests excellent prediction performance in empirical studies.29 Frequently, Xe includes more than ten variables; in this case, parameter estimation for a BN with inverse tree structure would be unreliable for undersampled data. Therefore, we adopt the naïve Bayes architecture.
We use leave-one-out cross-validation (LOOCV) to evaluate the generated classifier. In LOOCV, the data set D with K samples is partitioned into K subsets. For each subset, we generate a predictive model based on subjects in this subset, and assess the model based on the remaining subject. The overall performance is the average across all K subsets. To evaluate the classifier, we compute prediction accuracy, sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV).
Experimental Results
We applied BMCP to data from a DTI and rs-fMR study of normal aging.30 This study included 70 participants with normal cognitive function. Exclusion criteria were previous brain surgery, major CNS trauma, or prior documented history of stroke (not including transient ischemic attacks) that resulted in lasting sequelae, active neurological dysfunction, or the use of antipsychotic and/or antiepileptic medications with known neurological side effects. This data set included two groups: young and old. The young adult group had 23 participants, aged 18–38 years, 11 of which were male. The old adult group included 47 participants, ages 65–90 years, 22 of which were male. All subjects completed and signed a consent form, which was approved by the IRBs at Virgina Tech and the Wake Forest University School of Medicine. MR data were collected on a GE 1.5 T scanner with an eight-channel head coil. T1-weighted, rs-fMR, and DTI images were collected for each subject.
Several studies on healthy aging reported the default mode network and the salient network change with age.31 Therefore, we investigated whether the multidimensional signatures of connections involving these two networks can discriminate between the young and old adult subjects at the individual level. There were 350 connections involving the bilateral medial prefrontal gyrus (MFG), which is in the default mode network, and the bilateral anterior cingulate cortex (ACC), which is in the salient network. Our analysis focuses on these connections.
Using BMCP, we generated classifiers to classify C based on multidimensional signatures.
Table 1 lists the classification results. In this table, the label BMCP-s indicates that the top s ranked variables were used to construct the BMCP classifier. We found that BMCP achieved high accuracy for a wide range of s (s = 10, 20, 30, 40, 50). The mean [standard deviation (SD)] of accuracy, sensitivity, specificity, PPV, NPV were 0.900 (0.023), 0.927 (0.024), 0.843 (0.024), 0.924 (0.012), 0.852 (0.045). BMCP’s classification performance was stable across different values of s.
Table 1.
The classification performance for the aging study.
Accuracy | Sensitivity | Specificity | PPV | NPV | |
---|---|---|---|---|---|
BMCP-10 | 0.871 | 0.894 | 0.826 | 0.913 | 0.791 |
BMCP-20 | 0.886 | 0.914 | 0.826 | 0.915 | 0.826 |
BMCP-30 | 0.900 | 0.936 | 0.826 | 0.917 | 0.864 |
BMCP-40 | 0.928 | 0.957 | 0.870 | 0.938 | 0.909 |
BMCP-50 | 0.914 | 0.936 | 0.870 | 0.936 | 0.870 |
Naïve Bayes | 0.857 | 0.783 | 0.894 | 0.783 | 0.894 |
Gaussian Process | 0.786 | 0.783 | 0.787 | 0.643 | 0.881 |
SVM | 0.829 | 0.696 | 0.894 | 0.762 | 0.857 |
We computed the one-sample statistic for a binomial proportion to assess the statistical significance of our results. Random guessing achieved an accuracy of 0.67 (47/70). For our model, we tested the null hypothesis: accuracy = 0.67, versus the alternative hypothesis accuracy ≠ 0.67. P-values were <0.001 for s = 10, 20, 30, 40, 50; we found that our model’s performance was significantly different from that of random guessing.
We compared BMCP to three widely used classification methods: naïve Bayes,32 linear Gaussian process classifier,33 and linear support vector machines (SVMs).34 These classifiers do not have an embedded variable-selection mechanism. We used the filtering method19 for variable selection; in the filtering process, we use Fisher’s exact test to identify variables associated with C, and set the p-value cutoff to 0.01. Thus, if the p-value of a variable is below 0.01, we consider this variable to be predictive of C, and include it in the classifier. Given the selected predictors of C, the next step is to construct a model to predict C. Classifier parameters are tuned using internal fivefold cross-validation. Table 1 lists the classification results of these three classifiers. We found that BMCP achieved superior classification results relative to the other classifiers. The mean classification accuracy for these three classifiers was 0.824, whereas the mean accuracy of BMCP across s was 0.900.
Since BMCP’s classification performance was stable across s, we examined the detected biomarkers based on s = 30. During cross-validation, different connections were selected for different iterations. Among them, 11 were selected for all iterations. The regions involved in these selected connections included MFG, ACC, insula, the superior and middle temporal gyri, and the inferior frontal gyrus; Figure 3 shows these connections. For a connection, its association with C is measured by the p-value based on Fisher’s exact test. In this figure, the intensity of a connection is proportional to –log(p-value).
Figure 3.
Connections characterizing normal aging.
Discussion
We have proposed a method, BMCP, to analyze structural and functional connectivity patterns and use them to predict disease/cognitive states. BMCP has great potential in generating multimodal neuroimaging biomarkers that characterize a disease, clinical state, or cognitive process. We have shown that BMCP is applicable to integrated DTI/rs-fMR data analysis with good results, and that BMCP generates interpretable results.
BMCP combines finite mixture modeling and BN classification. Finite mixture modeling provides a framework to model unknown distribution shapes; and is widely used for clustering and density estimation. Finite mixture modeling has been successfully applied to problems in bioinformatics and medical imaging. One advantage of finite mixture modeling is computational efficiency. For the aging study, generating multidimensional connection signatures required approximately one minute for the algorithm to converge on a workstation with Intel Core 7 2.4 G CPU and 8 G memory. In this paper, we used the EM algorithm to estimate parameters of the mixture model; we used a model-based hierarchical clustering algorithm for initialization of EM parameters, which we found to produce good results.
BN classification has been used to generate classifiers based on a broad range of data, including neuroimaging data.35–37 BNCIT is a type of BN classifier particularly suited to high-dimensional data. In this paper, we incorporated ensemble learning to stabilize BNCIT model generation. Our evaluation demonstrated that BNCIT with ensemble learning can accurately distinguish between the young and old adult subjects at the individual level, based on multidimensional connectivity profiles.
One future extension of this work is to expand BMCP to handle high-dimensional neuroimaging connectivity scores. In the current version of BMCP, we focus on integrated analysis of DTI/rs-fMR data, because DTI and rs-fMR are among the most widely used MR techniques to study brain connectivity. Magnetoencephalography (MEG) is an alternative means of measuring functional connectivity.38 We will extend BMCP to accommodate DTI, rs-fMR, and MEG data. In that case, each connection would have three attributes: a functional connectivity score based on rs-fMR, a functional connectivity score based on MEG, and a structural connectivity score based on DTI. The generated mixture model would represent interactions among these connectivity measures.
Funding
This work is supported by the University of Maryland's Center for Health Informatics and Bioimaging, and the State of Maryland MPower initiative.
Conflict of interest
The authors declare no conflict of interest.
References
- 1.Berner ES. Clinical decision support systems: theory and practice (2nd edition). Springer; 2006.
- 2.Dosenbach NU, Nardos B, Cohen AL, et al. Prediction of individual brain maturity using fMRI. Science 2010; 329(5997): 1358–1361. doi: 10.1126/science.1194144. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Zhang D, Raichle ME. Disease and the brain's dark energy. Nat Rev Neurol 2010; 6(1): 15–28. doi: 10.1038/nrneurol.2009.198. [DOI] [PubMed] [Google Scholar]
- 4.Craddock RC, Jbabdi S, Yan CG, et al. Imaging human connectomes at the macroscale. Nat Methods 2013; 10(6): 524–539. doi: 10.1038/nmeth.2482. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Craddock RC, Holtzheimer 3rd, PE, Hu XP, et al. Disease state prediction from resting state functional connectivity. Magn Reson Med 2009; 62(6): 1619–1628. doi: 10.1002/mrm.22159. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Richiardi J, Eryilmaz H, Schwartz S, et al. Decoding brain states from fMRI connectivity graphs. Neuroimage 2011 May 15; 56(2): 616–26. doi: 10.1016/j.neuroimage.2010.05.081. [DOI] [PubMed] [Google Scholar]
- 7.Shao J, Myers N, Yang Q, et al. Prediction of Alzheimer's disease using individual structural connectivity networks. Neurobiol Aging 2012; 33(12): 2756–2765. doi: 10.1016/j.neurobiolaging.2012.01.017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Smith SM. Fast robust automated brain extraction. Human Brain Mapp 2002; 17(3): 143–155. doi: 10.1002/hbm.10062. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Jenkinson M, Beckmann CF, Behrens TE, et al. FSL. Neuroimage 2012; 62(2): 782–790. doi: 10.1016/j.neuroimage.2011.09.015. [DOI] [PubMed] [Google Scholar]
- 10.Tzourio-Mazoyer N, Landeau B, Papathanassiou D, et al. Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain. NeuroImage 2002; 15(1): 273–289. doi: 10.1006/nimg.2001.0978. [DOI] [PubMed] [Google Scholar]
- 11.Korgaonkar MS, Cooper NJ, Williams LM, et al. Mapping inter-regional connectivity of the entire cortex to characterize major depressive disorder: a whole-brain diffusion tensor imaging tractography study. Neuroreport 2012; 23(9): 566–571. doi: 10.1097/WNR.0b013e3283546264. [DOI] [PubMed] [Google Scholar]
- 12.Behrens TE, Berg HJ, Jbabdi S, et al. Probabilistic diffusion tractography with multiple fibre orientations: What can we gain? Neuroimage 2007; 34(1): 144–155. doi: 10.1016/j.neuroimage.2006.09.018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Rosner B. Fundamentals of biostatistics, Seventh edition Stamford: Brooks/Cole, 2010. [Google Scholar]
- 14.Fraley C, Raftery AE. Model-based clustering, discriminant analysis, and density estimation. J Am Stat Assoc 2002; 97(458): 611–631. doi: 10.1198/016214502760047131. [Google Scholar]
- 15.McLachlan G, Peel D. Finite mixture models: Wiley-Interscience. 2000. doi: 10.1002/0471721182.
- 16.Dempster AP, Laird NM, Rubin DB. Maximum likelihood from incomplete data via the EM algorithm. J Royal Stat Soc 1977; 39(1): 1–38. [Google Scholar]
- 17.Schwarz G. Estimating the dimension of a model. Ann Stat 1978; 6(2): 461–464. doi: 10.1214/aos/1176344136. [Google Scholar]
- 18.Pearl J. Probabilistic Reasoning in Intelligent Systems. Morgan Kaufmann; 1988.
- 19.Duda R, Hart R, Stork DG. Pattern classification, second edition New York: Wiley, 2001. [Google Scholar]
- 20.Chen R, Herskovits EH. A Bayesian network classifier with inverse tree structure for voxel-wise MR image analysis. In the 11th conference of SIGKDD; 2005.
- 21.Chen R, Herskovits EH. Clinical diagnosis based on bayesian classification of functional magnetic-resonance data. Neuroinformatics 2007; 5(3): 178–188. doi: 10.1007/s12021-007-0007-2. [DOI] [PubMed] [Google Scholar]
- 22.Chen R, Herskovits EH. Graphical-model based morphometric analysis. IEEE Trans Med Imaging 2005; 24(10): 1237–1248. doi: 10.1109/TMI.2005.854305. [DOI] [PubMed] [Google Scholar]
- 23.Cooper GF, Herskovits EH. A Bayesian method for the induction of probabilistic networks from data. Machine Learning 1992; 9(4): 309–347. doi: 10.1023/A:1022649401552. [Google Scholar]
- 24.Heckerman D, Chickering DM. Learning Bayesian Networks: The Combination of Knowledge and Statistical Data. Machine Learning 1995; 20: 197–243. [Google Scholar]
- 25.Fred A, Jain AK. Combining multiple clusterings using evidence accumulation. IEEE Trans Pattern Anal Mach Intell 2005; 27(6): 835–850. doi: 10.1109/TPAMI.2005.113. [DOI] [PubMed] [Google Scholar]
- 26.Breiman L. Random Forests. Machine Learning 2001; 45(1): 5–32. doi: 10.1023/A:1010933404324. [Google Scholar]
- 27.Breiman L. Bagging predictors. Machine learning 1996; 26(2): 123–140. doi: 10.1023/A:1018054314350. [Google Scholar]
- 28.Freund Y, Schapire R. Experiments with a new boosting algorithm. Machine Learning: Proceedings of the Thirteenth International Conference 1996, pp. 148–156. [Google Scholar]
- 29.Rash I. An empirical study of the naive Bayes classifier. IJCAI 2001 Workshop on Empirical Methods in Artificial Intelligence; 2001.
- 30.Wyatt CL, Laurienti PJ. Age related atrophy dataset. Virginia Tech and Wake Forest University School of Medicine.
- 31.Ferreira LK, Busatto GF. Resting-state functional connectivity in normal brain aging. Neurosci Biobehav Rev 2013; 37(3): 384–400. doi: 10.1016/j.neubiorev.2013.01.017. [DOI] [PubMed] [Google Scholar]
- 32.Friedman N, Geiger D, Goldszmidt M. Bayesian network classifiers. Machine learning 1997; 29: 131–163. doi: 10.1023/A:1007465528199. [Google Scholar]
- 33.Rasmussen CE, Williams CKI. Gaussian processes for machine learning, Cambridge: MA: The MIT Press, 2006. [Google Scholar]
- 34.Vapnik VN. The nature of statistical learning theory. Springer-Verlag; 1995. doi: 10.1007/978-1-4757-2440-0.
- 35.Chen R, Herskovits EH. Machine-learning techniques for building a diagnostic model for very mild dementia. Neuroimage 2010; 52(1): 234–244. doi: 10.1016/j.neuroimage.2010.03.084. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Chen R, Herskovits EH. Network analysis of mild cognitive impairment. NeuroImage 2006; 29(4): 1252–1259. doi: 10.1016/j.neuroimage.2005.08.020. [DOI] [PubMed] [Google Scholar]
- 37.Morales DA, Vives-Gilabert Y, Gomez-Anson B, et al. Predicting dementia development in Parkinson's disease using Bayesian network classifiers. Psychiatry Res 2013; 213(2): 92–98. doi: 10.1016/j.pscychresns.2012.06.001. [DOI] [PubMed] [Google Scholar]
- 38.Brookes MJ, Hale JR, Zumer JM, et al. Measuring functional connectivity using MEG: methodology and comparison with fcMRI. Neuroimage 2011; 56(3): 1082–1104. doi: 10.1016/j.neuroimage.2011.02.054. [DOI] [PMC free article] [PubMed] [Google Scholar]