Skip to main content
Frontiers in Human Neuroscience logoLink to Frontiers in Human Neuroscience
. 2017 Jul 25;11:380. doi: 10.3389/fnhum.2017.00380

Multi-Kernel Learning with Dartel Improves Combined MRI-PET Classification of Alzheimer’s Disease in AIBL Data: Group and Individual Analyses

Vahab Youssofzadeh 1,2,*, Bernadette McGuinness 3, Liam P Maguire 1, KongFatt Wong-Lin 1,*
PMCID: PMC5524673  PMID: 28790908

Abstract

Magnetic resonance imaging (MRI) and positron emission tomography (PET) are neuroimaging modalities typically used for evaluating brain changes in Alzheimer’s disease (AD). Due to their complementary nature, their combination can provide more accurate AD diagnosis or prognosis. In this work, we apply a multi-modal imaging machine-learning framework to enhance AD classification and prediction of diagnosis of subject-matched gray matter MRI and Pittsburgh compound B (PiB)-PET data related to 58 AD, 108 mild cognitive impairment (MCI) and 120 healthy elderly (HE) subjects from the Australian imaging, biomarkers and lifestyle (AIBL) dataset. Specifically, we combined a Dartel algorithm to enhance anatomical registration with multi-kernel learning (MKL) technique, yielding an average of >95% accuracy for three binary classification problems: AD-vs.-HE, MCI-vs.-HE and AD-vs.-MCI, a considerable improvement from individual modality approach. Consistent with t-contrasts, the MKL weight maps revealed known brain regions associated with AD, i.e., (para)hippocampus, posterior cingulate cortex and bilateral temporal gyrus. Importantly, MKL regression analysis provided excellent predictions of diagnosis of individuals by r2 = 0.86. In addition, we found significant correlations between the MKL classification and delayed memory recall scores with r2 = 0.62 (p < 0.01). Interestingly, outliers in the regression model for diagnosis were mainly converter samples with a higher likelihood of converting to the inclined diagnostic category. Overall, our work demonstrates the successful application of MKL with Dartel on combined neuromarkers from different neuroimaging modalities in the AIBL data. This lends further support in favor of machine learning approach in improving the diagnosis and risk prediction of AD.

Keywords: Alzheimer’s disease, classification, machine learning, multi-kernel learning, prediction, Australian imaging, biomarkers, lifestyle AIBL

Introduction

Magnetic resonance imaging (MRI) and positron emission tomography (PET) are two imaging data modalities that are routinely used for evaluating changes in the brain associated with Alzheimer’s disease (AD) (Ewers et al., 2011; Bateman et al., 2012). Biomarkers from these neuroimaging data are crucial for identifying early symptoms of AD pathology. For example, structural atrophy measured by MRI, and glucose metabolism or amyloid-β deposition measured by PET scans can be detected almost 15–20 years before the expected symptom onset as compared to 5 years in advance by clinical tests (Bateman et al., 2012; Weiner et al., 2013). In addition, biomarkers from neuroimaging data are more sensitive and reliable measures of AD progression than the cognitive and clinical assessments (Ye et al., 2008).

Previous MR-based studies have found brain atrophy in subcortical regions such as the hippocampal pathway (entorhinal cortex, hippocampus, and posterior cingulate cortex; Frisoni et al., 2010; Bateman et al., 2012) and cortical thickness reduction at vulnerable regions (Desikan et al., 2009; Dickerson et al., 2009) at the earliest stages of the disease. Moreover, longitudinal PET studies with [18F]-fluorodeoxyglucose (FDG) tracer have reported reduced cerebral metabolic rate of glucose (hypometabolism) in bilateral parietotemporal, frontal and posterior cingulate cortices in AD and mild cognitive impairment (MCI) participants with respect to healthy elderly (HE) participants (Mosconi et al., 2005; Nordberg et al., 2010) or in baseline-vs.-follow-up whole-group study (Apostolova et al., 2010). Other PET studies with [11C]-Pittsburgh compound B (PiB) tracer have found an increase of cortical PiB retention in areas known to significantly accumulate amyloid-beta [Aβ] deposits in AD and MCI subjects with respect to HE (Nordberg et al., 2010; Cohen and Klunk, 2014). However, in most cases, one biomarker is not sufficient for an accurate diagnosis or prognosis of the disease because each modality reveals information about different aspects of the underlying pathology (Hinrichs et al., 2011).

Recently, multi-modal imaging has gained popularity by integrating complementary AD characterization, and hence obtaining a more reliable AD biomarker (Cuingnet et al., 2011; Dukart et al., 2011; Zhang et al., 2011). In particular, multi-kernel learning (MKL) is a useful machine learning technique to enhance interpretability and classification accuracy of multi-modal imaging (Wang et al., 2008; Ye et al., 2008; Hinrichs et al., 2009; Zhang et al., 2011; Dai et al., 2012; Segovia et al., 2014). Specifically, the MKL forms an optimal kernel from a linear combination of kernels/features (Sonnenburg et al., 2006; Gönen and Alpaydin, 2011). Importantly, MKL can be easily embedded in a support vector machine (SVM) for high-dimensional pattern classification/recognition (Gönen and Alpaydin, 2011). The SVM relies on the assumption that (two) classes are separable by linear decision boundary (separating hyperplane) in a feature space (transformed features via a non-linear transformation function; Cortes and Vapnik, 1995). MKL simultaneously optimizes weights under a gradient descent algorithm and maximizes the margin in SVM (Rakotomamonjy et al., 2008). Previous works have utilized the MKL method on AD neuroimaging initiative (ADNI) data and reported substantial improvement in classification performance, an accuracy rate of above 90% (Hinrichs et al., 2009, 2011; Zhang et al., 2011; Zhang and Shen, 2012). Tested on ADNI dataset, MKL can outperform SVM by 3%–4% and enable early AD diagnosis e.g., separating converting vs. non-converting MCI (Hinrichs et al., 2011).

Dartel is a suite of tools to enhance inter-subject registration or spatial normalization of anatomical scans, allowing for less smoothing and improving MRI-PET coregistration (Ashburner, 2007). This leads to improved anatomical precision (Bergouignan et al., 2009; Klein et al., 2009; Yassa and Stark, 2009). In addition, improvements of AD classification of the ADNI dataset have been reported using Dartel registration as compared to unified segmentation (Cuingnet et al., 2011). However, previous work has only applied to structural neuroimaging data.

In the present work, we use a multimodal machine-learning framework, utilizing both MKL and Dartel techniques, to enhance multimodal classification accuracy of imaging scans from AIBL dataset. A flowchart of the framework is presented in Figure 1, where both MRI and PET imaging modalities after anatomical coregistration enhancement by the Dartel algorithm are jointly combined via a multi-kernel learning process. Contributions of ROIs are derived by whole-brain MKL weights, and results are compared with the t-contrasts of conventional general linear model (GLM) analysis. In addition, expected ranking values are computed for each data modality, indicating variability (stability) ranking of regions across the folds. Using the Australian imaging, biomarkers, and lifestyle (AIBL) data, our results showed considerable improvement of combined MRI-PET classification accuracy over single modal approach, and correlated strongly with the scores of a commonly used psychological test, the delayed memory recall test. Importantly, our results showed high diagnostic accuracy for individual samples and can potentially predict the likelihood of individual’s stability or conversion to another AD category.

Figure 1.

Figure 1

Schematic overview of the proposed multimodal (imaging) machine-learning framework. See text for detailed descriptions.

Materials and Methods

Data Characteristics

As shown in Table 1, for AD with respect to MCI and HE participants, the average Mini-Mental State Examination (MMSE) score was the lowest (20 ± 2.1), while the average clinical dementia rating (CDR) scores (0.85 ± 0.4) and average delayed memory recall (DeRecall) scores were the highest (1.28 ± 2.0). Expectedly, MCI participants showed lower average MMSE and DeRecall scores and higher average CDR scores with respect to HE subjects. These graded scores fall within the standard range of diagnosis, and hence confirmed the validity of the diagnosis, which will be used as targets in the supervised machine learning process.

Table 1.

Demographic and neuropsychological characteristics of the study population from 58 AD, 108 MCI and 120 HE samples.

Total scans: 286 AD MCI HE
(n = 58; 23 F/35 M) (n = 108; 47 F/61 M) (n = 120; 55 F/65 M)
Mean ± SD Range Mean ± SD Range Mean ± SD Range
Age 74.3 ± 4.1 55–93 75.1 ± 5.3 60–96 72.9 ± 5.1 55–93
MMSE 20 ± 2.1 2–29 27 ± 2.3 20–30 28.7 ± 2.2 24–30
CDR 0.85 ± 0.4 0.5–3 0.4 ± 0.2 0–0.5 0.02 ± 0.23 0–0.5
DeRecall 1.28 ± 2.08 0–8 4.83 ± 3.93 0–17 11.26 ± 3.95 0–22

The numbers refer to subjects. AD, Alzheimer’s Disease, MCI, Mild Cognitive Impairment, HE, Healthy Elderly, MMSE, Mini-Mental State Examination, CDR, Clinical Dementia Ratio; SD, standard deviation.

MRI and PET Data

The imaging data from the AIBL flagship study of ageing (Ellis et al., 2009) dataset was used in this study. Data was collected by the AIBL study group. AIBL study methodology has been reported previously (Ellis et al., 2009; Albrecht et al., 2015; Gupta et al., 2015). Demographic characteristics of the studied population of the AIBL data are shown in Table 1. To test the multi-modal machine-learning framework, subject-matched MRI and Pittsburgh compound B-positron emission tomography (PiB-PET) imaging data clinically diagnosed with AD (= 58), MCI (= 108), and HE (= 120) were analyzed. Among them, there were four MCI converters (2 MCI-to-AD and 2 MCI-to-HE), two AD converters (2 AD-to-MCI) and four HE converters (HE-to-MCI), from baseline (BL) to a later time (month 18 or later), which were used to partially evaluate the regression models.

Although our focus was on imaging data, the relationships with non-imaging data were also investigated. In particular, delayed memory recall (DeRecall) score was tested by the MKL regression model to check how well such non-categorical scores can correlate with the imaging features/kernels. As presented in Figure 2, by means of a one-way ANOVA test, we assessed 33 (29 non-imaging + 4 imaging) features to identify the most distinctive markers among three diagnostic groups of samples. The non-imaging features considered were gender, age, neuropsychology test scores (MMSE, CDR and delayed/immediate recall memory tests), blood test analysis (Apolipoprotein E or ApoE genotypes and hormones including thyroid stim, vitamin B12, red blood cells, nucleated red blood cells, platelets, hemoglobin, mean corpuscular hemoglobin (MCH), MCH concentration, urea nitrogen, serum glucose, cholesterol and creatinine), medical history (psychiatric, neurological, cardiovascular, hepatic, musculoskeletal, endocrine-metabolic, gastrointestinal, renal-genitourinary, smoking, malignancy), laboratory data (thyroid hormone, vitamin B12, red blood cell count, white blood cell count, platelets, hemoglobin, MCH, MCH concentration, urea nitrogen, serum glucose, cholesterol, creatinine). The imaging features were the mean intracranial volume of gray matter (GM), white matter (WM) and cerebrospinal fluid (CSF) brain tissues, and average voxel intensity of PIB-PET scans. The features were normalized in [0, 1] (each divided by its maximum) before applying the ANOVA test to select and rank the top features in terms of significance with respect to AD progression.

Figure 2.

Figure 2

Significant data features ranked by ANOVA test. A total of 8 out of 33 features were identified by ANOVA test from 58 Alzheimer’s disease (AD), 108 mild cognitive impairment (MCI) and 120 healthy elderly (HE) samples. Features/biomarkers were ranked based on their p-values. The biomarkers include clinical dementia ratio (CDR), Mini-Mental State Exam (MMSE), delayed memory recall (DeRecall), immediate memory recall (Immediate Recall), gray matter (GM), average intensity of PiB, cerebrospinal fluid (CSF), and Apolipoprotein E (ApoE) genotype 1. Error bars represent a standard error.

Processing of MRI and PET Scans

A total number of 286 (58 AD, 108 MCI and 120 HE) MRI scans were evaluated using a Dartel-enhanced voxel-based morphometry (VBM) analysis, as interfaced in SPM12 (Wellcome Trust Centre for Neuroimaging, London)1. Scans were acquired with a T1-weighted magnetization-prepared rapid gradient echo (MPRAGE) sequence using 1.5T scanners (repetition time/echo time/inversion time = 2300/3.04/900 ms and flip angle = 9°) in DICOM format. Steps to process the MRI scans were as follows: (1) origins of scans were set to the anterior commissure (AC) brain area; (2) MR scans were segmented into GM, WM and CSF brain tissues. Accordingly, “c1”, “c2” and “c3” NIFTI files were generated. In addition, Dartel imported versions of tissue class images for GM and WM i.e., “rc1” and “rc2” were obtained; (3) the “rc1” and “rc2” images were used to create a template from the mean of all scans/subjects. In addition, they were used to generate flow fields that contain deformation details of scans; (4) GMs (“c1”) were normalized into a Montreal Neurological Institute (MNI) space using the Dartel template (“Template_6”) and flow fields (“u_rc1”). To compensate for any residual due to inter-subject variability and, to increase the signal-to-noise ratio, spatially smoothing was applied to GMs (“c1”) images using a full-width-at-half-maximum (FWHM) Gaussian kernel with a common resolution of 8 mm. The resulted “smwc1” scans were then used for the unpaired two-sample t-test.

An equal number of PiB-PET scans were separately preprocessed. Scans were aligned, co-registered to the Dartel-enhanced average of GM scans (maximization of mutual information), normalized into MNI space and smoothed with FWHM = 8 mm. The resulted scans (image matrix dimension 121 × 145 × 121 with 1.5 × 1.5 × 1.5 mm spacing) were used for the unpaired two-sample t-test, and later for the multi-modal machine-learning process.

An unpaired two-sample t-test was separately applied to structural MR and PiB-PET scans to look for differences between three groups of subjects. A threshold of 0.05 with a family-wise error correction for multiple comparisons at the voxel-level was employed. Comparisons for MRI and PiB-PET were made based on the following contrasts: subtraction of AD from HE (AD-vs.-HE), MCI-vs.-HE and AD-vs.-MCI. In regards to t-statistics, we looked at regions of interests (ROIs) with maximum t-values. For MRI scans, we expected to identify significant differences in (para)hippocampus, temporal gyrus and posterior cingulate cortex regions (Frisoni et al., 2010; Bateman et al., 2012). For PiB-PET scans, we expected to see significant differences at bilateral parietotemporal, frontal and posterior cingulate cortices (Langbaum et al., 2009; Villain et al., 2012).

Multimodal Classification and Regression Analyses

Kernels per modality were built from subject matched whole-brain GM and PiB image scans (of 58 AD, 108 MCI, and 120 HE subjects), as implemented in PRoNTo.v2 (Schrouff et al., 2013b). Kernels were simply the pairwise similarity measures (dot product) between scan pairs per region (LaConte et al., 2005). An anatomical automatic labeling (AAL) atlas consisting of 90 ROIs was used to parcellate the regions (Tzourio-Mazoyer et al., 2002). An optimized MKL technique called “simpleMKL” or sMKL was used to combine multiple kernels of GM and PiB scans. The sMKL is an iterative method that results in a smooth and convex optimization problem (Rakotomamonjy et al., 2008). It works based on a weighted 2-norm regularization and sparsity (of linear combinations of the kernel) controlled by 1-norm constrains. This makes it superior to other similar algorithms in terms of convergence and efficiency. A whole-brain binary mask (provided by GLM analysis of GMs) was applied to both data modalities. The first-level masking discards all uninteresting features such as voxels outside the brain.

A nested cross-validation (CV) with hyperparameter optimization was used for assessment of the generalization error (Marquand et al., 2010). For the inner loop, 10-fold CV on subjects-per-group-out and for the outer loop, leave-one-subject-out techniques were used. The 10-fold CV was chosen for the inner loop since it had fewer folds and reduced the computational time. The sMKL employs a binary SVM for classification. We used a soft-margin hyper-parameter optimization with the best configuration among C = 0.1, 1, 10 and 100. All the C values were tested using a 10-fold cross validation (inner folds), then the best C value was used for the outer loop.

To have an unbiased classification, kernels were mean-centered and normalized) fi = (fif¯i)/σi, fi and σi are the mean and standard deviation of i-th sample, respectively). This is due to fact that kernels can be computed from samples with a different number of features (i.e., regions with different numbers of voxels). A balanced accuracy (BA) = 0.5 × (CA1 + CA2), where CA1 and CA1 were accuracies of class 1 and 2 was used to report the overall performance (Schrouff et al., 2013c). Due to inherent imbalance data, we also reported class accuracies. The AAL atlas was used to construct the sMKL weight maps. The weight maps were the spatial representation of the decision function that defined the level of ROIs contributions to the classification process. In addition to percentage contribution, expected ranking values were computed for each data modality, indicating variability (stability) ranking of the regions across folds (Schrouff et al., 2013a).

Following the classification, an sMKL regression analysis was applied to multimodal data to make predictions about age, diagnosis and psychological (DeRecall) scores of subjects. The regression accuracies were compared with individual data, which were modeled by kernel ridge regression (KRR) method (Shawe-Taylor and Cristianini, 2004). Similar to classification process, a nested cross validation technique was used to report the generalization error. Kernels were mean-centered and normalized. To assess the goodness-of-fit of the regression models the coefficient of determination (r2) based on Pearson’s correlation was computed. Confidence intervals (p-values) generated by non-parametric permutation testing with 5000 randomizations were used to assure low variability in the outputs of classification/regression models.

Results

Based on the ANOVA test, we identified eight features which were significantly linked to AD progression. Specifically, the ranking in terms of the highest significance to AD progression was CDR, DeRecall, immediate memory recall, MMSE, PiB-PET, GM-MRI, CSF-MRI and ApoE genotypes (Figure 2). We henceforth based our study on these relatively more significant features. In particular, with regard to imaging data, we considered GM-MRI and PiB-PET data modalities. By default, as a known key risk factor for AD, we also included age in our (regression) analyses.

GLM Analysis: Group Statistical Analysis of MRI and PET Scans

Group t-statistics of GMs in AD-vs.-HE contrast revealed significant changes in bilateral (left and right) subcortical regions: hippocampus (t = 11.0 and 8.7), parahippocampus (t = 7.0 and 9.5), fusiform gyrus (t = 10.8 and 10.34) as well as in bilateral cortical regions: middle temporal gyrus (t = 8.1 and 7.8), inferior temporal gyrus (t = 7.1 and 6.7) and posterior cingulate cortex (t = 7.4 and 6.8), as shown in Figure 3A. Similar ROIs were found in the hippocampus, fusiform gyrus, middle temporal gyrus and left posterior cingulate cortex for another two contrasts of MCI-vs.-HE and AD-vs.-MCI, but with lower t-values. Group t-statistics of PiB-PET scans for AD-vs.-HE contrast suggested significant differences in the majority of cortical regions e.g., parietotemporal, frontal gyrus and posterior cingulate cortex (Figure 3B). As expected, similar ROIs were suggested by the results of the other two contrasts and with lowest t-values for AD-vs.-MCI. This is consistent with results from several previous works (Frisoni et al., 2010; Nordberg et al., 2010; Hinrichs et al., 2011; Bateman et al., 2012; Cohen and Klunk, 2014; Stam, 2014; Hafkemeijer et al., 2015), and provides confidence in our subsequent analyses.

Figure 3.

Figure 3

Group statistical differences of the group pairs for GM-MRI and Pittsburgh compound B-positron emission tomography (PiB-PET) volumes. The rendered t-contrast maps obtained from unpaired two sample t-test factorial analysis of three group pairs, AD-vs.-HE, MCI-vs.-HE and AD-vs.-MCI for (A) GM and (B) PiB scans.

Multimodal MKL Analysis Enhanced Classification Accuracy

The sMKL was applied to multimodal GM-PiB of 58 AD, 108 MCI and 120 HE subjects. Figures 4A–D show scatter plots (prediction per fold) and their corresponding histograms of the function values obtained for the three contrasts. Note that scatter prediction plot in Figures 4A–C represent the predicted values (x-axis) against the real values or targets (y-axis) values, whereas histogram plots in Figure 4D are the smoothed density versions of the prediction plot that indicate how the function values were distributed. The performance curve with the frequency of selection of each hyper-parameter is shown in Figure 4E. In our data, margins with the SVM parameter C = 102 resulted in a stable model performance across three classification problems. Promisingly, multi-modal analysis yielded the balanced accuracies (BAs) of around 95% for all three classification problems: 95.7% (CA1 = 93.3, CA2 = 98.2) for AD-vs.-HE, 95.81% (CA1 = 91.6, CA2 = 100) for MCI-vs.-HE, and 95.1% (CA1 = 97.2, CA2 = 92.9) for AD-vs.-MCI contrasts, which with respect to single-modal analysis modeled by SVM was considerably higher (SVM for PiB scans yielded BAs of 79.59% in AD-vs.-MCI, 90.07% in AD-vs.-HE, and 81.6% in MCI-vs.-HE, and SVM for GM scans resulted BAs of 89.12% in AD-vs.-MCI, 92.48% in AD-vs.-HE, and 91.3% in MCI-vs.-HE). For ease of comparison, the BAs obtained based on single- (GM or PiB) and multi-modal (GM + PiB) analysis are summarized in Figure 4F supporting the superiority of the multi-modal classification to single-modal analysis.

Figure 4.

Figure 4

AD classification by simple multi-kernel learning (sKML) using combined GM-PiB imaging data. (A–C) Prediction plots (per fold). The decision threshold is displayed by a vertical line at the center of the plot. (D) Corresponding histograms of the function values of three groups modeled by sMKL. (E) Performance curve depending on the hyper-parameter values (C = 0.1, 1, 10, 100) with frequency of selection of each hyper-parameter, for three binary classification problems. (F) A summary of classification accuracies obtained by single and multi-modal data.

Whole-brain model weights obtained by sMKL (per region) from GM modality in the AD-vs.-HE contrast suggested ROIs at left hippocampus (1.85% ROI weight and 2221 voxels), right hippocampus (1.52%, 2296 voxels), left posterior cingulate cortex (1.61%, 4364 voxels), right posterior cingulate cortex (1.54%, 3557 voxels), right parahippocampus (1.43%, 2557 voxels), left parahippocampus (1.30%, 2344 voxels), left inferior occipital (1.22%, 2264 voxels), as in Figure 5A. In addition, sMKL weights from PiB-PET in the AD-vs.-MCI contrast suggested a bilateral temporal gyrus (left with 2.5% and 120 voxels and right with 2.05% and 132 voxels) and bilateral mid-frontal gyrus (left with 1.68%, 1206 voxels, right with 2.23%, 1100 voxels). The ROIs detected during MCI-vs.-HE and AD-vs.-MCI were analogous to that during AD-vs.-HE for each data modality, but with relatively lower weight values (Figures 5B,C). sMKL suggested a greater contribution by GM modality (a contribution level of 74.6 ± 3.1 (mean ± SD) and with an expected ranking of 0.9961 ± 0.3) than the PiB modality (a contribution level of 25.3 ± 4.9 and with an expected ranking of 1.99 ± 1.1) in classifying three groups of subjects.

Figure 5.

Figure 5

Weight (per region) maps modeled by multi-kernel learning (MKL) using GM-MRI data. Rendered MKL weights on a template. The results from a single modality GM-MRI data with an average of 75% contribution to (A) AD-vs.-HE, (B) MCI-vs.-HE and (C) AD-vs.-MCI classification problems. Weights for PiB-PET with lower (25% or less) contribution are not shown.

MKL Multi-Modal Analysis Improves Prediction Accuracy

The sMKL regression analysis applied to combined GM and PiB scans provided a correlation of r2 = 0.86 (p < 0.01) for the estimated diagnosis values, an improved prediction accuracy with respect to data based on individual participant (r2 = 0.72, p < 0.01 using GM and r2 = 0.61, p < 0.01 using PiB, modeled by KRR), as in Figure 6. Yet, we noticed some outliers in the predicted values or mispredicted samples (Figure 7A). Interestingly, 3/4 (2 MCI to AD and 1 MCI to HE) of MCI converters, 1/2 (AD to MCI) of AD converters and 1/4 (HE to MCI) of HE converters were correctly identified by the regression model. We hypothesized that the samples expected to be identified as HE but with predicted values closer to MCI had lower DeRecall scores than the average of the expected group i.e., HE. Conversely, the sample expected to be MCI but with estimated values closer to HE had cognitive scores relatively higher than the average of the MCI group. To test this hypothesis, we examined Pearson correlation between estimated values and their corresponding non-categorical variables, DeRecall and age values. We found a significant correlation between the estimated diagnostic values and the DeRecall scores (r2 = 0.62, p < 0.01), but no (significant) correlation with age values (r2 = 0.01, p = 0.7), as in Figures 7B,C.

Figure 6.

Figure 6

Predictions of diagnosis of individuals. A line prediction plot (predictions overlaid on targets) of diagnosis values of subjects derived from (A) GMs (B) PiB-PET scans modeled by kernel ridge regression (KRR) method and (C) combined GM-MRI and PiB-PET data modeled by sMKL. Proximity of sample data to any colored horizontal line denotes the likelihood of classifying under that particular diagnostic category associated with that line.

Figure 7.

Figure 7

Correlation of estimated diagnosis values (modeled by sMKL) with true DeRecall scores and age values. (A) Line plots of estimated diagnosis values (similar to Figure 6C), (B) true DeRecall scores and (C) true age values. Dashed lines: outlier (transition candidates) samples. Down/upside arrows: correctly detected transitions.

Finally, DeRecall scores were modeled by the sMKL regression for the combined GM-PiB data. Results were compared with a KRR regression model for individual modality data, GM and PiB (Figure 8). The multimodal analysis and sMKL provided the best regression accuracy for all three target values with r2 = 0.53 for DeRecall. This is compared to the lower values obtained from the individual modality data and KRR analysis with r2 = 0.46 and r2 = 0.36 for the DeRecall, using PiB and GM, respectively.

Figure 8.

Figure 8

Predictions of three target values using single and multi-modal data. (A–C) Line prediction plot of delay memory recall scores. Individual GM an PiB modeled by KRR while multimodal GM-PiB modeled by sMKL. The closer a particular predicted data (gray) to the targeted data (in green, red and blue), the better the accuracy of sample data.

Discussion

In the current work, we have applied a machine-learning framework based on multimodal analysis for AD classification and prediction of the AIBL data. Crucially, we utilized the Dartel algorithm to enhance coregistration of structural MRI and PET scans while the MKL technique combines complementary data information from different modalities to improve AD classification and prediction (Ye et al., 2008). In particular, we combined GM and PiB data from the AIBL dataset. At least for the AIBL data, our integrated Dartel-MKL multimodal approach revealed a very high classification accuracy (>95%) for the three binary diagnostic classification problems (Figure 4). Importantly, it could potentially predict the diagnosis of individuals, and their potential transition across diagnostic categories in the future (in 18 months or later), although more data are required to confirm this.

Although previous works have successfully tested the Dartel algorithm for AD classification in combination with MKL, there is no work that has tested on the AIBL data. Further, it was applied to only structural data (Cuingnet et al., 2011). Using voxel-based or cortical thickness features extracted from combined GM, WM and CSF, a classification rate of up to 81% sensitivity and 95% specificity was achieved for four groups of subjects including AD, MCI converters, MCI non-converters and HE. However, compared with GM alone, including all tissue maps provided slight improvement in the classification performance. In our work, we selected GM features as representative of the structural data, supported by the ANOVA test (Figure 2). Moreover, we included PET (PiB) scans to complement the analysis and improve the quality of diagnosis. In another work that used voxel-based features of GM and FDG data, an average classification accuracy of 94% was achieved for three groups of subjects including AD, frontotemporal lobar degeneration and control subjects (Dukart et al., 2011). However, the features used in their multimodal analysis were directly concatenated into a long feature vector and not formally combined. Importantly, the concatenation can increase the dimensionality of the feature space, making the classifier unstable and lead to overfitting. In contrast, MKL provides a unified way to combine heterogeneous data when different types of data cannot be directly concatenated. In addition, it provides more flexibility by using different weights on biomarkers of different modalities (Shen et al., 2014).

Our whole-brain investigations on group statistical differences using GMs and PiB scans (Figure 3) suggested MRI and PET changes in the regions that are known to be affected by AD e.g., temporal gyrus, (para)hippocampus, lingual gyrus, thalamus, posterior cingulate cortex and amygdala, consistent with previous findings (Frisoni et al., 2010; Nordberg et al., 2010; Hinrichs et al., 2011; Bateman et al., 2012; Cohen and Klunk, 2014; Stam, 2014; Hafkemeijer et al., 2015). Our MKL and multimodal analysis provided a high regression accuracy (r2 = 0.86) for AD classification (Figures 6, 7), as a rough measure of multi-class separation. We found high negative correlations between the estimated values and a well-known cognitive test score, DeRecall, with r2 = 0.62 (Figure 7B). However, we found no significant correlation (r2 = 0.02) between the estimated diagnostic category and age (Figure 7C). Importantly, those correlations can justify the sample outliers provided by the regression model. For example, samples that were expected to be identified as HE but estimated with values close to MCI, their DeRecall score were relatively lower than the average of the target group. Conversely, samples expected to be MCI but with estimated values closer to HE their cognitive scores were relatively higher than the average of the expected group values i.e., MCI (Figure 7A). These findings suggest that the proposed framework could potentially help in predicting individual baseline diagnoses and likelihood of stability or conversion. Future work will test this on a larger dataset or on follow-up analysis.

Another multimodal classification method, different from the MKL method, called multimodal support vector classification (SVC) had been introduced as a simple and effective way of combining various data sources (Zhang et al., 2011). Feature selection (not optimization) was performed under a coarse-grid searching via cross validation. Tested for combined GM, FDG and CSF biomarkers from the ADNI data, a high classification accuracy (93.2%) for AD-vs.-HE but a relatively low accuracy (76.4%) for MCI-vs.-HE subjects were achieved. In this work, the features were simply the volume of GM tissue and average intensity of FDG scans in 93 ROIs added with features from CSF biomarkers, i.e., a total of 93 + 93 + 3 features for each participant. In general, region-based features were more intuitive and suitable for post hoc analysis. However, cross similarities of the samples were indirectly generated by a (linear or Gaussian) kernel function. In comparison, the values of the similarity matrix were coarser, representing whole-brain activities, but directly account for the interactions among samples. Nevertheless, for a better evaluation, future work should compare the performance of two SVC and MKL methods under two region-based and whole-brain strategies, using a similar dataset e.g., ADNI or AIBL.

An extension of SVC called support vector regression (SVR) under a multi-modal multi-task learning (M3T) was introduced (Zhang and Shen, 2012). It was tested on the ADNI dataset (similar to Zhang et al., 2011) and provided improved correlations between biomarkers from multimodal data (GM + FDG + CSF) and MMSE (r = 0.69) and AD assessment scale-cognitive subscale (ADAS-Cog) score (r = 0.73), with respect to individual data. In comparison, our sMKL method applied to GM + PiB provided higher correlations between kernels and psychological test scores (r2 = 0.62 for DeRecall). Again, for a better evaluation, future works should compare the two methods under similar conditions and over a similar dataset. Another multimodal approach for AD classification is a classical multidimensional scaling (to generate joint embedding) and random forest-based algorithm (for classification), which was tested on four different modalities including MRI, FDG-PET, CSF biomarkers and categorical genetic information in the ADNI data (Gray et al., 2013). It would perhaps be worthwhile to more closely compare this method with the MKL and multi-kernel SVM methods.

In our study, PiB-PET scans were masked with a whole brain template where only non-informative features from outside the brain were discarded. However, it is known that brain tissues (GM, WM and CSF) can have different uptake patterns (Klunk et al., 2004). Hence this is clearly a limitation of our method. As a solution, an additional normalization, standardized uptake value ratio (SUVR) has been suggested (Raniga et al., 2008; Dore et al., 2012). Under PiB-SUVR normalization, uptake values are normalized by the mean uptake value within a region containing nonspecific binding e.g., cerebellar gray matter (Perani et al., 2014). This leads to better inter-subject variability, suitable for group analysis. However, this normalization requires a careful (tissue) segmentation and brain parcellation. In future work, this will be incorporated as an extension to our current framework to improve performance of the multimodal analysis.

Our study can also be extended in two aspects. First, due to a high level of accuracy offered by our multimodal analysis, subtle changes (mild symptoms) that are evident at an early stage of the disease can be investigated (Davatzikos et al., 2011). This may eventually improve the diagnostic quality. Second, as significant correlation was found between imaging measures with known psychological test scores such as DeRecall, incorporating such features into the machine-learning framework (Figure 1) may potentially further improve the diagnosis. This can be achieved as a contributing (weight) factor for the decision function of the (SVM) classifier or as an independent data modality, similar to the CSF biomarkers in Zhang et al. (2011) to enhance the multimodal analysis.

To conclude, our work has shown that a proper combination of Dartel and sMKL methods applied to the multimodal neuroimaging data in AIBL can substantially improve the classification of AD of not only group samples but also individual samples. There is also a potential application for predicting individual conversions. Hence, this could likely lead to direct clinical applications, for example, in the form of a clinical decision support tool. Our work provides further support in favor of machine learning, particularly MKL-based, approach in improving the diagnosis and risk prediction of AD. The computational approach is also sufficiently general to allow incorporating other e.g., non-imaging data features to further improve the performance.

Software Note

Part of the implementation (preprocessing, DARTEL and MKL analyses) have been unified under a publicly available MATLAB graphical user interface, named Software Integrating NEuroimaging And other Data (SINEAD) accessible at github.com/vyoussofzadeh/SINEAD_tool.

Author Contributions

VY conceptualized the methodological framework and carried out the bulk of analyses for the current study. He prepared the initial manuscript and carried out revisions after consulting with coauthors. BM and LPM advised on analyses, choice of metrics and provided edits on the manuscripts during development. KFW-L led the project and oversaw all aspects of the study, from conceptualization, data analysis, writing and editing this manuscript.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

This work was performed under the Northern Ireland International Health Analytics Centre (IHAC) collaborative network project funded by Invest NI through Northern Ireland Science Park (Catalyst Inc.). KFW-L and LPM were additionally supported by the Northern Ireland Functional Brain Mapping Facility (1303/101154803) funded by Invest NI and the University of Ulster, and KFW-L by COST Action Open Multiscale Systems Medicine (OpenMultiMed) supported by COST (European Cooperation in Science and Technology). The authors would wish to thank Jose Sanchez-Bornot, Xuemei Ding, and the IHAC collaborative network especially Le Roy Dowey, for helpful discussions, and Stephen Lusty and Peter Devine for administrative support. The data used in the preparation of this article was obtained from the Australian Imaging Biomarkers and Lifestyle flagship study of ageing (AIBL) funded by the Commonwealth Scientific and Industrial Research Organization (CSIRO) which was made available at the ADNI database (www.loni.usc.edu/ADNI). The AIBL researchers contributed data but did not participate in analysis or writing of this report. AIBL researchers are listed at www.aibl.csiro.au.

Footnotes

References

  1. Albrecht M. A., Szoeke C., Maruff P., Savage G., Lautenschlager N. T., Ellis K. A., et al. (2015). Longitudinal cognitive decline in the AIBL cohort: the role of APOEε4 status. Neuropsychologia 75, 411–419. 10.1016/j.neuropsychologia.2015.06.008 [DOI] [PubMed] [Google Scholar]
  2. Apostolova L. G., Thompson P. M., Rogers S. A., Dinov I. D., Zoumalan C., Steiner C. A., et al. (2010). Surface feature-guided mapping of cerebral metabolic changes in cognitively normal and mildly impaired elderly. Mol. Imaging Biol. 12, 218–224. 10.1007/s11307-009-0247-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Ashburner J. (2007). A fast diffeomorphic image registration algorithm. Neuroimage 38, 95–113. 10.1016/j.neuroimage.2007.07.007 [DOI] [PubMed] [Google Scholar]
  4. Bateman R. J., Xiong C., Benzinger T. L. S., Fagan A. M., Goate A., Fox N. C., et al. (2012). Clinical and biomarker changes in dominantly inherited Alzheimer’s disease. N. Engl. J. Med. 367, 795–804. 10.1056/NEJMoa1202753 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Bergouignan L., Chupin M., Czechowska Y., Kinkingnéhun S., Lemogne C., Le Bastard G., et al. (2009). Can voxel based morphometry, manual segmentation and automated segmentation equally detect hippocampal volume differences in acute depression? Neuroimage 45, 29–37. 10.1016/j.neuroimage.2008.11.006 [DOI] [PubMed] [Google Scholar]
  6. Cohen A. D., Klunk W. E. (2014). Early detection of Alzheimer’s disease using PiB and FDG PET. Neurobiol. Dis. 72, 117–122. 10.1016/j.nbd.2014.05.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Cortes C., Vapnik V. (1995). Support-vector networks. Mach. Learn. 20, 273–297. 10.1023/A:1022627411411 [DOI] [Google Scholar]
  8. Cuingnet R., Gerardin E., Tessieras J., Auzias G., Lehéricy S., Habert M.-O., et al. (2011). Automatic classification of patients with Alzheimer’s disease from structural MRI: a comparison of ten methods using the ADNI database. Neuroimage 56, 766–781. 10.1016/j.neuroimage.2010.06.013 [DOI] [PubMed] [Google Scholar]
  9. Dai D., Wang J., Hua J., He H. (2012). Classification of ADHD children through multimodal magnetic resonance imaging. Front. Syst. Neurosci. 6:63. 10.3389/fnsys.2012.00063 [DOI] [PMC free article] [PubMed] [Google Scholar]
  10. Davatzikos C., Bhatt P., Shaw L. M., Batmanghelich K. N., Trojanowski J. Q. (2011). Prediction of MCI to AD conversion, via MRI, CSF biomarkers, and pattern classification. Neurobiol. Aging 32, 2322.e19–2322.e27. 10.1016/j.neurobiolaging.2010.05.023 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Desikan R. S., Cabral H. J., Hess C. P., Dillon W. P., Glastonbury C. M., Weiner M. W., et al. (2009). Automated MRI measures identify individuals with mild cognitive impairment and Alzheimer’s disease. Brain 132, 2048–2057. 10.1093/brain/awp123 [DOI] [PMC free article] [PubMed] [Google Scholar]
  12. Dickerson B. C., Bakkour A., Salat D. H., Feczko E., Pacheco J., Greve D. N., et al. (2009). The cortical signature of Alzheimer’s disease: regionally specific cortical thinning relates to symptom severity in very mild to mild AD dementia and is detectable in asymptomatic amyloid-positive individuals. Cereb. Cortex 19, 497–510. 10.1093/cercor/bhn113 [DOI] [PMC free article] [PubMed] [Google Scholar]
  13. Dore V., Fripp J., Bourgeat P., Acosta O., Zhou L., Raniga P., et al. (2012). Longitudinal analysis of cortical thickness in PiB+ and PiB- healthy elderly controls. Alzheimers Dement. 8:P34 10.1016/j.jalz.2012.05.088 [DOI] [Google Scholar]
  14. Dukart J., Mueller K., Horstmann A., Barthel H., Möller H. E., Villringer A., et al. (2011). Combined evaluation of FDG-PET and MRI improves detection and differentiation of dementia. PLoS One 6:e18111. 10.1371/journal.pone.0018111 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Ellis K. A., Bush A. I., Darby D., De Fazio D., Foster J., Hudson P., et al. (2009). The Australian imaging, biomarkers and lifestyle (AIBL) study of aging: methodology and baseline characteristics of 1112 individuals recruited for a longitudinal study of Alzheimer’s disease. Int. Psychogeriatr. 21:672. 10.1017/S1041610209009405 [DOI] [PubMed] [Google Scholar]
  16. Ewers M., Frisoni G. B., Teipel S. J., Grinberg L. T., Amaro E., Jr., Heinsen H., et al. (2011). Staging Alzheimer’s disease progression with multimodality neuroimaging. Prog. Neurobiol. 95, 535–546. 10.1016/j.pneurobio.2011.06.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Frisoni G. B., Fox N. C., Jack C. R., Jr., Scheltens P., Thompson P. M. (2010). The clinical use of structural MRI in Alzheimer disease. Nat. Rev. Neurol. 6, 67–77. 10.1038/nrneurol.2009.215 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Gönen M., Alpaydin E. (2011). Multiple kernel learning algorithms. J. Mach. Learn. Res. 12, 2211–2268. [Google Scholar]
  19. Gray K. R., Aljabar P., Heckemann R. A., Hammers A., Rueckert D., Alzheimer’s Disease Neuroimaging Initiative . (2013). Random forest-based similarity measures for multi-modal classification of Alzheimer’s disease. Neuroimage 65, 167–175. 10.1016/j.neuroimage.2012.09.065 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Gupta V. B., Wilson A. C., Burnham S., Hone E., Pedrini S., Laws S. M., et al. (2015). Follow-up plasma apolipoprotein E levels in the Australian imaging, biomarkers and lifestyle flagship study of ageing (AIBL) cohort. Alzheimers Res. Ther. 7:16. 10.1186/s13195-015-0105-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Hafkemeijer A., Möller C., Dopper E. G. P., Jiskoot L. C., Schouten T. M., van Swieten J. C., et al. (2015). Resting state functional connectivity differences between behavioral variant frontotemporal dementia and Alzheimer’s disease. Front. Hum. Neurosci. 9:474. 10.3389/fnhum.2015.00474 [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Hinrichs C., Singh V., Mukherjee L., Xu G., Chung M. K., Johnson S. C. (2009). Spatially augmented LP boosting for AD classification with evaluations on the ADNI dataset. Neuroimage 48, 138–149. 10.1016/j.neuroimage.2009.05.056 [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Hinrichs C., Singh V., Xu G., Johnson S. C., Alzheimers Disease Neuroimaging Initiative . (2011). Predictive markers for AD in a multi-modality framework: an analysis of MCI progression in the ADNI population. Neuroimage 55, 574–589. 10.1016/j.neuroimage.2010.10.081 [DOI] [PMC free article] [PubMed] [Google Scholar]
  24. Klein A., Andersson J., Ardekani B. A., Ashburner J., Avants B., Chiang M. C., et al. (2009). Evaluation of 14 nonlinear deformation algorithms applied to human brain MRI registration. Neuroimage 46, 786–802. 10.1016/j.neuroimage.2008.12.037 [DOI] [PMC free article] [PubMed] [Google Scholar]
  25. Klunk W. E., Engler H., Nordberg A., Wang Y., Blomqvist G., Holt D. P., et al. (2004). Imaging brain amyloid in Alzheimer’s disease with pittsburgh compound-B. Ann. Neurol. 55, 306–319. 10.1002/ana.20009 [DOI] [PubMed] [Google Scholar]
  26. LaConte S., Strother S., Cherkassky V., Anderson J., Hu X. (2005). Support vector machines for temporal classification of block design fMRI data. Neuroimage 26, 317–329. 10.1016/j.neuroimage.2005.01.048 [DOI] [PubMed] [Google Scholar]
  27. Langbaum J. B. S., Chen K., Lee W., Reschke C., Bandy D., Fleisher A. S., et al. (2009). Categorical and correlational analyses of baseline fluorodeoxyglucose positron emission tomography images from the Alzheimer’s disease neuroimaging initiative (ADNI). Neuroimage 45, 1107–1116. 10.1016/j.neuroimage.2008.12.072 [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Marquand A., Howard M., Brammer M., Chu C., Coen S., Mourão-Miranda J. (2010). Quantitative prediction of subjective pain intensity from whole-brain fMRI data using Gaussian processes. Neuroimage 49, 2178–2189. 10.1016/j.neuroimage.2009.10.072 [DOI] [PubMed] [Google Scholar]
  29. Mosconi L., Tsui W. H., De Santi S., Li J., Rusinek H., Convit A., et al. (2005). Reduced hippocampal metabolism in MCI and AD: automated FDG-PET image analysis. Neurology 64, 1860–1867. 10.1212/01.WNL.0000163856.13524.08 [DOI] [PubMed] [Google Scholar]
  30. Nordberg A., Rinne J. O., Kadir A., Långström B. (2010). The use of PET in Alzheimer disease. Nat. Rev. Neurol. 6, 78–87. 10.1038/nrneurol.2009.217 [DOI] [PubMed] [Google Scholar]
  31. Perani D., Schillaci O., Padovani A., Nobili F. M., Iaccarino L., Della Rosa P. A., et al. (2014). A survey of FDG- and amyloid-PET imaging in dementia and GRADE analysis. Biomed. Res. Int. 2014:785039. 10.1155/2014/785039 [DOI] [PMC free article] [PubMed] [Google Scholar]
  32. Rakotomamonjy A., Bach F. R., Canu S., Grandvalet Y. (2008). SimpleMKL. J. Mach. Learn. Res. 9, 2491–2521. [Google Scholar]
  33. Raniga P., Bourgeat P., Fripp J., Acosta O., Villemagne V. L., Rowe C., et al. (2008). Automated 11C-PiB standardized uptake value ratio. Acad. Radiol. 15, 1376–1389. 10.1016/j.acra.2008.07.006 [DOI] [PubMed] [Google Scholar]
  34. Schrouff J., Cremers J., Garraux G., Baldassarre L., Mourao-Miranda J., Phillips C. (2013a). “Localizing and comparing weight maps generated from linear kernel machine learning models,” in Proceedings—2013 3rd International Workshop on Pattern Recognition in Neuroimaging, PRNI 2013, (Philadelphia, PA: ), 124–127. [Google Scholar]
  35. Schrouff J., Monteiro J., Rosa M. J., Phillips C., Mourao-miranda J., Neurology C., et al. (2013b). “Can we interpret linear kernel machine learning models using anatomically labelled regions?,” in 20th Annual Meeting of the Organization for Human Brain Mapping, (Hamburg: Germany; ). Available online at: http://hdl.handle.net/2268/170848 [Google Scholar]
  36. Schrouff J., Rosa M. J., Rondina J. M., Marquand A. F., Chu C., Ashburner J., et al. (2013c). PRoNTo: pattern recognition for neuroimaging toolbox. Neuroinformatics 11, 319–337. 10.1007/s12021-013-9178-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  37. Segovia F., Bastin C., Salmon E., Górriz J. M., Ramírez J., Phillips C. (2014). Combining PET images and neuropsychological test data for automatic diagnosis of Alzheimer’s disease. PLoS One 9:e88687. 10.1371/journal.pone.0088687 [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Shawe-Taylor J., Cristianini N. (2004). Kernel Methods for Pattern Analysis, Elements. Cambridge, MA: Cambridge University Press. [Google Scholar]
  39. Shen D., Wee C.-Y., Zhang D., Zhou L., Yap P.-T. (2014). “Machine learning techniques for AD/MCI diagnosis and prognosis,” in Machine Learning in Healthcare Informatics, eds Dua S., Acharya U., Dua P. (Berlin, Heidelberg: Springer; ), 147–179. 10.1007/978-3-642-40017-9_8 [DOI] [Google Scholar]
  40. Sonnenburg S., Rätsch G., Schölkopf B. (2006). Large scale multiple kernel learning. J. Mach. Learn. Res. 7, 1531–1565. [Google Scholar]
  41. Stam C. J. (2014). Modern network science of neurological disorders. Nat. Rev. Neurosci. 15, 683–695. 10.1038/nrn3801 [DOI] [PubMed] [Google Scholar]
  42. Tzourio-Mazoyer N., Landeau B., Papathanassiou D., Crivello F., Etard O., Delcroix N., et al. (2002). Automated anatomical labeling of activations in SPM using a macroscopic anatomical parcellation of the MNI MRI single-subject brain. Neuroimage 15, 273–289. 10.1006/nimg.2001.0978 [DOI] [PubMed] [Google Scholar]
  43. Villain N., Chételat G., Grassiot B., Bourgeat P., Jones G., Ellis K. A., et al. (2012). Regional dynamics of amyloid-β deposition in healthy elderly, mild cognitive impairment and Alzheimer’s disease: a voxelwise PiB-PET longitudinal study. Brain 135, 2126–2139. 10.1093/brain/aws125 [DOI] [PubMed] [Google Scholar]
  44. Wang Z., Chen S., Sun T. (2008). MultiK-MHKS: a novel multiple kernel learning algorithm. IEEE Trans. Pattern Anal. Mach. Intell. 30, 348–353. 10.1109/tpami.2007.70786 [DOI] [PubMed] [Google Scholar]
  45. Weiner M. W., Veitch D. P., Aisen P. S., Beckett L. A., Cairns N. J., Green R. C., et al. (2013). The Alzheimer’s disease neuroimaging initiative: a review of papers published since its inception. Alzheimers Dement. 9, e111–e194. 10.1016/j.jalz.2013.05.1769 [DOI] [PMC free article] [PubMed] [Google Scholar]
  46. Yassa M. A., Stark C. E. L. (2009). A quantitative evaluation of cross-participant registration techniques for MRI studies of the medial temporal lobe. Neuroimage 44, 319–327. 10.1016/j.neuroimage.2008.09.016 [DOI] [PubMed] [Google Scholar]
  47. Ye J., Chen K., Wu T., Li J., Zhao Z., Patel R. (2008). “Heterogeneous data fusion for Alzheimer’s disease study,” in 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, (Las Vegas, NV: ), 1025–1033. [Google Scholar]
  48. Zhang D., Shen D., Alzheimer’s Disease Neuroimaging Initiative . (2012). Multi-modal multi-task learning for joint prediction of multiple regression and classification variables in Alzheimer’s disease. Neuroimage 59, 895–907. 10.1016/j.neuroimage.2011.09.069 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Zhang D., Wang Y., Zhou L., Yuan H., Shen D. (2011). Multimodal classification of Alzheimer’s disease and mild cognitive impairment. Neuroimage 55, 856–867. 10.1016/j.neuroimage.2011.01.008 [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Frontiers in Human Neuroscience are provided here courtesy of Frontiers Media SA

RESOURCES