Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2022 Feb 1.
Published in final edited form as: Brain Imaging Behav. 2021 Feb;15(1):327–345. doi: 10.1007/s11682-020-00260-3

Brain-predicted age difference score is related to specific cognitive functions: A multi-site replication analysis

R Boyle 1, L Jollans 2, LM Rueda-Delgado 1, R Rizzo 3, GG Yener 4,5,6, JP McMorrow 7,8, SP Knight 8,9, D Carey 9,10, IH Robertson 1,11, DD Emek-Savaş 4,11,12, Y Stern 13, RA Kenny 8,9,14, R Whelan 1,11,*
PMCID: PMC7500565  NIHMSID: NIHMS1625930  PMID: 32141032

Abstract

Brain-predicted age difference scores are calculated by subtracting chronological age from ‘brain’ age, which is estimated using neuroimaging data. Positive scores reflect accelerated ageing and are associated with increased mortality risk and poorer physical function. To date, however, the relationship between brain-predicted age difference scores and specific cognitive functions has not been systematically examined using appropriate statistical methods. First, applying machine learning to 1,359 T1-weighted MRI scans, we predicted the relationship between chronological age and voxel-wise grey matter data. This model was then applied to MRI data from three independent datasets, significantly predicting chronological age in each dataset: Dokuz Eylül University (n=175), the Cognitive Reserve/Reference Ability Neural Network study (n=380), and The Irish Longitudinal Study on Ageing (n=487). Each independent dataset had rich neuropsychological data. Brain-predicted age difference scores were significantly negatively correlated with performance on measures of general cognitive status (two datasets); processing speed, visual attention, and cognitive flexibility (three datasets); visual attention and cognitive flexibility (two datasets); and semantic verbal fluency (two datasets). As such, there is firm evidence of correlations between increased brain-predicted age differences and reduced cognitive function in some domains that are implicated in cognitive ageing.

Keywords: MRI, Brain Ageing, Cognitive Ageing, Cognitive Function, Machine Learning, Biomarkers

Introduction

Longitudinal neuropsychological testing in older adults can be used to detect cognitive decline. However, practice effects can obscure assessment of cognitive ability (Elman et al., 2018), and test performance is affected by subject-level factors such as the individual’s level of comprehension, reading ability, self-efficacy, motivation, fatigue, and fluctuations in concentration (McCaffrey & Westervelt, 1995). In contrast, objective biomarkers are not subject to such biases or patients’ physical limitations (Jollans & Whelan, 2016). An objective biomarker of cognitive ageing would therefore be useful for the timely identification of cognitive decline outside of age-related norms.

Ageing is a process with significant heterogeneity across individuals (McCrory & Kenny, 2018). Consequently, chronological age is not the most accurate marker of an individual’s rate of biological ageing (Sprott, 2010). Ageing biomarkers have been developed that provide additional information about an individual’s health status and life expectancy (Dean & Morgan, 1988). For example, DNA methylation data can estimate epigenetic ageing (‘epigenetic clocks’), reflecting the age of an individual’s tissues or blood cells (Fiorito et al., 2019). Subtracting chronological age from the biological age results in a biologically informative summary score – the predicted age difference – for each individual, which reflects the deviation from typical lifespan trajectories (Richard et al., 2018). This approach has also been applied in neuroimaging, where machine learning can be used to quantify the relationship between structural MRI data and chronological age, in order to estimate an individual’s ‘brain age’. Subtracting chronological age from the estimated ‘brain age’ results in a brain predicted-age difference score (brainPAD, also referred to as brain age gap, brainAGE, Brain-Age Score; (Beheshti, Maikusa, & Matsuda, 2018; Franke, Ziegler, Klöppel, & Gaser, 2010; Schnack et al., 2016) which quantifies how a person’s brain health differs from what would be expected for their chronological age.

BrainPAD is a promising biomarker of general brain ageing as it already satisfies several criteria for ageing biomarkers (Butler et al., 2004). BrainPAD is predictive of mortality (Cole et al., 2018) and of age-sensitive physiological measures, including grip strength, lung function, walking speed and allostatic load (Cole et al., 2018). Moreover, brainPAD has been associated with cognitive impairment (Liem et al., 2017), is negatively correlated with fluid cognitive performance (Cole et al., 2018) and is significantly increased in Alzheimer's disease (AD) and mild cognitive impairment (MCI) (Franke & Gaser, 2012; Gaser et al., 2013; Löwe, Gaser, & Franke, 2016). As such, brainPAD could serve as a cognitive ageing biomarker. However, this potential use of brainPAD is currently limited by a lack of knowledge regarding the exact relationship between brainPAD and specific cognitive functions in healthy individuals.

Studies relating specific cognitive functions and brainPAD have been assessed in solely clinical samples (e.g., Cole et al. (2015), traumatic brain injury), or in mixed samples of clinical groups and healthy controls (e.g., Beheshti et al. (2018); AD, MCI, and healthy controls) and not samples comprised only of healthy adults. As such, the reported associations between brainPAD and specific domains of cognitive function in such studies (Beheshti et al., 2018; Cole et al., 2015) may be skewed towards significance by the inclusion of the clinical samples with typically higher brainPADs. Consequently, these findings may not represent the brainPAD-cognition relationship in normal ageing. For example, Le and colleagues (2018) reported a significant negative correlation between brainPAD and response inhibition and selective attention in a sample of individuals comprised of healthy controls and patients with mood or anxiety disorders, substance use disorder and/or eating disorders. However, significantly increased brainPADs have been reported in mood disorders such as major depression (Koutsouleris et al., 2014) and in substance use disorders such as alcohol dependence (Guggenmos et al., 2017). As both major depression and alcohol dependence are associated with cognitive impairments (Chanraud et al., 2007; McIntyre et al., 2013), the significant brainPAD-cognitive function correlations reported across samples including such populations could be driven by the inclusion of such clinical groups.

The relationship between specific cognitive functions and BrainPAD has also been somewhat obscured by statistical considerations. Recent work has empirically demonstrated that chronological age must be controlled for when testing relationships between brainPAD and cognitive functions (Le et al., 2018; Smith, Vidaurre, Alfaro-Almagro, Nichols, & Miller, 2019). Failure to correct for chronological age can result in false positive findings because some cognitive variables are correlated with chronological age – but not brain ageing – and brainPAD is typically correlated with chronological age (Le et al., 2018). In light of this recent work, it is difficult to interpret studies that did not control for chronological age when investigating the brainPAD-cognition relationship in healthy controls (Franke, Gaser, Manor, & Novak, 2013; Löwe et al., 2016). A second statistical issue is a failure to correct for multiple comparisons. Researchers testing the brainPAD-cognition relationship have tended to carry out multiple statistical tests of the correlation between brainPAD and various cognitive measures. The performance of multiple statistical tests can increase the Type I error and result in false positive findings (Ranganathan, Pramesh, & Buyse, 2016). However, some papers did not control for multiple comparisons when investigating the brainPAD-cognition relationship (Beheshti et al., 2018; Cole, Underwood, et al., 2017). Other studies have investigated the relationship between brainPAD and specific domains of cognitive function while controlling for chronological age and multiple comparisons, but there are conflicting results for most cognitive domains. For example, a significant correlation between verbal fluency and brainPAD was reported by Franke and colleagues (2013) whereas Richard and colleagues (2018) found no association between verbal fluency and brainPAD. We have summarized the brainPAD-cognition findings in Table 1.

Table 1.

Summary of findings on the relationship between brainPAD and cognitive function.

Cognitive Domain Measure Reference Sample n Sig. Sig. in HCs Age adj. MC corr.
General Cognitive Status MOCA (Richard et al., 2018) HC 265
MMSE (Kaufmann et al., 2019) MCI; DEM 921; 707
MMSE (Gaser et al., 2013) MCI 195
CDR
ADAS
MMSE (Löwe et al., 2016) APOE; Non APOE 219; 186 ?1
CDR APOE; Non APOE 219; 186 ?1
ADAS APOE; Non APOE 219; 186 ?1
MMSE (Beheshti et al., 2018) AD; pMCI; sMCI; HC 147; 112; 102; 146 2
CDR 2
ADAS 2
Composite measure3 (Cole, Underwood, et al., 2017) HIVp; HC 161; 102 4 2
Verbal Fluency Composite measure5 (Cole, Underwood, et al., 2017) HIVp; HC 161; 102 4
Composite measure6 (Richard et al., 2018) HC 265
Semantic (Category Fluency Test) (Franke et al., 2013) DM2; HC 98; 87
Phonemic (Letter Fluency Test) (Cole et al., 2015) TBI 89
Processing Speed Composite measure7 (Cole, Underwood, et al., 2017) HIVp; HC 161; 102 4
Composite measure8 (Richard et al., 2018) HC 265
TMT-A (Cole et al., 2015) TBI 90
CRT TBI 66
Processing Speed, Visual Attention, and Cognitive Flexibility Composite measure9 (Cole, Underwood, et al., 2017) HIVp; HC 161; 102 4
TMT-B (Cole et al., 2015) TBI 90
Visual Attention and Cognitive Flexibility TMT-B minus TMT-A (Cole et al., 2015) TBI 90
Response Inhibition and Selective Attention D-KEFS CWIT Composite measure 10 (Richard et al., 2018) HC 265
D-KEFS CWIT (Inhibition vs Color Naming – scaled) (Le et al., 2018) HC, MOOD/ANX, SUD, ED 489
D-KEFS CWIT (Inhibition/Switching) TBI 89
D-KEFS CWIT (Inhibition/Switching minus Baseline Stroop performance) TBI 89
Sustained Attention Composite measure9 (Cole, Underwood, et al., 2017) HIVp; HC 161; 102 2
Verbal Episodic Memory Composite measure11 (General) (Cole, Underwood, et al., 2017) HIVp; HC 161; 102 2
CVLT Immediate Recall, CVLT Delayed Recall, CVLT Learning 1-5, (all tested separately) (Richard et al., 2018) HC 265
People Test (Immediate) (Cole et al., 2015) TBI 90
Working Memory Composite measure12 (Richard et al., 2018) HC 265
Blocked Verbal N-back Task (Scheller et al., 2018) HC 34 n/a13
Motor Function Composite measure14 (Cole, Underwood, et al., 2017) HIVp; HC 161; 102 2
Intelligence WASI Similarities (Abstract verbal reasoning) (Cole et al., 2015) TBI 90
WASI Matrix Reasoning (Non-verbal reasoning) TBI 88
Composite measure15 (Fluid-type intelligence) (Cole et al., 2018) HC 669
Moray House Test (Childhood IQ)

Note: Sig: results were statistically significant; Sig. in HC: results were statistically significant in healthy controls; Age adj.: results were adjusted for age; MC corr.: results were corrected for multiple comparisons. MOCA = Montreal Cognitive Assessment; MMSE = Mini Mental State Examination, MCI = Mild Cognitive Impairment, DEM = Dementia; CDR = Clinical Dementia Rating Scale, AD = Alzheimer’s Disease, ADAS = Alzheimer’s Disease Assessment Scale, HC = Healthy Controls, APOE = APOE e4 carrier, Non APOE = APOE e4 non-carrier, sMCI = Stable MCI, pMCI = Progressive MCI, HIVp = HIV-positive, DM2 = Diabetes Mellitus Type 2, TBI = Traumatic Brain Injury, TMT-A = Trail Making Test A (Time to complete), CRT = Choice Reaction Time Task (Median reaction time), TMT-B = Trail Making Test B (Time to complete), D-KEFS CWIT = Delis-Kaplan Executive Function System D Color-Word Interference Test, MOOD/ANX = Mood/Anxiety Disorder, SUD = Substance use Disorder, ED = Eating Disorder, CVLT = California Verbal Learning Test, WASI = Weschler Abbreviated Scale of Intelligence.

1

Awaiting clarification from authors – Multiple comparison correction not outlined in relation to brainPAD-cognition tests but used elsewhere in paper.

2

Finding not corrected for multiple comparison but likely would have survived Bonferroni correction so not affected by lack of correction.

3

Average of average standardised t-scores (adjusted for age, sex, education) across domains of verbal fluency, processing speed, executive function, memory, attention, and motor function

4

T-scores controlled for effect of age on cognitive scores, but relationship between brainPAD and age was not controlled for, so not fully adjusted for age.

5

Average of standardised t-scores (adjusted for age, sex, education) from Category Fluency and Letter Fluency tests

6

Cluster measure combining Phonological Flow and Semantic Flow measures from CABPad (Willer, Pedersen, Forchhammer, & Christensen, 2016)

7

Average of standardised t-scores (adjusted for age, sex, education) from TMT-A, WAIS-III Digit Symbol and Symbol Search, and Stroop Colour-Word Test

8

Cluster measure combining processing speed parameters based on the Theory of Visual Attention obtained from test battery using CABPad

9

Average of standardised t-scores (adjusted for age, sex, education) from TMT-B and WCST (Number of total errors, perseverative errors and responses)

10

Cluster measure combining scores from the Colour-Naming, Reading, Inhibition, and Inhibition/Switching trails of the D-KEFS CWIT

11

Average of standardised t-scores (adjusted for age, sex, education) from Rey Auditory Verbal Learning test and WMS-IV Visual Reproduction

12

Cluster measure combining measures from working memory test of CABPad

13

Only one test of brainPAD-cognition relationship conducted so multiple comparison correction not necessary

14

Average of standardised t-scores (adjusted for age, sex, education) from Grooved Pegboard and Finger Tapping tasks

15

Index derived from a principal components analysis of WASI-III Letter-number sequencing, digit span backwards, matrix reasoning, block design, digit symbol coding, symbol search

The first step in generating a brainPAD score is creating a feature set of neuroimaging data which is correlated with chronological age. Neuroimaging data have high dimensionality, which can result in overfitting and overoptimistic predictions (Whelan & Garavan, 2014). Brain age prediction models thus rely on feature engineering techniques such as principal components analysis (PCA; Franke et al., 2010; Gutierrez Becker, Klein, & Wachinger, 2018) or even dot products of different features (e.g. vectors of GM and white matter (WM) voxels as in Cole et al., 2015; Cole, Ritchie, et al., 2018; Cole, Underwood, et al., 2017) in order to reduce the dimensionality (Mwangi, Tian, & Soares, 2014). These techniques map the original variables onto a feature space (in effect, ‘new’ variables) typically using linear transformations in the case of dot products (Snyder, Mika, Burke, & Müller, 2013), although non-linear transformation may also be used for kernel methods (Honeine & Richard, 2009; Kwok & Tsang, 2004). While these models create generalizable and accurate predictions, this may come at the cost of reduced interpretability of the contributions of the features (Bunea et al., 2011; Mateos-Pérez et al., 2018), which is important for assessing the neurobiological validity of the model (Woo, Chang, Lindquist, & Wager, 2017) and to identify specific brain areas for further investigation (Scheinost et al., 2019). Due to the importance of interpretability in neuroimaging, unlike with other data (e.g. credit card transactions for fraud detection), the application of machine learning to MRI does not necessarily involve the goal of achieving the highest accuracy (Mateos-Pérez et al., 2018). While methods do exist for projecting the ‘new’ variables back from the feature space to the input space (Honeine & Richard, 2009; Kwok & Tsang, 2004; Snyder et al., 2013) thus enabling interpretability of models employing dot products, PCA or kernel methods, these methods are not always implemented and/or reported in brain-age papers (Cole et al., 2015, 2018; Cole, Poudel, et al., 2017; Gaser et al., 2013; Gutierrez Becker et al., 2018; Nenadić, Dietzek, Langbein, Sauer, & Gaser, 2017). In contrast, penalized regression methods (e.g., the Elastic Net; Zou & Hastie, 2005) do not require the back-projection of coefficients from feature space to input space and therefore have good interpretability, particularly when less complex feature sets are used (Luo et al., 2019). GM data is particularly well-suited for age prediction as GM volume linearly declines with age (but cf. Fjell et al., 2013) whereas WM volume has a less straightforward relationship with age, as it doesn’t decline significantly until middle age (Farokhian, Yang, Beheshti, Matsuda, & Wu, 2017; Ge et al., 2002). The Elastic Net is a machine learning model well-suited to the high dimensionality and multicollinearity inherent in neuroimaging data as shown by the finding that it produced the most consistent predictions as compared to various other models over datasets with varying sample-, feature set-, and effect-sizes (Jollans et al., 2019).

A final challenge in the development of neuroimaging biomarkers, or neuromarkers, is ensuring the generalisability of the neuromarker to new data. For practical reasons, cross-validation, where a dataset is split into a training set and a test set (Varoquaux et al., 2017), is often used as an estimate of model accuracy for new data (Jollans & Whelan, 2018; Scheinost et al., 2019). However, cross-validation accuracy estimates are often optimistically biased and can vary considerably (Varoquaux et al., 2017), particularly when preprocessing and feature selection are carried out on the entire dataset before splitting it into training and test sets (Dwyer, Falkai, & Koutsouleris, 2018; Woo et al., 2017). As such, the gold-standard for assessing the external validity and generalisability of a neuromarker is by testing how the model performs on a completely independent held-out dataset (Jollans & Whelan, 2018). While various brainPAD studies have externally validated their models (Beheshti et al., 2018; Cole et al., 2015, 2018; Cole, Underwood, et al., 2017; Franke et al., 2010; Gutierrez Becker et al., 2018; Lancaster, Lorenz, Leech, & Cole, 2018; Liem et al., 2017; Madan & Kensinger, 2018; Varikuti et al., 2018), only a few studies have reported model performance in terms of accuracy (i.e., correlation or mean absolute error between brain-predicted age and chronological age) on the external validation dataset (Cole et al., 2015; Lancaster et al., 2018; Liem et al., 2017; Madan & Kensinger, 2018). This does not necessarily cast doubt on the validity of the models whose accuracy is reported in terms of internal cross-validation performance. However, not reporting the external validation performance limits the interpretation of the accuracy and generalisability of various brainPAD models as typically performance will be lower in the external validation dataset.

In order to clarify the unclear relationship between brainPAD and specific domains of cognitive function, we aimed to 1) establish an interpretable model of brainPAD using the Elastic Net with GM voxel-wise data, 2) externally validate this model in three independent datasets, and 3) to establish the domains of cognitive function that are reliably correlated with brainPAD across different datasets.

Methods

Study Design

The present study used data from open-access repositories to form a training set in which a machine learning model was trained. Data from three separate datasets (Dokuz Eylül University (DEU); Cognitive Reserve/Reference Ability Neural Network (CR/RANN) and The Irish Longitudinal Study on Ageing (TILDA)) were then used to form three external validation sets in which the machine learning model was validated and the relationship between brainPAD and cognitive function was investigated. In all cases, the data was collected prior to conception and design of the present study. The target population were healthy adults.

Participants

Training Set

The data were comprised of MRI scans from 1,359 healthy adults (mean age 40.04 years, SD = 17.78 years, range = 18.00 – 88.36 years; 855 females) drawn from various open-access data repositories (see Table S.1 in Supplementary Info). Inclusion criteria for the training cohort were: over 18 years old, age and gender data available, and not diagnosed with any neurological, psychiatric or major medical conditions.

Test Set 1 –DEU

The first test set was comprised of 175 community-dwelling adults (mean age = 68.95 years, SD = 8.59 years; range = 47.56 – 93.51 years; 104 females) recruited as part of a study conducted at Dokuz Eylül University, Izmir, Turkey. Exclusion criteria included history of neurological or psychiatric diseases, use of psychotropic drugs including cholinesterase inhibitors, traumatic brain injury, history of stroke, drug and/or alcohol addiction and uncontrolled systemic diseases.

Test Set 2 – CR/RANN

The second test set was comprised of 380 community-dwelling adults (mean age = 52.41 years, SD = 17.09 years; range = 19 – 80 years; 210 females) who participated in the Cognitive Reserve/Reference Ability Neural Network study (CR/RANN; (Stern, Gazes, Razlighi, Steffener, & Habeck, 2018; Stern et al., 2014). These participants were screened for MRI contraindications, hearing and visual impairments, medical or psychiatric conditions, and dementia and MCI. Further inclusion criteria were a score of over 135 on the Mattis Dementia Rating Scale (Jurica, Leitten, & Mattis, 2001), a reading level at least equivalent to the US 4th grade, and minimal complaints of functional impairment.

Test Set 3 –TILDA

The third test set was comprised of an MRI subset of a nationally representative longitudinal cohort study of community-dwelling adults in Ireland (B. J. Whelan & Savva, 2013). From an initial subset of 502 participants, participants were excluded due to missing a portion of the cerebellum (n = 2), a history of Parkinson’s disease, stroke, or transient ischemic attack (n = 11) and no cognitive data (n= 2). The final test set was comprised of MRI data from 487 participants (mean age = 68.6 years, SD = 7.21 years; range = 50 – 88 years; 260 females).

MRI data acquisition

Training Set

A range of T1-weighted MRI scans from different scanners and using different protocols were used as the training set (see Table S.1 in Supplementary Info).

Test Set 1 – DEU

DEU participants underwent a 10 minute T1 scan in a 1.5 T Philips Achieva scanner as part of a larger 20-min MRI battery. Two separate protocols were used for scans included here. The Alzheimer’s Disease Neuroimaging Initiative (ADNI) T1 protocol was followed for 126 scans using the turbo field echo sequence with the following parameters: number of slices = 166, FOV = 240mm3, matrix size = 256×256, slice thickness = 1 mm, slice gap = 0 mm, TR = 9 ms, TE = 4 ms. For 49 scans, a local protocol using a gradient echo sequence was followed with the following parameters: FOV = 230mm3, matrix size = 400×512, slice thickness = 1 mm, slice gap = 0 mm, TR = 25 ms, TE = 6 ms.

Test Set 2 – CR/RANN

CR/RANN participants underwent a 5 minute T1 MPRAGE scan in a 3T Philips Achieva scanner as part of a larger 2-hr imaging battery. The following parameters were used: FOV = 256×256×180 mm, matrix size = 256×256, slice thickness = 1 mm, slice gap = 0 mm, TR = 6.5 ms, TE = 3 ms.

Test Set 3 – TILDA

TILDA participants underwent a 5 minute 24 seconds T1 MPRAGE scan in a 3T Philips Achieva scanner as part of a larger 45-min MRI battery. The following parameters were used: FOV = 240×240×162mm3, matrix size = 288×288, slice thickness = 0.9 mm, slice gap = 0 mm, TR = 6.7 ms, TE = 3.1 ms.

MRI pre-processing

All images were preprocessed using SPM12 (University College London, London, UK). Prior to processing, all scans were automatically approximately reoriented (see Supplemental Information; MRI pre-processing) to a canonical SPM template. All scans were then visually inspected for good orientation and gross artefacts before preprocessing. In the test set, badly oriented scans were manually reoriented before preprocessing. In both training and test sets, each individual dataset was preprocessed in a separate batch. Bias correction was applied to image which were then segmented into GM, WM, and CSF. Segmented GM images were non-linearly registered to a custom template, using SPM’s DARTEL. Images were then affine registered to MNI space (1 mm3) and resampled with modulation to preserve the total amount of signal from each voxel. Images were smoothed with a 4 mm full-width at half maximum Gaussian kernel. Finally, images were visually inspected for accurate segmentation. The code used to auto-reorient and preprocess the MRI data is available at https://github.com/rorytboyle/brainPAD.

Machine learning

Data preparation

GM images were resized to 2 mm3 voxels and individual voxel values were extracted from each image. A threshold was applied such that a voxel was retained if it had GM density > 0.2 in that voxel across all 1,359 training set images. After thresholding, the training data consisted of 1,359 images, each with 54,869 voxels.

Machine learning model

The goal of the training phase was to construct a generalizable model that could predict chronological age from GM data. In order to increase generalizability, a data resampling ensemble approach was used. That is, 500 participants, with a 50:50 gender ratio, were randomly sampled without replacement from the training data to form a nested training set. This process was repeated 25 times, creating 25 nested training sets. Each nested training set (500 participants x 54,869 voxels), was used as the input to a regularized linear regression model (Elastic Net), with 10-fold cross-validation (CV), to predict the chronological age of each participant (see Supplementary Info for further information on the machine learning model). The performance of the model was quantified using the mean of each of the 25 nested models’ Pearson’s correlation between chronological age and predicted age (r), total variance explained (R2), mean absolute error (MAE), and the weighted MAE. The weighted MAE is equal to the MAE divided by the age range of the sample tested and is a more suitable metric for comparing the MAE of brainPAD models across studies as it accounts for the impact of a sample’s age range on prediction accuracy (Cole, Franke, & Cherbuin, 2019). A lower weighted MAE reflects greater accuracy.

Application to independent test sets

First, the average coefficient value for each voxel across all folds in all 25 training models was calculated, resulting in a vector of length 54,869. For each independent test set, the mean coefficient values were multiplied by the voxels’ GM density values and the product was summed to create a brain-age prediction for each participant. To correct for the proportional bias in the model, the prediction was added to the intercept of the training set, and the result was then divided by the slope of the training set. This correction does not affect the relationship between brainPAD and outcome measures but scales the data correctly so that brainPAD scores can be interpreted in units of years proportional to a person’s chronological age. Similar corrections have been applied in other brainPAD models (Cole et al., 2018). BrainPAD was calculated by subtracting chronological age from the corrected predicted age, hence, a positive brainPAD value indicates a brain-predicted age that exceeds the participant’s chronological age, suggesting accelerated brain ageing. The code used to make brain-age predictions and calculated brainPAD scores for independent test sets is available at https://github.com/rorytboyle/brainPAD.

Cognitive function measures

Each of the three datasets contained a wide range of cognitive measures. However, as the three datasets were completely independent of one another, and all data collection was completed prior to conception and design of the present study, different cognitive measures were used across the datasets. For the purposes of the present study, a cognitive measure was selected for analysis if it assayed a cognitive domain that was assessed in at least one other dataset. For example, the AMNART and NART assessed premorbid intelligence in CR/RANN and TILDA respectively so both measures were selected for analysis and considered as ‘comparable’ measures. The cognitive domains assessed by each measure were decided with reference to the literature as outlined in the Supplementary Information. Across all three datasets, 17 common cognitive domains were identified (see Table 2 for list of cognitive domains and cognitive measures used and Supplementary Information for detailed descriptions of each cognitive measure).

Table 2.

Cognitive measures available across each dataset in comparable cognitive domains

Cognitive Domain(s) DEU Measure (N) CR/RANN Measure (N) TILDA Measure (N)
General Cognitive Status MMSE (total score) (172) DRS (370) MMSE (485)
Premorbid Intelligence n/a AMNART (362) NART (486)
Phonemic Verbal Fluency KAS Test (137) CFL Test (360) n/a
Semantic Verbal Fluency Animals Test (175) Animals Test (361) Animals Test (487)
Processing Speed TMT A (93) TMT A (361) CTT 1 (487)
Processing Speed, Visual Attention, Cognitive Flexibility TMT B (84) TMT B (357) CTT 2 (482)
Visual Attention, Cognitive Flexibility TMT B minus TMT A (84) TMT B minus TMT A (357) CTT 2 minus CTT 1 (482)
Cognitive Flexibility WCST Perseverative Error Percentage (50) WCST Perseverative Error Percentage (327) n/a
Response Inhibition, Selective Attention Stroop (Turkish Capa version; (Emek-Savaş, Yerlikaya, Yener, & Öktem, 2019) Interference Score - Time (150) Stroop (Golden version; Golden, 1978) Interference Score - Words (359) n/a
Sustained Attention (Errors of Commission) n/a PVT False Alarms (176) SART Errors of Commission (482)
Sustained Attention (Reaction Time) n/a PVT Median Reaction Time (176) SART Coefficient of Variation in Reaction Time (479)
Verbal Episodic Memory (Immediate) OVMPT Immediate Recall (175) SRT Total Score (360) Immediate Recall (487)
Verbal Episodic Memory (Delayed) OVMPT Delayed Recall (175) SRT Delayed Recall (360) Delayed Recall (487)
Verbal Episodic Memory (Learning) OVMPT Total Learning Score (175) SRT Consistent Long Term Retrieval (360) n/a
Working Memory WMS-R Digit Span Forward Test (171) WAIS-III Letter Number Sequencing Test (360) n/a
WMS-R Digit Span Backward Test (170)
Visuospatial Ability BLOT (80) WAIS-III Block Design Test (356) n/a

Note: see Table S.4 for full information on each measure; MMSE = Mini-mental state examination (Folstein, Folstein, & McHugh, 1975); DRS Total Score = Mattis Dementia Rating Scale-2 – Total Score (Jurica et al., 2001); NART = National Adult Reading Test (Nelson & Willinson, 1982); AMNART = American National Adult Reading Test (Grober & Sliwinski, 1991); CTT = Colour Trails Test (D’Elia, Satz, Uchiyama, & White, 1996); TMT = Trail Making Test (Reitan, 1955); WCST = Wisconsin Card Sorting Test (Heaton, Chelune, Talley, Kay, & Curtiss, 1993); SART = Sustained Attention to Response Test (Robertson, Manly, Andrade, Baddeley, & Yiend, 1997); PVT = Psychomotor Vigilance Task (Dorrian, Rogers, & Dinges, 2005); OVMPT = Öktem Verbal Memory Processes Test (Öktem, 1992); SRT = Selective Reminding Test (Buschke & Fuld, 1974); WMS-R = Wechsler Memory Scale (Wechsler, 1987); WAIS-III = Wechsler Adult Intelligence Scale – Third Edition (Wechsler, 1997); BLOT = Benton’s Judgement of Line Orientation Test (Benton, Varney, & Hamsher, 1978)

Statistical Analysis

The statistical analysis was conducted using the following procedure:

  1. Correlate. Within each independent test set, partial Spearman’s rank order correlations were conducted between brainPAD scores and cognitive measures, controlling for chronological age and sex. Sex was adjusted for to account for a significant sex difference in brainPADs (p < 0. 0001), see Supplementary Results for further detail.

  2. Replicate. For findings replicated in multiple datasets, the probability of obtaining p-values by chance was calculated by random-label permutation (see Supplementary Methods for further detail). Briefly, this involved randomly shuffling brainPAD scores, conducting Spearman’s partial correlations between randomly shuffled brainPAD scores and the cognitive dependent variables, controlling for age and sex. This process was repeated one million times. The number of times in which all random p-values were more extreme (i.e. smaller) than the actual p-values was summed and divided by one million to obtain the probability of the finding replicating across multiple datasets by chance. Replicated findings were deemed significant if this probability was less than .05.

  3. Correct for multiple comparisons. All other correlations were then corrected for multiple comparisons, while allowing for correlations among dependent cognitive variables, using a maximum statistic approach (see Supplementary Methods for further detail). Briefly, in each test set, brainPAD scores were randomly shuffled and then Spearman’s partial correlations were conducted between the randomly shuffled brainPAD scores and the cognitive dependent variables, controlling for age and sex. This process was repeated one million times and the maximum rho value was stored each time. Correlations between actual brainPAD scores and cognitive variables were deemed significant if they exceeded the 95th percentile of the maximum rho values.

Results

Brain age prediction

Training set

The model accurately predicted chronological age (r = 0.85, R2 = 67.24%, MAE = 7.28 years, weighted MAE = 0.10, p < 0.0001). As with other brain PAD models (e.g., Cole et al., 2018), a proportional bias was observed in this model where chronological age correlated with prediction error (r = - 0.4452, p = 1.1036e-10).

Independent test sets

Sex differences in brainPAD

Mean brainPAD differed significantly by sex in all datasets, Welch’s t(1009.55) = - 5.79, p < .0001. Males (M = -1.81, SD = 9.92) had significantly lower brainPADs than females (M = 1.81, SD = 10.23; see Fig 1). Within individual test sets, males had significantly lower brainPADs, compared to females, in in CR/RANN (p < .0001) and TILDA (p < .0001) but not in DEU (p = 0.148; see Fig 2).

Fig 1.

Fig 1.

Violin plot comparing distributions of brainPADs between sexes across all datasets

Fig 2.

Fig 2.

Violin plots comparing distributions of brainPADs between sexes within datasets

Brain regions involved in brain age prediction

The voxel-wise method used here to predict brain age resulted in individual coefficient values for each voxel. Voxels with positive coefficient values contributed to older brain age predictions and voxels with negative coefficient values contributed to younger brain age predictions. Figure 3 shows all voxels with binarised negative and positive coefficient values, respectively. However, as it is difficult to visualize a 3D object containing a large number of small voxels, a .nii file of the regression coefficients can be downloaded here: https://osf.io/5n6t8/. Overlaying this .nii file in a viewer such as mricroGL will allow for a more accurate view of the regression coefficients. Moreover, an .xlsx file containing the absolute value of each regression coefficient, the sign direction of that coefficient (i.e. positive/negative), the coefficient rank in terms of absolute values, as well as the MNI coordinates and anatomical labels of the coefficients, is available here: https://osf.io/dkz67/.

Fig 3.

Fig 3.

Binarised regression coefficients (positive coefficients shown in pink, negative coefficients shown in yellow) overlaid on 5 coronal slices. A: No threshold applied; B: thresholded at 25th percentile of absolute value of regression coefficients; C: thresholded at 50th percentile of absolute value of regression coefficients; D: thresholded at 75th percentile of absolute value of regression coefficients; E: thresholded at 95th percentile of absolute value of regression coefficients.

BrainPAD and Cognitive Function

Across multiple datasets, higher brainPAD scores were significantly correlated with reduced performance on measures of general cognitive status, semantic verbal fluency, processing speed, cognitive flexibility, and visual attention (see Figure 4 and Table 4).

Fig. 4.

Fig. 4

Scatterplots of replicated correlations between the residuals of brainPAD and cognitive measures after regressing brainPAD on age and sex, and each cognitive measure on age and sex. A: General cognitive status; B: Semantic verbal fluency; C: Processing speed, visual attention, and cognitive flexibility; D: Visual attention and cognitive flexibility. For scatterplots of non-replicated correlations, see Supplementary Info, Figure S.4.

Table 4.

Results of Spearman’s partial correlations between brainPAD and 17 cognitive domains.

Cognitive Domain DEU CR/RANN TILDA Probability of replicating by chance Sig. by max statistic correction (where finding not replicated)
rho df p rho df p rho df p
General Cognitive Status −0.3199 168 <0.0001 −0.1449 366 0.0275 −0.0333 481 0.4655 <0.00001 n/a
Premorbid Intelligence n/a −0.2322 358 <0.0001 0.0485 482 0.2873 n/a CR/RANN
Phonemic Verbal Fluency −0.3259 134 0.0001 −0.0771 356 0.1454 n/a n/a DEU
Semantic Verbal Fluency −0.2507 171 0.0009 −0.2019 357 0.0001 −0.0615 483 0.1765 <0.00001 n/a
Processing Speed 0.1232 89 0.2448 0.0595 357 0.2610 0.1208 483 0.0077 n/a None
Processing Speed, Visual Attention, Cognitive Flexibility 0.2662 80 0.0156 0.1167 353 0.0279 0.0904 478 0.0478 0.00005 n/a
Visual Attention, Cognitive Flexibility 0.2702 80 0.0141 0.1211 353 0.0225 0.0166 478 0.7165 0.00097 n/a
Cognitive Flexibility 0.0722 46 0.6258 0.0429 323 0.4411 n/a n/a None
Response Inhibition, Selective Attention 0.0854 146 0.3019 −0.1755 355 0.0009 n/a n/a None
Sustained Attention (Errors of Commission) n/a 0.0203 172 0.7902 0.0499 478 0.2752 n/a None
Sustained Attention (Reaction Time) n/a −0.0212 172 0.7813 0.0436 475 0.3425 n/a None
Verbal Episodic Memory (Immediate) 0.2194 171 0.0037 −0.0407 356 0.4428 −0.0347 483 0.4114 n/a None
Verbal Episodic Memory (Delayed) 0.2797 171 0.0002 0.0343 356 0.5173 0.0122 483 0.7887 n/a None
Verbal Episodic Memory (Learning) −0.3196 171 <0.0001 0.0657 356 0.2151 n/a n/a DEU
Working Memory −0.1310
−0.2974
167
166
0.0895a
0.0001b
−0.0469 360 0.3759 n/a n/a None
Visuospatial Ability −0.0809 76 0.4815 −0.1824 352 0.0006 n/a n/a CR/RANN

Note:

a

Digit Span Forwards;

b

Digit Span Backwards

Discussion

A penalized regression approach was able to produce accurate brain-age predictions from T1 MRI data in three independent datasets. In non-demented adults, brain predicted-age difference (brainPAD), calculated by subtracting these brain-age predictions from chronological age, was negatively correlated with general cognitive status, semantic verbal fluency, processing speed, visual attention, and cognitive flexibility; and visual attention and cognitive flexibility in multiple datasets. BrainPAD was significantly correlated with phonemic verbal fluency, premorbid intelligence, verbal episodic memory (learning score), and visuospatial ability in single datasets after controlling for multiple comparisons; however, these correlations were not replicated in another dataset so we do not have strong evidence here in support of these relationships. BrainPAD was not significantly correlated with processing speed, cognitive flexibility, response inhibition and selective attention, sustained attention, verbal episodic memory (immediate recall or delayed recall), or working memory in any dataset.

BrainPAD and Cognition

General Cognitive Status

BrainPAD was negatively correlated with general cognitive status, as measured using the MMSE and DRS, in DEU and CR/RANN, and the replication of this result across both datasets was statistically significant. However, brainPAD was not significantly correlated with the MMSE in TILDA. Nonetheless, given the statistically significant replication across two of the three datasets, there is reliable evidence in support of the correlation between brainPAD and general cognitive status in healthy older adults. Previous studies have reported that brainPAD is related to general cognitive status, albeit in samples including individuals with MCI, AD, or dementia (Beheshti et al., 2018; Kaufmann et al., 2019), and without adjusting for the effect of age or controlling for multiple comparisons (Beheshti et al., 2018; Cole, Underwood, et al., 2017; but see Table 1 Footnote 4 for information about adjusting for age in Cole, Underwood et al., 2017). In contrast to our findings, Gaser et al. (2013) reported that brainPAD was correlated with the CDR and ADAS but not the MMSE in an MCI sample. However, Gaser et al. (2013) did not account for the effect of age. While Löwe et al. (2016) reported that brainPAD was negatively correlated with the MMSE across mixed samples of APOE e4 carriers and non-carriers (including healthy controls, MCI, and AD), it was not significantly correlated with the MMSE within healthy control and MCI subgroups. Sample sizes within these subgroups were relatively small, ranging from 14 to 81 participants. Consequently, the correlations between brainPAD and the MMSE in these participants may not have been adequately powered to reach significance. Our study is the first to report a relationship between brainPAD and measures of general cognitive status in healthy adults while controlling for the effects of age and correcting for multiple comparisons. This is also the first study to investigate the relationship between brainPAD and DRS score. Our findings provide strong support for the existence of a significant negative relationship between brainPAD and general cognitive status. As such, this finding provides some preliminary support in favour of brainPAD as an objective measure of general cognitive function given that brainPAD is not subject to the various biases and effects (e.g. low reliability, practice effects) that limit the MMSE (Galasko, Abramson, Corey-Bloom, & Thal, 1993; Pfeffer, Kurosaki, Chance, Filos, & Bates, 1984; Tombaugh & McIntyre, 1992) and the DRS (Emery, Gillie, & Smith, 1996; Green, Woodard, & Green, 1995).

Semantic Verbal Fluency

BrainPAD was significantly negatively correlated with semantic verbal fluency, as measured using the Animals task, in both DEU and CR/RANN but not in TILDA. Regardless, the replication of this result across both DEU and CR/RANN was statistically significant. This finding contradicts non-significant correlations between brainPAD and composite measures of semantic and phonemic verbal fluency (Cole, Underwood, et al., 2017; Richard et al., 2018), although the former study used age-adjusted t-scores to control for the age-cognition relationship rather than adding age as a covariate to the brainPAD-fluency measure (cf. Le et al., 2018). As semantic verbal fluency is associated with age (Clark et al., 2009; Santos Nogueira, Azevedo Reis, & Vieira, 2016), the failure to adjust for age may have obscured a significant effect. Alternatively, these previously reported non-significant correlations could be explained by the use of composite measures of both phonemic and semantic fluency as we did not find strong evidence for a relationship between phonemic verbal fluency and brainPAD (although it was significant in DEU, this correlation was not replicated in CR/RANN). Therefore, it is possible that a non-significant relationship between phonemic fluency and brainPAD in the Cole et al. (2017) and Richard et al. (2018) study may have diluted a possible significant relationship between semantic fluency and brainPAD. In a study controlling for age, brainPAD was found to significantly negatively correlate with semantic verbal fluency (Franke et al., 2013). Although the Animals task has been described as an optimal test of neuropsychological function (Ardila, Ostrosky‐Solís, & Bernal, 2006), scores on this task are affected by various factors, including scoring and administration procedures (Woods, Wyma, Herron, & Yund, 2016) and practice effects (Cooper et al., 2001; Harrison, Buxton, Husain, & Wise, 2000; Wilson, Watson, Baddeley, Emslie, & Evans, 2000). As such, brainPAD, as an objective marker of general brain health and global cognitive function, could be a viable alternative to the Animals task. In sum, our results provide further evidence in support of a correlation between brainPAD and semantic verbal fluency.

Processing speed, visual attention, and cognitive flexibility

Across all three datasets, brainPAD was negatively correlated with processing speed, visual attention, and cognitive flexibility as measured by trail-making tests (TMT B or CTT 2). The TMT B is a relatively sensitive measure of cognitive decline: completion times were shown to be significantly different between healthy controls, MCI, and AD (Ashendorf et al., 2008). Likewise, the CTT 2 is sensitive to cognitive decline, with differences between AD and healthy controls (Lin et al., 2014), and between healthy controls, MCI, and AD (Guo et al., 2010). Therefore, it is no surprise that processing speed, visual attention, and cognitive flexibility were also negatively correlated with an index of accelerated brain ageing. Indeed, previous studies have reported similar results for trail-making versus brainPAD; however, these studies did not correct for multiple comparisons (Cole, Underwood, et al., 2017) or used clinical samples (TBI; Cole et al., 2015). Our data therefore augment these findings by replicating this result across three independent datasets. This evidence could suggest that brainPAD may be a potential objective measure of cognitive decline as it is not subject to the same factors which bias trail-making performance, including to practice effects (Bartels, Wegrzyn, Wiedl, Ackermann, & Ehrenreich, 2010), rater effects (Feeney et al., 2016) and participant literacy (Vaucher et al., 2014).

Visual attention and cognitive flexibility

BrainPAD was also negatively correlated with visual attention and cognitive flexibility (TMT B minus A), in DEU and CR/RANN, but not in TILDA (CTT 2 minus 1). Replication of this finding (albeit with relatively small rho values) in DEU and CR/RANN suggests a modest association between visual attention and cognitive flexibility. The relationship between brainPAD and TMT B minus A was only investigated in one previous study, in a TBI sample, (Cole et al., 2015) where a significant positive correlation was reported. Although the TMT B minus A can distinguish between stable and progressive MCI on a group level (Zanetti et al., 2006), and is associated with reduced mobility, increased mortality risk (Vazzana et al., 2010) and slower walking speed (Ble et al., 2005), as a derived measure of the TMT, the TMT B minus A index is similarly affected by the various factors that can limit interpretation of the TMT B scores. Therefore, given the correlation shown here between TMT B minus A and brainPAD, brainPAD may be a potential objective measure of general cognitive function.

It is notable that several significant brainPAD-cognition relationships were observed in the DEU and CR/RANN datasets, but not in TILDA. We tentatively offer some suggestions for this pattern of results. Confounding factors obscuring the brainPAD-general cognitive status relationship may have been uniquely present in TILDA. Whereas the DEU and CR/RANN cohorts were part of neuroimaging research studies, which have typically strict inclusion criteria, the TILDA MRI sample were a subset of a large nationally representative longitudinal study encompassing health, economic and social research (B. J. Whelan & Savva, 2013). TILDA therefore had few inclusion criteria: being at least 50 years old, having a residential address, and absence of dementia at baseline (Kearney et al., 2011; Savva, Maty, Setti, & Feeney, 2013). TILDA’s MRI sample were screened for MRI contraindications and were on average healthier than the full sample, but it is likely that the TILDA sample included participants who might normally be excluded from neuroimaging research studies (e.g., those using psychotropic or cardiovascular medication). Moreover, the range of some cognitive measures in TILDA was also smaller than DEU and CR/RANN in some cases (see Supplemental Information: Table S.5): notably for general cognitive status, and visual attention and cognitive flexibility, where the brainPAD-cognition correlations were not replicated within TILDA. Restricted range of scores on these measures in TILDA may have contributed to smaller correlation coefficients (Bland & Altman, 2011; Mendoza & Mumford, 1987). Additionally, the age range within TILDA was smaller than both DEU and CR/RANN which may have reduced the statistical power of the brainPAD-cognition correlations within TILDA as range restriction on covariates has also been shown to reduce power (Miciak, Taylor, Stuebing, Fletcher, & Vaughn, 2016) and decrease the magnitude of correlation coefficients (Sackett & Yang, 2000).

The smaller age range within TILDA (38 years) as compared to DEU (45.95 years) and CR/RANN (70 years) might also have contributed to the weaker correlation between chronological age and brain-predicted age in TILDA, as range restriction will reduce the size of correlation coefficients (Goodwin & Leech, 2006). Moreover, a negative mean brainPAD was reported in TILDA (-6.97 years) whereas both DEU and CR/RANN had positive mean brainPADs, +6.6 and +6.39 years respectively. Various factors, including intelligence, educational attainment, and environmental factors, have been proposed to affect brain ageing (Irimia, Torgerson, Goh, & Van Horn, 2015). TILDA had significantly higher levels of education versus both CR/RANN and DEU (see Supplementary Results). Steffener and colleagues (2016) reported that brainPAD was significantly related to education, with higher education associated with younger brains (or smaller/more negative brainPADs). This association with education could be one reason why much lower mean brainPADs were observed for the TILDA dataset. As the cohorts are each from different countries, there could be various other environmental factors that could further explain this relationship.

Model evaluation

We evaluated our model based on its predictive accuracy in three independent test sets, as proposed by Madan and Kensinger (2018). While internal cross-validation is a valuable and widely used technique that can attenuate overfitting (Arlot & Celisse, 2010); the use of cross-validation in certain situations and when it is not implemented correctly, can result in overestimated prediction accuracy and overfitting (Saeb, Lonini, Jayaraman, Mohr, & Kording, 2016; Skocik, Collins, Callahan-Flintoft, Bowman, & Wyble, 2016; Varoquaux et al., 2017). For brainPAD to be considered for clinical use, it must perform accurately with MRIs acquired in different scanners and under different protocols. However, in most instances of cross-validation, while the test set is split and held completely independent from the training set, factors common to both sets, such as scanner and protocol, could influence model performance. As such, the gold-standard evaluation for brainPAD should be accurate performance on independent external datasets.

The significant correlations between chronological age and brain-predicted age in all three external datasets shows that our model is accurate and generalizable (0.65, 0.78, and 0.87 for external datasets). Although the magnitude of these correlations is lower than correlations reported elsewhere, ranging from 0.91 to 0.94 (Cole et al., 2015; Cole, Poudel, et al., 2017; Franke et al., 2010; Lancaster et al., 2018; Liem et al., 2017), it exceeds other externally validated brain-predicted age studies, ranging from 0.65 to 0.85 (Beheshti et al., 2018; Madan & Kensinger, 2018; Varikuti et al., 2018).

With respect to mean absolute error (MAE), our model did not perform as well as other externally validated studies, ranging from 4.28 to 7.5 years (Beheshti et al., 2018; Cole et al., 2018; Franke et al., 2010; Lancaster et al., 2018; Madan & Kensinger, 2018). As a result, it could be possible that we may have lost some precision by not integrating WM information as input in the model, as was done by Cole et al. (2018), for example. Another potential reason is that other studies centered the age predictions using the mean of the ages from the test set. Although this correction is typically not explicitly described in method sections, Madan and Kensinger (2018) note that this is a standard correction in brain age prediction. Moreover, some studies also match the variance in predicted age in the test set with the variance of the training data (Madan & Kensinger, 2018). Both corrections are principled and acceptable methods of correcting for the regression to the mean artefact in brain age predictions but they result in biased age predictions in the test set. These corrections also limit the use of brainPAD to make single subject predictions, as both the test set mean and variance are used in the prediction. Our method used only training set information and therefore produced slightly less accurate but less biased predictions. Finally, our model may also appear to be less precise in terms of MAE as an artefact of the greater age range of our sample in comparison to most brainPAD studies. An alternative metric, the weighted MAE (calculated by dividing the MAE by the age range of the sample), may enable better comparisons across studies with different age ranges (Cole et al., 2019). While our weighted MAE is higher than some studies, ranging from 0.072 to 0.087 (Lancaster et al., 2018; Liem et al., 2017), the lowest weighted MAE in our sample (0.14 in CR/RANN) outperformed this metric when calculated for other studies, 0.178 (Beheshti et al., 2018), and 0.18 (Varikuti et al., 2018) and is comparable to 0.139 (Franke et al., 2010, 'Test 4' external test set). As such, the predictive accuracy of our model is comparable to the rest of the literature and is arguably less biased as only training set information is used.

Sex differences in brainPAD

There were significantly higher mean brainPADs in females in two of the three datasets in this study (TILDA and CR/RANN). There is mixed evidence in relation to sex differences in other brain age prediction studies, with some studies reporting significantly higher mean brainPADs in males (Cole et al., 2018; Franke et al., 2013; Luders, Cherbuin, & Gaser, 2016), some reporting no significant sex differences (Azor et al., 2019; Cruz-Almeida et al., 2019; Franke, Ristow, Gaser, & Alzheimer’s Disease Neuroimaging Initiative, 2014; Han et al., 2019), and another study, with a notably large sample size of 19,000, reporting higher mean brainPADs in females (Smith et al., 2019). Even studies using the same training sets have contrasting results in terms of sex effects. For example, one training set reported significantly higher male brainPADs in two studies (Franke et al., 2013; Luders et al., 2016) but no sex differences in another study (Franke et al., 2014): however, this divergence could be due to the likely mean centering of both brainPADs in both sexes in the latter study (i.e., male and female groups had mean brainPADs of 0 years). This was also the case in another training set used in multiple studies, with one study reporting significantly higher male brainPADs (Cole et al., 2018) but another reporting no significant differences (Azor et al., 2019). As such, it is likely that sex differences in brainPAD reflect the characteristics of the test sample. This is apparent in the present study with two out of the three datasets showing higher mean female brainPAD but one dataset showing no significant differences. We therefore recommend that future brain age studies report sex differences.

Model interpretation

Model interpretability

The interpretability of machine learning models is an important and widely discussed problem (Doshi-Velez & Kim, 2017), and although it is poorly defined (Lipton, 2018) it has been described as “the ability to explain or to present in understandable terms to a human” (Doshi-Velez & Kim, 2017, p. 2) and elsewhere as the ability to “understand the contribution of individual features in the model” (Lou, Caruana, & Gehrke, 2012, p. 1). Additionally, Lipton (2018) argued that for a model to be considered truly interpretable, it should possess the following three properties: algorithmic transparency (i.e. it should be possible to understand the mechanism by which the model works), decomposability (each part of the model, such as the model input and parameters, should have an intuitive explanation), and simulatability (a person should be able to consider the entire model at once). We contend that our model possesses these three properties as well as conforming to the definitions proposed above. First, our model possesses algorithmic transparency in that the Elastic Net is a penalized linear regression. Second, our model possesses decomposability. The inputs to the model were GM voxel density values and the parameters, or beta coefficient values, weighted the contribution of each individual value to the model output, which is brain predicted age. Third, our model possesses simulatability as the entire model can be considered as follows: summing the multiplication of GM voxel density values by the average contribution of these voxels to the prediction of chronological age in the training set (i.e., the beta coefficient values) resulted in a prediction of a new individual’s brain age.

Biological interpretability

Our statistical model of brain age contains many adjacent voxels that have opposite signs. The negative weights represent those areas with less volume (associated with older age). The positive weights show areas that have more volume associated with older age, which may seem counterintuitive: we propose that this is because GM in these areas represents a shift away from the cortex or periventricular regions (i.e., younger participants would have WM or CSF in those regions). An example of a similar result can be seen in an Alzheimer’s disease classification study (Dubois et al., 2014), which also used penalized regression.

Limitations

While we argue that the current model has good biological interpretability, this could be further improved by forcing sparsity to limit the number of voxels making significant contributions to brain age predictions. Modified Elastic Net algorithms, such as Enet-BETA (Liu & Li, 2017), can obtain sparser models which would reduce the number of predictive voxels, thereby further improving interpretability. However, as the Elastic Net’s prediction accuracy can increase with feature set size (Jollans et al., 2019), further limiting the feature set size could reduce model accuracy. As such, it might be difficult to achieve the right balance between interpretability and accuracy. An alternative approach could be to incorporate a penalty such as Total Variation within the Elastic Net in order to take into account the spatial structure of MRI data and produce weight maps that show the predictive voxels clustered in regions rather than dispersed across the brain (Dubois et al., 2014). These algorithms have been shown to produce models with greater biological interpretability (i.e. spatially organized weight maps) and comparable predictive accuracy to regular Elastic Net models for classification problems (Dubois et al., 2014). However, the technical implementation of such algorithms can be difficult and computationally expensive, although solutions such as early stopping and feature screening, have been proposed (Dohmatob, Eickenberg, Thirion, & Varoquaux, 2015).

Another possible limitation of the current model is that it uses only voxel-wise GM density data and thus our model may have lower accuracy due to this restricted feature set. Other brain age models have used feature sets including combinations of cortical and subcortical GM regional volumes (Steffener et al., 2016); combinations of GM voxel density values, cortical thickness, and regional volume data (Gutierrez Becker et al., 2018); combinations of cortical thickness, cortical surface area, subcortical volume, and functional connectivity information (Liem et al., 2017); and combinations of GM and WM voxel-wise density information (Cole et al., 2015, 2018; Cole, Underwood, et al., 2017). Notably, diffusion tensor imaging metrics and cortical thickness have been related to the cognitive domains assessed here, including semantic verbal fluency (Eastman et al., 2013; Rodríguez-Aranda et al., 2016) and processing speed, visual attention, and cognitive flexibility (Ciulli et al., 2016; Dickerson et al., 2008). More complex feature sets, which combined different feature types and imaging modalities, resulted in higher predictive accuracy versus single feature sets (Liem et al., 2017). As such, it is possible that a more accurate model using a more complex feature set would strengthen the brainPAD-cognition correlations reported here. However, such feature sets typically require dimension reduction such as PCA (Gutierrez Becker et al., 2018) or even dot products to combine GM and WM data (Cole et al., 2015, 2018; Cole, Underwood, et al., 2017). These steps can reduce the interpretability of the relationship between the original feature and brain age (Mateos-Pérez et al., 2018), although methods exist for making such feature sets interpretable (Honeine & Richard, 2009; Kwok & Tsang, 2004; Snyder et al., 2013). However, our aim was to produce an interpretable model with a relatively straightforward method, an aim which required a simple feature set. While this approach may have limited our model’s accuracy as larger and more complex feature sets often produce more accurate predictions (Scheinost et al., 2019), our model’s accuracy is still comparable to other models reported to-date in the literature.

The major limitation of our study is that for the majority of the cognitive domains investigated here, we used different cognitive measures to assess the putatively same cognitive processes. For example, although we considered the CTT 2 as a direct ‘culture-free’ analogue of the TMT B, as it is widely described (Elkin-Frankston, Lebowitz, Kapust, Hollis, & O’Connor, 2007; Messinis, Malegiannaki, Christodoulou, Panagiotopoulos, & Papathanasopoulos, 2011), the CTT 2 has different stimuli (shapes and colors vs numbers and letters) and takes longer because it has more stimuli (Mitrushina, Boone, Razani, & D’Elia, 2005). Consequently, some have argued, based on findings of significant difference in mean scores on CTT 2 and TMT B, that the tests are not direct equivalents (Dugbartey, Townes, & Mahurin, 2000; Strauss, Sherman, & Spreen, 2006). However, mean scores for both measures are calculated as time to completion and thus a difference in means between both measures reflects a difference primarily in test length. A more appropriate measure of test equivalence would be correlations between mean scores, and various studies report significant correlations between both measures (Dugbartey et al., 2000; Elkin-Frankston et al., 2007; Lee, Cheung, Chan, & Chan, 2000; Messinis et al., 2011). Similar arguments might be made for the other tests (e.g. the MMSE and DRS) that we used to assess the same cognitive constructs (e.g. general cognitive status). While it would be preferable to use the identical measures across datasets, our study used existing data and was designed after data collection. As a result, this approach was not possible here. Nonetheless, the measures used here were broadly comparable in that they are apparent measures of the same underlying cognitive constructs and it is these constructs which we are most interested in, more so than the actual measures.

Conclusions

The brain age model presented here is accurate and generalizable as it significantly predicts chronological age in 3 independent datasets. Furthermore, this model is interpretable and biologically plausible as older brain age is driven by decreased GM density in voxels that have been previously shown to be vulnerable to GM atrophy and volume loss. Finally, brainPAD scores, calculated using this model, are associated with reduced cognitive performance within the domains of general cognitive status; semantic verbal fluency; processing speed, visual attention, and cognitive flexibility; and visual attention and cognitive flexibility. The replication of these correlations in multiple datasets demonstrates that the relationship between brainPAD and these domains of cognitive function is robust to cultural- and site/scanner effects. As such, given that brainPAD is also not limited by task effects which can hinder neuropsychological assessment, these findings provide support for the use of brainPAD as an objective measure of general cognitive function with applications as a general measure of brain health and cognitive performance in the clinic and as a summary outcome measure for intervention studies in research settings.

Supplementary Material

Supplementary Information

Table 3.

Results of application of trained model parameters to 3 independent test sets.

Test Set Pearson’s r Mean brainPAD SD brainPAD MAE Weighted MAE
Test Set 1 – DEU 0.78* +6.60 6.44 7.60 0.17
Test Set 2 – CR/RANN 0.87* +6.39 8.57 8.56 0.14
Test Set 3 - TILDA 0.65* −6.97 7.52 8.42 0.22

Note: * = p < 10–37. Pearson’s r between brain age and chronological age. Weighted MAE = MAE divided by age range.

Acknowledgements

The authors would like to thank all participants who participated in the various studies which are used here.

Funding Sources:

RB is supported by the Irish Research Council under grant number EPSPG/2017/277. LMRD and RW are supported by the Science Foundation Ireland under grant number 16/ERCD/3797. RR is supported by a PhD scholarship funded by the Region Calabria. Data collection in Dokuz Eylul University, managed and supervised by GGY and DDS, was partially supported by the Turkish National Science and Research Council (TUBITAK, Grant number: 112S459) and the Dokuz Eylul University Scientific Research Projects (Grant number: 2018.KB.SAG.084). The Irish Longitudinal Study on Ageing is funded by core grants from the Health Research Board, Atlantic Philanthropies and Irish Life. MRI data collection in TILDA was supported by the Centre for Advanced Medical Imaging (CAMI) at St. James’ Hospital, Dublin. IHR thanks The Atlantic Philanthropies for their grant to the Global Brain Health Institute. YS is supported by NIA RF1 AG038465 and R01 AG026158. The funding agencies had no involvement in the conduct of the research or preparation of the article.

Footnotes

This is a post-peer-review, pre-copyedit version of an article published in Brain Imaging and Behavior. The final authenticated version is available online at: http://dx.doi.org/10.1007/s11682-020-00260-3.

Compliance with Ethical Standards

All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki declaration and its later amendments or comparable ethical standards. Informed consent was obtained from all individual participants included in the study.

Conflict of Interest

The authors report no conflict of interes

References

  1. Ardila A, Ostrosky-Solís F, & Bernal B (2006). Cognitive testing toward the future: The example of Semantic Verbal Fluency (ANIMALS). Int. J. Psychol, 41(5), 324–332. 10.1080/00207590500345542 [DOI] [Google Scholar]
  2. Arlot S, & Celisse A (2010). A survey of cross-validation procedures for model selection. Stat. Surv, 4(0), 40–79. 10.1214/09-SS054 [DOI] [Google Scholar]
  3. Ashendorf L, Jefferson AL, O’connor MK, CHAISSON C, Green RC, Stern RA, … Stern RA (2008). Trail Making Test errors in normal aging, mild cognitive impairment, and dementia. Arch. Clin. Neuropsychol, 23(2), 129–137. 10.1016/j.acn.2007.11.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  4. Azor AM, Cole JH, Holland AJ, Dumba M, Patel MC, Sadlon A, … Manning KE (2019). Increased brain age in adults with Prader-Willi syndrome. Neuroimage Clin, 21, 101664 [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Bartels C, Wegrzyn M, Wiedl A, Ackermann V, & Ehrenreich H (2010). Practice effects in healthy adults: a longitudinal study on frequent repetitive cognitive testing. BMC Neurosci., 11, 118 10.1186/1471-2202-11-118 [DOI] [PMC free article] [PubMed] [Google Scholar]
  6. Beheshti I, Maikusa N, & Matsuda H (2018). The association between “Brain-Age Score” (BAS) and traditional neuropsychological screening tools in Alzheimer’s disease. Brain. Behav, 8(8), e01020 10.1002/brb3.1020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  7. Benton AL, Varney NR, & Hamsher KD (1978). Visuospatial judgment. A clinical test. Arch. Neurol, 35(6), 364–367. Retrieved from http://www.ncbi.nlm.nih.gov/pubmed/655909 [DOI] [PubMed] [Google Scholar]
  8. Bland JM, & Altman DG (2011). Correlation in restricted ranges of data. BMJ, 342, d556 10.1136/bmj.d556 [DOI] [PubMed] [Google Scholar]
  9. Ble A, Volpato S, Zuliani G, Guralnik JM, Bandinelli S, Lauretani F, … Ferrucci L (2005). Executive Function Correlates with Walking Speed in Older Persons: The InCHIANTI Study. J. Am. Geriatr. Soc, 53(3), 410–415. 10.1111/j.1532-5415.2005.53157.x [DOI] [PubMed] [Google Scholar]
  10. Bunea F, She Y, Ombao H, Gongvatana A, Devlin K, & Cohen R (2011). Penalized least squares regression methods and applications to neuroimaging. NeuroImage, 55(4), 1519–1527. 10.1016/j.neuroimage.2010.12.028 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Buschke H, & Fuld PA (1974). Evaluating storage, retention, and retrieval in disordered memory and learning. Neurology, 24(11), 1019–1025. 10.1212/WNL.24.11.1019 [DOI] [PubMed] [Google Scholar]
  12. Butler RN, Sprott R, Warner H, Bland J, Feuers R, Forster M, … Wolf N (2004). Aging: The Reality: Biomarkers of Aging: From Primitive Organisms to Humans. J. Gerontol. A. Biol. Sci. Med. Sci, 59(6), B560–B567. 10.1093/gerona/59.6.B560 [DOI] [PubMed] [Google Scholar]
  13. Chanraud S, Martelli C, Delain F, Kostogianni N, Douaud G, Aubin HJ, … Martinot JL (2007). Brain Morphometry and Cognitive Performance in Detoxified Alcohol-Dependents with Preserved Psychosocial Functioning. Neuropsychopharmacology, 32(2), 429–438. 10.1038/sj.npp.1301219 [DOI] [PubMed] [Google Scholar]
  14. Ciulli S, Citi L, Salvadori E, Valenti R, Poggesi A, Inzitari D, … Diciotti S (2016). Prediction of Impaired Performance in Trail Making Test in MCI Patients With Small Vessel Disease Using DTI Data. IEEE J. Biomed. Health, 20(4), 1026–1033. 10.1109/JBHI.2016.2537808 [DOI] [PubMed] [Google Scholar]
  15. Clark LJ, Gatz M, Zheng L, Chen Y-L, McCleary C, & Mack WJ (2009). Longitudinal Verbal Fluency in Normal Aging, Preclinical, and Prevalent Alzheimer’s Disease. Am. J. Alzheimers Dis. Other Demen, 24(6), 461–468. 10.1177/1533317509345154 [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Cole JH, Franke K, & Cherbuin N (2019). Quantification of the Biological Age of the Brain Using Neuroimaging In Moskalev A (Ed.), Biomarkers of Human Aging (pp. 293–328). Cham: Springer International Publishing; 10.1007/978-3-030-24970-0_19 [DOI] [Google Scholar]
  17. Cole JH, Leech R, & Sharp DJ (2015). Prediction of brain age suggests accelerated atrophy after traumatic brain injury. Ann. Neurol, 77(4), 571–581. 10.1002/ana.24367 [DOI] [PMC free article] [PubMed] [Google Scholar]
  18. Cole JH, Poudel RPK, Tsagkrasoulis D, Caan MWA, Steves C, Spector TD, & Montana G (2017). Predicting brain age with deep learning from raw imaging data results in a reliable and heritable biomarker. NeuroImage, 163, 115–124. 10.1016/j.neuroimage.2017.07.059 [DOI] [PubMed] [Google Scholar]
  19. Cole JH, Ritchie SJ, Bastin ME, Valdés Hernández MC, Muñoz Maniega S, Royle N, … Deary IJ (2018). Brain age predicts mortality. Mol. Psychiatry, 23(5), 1385–1392. 10.1038/mp.2017.62 [DOI] [PMC free article] [PubMed] [Google Scholar]
  20. Cole JH, Underwood J, Caan MWAA, De Francesco D, van Zoest RA, Leech R, … COBRA collaboration. (2017). Increased brain-predicted aging in treated HIV disease. Neurology, 88(14), 1349–1357. 10.1212/WNL.0000000000003790 [DOI] [PMC free article] [PubMed] [Google Scholar]
  21. Cooper DB, Epker M, Lacritz L, Weiner M, Rosenberg RN, Honig L, & Cullum CM (2001). Effects of Practice on Category Fluency in Alzheimers Disease*. Clin. Neuropsychol, 15(1), 125–128. 10.1076/clin.15.1.125.1914 [DOI] [PubMed] [Google Scholar]
  22. Cruz-Almeida Y, Fillingim RB, Riley JLIII, Woods AJ, Porges E, Cohen R, & Cole J (2019). Chronic pain is associated with a brain aging biomarker in community-dwelling older adults. PAIN, 160(5). [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. D’Elia LF, Satz P, Uchiyama CL, & White T (1996). Color Trials Test. Professional Manual. Odessa, FL: Psychological Assessment Resources. [Google Scholar]
  24. Dean W, & Morgan RF (1988). In defense of the concept of biological aging measurement--current status. Arch. Gerontol. Geriatr, 7(3), 191–210. [DOI] [PubMed] [Google Scholar]
  25. Dickerson BC, Fenstermacher E, Salat DH, Wolk DA, Maguire RP, Desikan R, … Fischl B (2008). Detection of cortical thickness correlates of cognitive performance: Reliability across MRI scan sessions, scanners, and field strengths. NeuroImage, 39(1), 10–18. 10.1016/j.neuroimage.2007.08.042 [DOI] [PMC free article] [PubMed] [Google Scholar]
  26. Dohmatob E, Eickenberg M, Thirion B, & Varoquaux G (2015). Speeding-Up Model-Selection in Graphnet via Early-Stopping and Univariate Feature-Screening In 2015 International Workshop on Pattern Recognition in NeuroImaging (pp. 17–20). IEEE; 10.1109/PRNI.2015.19 [DOI] [Google Scholar]
  27. Dorrian J, Rogers N, & Dinges D (2005). Psychomotor Vigilance Performance: Neurocognitive Assay Sensitive to Sleep Loss In Kushida CA (Ed.), Sleep Deprivation: Clinical Issues, Pharmacology, and Sleep Loss Effects (Vol. 193, pp. 39–70). Boca Raton, FL: CRC Press. [Google Scholar]
  28. Doshi-Velez F, & Kim B (2017). Towards A Rigorous Science of Interpretable Machine Learning. Retrieved from http://arxiv.org/abs/1702.08608
  29. Dubois M, Hadj-Selem F, Lofstedt T, Perrot M, Fischer C, Frouin V, & Duchesnay E (2014). Predictive support recovery with TV-Elastic Net penalty and logistic regression: An application to structural MRI In Proceedings - 2014 International Workshop on Pattern Recognition in Neuroimaging, PRNI 2014 (pp. 1–4). Tubingen; 10.1109/PRNI.2014.6858517 [DOI] [Google Scholar]
  30. Dugbartey AT, Townes BD, & Mahurin RK (2000). Equivalence of the Color Trails Test and Trail Making Test in Nonnative English-Speakers. Arch. Clin. Neuropsychol, 15(5), 425–431. 10.1016/S0887-6177(99)00034-7 [DOI] [PubMed] [Google Scholar]
  31. Dwyer DB, Falkai P, & Koutsouleris N (2018). Machine Learning Approaches for Clinical Psychology and Psychiatry. Annu. Rev. Clin. Psychol, 14, 91–118. 10.1146/annurev-clinpsy-032816-045037 [DOI] [PubMed] [Google Scholar]
  32. Eastman JA, Hwang KS, Lazaris A, Chow N, Ramirez L, Babakchanian S, … Apostolova LG (2013). Cortical thickness and semantic fluency in Alzheimer’s disease and mild cognitive impairment. Am. J. Alzheimers Dis. (Columbia), 1(2), 81–92. 10.7726/ajad.2013.1006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Elkin-Frankston S, Lebowitz BK, Kapust LR, Hollis HH, & O’Connor MG (2007). The use of the Color Trails Test in the assessment of driver competence: Preliminary report of a culture-fair instrument. Arch. Clin. Neuropsychol, 22(5), 631–635. 10.1016/j.acn.2007.04.004 [DOI] [PubMed] [Google Scholar]
  34. Elman JA, Jak AJ, Panizzon MS, Tu XM, Chen T, Reynolds CA, … Kremen WS (2018). Underdiagnosis of mild cognitive impairment: A consequence of ignoring practice effects. Alzheimers Dement. (Amst), 10, 372–381. 10.1016/J.DADM.2018.04.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  35. Emek-Savaş D, Yerlikaya D, Yener G, & Öktem Ö (2019). Validity, reliability and norm scores of the Stroop Test Çapa Version. Turk Psikiyatri Derg, in press. 10.5080/u23549 [DOI] [PubMed] [Google Scholar]
  36. Emery VOB, Gillie EX, & Smith JA (1996). Reclassification of the Vascular Dementias: Comparisons of Infarct and Noninfarct Vascular Dementias. Int. Psychogeriatr, 8(1), 33–61. 10.1017/S1041610296002475 [DOI] [PubMed] [Google Scholar]
  37. Farokhian F, Yang C, Beheshti I, Matsuda H, & Wu S (2017). Age-Related Gray and White Matter Changes in Normal Adult Brains. Aging Dis, 8(6), 899–909. 10.14336/AD.2017.0502 [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Feeney J, Savva GM, O’Regan C, King-Kallimanis B, Cronin H, & Kenny RA (2016). Measurement Error, Reliability, and Minimum Detectable Change in the Mini-Mental State Examination, Montreal Cognitive Assessment, and Color Trails Test among Community Living Middle-Aged and Older Adults. J. Alzheimers Dis, 53(3), 1107–1114. 10.3233/JAD-160248 [DOI] [PubMed] [Google Scholar]
  39. Fiorito G, McCrory C, Robinson O, Carmeli C, Rosales CO, Zhang Y, … Zins M (2019). Socioeconomic position, lifestyle habits and biomarkers of epigenetic aging: A multi-cohort analysis. Aging, 11(7), 2045–2070. 10.18632/aging.101900 [DOI] [PMC free article] [PubMed] [Google Scholar]
  40. Fjell AM, Westlye LT, Grydeland H, Amlien I, Espeseth T, Reinvang I, … Alzheimer Disease Neuroimaging Initiative. (2013). Critical ages in the life course of the adult brain: nonlinear subcortical aging. Neurobiol. Aging, 34(10), 2239–2247. 10.1016/j.neurobiolaging.2013.04.006 [DOI] [PMC free article] [PubMed] [Google Scholar]
  41. Folstein MF, Folstein SE, & McHugh PR (1975). “Mini-mental state”. A practical method for grading the cognitive state of patients for the clinician. J. Psychiatr. Res, 12(3), 189–198. 10.1016/0022-3956(75)90026-6 [DOI] [PubMed] [Google Scholar]
  42. Franke K, & Gaser C (2012). Longitudinal Changes in Individual BrainAGE in Healthy Aging, Mild Cognitive Impairment, and Alzheimer’s Disease. GeroPsych, 25(4), 235–245. 10.1024/1662-9647/a000074 [DOI] [Google Scholar]
  43. Franke K, Gaser C, Manor B, & Novak V (2013). Advanced BrainAGE in older adults with type 2 diabetes mellitus. Front. Aging Neurosci, 5, 90 10.3389/fnagi.2013.00090 [DOI] [PMC free article] [PubMed] [Google Scholar]
  44. Franke K, Ristow M, Gaser C, & Alzheimer’s Disease Neuroimaging Initiative. (2014). Gender-specific impact of personal health parameters on individual brain aging in cognitively unimpaired elderly subjects. Front. Aging Neurosci, 6, 94 10.3389/fnagi.2014.00094 [DOI] [PMC free article] [PubMed] [Google Scholar]
  45. Franke K, Ziegler G, Klöppel S, & Gaser C (2010). Estimating the age of healthy subjects from T1-weighted MRI scans using kernel methods: Exploring the influence of various parameters. Neuroimage, 50(3), 883–892. 10.1016/j.neuroimage.2010.01.005 [DOI] [PubMed] [Google Scholar]
  46. Galasko D, Abramson I, Corey-Bloom J, & Thal LJ (1993). Repeated exposure to the Mini-Mental State Examination and the Information-Memory-Concentration Test results in a practice effect in Alzheimer’s disease. Neurology, 43(8), 1559–1563. 10.1212/wnl.43.8.1559 [DOI] [PubMed] [Google Scholar]
  47. Gaser C, Franke K, Klöppel S, Koutsouleris N, Sauer H, & Alzheimer’s Disease Neuroimaging Initiative. (2013). BrainAGE in Mild Cognitive Impaired Patients: Predicting the Conversion to Alzheimer’s Disease. PLoS One, 8(6), e67346 10.1371/journal.pone.0067346 [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Ge Y, Grossman RI, Babb JS, Rabin ML, Mannon LJ, & Kolson DL (2002). Age-related total gray matter and white matter changes in normal adult brain. Part I: volumetric MR imaging analysis. AJNR Am. J. Neuroradiol, 23(8), 1327–1333. Retrieved from http://www.ncbi.nlm.nih.gov/pubmed/12223373 [PMC free article] [PubMed] [Google Scholar]
  49. Golden CJ (1978). Stroop Color and Word Test: A manual for clinical and experimental uses. Chicago: Stoelting; https://doi.org/10.1002 [Google Scholar]
  50. Goodwin LD, & Leech NL (2006). Understanding Correlation: Factors That Affect the Size of r. J. Exp. Educ, 74(3), 249–266. 10.3200/JEXE.74.3.249-266 [DOI] [Google Scholar]
  51. Green RC, Woodard JL, & Green J (1995). Validity of the Mattis Dementia Rating Scale for detection of cognitive impairment in the elderly. J. Neuropsychiatry Clin. Neurosci, 7(3), 357–360. 10.1176/jnp.7.3.357 [DOI] [PubMed] [Google Scholar]
  52. Grober E, & Sliwinski M (1991). Development and validation of a model for estimating premorbid verbal intelligence in the elderly. J. Clin. Exp. Neuropsychol. 10.1080/01688639108405109 [DOI] [PubMed] [Google Scholar]
  53. Guggenmos M, Schmack K, Sekutowicz M, Garbusow M, Sebold M, Sommer C, … Sterzer P (2017). Quantitative neurobiological evidence for accelerated brain aging in alcohol dependence. Transl. Psychiatry, 7(12), 1279 10.1038/s41398-017-0037-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  54. Guo QH, Cao XY, Zhou Y, Zhao QH, Ding D, & Hong Z (2010). Application study of quick cognitive screening test in identifying mild cognitive impairment. Neurosci. Bull, 26(1), 47–54. 10.1007/s12264-010-0816-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Gutierrez Becker B, Klein T, & Wachinger C (2018). Gaussian process uncertainty in age estimation as a measure of brain abnormality. Neuroimage, 175, 246–258. 10.1016/J.NEUROIMAGE.2018.03.075 [DOI] [PubMed] [Google Scholar]
  56. Han LKM, Dinga R, Hahn T, Ching C, Eyler L, Aftanas L, … Schmaal L (2019). Brain Aging in Major Depressive Disorder: Results from the ENIGMA Major Depressive Disorder working group. BioRxiv, 560623. 10.1101/560623 [DOI] [PMC free article] [PubMed] [Google Scholar]
  57. Harrison JE, Buxton P, Husain M, & Wise R (2000). Short test of semantic and phonological fluency: normal performance, validity and test-retest reliability. Br. J. Clin. Psychol, (2), 181–191. 10.1348/014466500163202 [DOI] [PubMed] [Google Scholar]
  58. Heaton RKK, Chelune GJ, Talley JL, Kay GG, & Curtiss G (1993). Wisconsin Card Sorting Test Manual: Revised and expanded Psychological Reports. Odessa, FL: Psychological Assessment Resources; 10.2466/pr0.1995.76.2.623 [DOI] [Google Scholar]
  59. Honeine P, & Richard C (2009). Solving the pre-image problem in kernel machines: A direct method In 2009 IEEE International Workshop on Machine Learning for Signal Processing (pp. 1–6). 10.1109/MLSP.2009.5306204 [DOI] [Google Scholar]
  60. Irimia A, Torgerson CM, Goh S-YM, & Van Horn JD (2015). Statistical estimation of physiological brain age as a descriptor of senescence rate during adulthood. Brain Imaging Behav., 9(4), 678–689. 10.1007/s11682-014-9321-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  61. Jollans L, Boyle R, Artiges E, Banaschewski T, Desrivières S, Grigis A, … Whelan R (2019). Quantifying performance of machine learning methods for neuroimaging data. Neuroimage. 10.1016/J.NEUROIMAGE.2019.05.082 [DOI] [PMC free article] [PubMed] [Google Scholar]
  62. Jollans L, & Whelan R (2016). The Clinical Added Value of Imaging: A Perspective From Outcome Prediction. Biol. Psychiatry Cogn. Neurosci. Neuroimaging, 1(5), 423–432. 10.1016/j.bpsc.2016.04.005 [DOI] [PubMed] [Google Scholar]
  63. Jollans L, & Whelan R (2018). Neuromarkers for Mental Disorders: Harnessing Population Neuroscience. Front. Psychiatry, 9, 242 10.3389/fpsyt.2018.00242 [DOI] [PMC free article] [PubMed] [Google Scholar]
  64. Jurica PJ, Leitten CL, & Mattis S (2001). DRS-2: Dementia Rating Scale-2: Professional Manual. Psychological Assessment Resources. Retrieved from https://books.google.ie/books?id=tovFPwAACAAJ [Google Scholar]
  65. Kaufmann T, van der Meer D, Doan NT, Schwarz E, Lund MJ, Agartz I, … (KaSP) KSP (2019). Common brain disorders are associated with heritable patterns of apparent aging of the brain. Nat. Neurosci, 22(10), 1617–1623. 10.1038/s41593-019-0471-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  66. Kearney PM, Cronin H, O’Regan C, Kamiya Y, Savva GM, Whelan B, & Kenny R (2011). Cohort Profile: The Irish Longitudinal Study on Ageing. Int. J. Epidemiol, 40(4), 877–884. 10.1093/ije/dyr116 [DOI] [PubMed] [Google Scholar]
  67. Koutsouleris N, Davatzikos C, Borgwardt S, Gaser C, Bottlender R, Frodl T, … Meisenzahl E (2014). Accelerated Brain Aging in Schizophrenia and Beyond: A Neuroanatomical Marker of Psychiatric Disorders. Schizophr. Bull, 40(5), 1140–1153. 10.1093/schbul/sbt142 [DOI] [PMC free article] [PubMed] [Google Scholar]
  68. Kwok JT-., & Tsang IW. (2004). The pre-image problem in kernel methods. IEEE Transactions on Neural Networks, 15(6), 1517–1525. 10.1109/TNN.2004.837781 [DOI] [PubMed] [Google Scholar]
  69. Lancaster J, Lorenz R, Leech R, & Cole JH (2018). Bayesian optimization for neuroimaging pre-processing in brain age classification and prediction. Front. Aging Neurosci, 10(FEB), 1–10. 10.3389/fnagi.2018.00028 [DOI] [PMC free article] [PubMed] [Google Scholar]
  70. Le TT, Kuplicki RT, McKinney BA, Yeh H-W, Thompson WK, Paulus MP, & Tulsa 1000 Investigators, T. 1000. (2018). A Nonlinear Simulation Framework Supports Adjusting for Age When Analyzing BrainAGE. Front. Aging Neurosci, 10, 317 10.3389/fnagi.2018.00317 [DOI] [PMC free article] [PubMed] [Google Scholar]
  71. Lee TMC, Cheung CCY, Chan JKP, & Chan CCH (2000). Trail Making Across Languages. J. Clin. Exp. Neuropsychol, 22(6), 772–778. 10.1076/jcen.22.6.772.954 [DOI] [PubMed] [Google Scholar]
  72. Liem F, Varoquaux G, Kynast J, Beyer F, Kharabian Masouleh S, Huntenburg JM, … Margulies DS (2017). Predicting brain-age from multimodal imaging data captures cognitive impairment. Neuroimage, 148, 179–188. 10.1016/j.neuroimage.2016.11.005 [DOI] [PubMed] [Google Scholar]
  73. Lin YC, Shih YC, Tseng WYI, Chu YH, Wu MT, Chen TF, … Chiu MJ (2014). Cingulum Correlates of Cognitive Functions in Patients with Mild Cognitive Impairment and Early Alzheimer’s Disease: A Diffusion Spectrum Imaging Study. Brain Topogr., 27(3), 393–402. 10.1007/s10548-013-0346-2 [DOI] [PubMed] [Google Scholar]
  74. Lipton ZC (2018). The Mythos of Model Interpretability. ACM Queue, 16(3), 30:31–30:57. 10.1145/3236386.3241340 [DOI] [Google Scholar]
  75. Liu W, & Li Q (2017). An Efficient Elastic Net with Regression Coefficients Method for Variable Selection of Spectrum Data. PLoS One, 12(2), e0171122 10.1371/journal.pone.0171122 [DOI] [PMC free article] [PubMed] [Google Scholar]
  76. Lou Y, Caruana R, & Gehrke J (2012). Intelligible Models for Classification and Regression. In Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 150–158). New York, NY, USA: ACM; 10.1145/2339530.2339556 [DOI] [Google Scholar]
  77. Löwe LC, Gaser C, & Franke K (2016). The Effect of the APOE Genotype on Individual BrainAGE in Normal Aging, Mild Cognitive Impairment, and Alzheimer’s Disease. PLoS One, 11(7), e0157514 10.1371/journal.pone.0157514 [DOI] [PMC free article] [PubMed] [Google Scholar]
  78. Luders E, Cherbuin N, & Gaser C (2016). Estimating brain age using high-resolution pattern recognition: Younger brains in long-term meditation practitioners. Neuroimage, 134, 508–513. 10.1016/j.neuroimage.2016.04.007 [DOI] [PubMed] [Google Scholar]
  79. Luo Y, Tseng H-H, Cui S, Wei L, Ten Haken RK, & El Naqa I (2019). Balancing accuracy and interpretability of machine learning approaches for radiation treatment outcomes modeling. BJR|Open. 10.1259/bjro.20190021 [DOI] [PMC free article] [PubMed] [Google Scholar]
  80. Madan CR, & Kensinger EA (2018). Predicting age from cortical structure across the lifespan. Eur. J. Neurosci, 47(5), 399–416. 10.1111/ejn.13835 [DOI] [PMC free article] [PubMed] [Google Scholar]
  81. Mateos-Pérez JM, Dadar M, Lacalle-Aurioles M, Iturria-Medina Y, Zeighami Y, & Evans AC (2018). Structural neuroimaging as clinical predictor: A review of machine learning applications. Neuroimage Clin, 20, 506–522. 10.1016/j.nicl.2018.08.019 [DOI] [PMC free article] [PubMed] [Google Scholar]
  82. McCaffrey RJ, & Westervelt HJ (1995). Issues associated with repeated neuropsychological assessments. Neuropsychol. Rev, 5(3), 203–221. 10.1007/BF02214762 [DOI] [PubMed] [Google Scholar]
  83. McCrory C, & Kenny RA (2018). Rebuking the concept of ageing as a disease. Lancet Diabetes Endocrinol., 6(10), 768 10.1016/S2213-8587(18)30266-3 [DOI] [PubMed] [Google Scholar]
  84. McIntyre RS, Cha DS, Soczynska JK, Woldeyohannes HO, Gallaugher LA, Kudlow P, … Baskaran A (2013). Cognitive deficits and functional outcomes in major depressive disorder: Determinants, substrates, and treatment interventions. Depress. Anxiety, 30(6), 515–527. 10.1002/da.22063 [DOI] [PubMed] [Google Scholar]
  85. Mendoza JL, & Mumford M (1987). Corrections for Attenuation and Range Restriction on the Predictor. J. Educ. Stat, 12(3), 282 10.2307/1164688 [DOI] [Google Scholar]
  86. Messinis L, Malegiannaki A-C, Christodoulou T, Panagiotopoulos V, & Papathanasopoulos P (2011). Color Trails Test: Normative Data and Criterion Validity for the Greek Adult Population. Arch. Clin. Neuropsychol, 26(4), 322–330. 10.1093/arclin/acr027 [DOI] [PubMed] [Google Scholar]
  87. Miciak J, Taylor WP, Stuebing KK, Fletcher JM, & Vaughn S (2016). Designing Intervention Studies: Selected Populations, Range Restrictions, and Statistical Power. J. Res. Educ. Eff, 9(4), 556–569. 10.1080/19345747.2015.1086916 [DOI] [PMC free article] [PubMed] [Google Scholar]
  88. Mitrushina MN, Boone KB, Razani JL, & D’Elia LF (2005). Handbook of normative data for neuropsychological assessment (Second). New York: Oxford University Press; Retrieved from https://global.oup.com/academic/product/handbook-of-normative-data-for-neuropsychological-assessment-9780195169300?cc=ie&lang=en& [Google Scholar]
  89. Mwangi B, Tian TS, & Soares JC (2014). A review of feature reduction techniques in neuroimaging. Neuroinformatics, 12(2), 229–244. 10.1007/s12021-013-9204-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  90. Nelson HE, & Willinson J (1982). The National Adult Reading Test (NART): Test Manual. Windsor, UK: NFER: Nelson. [Google Scholar]
  91. Nenadić I, Dietzek M, Langbein K, Sauer H, & Gaser C (2017). BrainAGE score indicates accelerated brain aging in schizophrenia, but not bipolar disorder. Psychiatry Res. Neuroimaging, 266, 86–89. 10.1016/j.pscychresns.2017.05.006 [DOI] [PubMed] [Google Scholar]
  92. Öktem O (1992). A verbal test of memory processes: A preliminary study. Noro. Psikiyatr. Ars, 29(4), 196–206. [Google Scholar]
  93. Pfeffer RI, Kurosaki TT, Chance JM, Filos S, & Bates D (1984). Use of the mental function index in older adults: Reliability, validity, and measurement of change over time. Am. J. Epidemiol, 120(6), 922–935. 10.1093/oxfordjournals.aje.a113963 [DOI] [PubMed] [Google Scholar]
  94. Ranganathan P, Pramesh CS, & Buyse M (2016). Common pitfalls in statistical analysis: The perils of multiple testing. Perspect. Clin. Res, 7(2), 106–107. 10.4103/2229-3485.179436 [DOI] [PMC free article] [PubMed] [Google Scholar]
  95. Reitan RM (1955). The relation of the Trail Making Test to organic brain damage. J. Consult. Psychol. 10.1037/h0044509 [DOI] [PubMed] [Google Scholar]
  96. Richard G, Kolskår K, Sanders A-M, Kaufmann T, Petersen A, Doan NT, … Westlye LT (2018). Assessing distinct patterns of cognitive aging using tissue-specific brain age prediction based on diffusion tensor imaging and brain morphometry. PeerJ, 6, e5908 10.7717/peerj.5908 [DOI] [PMC free article] [PubMed] [Google Scholar]
  97. Robertson IH, Manly T, Andrade J, Baddeley BT, & Yiend J (1997). Òops!’: Performance correlates of everyday attentional failures in traumatic brain injured and normal subjects. Neuropsychologia, 35(6), 747–758. 10.1016/S0028-3932(97)00015-8 [DOI] [PubMed] [Google Scholar]
  98. Rodríguez-Aranda C, Waterloo K, Johnsen SH, Eldevik P, Sparr S, Wikran GC, … Vangberg TR (2016). Neuroanatomical correlates of verbal fluency in early Alzheimer’s disease and normal aging. Brain Lang., 155–156, 24–35. 10.1016/J.BANDL.2016.03.001 [DOI] [PubMed] [Google Scholar]
  99. Sackett PR, & Yang H (2000). Correction for range restriction: An expanded typology. J. Appl. Psychol, 85(1), 112–118. 10.1037/0021-9010.85.1.112 [DOI] [PubMed] [Google Scholar]
  100. Saeb S, Lonini L, Jayaraman A, Mohr DC, & Kording KP (2016). Voodoo Machine Learning for Clinical Predictions. BioRxiv, 059774 10.1101/059774 [DOI] [Google Scholar]
  101. Santos Nogueira D, Azevedo Reis E, & Vieira A (2016). Verbal Fluency Tasks: Effects of Age, Gender, and Education. Folia Phoniatr. Logo, 68(3), 124–133. 10.1159/000450640 [DOI] [PubMed] [Google Scholar]
  102. Savva GM, Maty SC, Setti A, & Feeney J (2013). Cognitive and Physical Health of the Older Populations of England, the United States, and Ireland: International Comparability of The Irish Longitudinal Study on Ageing. J. Am. Geriatr. Soc, 61, S291–S298. 10.1111/jgs.12196 [DOI] [PubMed] [Google Scholar]
  103. Scheinost D, Noble S, Horien C, Greene AS, Lake EM, Salehi M, … Constable RT (2019). Ten simple rules for predictive modeling of individual differences in neuroimaging. Neuroimage, 193, 35–45. 10.1016/J.NEUROIMAGE.2019.02.057 [DOI] [PMC free article] [PubMed] [Google Scholar]
  104. Scheller E, Schumacher LV, Peter J, Lahr J, Wehrle J, Kaller CP, … Klöppel S (2018). Brain Aging and APOE ε4 Interact to Reveal Potential Neuronal Compensation in Healthy Older Adults. Front. Aging Neurosci, 10, 74 10.3389/fnagi.2018.00074 [DOI] [PMC free article] [PubMed] [Google Scholar]
  105. Schnack HG, van Haren NEM, Nieuwenhuis M, Hulshoff Pol HE, Cahn W, & Kahn RS (2016). Accelerated Brain Aging in Schizophrenia: A Longitudinal Pattern Recognition Study. Am. J. Psychiatry, 173(6), 607–616. 10.1176/appi.ajp.2015.15070922 [DOI] [PubMed] [Google Scholar]
  106. Skocik M, Collins J, Callahan-Flintoft C, Bowman H, & Wyble B (2016). I tried a bunch of things: The dangers of unexpected overfitting in classification. BioRxiv, 078816 10.1101/078816 [DOI] [PubMed] [Google Scholar]
  107. Smith SM, Vidaurre D, Alfaro-Almagro F, Nichols TE, & Miller KL (2019). Estimation of Brain Age Delta from Brain Imaging. Neuroimage. 10.1016/j.neuroimage.2019.06.017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  108. Snyder JC, Mika S, Burke K, & Müller K-R (2013). Kernels, Pre-images and Optimization In Schölkopf B, Luo Z, & Vovk V (Eds.), Empirical Inference: Festschrift in Honor of Vladimir N. Vapnik (pp. 245–259). Berlin, Heidelberg: Springer Berlin Heidelberg; 10.1007/978-3-642-41136-6_21 [DOI] [Google Scholar]
  109. Sprott RL (2010). Biomarkers of aging and disease: Introduction and definitions. Exp. Gerontol, 45(1), 2–4. 10.1016/J.EXGER.2009.07.008 [DOI] [PubMed] [Google Scholar]
  110. Steffener J, Habeck C, O’Shea D, Razlighi Q, Bherer L, & Stern Y (2016). Differences between chronological and brain age are related to education and self-reported physical activity. Neurobiol. Aging, 40, 138–144. 10.1038/nn.3945.Dopaminergic [DOI] [PMC free article] [PubMed] [Google Scholar]
  111. Stern Y, Gazes Y, Razlighi Q, Steffener J, & Habeck C (2018). A task-invariant cognitive reserve network. Neuroimage, 178, 36–45. 10.1016/J.NEUROIMAGE.2018.05.033 [DOI] [PMC free article] [PubMed] [Google Scholar]
  112. Stern Y, Habeck C, Steffener J, Barulli D, Gazes Y, Razlighi Q, … Salthouse T (2014). The Reference Ability Neural Network Study: Motivation, design, and initial feasibility analyses. Neuroimage, 103, 139–151. 10.1016/J.NEUROIMAGE.2014.09.029 [DOI] [PMC free article] [PubMed] [Google Scholar]
  113. Strauss EH, Sherman EMS, & Spreen O (2006). A Compendium of Neuropsychological Tests; Administration Norms And Commentary (3rd ed.). New York, NY: Oxford University Press; 10.1016/j.jvolgeores.2008.06.015 [DOI] [Google Scholar]
  114. Tombaugh TN, & McIntyre NJ (1992). The mini-mental state examination: a comprehensive review. J. Am. Geriatr. Soc, 40(9), 922–935. 10.1111/j.1532-5415.1992.tb01992.x [DOI] [PubMed] [Google Scholar]
  115. Varikuti DP, Genon S, Sotiras A, Schwender H, Hoffstaedter F, Patil KR, … Eickhoff SB (2018). Evaluation of non-negative matrix factorization of grey matter in age prediction. Neuroimage, 173(March), 394–410. 10.1016/j.neuroimage.2018.03.007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  116. Varoquaux G, Raamana PR, Engemann DA, Hoyos-Idrobo A, Schwartz Y, & Thirion B (2017). Assessing and tuning brain decoders: Cross-validation, caveats, and guidelines. Neuroimage, 145, 166–179. 10.1016/J.NEUROIMAGE.2016.10.038 [DOI] [PubMed] [Google Scholar]
  117. Vaucher P, Herzig D, Cardoso I, Herzog MH, Mangin P, & Favrat B (2014). The trail making test as a screening instrument for driving performance in older drivers; a translational research. BMC Geriatrics, 14(1), 123 10.1186/1471-2318-14-123 [DOI] [PMC free article] [PubMed] [Google Scholar]
  118. Vazzana R, Bandinelli S, Lauretani F, Volpato S, Lauretani F, Di Iorio A, … Ferrucci L (2010). Trail Making Test predicts physical impairment and mortality in older persons. J. Am. Geriatr. Soc, 58(4), 719–723. 10.1111/j.1532-5415.2010.02780.x [DOI] [PMC free article] [PubMed] [Google Scholar]
  119. Wechsler D (1987). Wechsler Memory Scale - Revised manual. San Antonio, Texas: Psychological Corporation. [Google Scholar]
  120. Wechsler D (1997). WAIS-III administration and scoring manual. The Psychological Corporation, San Antonio, Texas: 10.1177/1073191102009001003 [DOI] [Google Scholar]
  121. Whelan BJ, & Savva GM (2013). Design and Methodology of The Irish Longitudinal Study on Ageing. J. Am. Geriatr. Soc, 61, S265–S268. 10.1111/jgs.12199 [DOI] [PubMed] [Google Scholar]
  122. Whelan R, & Garavan H (2014). When Optimism Hurts: Inflated Predictions in Psychiatric Neuroimaging. Biol. Psychiatry, 75(9), 746–748. 10.1016/j.biopsych.2013.05.014 [DOI] [PubMed] [Google Scholar]
  123. Willer L, Pedersen PM, Forchhammer HB, & Christensen H (2016). Cognitive assessment at bedside for iPad: A preliminary validation of a novel cognitive test for stroke patients. Eur. Stroke J, 1(4), 294–301. 10.1177/2396987316665233 [DOI] [PMC free article] [PubMed] [Google Scholar]
  124. Wilson BA, Watson PC, Baddeley AD, Emslie H, & Evans JJ (2000). Improvement or simply practice? The effects of twenty repeated assessments on people with and without brain injury. J. Int. Neuropsychol. Soc, 6(4), 469–479. Retrieved from http://www.ncbi.nlm.nih.gov/pubmed/10902416 [DOI] [PubMed] [Google Scholar]
  125. Woo CW, Chang LJ, Lindquist MA, & Wager TD (2017). Building better biomarkers: brain models in translational neuroimaging. Nat. Neurosci, 20(3), 365–377. 10.1038/nn.4478 [DOI] [PMC free article] [PubMed] [Google Scholar]
  126. Woods DL, Wyma JM, Herron TJ, & Yund EW (2016). Computerized Analysis of Verbal Fluency: Normative Data and the Effects of Repeated Testing, Simulated Malingering, and Traumatic Brain Injury. PLoS One, 11(12), e0166439 10.1371/journal.pone.0166439 [DOI] [PMC free article] [PubMed] [Google Scholar]
  127. Zanetti M, Ballabio C, Abbate C, Cutaia C, Vergani C, & Bergamaschini L (2006). Mild Cognitive Impairment Subtypes and Vascular Dementia in Community-Dwelling Elderly People: A 3-Year Follow-Up Study. J. Am. Geriatr. Soc, 54(4), 580–586. 10.1111/j.1532-5415.2006.00658.x [DOI] [PubMed] [Google Scholar]
  128. Zou H, & Hastie T (2005). Regularization and variable selection via the elastic net. J. R. Stat. Soc. Series B Stat. Methodol, 67(2), 301–320. 10.1111/j.1467-9868.2005.00503.x [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Information

RESOURCES