Abstract
Introduction
The NIH Toolbox Cognition Battery (NIHTB-CB) is a computer-based protocol not yet validated for clinical assessment.
Methods
We administered the NIHTB-CB and traditional neuropsychological tests to 247 Memory Disorders and Alzheimer's Prevention Clinic patients with subjective cognitive decline, mild cognitive impairment, mild dementia due to Alzheimer's disease, and normal cognition. Principal component analysis, partial correlations, and univariate general linear model tests were performed to assess construct validity. Discriminant function analyses compared classification accuracy.
Results
Principal component analysis identified three conceptually coherent factors: memory (MEMNIH), executive function (EFNIH), and crystallized intelligence (CINIH). These factors were strongly associated with corresponding traditional tests and differed across diagnostic groups as expected. Both NIHTB and traditional batteries yielded strong overall discriminative ability (>80%).
Discussion
The NIHTB-CB is a valid method to assess neurocognitive domains pertinent to aging and dementia and has utility for applications in a memory clinic setting.
Keywords: Neuropsychology, Clinical assessment, Validation, Dementia, Memory, Technology
1. Introduction
Alzheimer's disease (AD) affects over 26 million people worldwide, a prevalence expected to quadruple by 2050 [1]. The pathophysiological process of AD begins years before clinical symptoms emerge, providing an opportunity for intervention in at-risk individuals [2]. Neuropsychological assessment facilitates diagnosis of AD and improves identification of individuals at risk for dementia [2], but neuropsychological testing can be a lengthy and resource-intensive process, which is not feasible in all clinics. As preventive therapies become available [3], [4], [5], efficient and sensitive assessments will be needed for longitudinal evaluations of large samples. Computerized measures offer standardized administration, streamlined scoring and analysis, access to large normative data sets, and automated capture of reaction time and other sensitive response parameters, making them highly effective for detection of subtle cognitive impairment [2], [6].
The NIH Toolbox Cognition Battery (NIHTB-CB) is a computer-administered protocol that assesses key cognitive domains [7]. It demonstrates strong convergent validity against traditional neuropsychological tests in cognitively healthy individuals [8], with normative data available across the life span [9], [10]. The NIHTB-CB was not designed for clinical assessment, although investigations are underway to explore its validity in traumatic brain injury, spinal cord injury, and stroke [11], [12], [13], [14]. In clinically normal older adults, the NIHTB-CB correlated with traditional tests and was able to identify subtle cognitive impairment [15]. To our knowledge, the NIHTB-CB has not yet been validated in samples with or at risk for AD, a critical gap in investigating its utility in the context of AD staging.
The aim of this study was to investigate the validity and utility of the NIHTB-CB in a clinical setting with patients across the continuum of cognitive decline, including patients seeking risk reduction care. The factor structure of the NIHTB-CB was examined and validated against traditional neuropsychological tests. The discriminative accuracy of the NIHTB-CB for diagnostic grouping was also evaluated. Finally, given the importance of episodic memory assessment for AD, we examined the validity and utility of the NIHTB-CB protocol with and without the delayed recall measure of the Rey Auditory Verbal Learning Test (RAVLT-DR), which is not included in the standard NIHTB-CB protocol.
2. Methods
2.1. Participants
Participants provided informed consent for the IRB-approved Comparative Effectiveness Dementia & Alzheimer's Registry at Weill Cornell Medicine/NewYork-Presbyterian Memory Disorders and Alzheimer's Prevention Clinic. The sample included patients with a range of cognitive conditions from those with family history of AD and no cognitive complaint seeking risk assessment for early intervention to those with advanced cognitive decline. Participants underwent a clinical interview (with informant), medical and neurological examinations, anthropometric and laboratory measures, neuropsychological testing, and structural brain MRI when indicated. Diagnoses were assigned following a consensus conference including a neurologist, family nurse practitioner, and multidisciplinary health care team members who integrated results from (1) clinical history and informant/self-report questionnaires to determine onset and type of cognitive and/or non-cognitive symptoms, ability to perform activities of daily living, and subjective assessment of cognitive function; (2) physical and neurological examination, including brain MRI when available; and (3) the Mini–Mental State Examination [16]. Diagnosis of subjective cognitive decline (SCD) required self-reported decline in memory over the past year according to a standardized questionnaire [17] paired with expressed concern about cognitive difficulties to the clinician, in the absence of objective evidence of impaired cognition [18]. Diagnoses of mild cognitive impairment (MCI) and dementia due to AD were determined using the most recent clinical diagnostic criteria [19], [20]. Amnestic and nonamnestic presentations of MCI were classified after consideration of scores on the delayed recall subtest of the Mini–Mental State Examination, along with description and type of cognitive complaint (problems with memory vs. attention, language, or processing speed) [21].
The current sample included 247 patients with the following diagnoses: cognitively normal (CN; n = 129), SCD (n = 46), amnestic MCI (aMCI; n = 27), nonamnestic MCI (naMCI; n = 19), and mild dementia due to AD (n = 26).
2.2. Cognitive measures
All tests were administered during a single visit (under 75 minutes) by research assistants. Five core and two supplemental NIHTB-CB measures [22] were administered, along with traditional neuropsychological measures [16], [23], [24], [25], [26], [27], [28] commonly used in dementia evaluations [28], [29], [30], [31], [32] (see Table 1).
Table 1.
Cognitive domain | NIHTB-CB tests | Traditional tests |
---|---|---|
Learning/Memory | RAVLT 1, 2, 3∗ | MMSE-DR∗ |
RAVLT-DR∗ | Logical Memory immediate recall∗ | |
Logical Memory delayed recall∗ | ||
FNAME∗ | ||
Executive function/Attention/Processing speed | DCCS† | FAS∗ |
Flanker† | ANT∗ | |
Pattern Comparison∗ | MMSE-attention∗ | |
ODS∗ | Trail-Making Test Part B∗ | |
Crystallized intelligence | Picture Vocabulary† | |
Oral Reading Recognition† |
Abbreviations: NIHTB-CB, NIH Toolbox Cognition Battery; RAVLT 1, 2, 3, Rey Auditory Verbal Learning Task immediate recall trials 1–3; RAVLT-DR, Rey Auditory Verbal Learning Task delayed recall; DCCS, Dimensional Change Card Sort; Flanker, Flanker Inhibitory Control/Attention; Pattern Comparison, Pattern Comparison Processing Speed; ODS, Oral Digit Symbol; MMSE-DR, Mini–Mental State Examination delayed recall subscore; FNAME, Face Name Associative Memory-cued first letter; FAS, verbal fluency under phonemic constraint to letters F-A-S; ANT, verbal fluency under categorical constraint (animals); MMSE-attention, Mini–Mental State Examination attention subscore.
NOTE. Trail-Making Test Part B score represents time to completion (seconds).
NOTE. Raw and computed scores are unadjusted for demographics.
Raw score.
Computed score (provided by the NIH toolbox, used for computer adaptive tests and tests whose score requires combination of accuracy and reaction time vectors).
2.3. Statistical methods
The NIHTB Picture Sequence Memory and List Sorting Working Memory subtests [22] were too challenging for participants with cognitive impairment and were not included due to low completion rates. Of the 247 patients, 52 (21%) had some missing data on the remaining NIHTB-CB protocol. Completion rates did not differ across groups (CN 81%, SCD 89%, MCI 78%) with the exception of AD (53%), where some participants completed abbreviated protocols due to time constraints, inability to complete tests, or because comprehensive testing was not clinically indicated. Missing data for the NIHTB-CB were handled with list-wise deletion.
Principal component analysis (PCA) was performed to examine the factor structure of the NIHTB tests. To test the model's applicability in people without dementia, PCA was repeated without the AD group. To test the model's utility without the additional RAVLT-DR subtest, PCA was repeated without RAVLT-DR. Factor-based scores for the NIHTB and traditional tests were computed using averaged z-scores of tests that strongly loaded onto each factor. Most factor score distributions were not normally distributed according to tests (Shapiro-Wilk, P < .001; Kolmogorov-Smirnov, P < .001) and associated histograms. Therefore, both nonparametric and parametric tests were performed.
Nonparametric Spearman's partial correlations were performed to assess relations between NIHTB and traditional factor scores within and across corresponding cognitive domains, controlling for age, sex, and education.
Univariate general linear model tests with Fisher's least significant difference pairwise comparisons of estimated marginal means were used to test group differences on continuous demographic variables and on NIHTB and traditional factor scores. χ2 tests were used to compare categorical variables (sex and race) across groups.
Discriminant function analyses examined cognitive factor scores as predictors of diagnostic group reclassification (including CN, aMCI, naMCI, and AD). SCD was not included as we did not expect correspondence with consensus diagnoses when classifying SCD using cognitive scores alone, and including this group would limit overall predictive power. Factor scores from the three unique PCAs were entered as independent predictors in three separate discriminant analyses: (1) scores from PCA 1a (NIHTB-CB including RAVLT-DR); (2) scores from PCA 2a (NIHTB-CB excluding RAVLT-DR); and (3) scores from PCA 3 (traditional tests). Cohen's κ was calculated to account for chance agreement due to differences in diagnostic group sizes. Age, sex, and education were also entered as independent variables in all three analyses. Results were considered significant at P < .05.
3. Results
Sample characteristics are shown in Table 2. Participants were 27 to 94 years old, with a mean age of 61 (±15 years). AD, aMCI, and naMCI groups were older than CN and SCD groups (P's < 0.01).
Table 2.
Total (N = 247) | CN (N = 129) | SCD (N = 46) | naMCI (N = 19) | aMCI (N = 27) | AD (N = 26) | P | |
---|---|---|---|---|---|---|---|
Age (mean, SD), years | 61.0 (14.7) | 52.5 (12.0) | 62.9 (11.2) | 71.8 (8.1) | 74.3 (7.2) | 78.4 (8.5) | <.001∗ |
Age range | 27–94 | 27–82 | 34–80 | 56–88 | 58–86 | 66–94 | |
Sex (% female) | 52% | 55% | 63% | 32% | 48% | 38% | n.s. |
Education (mean, SD) | 15.6 (1.2) | 15.9 (1.0) | 15.6 (1.0) | 15.4 (1.3) | 15.4 (1.7) | 14.7 (1.7) | .001† |
Race (% white) | 65% | 62% | 72% | 63% | 67% | 69% | n.s. |
Total MMSE | 28.2 (3.3) | 29.5 (.8) | 29.4 (1.4) | 28.7 (1.4) | 26.3 (2.2) | 19.8 (4.4) | <.001∗ |
Abbreviations: CN, cognitively normal; SCD, subjective cognitive decline; naMCI, nonamnestic MCI; aMCI, amnestic MCI; AD, Alzheimer's disease; MMSE, Mini–Mental State Examination; n.s., nonsignificant.
NOTE. MMSE P values reflect General Linear Model test of between-subjects effect after covarying for age, sex, and education.
P < .001.
P < .01.
3.1. PCA of NIHTB-CB
PCA was conducted on all 10 NIHTB tests with orthogonal rotation (PCA 1a; Table 3). Three factors were identified, explaining 83% of the total variance: (1) learning/memory (MEMNIH; where NIH = NIHTB); (2) executive function (EFNIH); and (3) crystallized intelligence (CINIH). The factor structure remained unchanged after excluding participants with AD (PCA 1b; Supplementary Table 1).
Table 3.
PCA 1a (including RAVLT-DR, N = 197) |
PCA 2a (excluding RAVLT-DR, N = 198) |
||||
---|---|---|---|---|---|
Component |
Component |
||||
1 |
2 |
3 |
1 |
2 |
|
Memory (MEMNIH) | Executive function (EFNIH) | Crystallized intelligence (CINIH) | Executive function/working memory (EF/WMNIH) | Crystallized intelligence (CINIH) | |
RAVLT Trial 2 | 0.886 | 0.823 | |||
RAVLT-DR | 0.853 | ||||
RAVLT Trial 3 | 0.852 | 0.838 | |||
RAVLT Trial 1 | 0.818 | 0.711 | |||
Flanker | 0.897 | 0.842 | |||
DCCS | 0.874 | 0.84 | |||
Pattern Comparison | 0.833 | 0.867 | |||
ODS | 0.741 | 0.876 | |||
Oral Reading Recognition | 0.89 | 0.891 | |||
Picture vocabulary | 0.702 | 0.714 | |||
% Explained variance | 35% | 33% | 15% | 55% | 17% |
Total % explained variance | 83% | 72% |
Abbreviations: NIHTB-CB, NIH Toolbox Cognition Battery; PCA, principal component analysis; RAVLT, Rey Auditory Verbal Learning Test immediate recall; RAVLT-DR, Rey Auditory Verbal Learning Task delayed recall; Flanker, Flanker Inhibitory Control/Attention; DCCS, Dimensional Change Card Sort; Pattern Comparison, Pattern Comparison Processing Speed; ODS, Oral Digit Symbol.
NOTE. PCA was conducted using varimax with Kaiser Normalization. Factors with eigenvalues greater than 1, with a maximum iteration of 25, were extracted—factor loadings shown after orthogonal rotation.
NOTE. PCA 1a rotation converged in 5 iterations. The Kaiser-Meyer-Olkin (KMO) measure of sampling adequacy = 0.90.
NOTE. PCA 2a rotation converged in 3 iterations. KMO = 0.89.
PCA conducted on the NIHTB tests excluding RAVLT-DR identified a solution with only two factors explaining 72% of the variance (PCA 2a; Table 3): (1) EF/working memory (EF/WMNIH) and (2) CINIH. A final PCA without RAVLT-DR and excluding AD participants yielded a three-factor solution, similar to PCA 1a, explaining 79% of the variance (PCA 2b; Supplementary Table 2).
As we were interested in assessing the utility of the NIHTB-CB in a clinical population, factor-based scores from PCAs including all diagnostic groups were calculated (PCAs 1a and 2a). All tests had adequate unidimensionality (factor loadings at or above 0.70 for a single latent construct) [33] and Kaiser-Meyer-Olkin sampling adequacy [34].
3.2. PCA of traditional tests
PCA of traditional tests identified two factors, explaining 67% of the total variance (PCA 3; Table 4): (1) memory (MEMT; where T = traditional tests) and (2) executive function (EFT). All tests demonstrated adequate unidimensionality and Kaiser Meyer Olkin sampling adequacy [33], [34]. Z-scores for each test uniquely loaded onto each component were averaged to calculate factor-based scores for MEMT and EFT from PCA 3. Descriptive data for all factor scores are shown in Supplementary Fig. 1.
Table 4.
PCA 3 |
||
---|---|---|
Component | ||
1 |
2 |
|
Memory (MEMT) | Executive function (EFT) | |
Logical Memory immediate recall | 0.875 | |
Logical Memory delayed recall | 0.807 | |
FNAME | 0.727 | |
MMSE-DR | 0.626 | |
FAS | 0.83 | |
ANT | 0.77 | |
MMSE-attention | 0.718 | |
Trails B | −0.648 | |
% Explained variance | 34% | 33% |
Total % explained variance | 67% |
Abbreviations: PCA, principal component analysis; FNAME, Face Name Associative Memory-cued first letter; MMSE-DR, Mini–Mental State Examination delayed recall subscore; FAS, verbal fluency under phonemic constraint to letters F-A-S; ANT, verbal fluency under categorical constraint (animals); MMSE-attention, Mini–Mental State Examination attention subscore.
NOTE. PCA was conducted using varimax with Kaiser Normalization. Factors with eigenvalues greater than 1, with a maximum iteration of 25, were extracted—factor loadings shown after orthogonal rotation.
NOTE. PCA 3 rotation converged in 3 iterations. The Kaiser-Meyer-Olkin (KMO) measure of sampling adequacy = 0.75.
NOTE. Owing to missing data for FNAME and Logical Memory tests added to the protocol at a later date, alternate factor-based scores for the traditional memory domain were calculated leveraging the most available data (i.e., average of MMSE-DR and Logical Memory when FNAME was missing). This factor-based score is noted as MEMT1 and is considered in certain analyses by clinical group.
3.3. Associations between NIHTB and traditional tests
Nonparametric partial correlations controlling for age, sex, and education indicated that NIHTB factor scores were significantly correlated with corresponding traditional factor scores assessing the same cognitive domain (P's < 0.01), demonstrating convergent validity (Table 5). Evidence of discriminant validity included lower correlations with traditional factor scores of different cognitive domains. Similar associations were found when asymptomatic (CN and SCD) versus symptomatic (MCI and AD) groups were analyzed separately (Supplementary Table 3).
Table 5.
PCA 1 (NIHTB-CB with RAVLT-DR) |
PCA 2a (NIHTB-CB without RAVLT-DR) |
|||||
---|---|---|---|---|---|---|
MEMNIH | EFNIH | CINIH | EF/WMNIH | CINIH | ||
PCA 3 (Traditional tests) | MEMT | 0.471∗ | 0.320∗ | 0.196 | 0.500† | 0.196 |
MEMT1 | 0.389† | 0.270† | 0.314† | 0.348† | 0.314† | |
EFT | 0.423† | 0.510† | 0.519† | 0.547† | 0.519† | |
EDUCATION | 0.312† | 0.171∗ | 0.468† | 0.253† | 0.468† |
Abbreviations: NIHTB-CB, NIH Toolbox Cognition Battery; PCA, principal component analysis; MEMNIH, NIHTB-CB memory factor; EFNIH, NIHTB-CB executive function factor; CINIH, NIHTB-CB crystallized intelligence factor; EF/WMNIH, NIHTB-CB executive function/working memory factor; MEMT, traditional memory factor; MEMT1, traditional memory factor with most available data; EFT, traditional executive function factor.
NOTE. Partial correlations controlled for age, sex, and education.
P < .05.
P < .001.
We evaluated the construct validity of CINIH against education, a common proxy for IQ. Bivariate nonparametric Spearman's correlations showed a significant relation between CINIH and education (rs = 0.468, P < .001), and Fisher's Z tests showed that relation was stronger than the relation between education and all other NIHTB factors (rs = 0.312 with MEMNIH; rs = 0.171 with EFNIH; rs = 0.253 with EF/WMNIH; P's < 0.05). Correlations between CINIH and education also were observed within the asymptomatic versus symptomatic groups (P's < 0.05; Supplementary Table 3).
3.4. Group differences in factor scores
Independent-samples Kruskal-Wallis tests for nonparametric data demonstrated significant differences on each factor score (NIHTB-CB and traditional) across diagnostic groups (P's < 0.01). Additional parametric analyses (univariate general linear model) were performed to covary for age, sex, and education when examining differences between groups (Supplementary Fig. 1). Overall, analyses showed a significant effect of group for each factor score, and pairwise comparisons demonstrated expected relative performance trends according to type and level of cognitive impairment (CN/SCD < MCI < AD). Detailed differential performance data on factor scores between diagnostic groups are found in the Supplementary Fig. 1.
3.5. Prediction of group membership
Classification results from discriminant function analyses are presented in Fig. 1.
In the first discriminant function analysis (NIHTB PCA 1a, including RAVLT-DR), the overall χ2 test was significant (Wilks λ = 0.225, χ2 = 187.84, df = 18, P < .001). The first function accounted for 94.2% of the variance in diagnostic group membership, with a Canonical correlation of 0.857. Reclassification of cases based on new canonical variables was successful at an overall rate of 84.1% and remained substantial [35] after adjusting for chance (Cohen's κ = 0.66).
In the second discriminant function analysis (NIHTB PCA 2a, excluding RAVLT-DR), the overall χ2 test was significant (Wilks λ = 0.258, χ2 = 172.89, df = 15, P < .001). The first function accounted for 96.7% of the variance in diagnostic group membership, with a Canonical correlation of 0.848. Reclassification of cases based on new canonical variables was successful at an overall rate of 79.7% yet markedly decreased after adjusting for chance (Cohen's κ = 0.58).
The third discriminant function analysis (traditional tests from PCA 3) demonstrated a significant overall χ2 test (Wilks λ = 0.214, χ2 = 173.53, df = 15, P < .001). The first function accounted for 90.7% of the variance in diagnostic group membership, with a Canonical correlation of 0.852. Reclassification of cases based on new canonical variables was successful at an overall rate of 84.7% and remained substantial after adjusting for chance (Cohen's κ = 0.66).
Regarding specific diagnostic groups, tests from PCA 1a were superior at classifying CN and naMCI, whereas tests from PCA 3 were superior at classifying aMCI and AD. Overall, tests from the NIHTB-CB including RAVLT-DR (PCA 1a) were as good as traditional tests (PCA 3) in classifying clinical groups, and the NIHTB-CB with RAVLT-DR (PCA 1a) performed better than without RAVLT-DR (PCA 2a). Both the NIHTB and traditional batteries demonstrated relative weaknesses in distinguishing MCI. Upon review of the 55%–63% of MCI participants incorrectly classified using the NIHTB-CB, individuals were more likely to be grouped CN (25%–27%) than as the alternate aMCI/naMCI group (13%–20%) or as AD (16%). Among the 52% incorrectly classified using traditional tests, aMCI individuals were equally likely to be grouped as AD or CN (17%), whereas naMCI were equally likely to be grouped as CN or aMCI (31%).
4. Discussion
Our findings support the validity and utility of the NIHTB-CB augmented with a delayed recall subtest for assessment of cognitive status in a clinic setting including younger, middle-aged, and older adults at different stages of cognitive health. We found that (1) the factor structure supported the domains the NIHTB-CB was designed to measure; (2) performance on the NIHTB-CB varied in a manner consistent with performance on traditional neuropsychological tests; (3) the NIHTB-CB explained more variance in cognitive performance and demonstrated a higher agreement rate with consensus diagnoses when RAVLT-DR was included; and (4) the NIHTB-CB with RAVLT-DR demonstrated classification agreement similar to that of the traditional tests. Although the NIHTB-CB was not designed to substitute comprehensive neuropsychological assessment, these data indicate it is a valid tool for broad assessment of cognition in dementia and predementia conditions.
PCA analyses confirmed the convergent validity of the NIHTB tests with strong factor loadings following a pattern consistent with the construct each test was designed to measure [33], similar to that found in the initial validation study in a cognitively unimpaired adult sample [36]. Executive function, learning/memory, and crystallized intelligence tests consistently clustered together across all PCAs both with and without RAVLT-DR and with and without AD participants. When RAVLT-DR was excluded in PCA 2a (including all groups), a two-factor structure emerged, explaining a lower percentage of total variance. These results suggest that when using the NIHTB-CB to measure cognition in a heterogeneous sample with memory impairment, the inclusion of a delayed memory test is necessary to explain maximum variance and to fully inform patterns of cognitive performance across multiple domains. This concept is further supported by the higher correspondence with consensus diagnoses resulting from the NIHTB discriminant function model that includes RAVLT-DR, overall and especially for reclassification of dementia due to AD. Although in a nonclinical sample, learning is highly associated with recall performance, this will tend not to be the case in individuals with memory deficits such as those with aMCI and AD. The addition of the RAVLT-DR subtest was also of particular importance in our sample given the fact that many MCI and AD patients were unable to complete the standard NIHTB memory tests (Picture Sequencing and List Sorting), an important consideration for future development of the NIHTB-CB in an AD context.
Our second aim was to compare NIHTB tests to traditional neuropsychological tests. NIHTB factor scores were significantly correlated with traditional factor scores of the same cognitive domains, supporting the convergent validity of the NIHTB-CB. These results held when asymptomatic (CN and SCD) and symptomatic (MCI and AD) groups were analyzed separately. Between-group differences on the NIHTB factors patterned as expected and similarly to traditional tests, which also supported the construct validity of the NIHTB-CB.
As compared with available traditional measures, the NIHTB-CB including RAVLT-DR demonstrated similar classification agreement with clinical groups, even after adjusting for chance. The NIHTB-CB without RAVLT-DR demonstrated the lowest rates of classification agreement. These results suggest that the NIHTB-CB with RAVLT-DR may be just as effective at classifying clinical groups as traditional tests and has the potential to do so more efficiently. Neither the NIHTB nor traditional batteries were highly effective at classifying aMCI and naMCI, which speaks to the importance of measures that can accurately distinguish prodromal stages of AD. The error trend of the NIHTB tests, which were more likely to classify MCI as CN, may suggest the difficulty level needs to be increased to detect mild impairment seen in MCI. This type of systematic error is preferable to the unsystematic error demonstrated by the traditional tests (equally likely to misclassify aMCI as AD or NC, and naMCI as CN or aMCI), as it can be corrected using data modification techniques.
Considering the influence of demographics on cognitive test performance [37], [38], it is worthwhile to note we did not use fully-adjusted scores available through the NIHTB-CB software, which account for age, sex, race, ethnicity, and education [9], [10]. Instead, we used nonadjusted raw and computed scores of the NIHTB and traditional tests when calculating factor scores and subsequently covaried for demographic variables concurrently in all other analyses. We chose this method because the full range of demographic adjustments was not available for all traditional tests. Although our methodological decision was important for establishing the convergent validity of the NIHTB-CB, the normative adjustments available through the NIHTB-CB should be applied in clinical settings to account for differences in demographic characteristics such as race and education that may influence cognitive test results.
A future research direction will be validation of the NIHTB-CB in longitudinal assessment and with AD biomarkers, including positron emission tomography, cerebrospinal fluid amyloid, and tau scans [39]. Future studies and accumulation of normative data within a diverse population, including low education groups, will enable identification and dissemination of cut-scores for more efficient classification of different stages. An advantageous feature of the NIHTB-CB is incorporation of a crystallized intelligence composite, which others have noted may lend a unique method to efficiently measure cognitive reserve and increase sensitivity to longitudinal decline [15]. We found that the crystallized intelligence factor (CINIH) correlated with education across both asymptomatic and symptomatic groups. The utility of adopting criteria for detection of cognitive impairment that considers decrements from estimated premorbid functioning has been proposed [14] and explored in the context of preclinical AD [40]. Further investigation of cognitive profiles seen at preclinical stages of AD [40], [41] will be a critical next step toward examining the utility of the NIHTB-CB for earlier detection.
A limitation of this study includes the relatively low agreement rate for classification of MCI. Although the consensus diagnoses represent those that would be expected to be assigned in a neurology clinic setting, it is important to consider that detailed analysis of comprehensive neuropsychological test results has been shown to improve diagnostic accuracy [42]. Although the consensus diagnosis process used here is conventional in many clinical settings, these diagnoses are not considered “gold standard” and may not meet research criteria. However, our intent was to compare typical clinical practice methodology by primary care physicians and general neurologists (among other specialists) with the NIHTB tests to differentiate groups along the AD spectrum. Furthermore, the pattern of scores on both the NIHTB and traditional batteries, which demonstrated increasing impairment across groups (CN < MCI < AD), provides support for overall accuracy of the original diagnoses.
Another limitation resulting from our observational design includes a cohort with diagnostic groups of varying sample sizes. We attempted to address this by examining the consistency of results within asymptomatic and symptomatic groups; however, larger samples and more consistent group sizes will be needed to replicate these results. Strengths include standardized administration of the battery for all clinic patients, and the wide age range and heterogeneous makeup of participants.
Protocols for applying the NIHTB-CB outside of a research context may depend on the setting. Although the NIHTB-CB can be administered by technicians and scored by the program, interpretation should continue to be the domain of clinicians experienced in understanding the application of such tests in combination with clinical history, particularly if diagnoses are being shared or clinical recommendations made. If cut-scores are established in the future, individuals exceeding an asymptomatic threshold or those with borderline scores may proceed to complete comprehensive neuropsychological evaluation, saving resources and avoiding costly evaluations for individuals identified as CN.
In conclusion, we have shown that the NIHTB-CB augmented with a delayed recall subtest is valid and useful for assessing cognitive capability in individuals at risk for, and those diagnosed with, age-related cognitive disorders such as AD. Most participants had little trouble acclimating to the computer-based test procedure, which shows, similar to other studies [15], [43], [44], that computerized cognitive testing is a feasible method to test younger, middle-aged, and older adults. Data collection was automated, allowing for score reports to be generated and interpreted rapidly. The protocol demonstrated efficiency in the context of a memory clinic and produced a relatively comprehensive evaluation after a brief assessment. Although a traditional neuropsychological evaluation often requires several hours, administration time for our version of the NIHTB-CB ranged from 35 to 40 minutes. These results support the usefulness of the NIHTB-CB as a key part of the clinical diagnostic evaluation and a promising tool for neuroepidemiological studies.
Research in Context.
-
1.
Systematic review: Search engines (PubMed, Web of Science, etc.) were used to identify literature on the development, use, and validity of the NIH Toolbox Cognition Battery (NIHTB-CB).
-
2.
Interpretation: The NIHTB-CB was found to be comparable to traditional neuropsychological measures in characterizing cognitive performance, and the inclusion of a delayed recall subtest substantially improved its validity and utility in this population. The NIHTB-CB is a valid and useful computerized test battery for brief assessment of cognition in a memory clinic and Alzheimer's risk reduction setting.
-
3.
Future directions: Some NIHTB-CB subtests are too difficult for individuals with dementia due to AD; delayed list learning should be added for these patients. Longitudinal evaluation should explore the predictive validity of the NIHTB-CB against AD biomarkers. NIHTB-CB cut-scores for MCI and dementia subgroups should be established. Measures of crystallized intelligence available in the NIHTB-CB may be leveraged to characterize cognitive abilities relative to premorbid intelligence.
Acknowledgments
The authors would like to thank Chiashin Shih, Jessica Shum, Pedja Stevanovic, and Hemali Patel for assistance with data management and editorial advice.
This study was funded by philanthropic support (Zuckerman Family Foundation, proceeds from the Annual Memories for Mary fundraiser organized by David and Kathy Twardock, the Rimora Foundation, the Washkowitz Family in Memory of Alan Washkowitz, and contributions from grateful patients of the Alzheimer's Prevention Clinic, Weill Cornell Memory Disorders Program), the Women's Alzheimer's Movement, Hilarity for Charity, the Leon Levy Foundation, the Samuel I. Newhouse Foundation, the Weill Cornell Medical College Clinical and Translational Science Center (NIH/NCATS #UL1TR002384), and NIH PO1AG026572. The funders had no role in the study design, data collection, decision to publish, or manuscript preparation.
Footnotes
The authors have declared that no conflict of interest exists.
Supplementary data related to this article can be found at https://doi.org/10.1016/j.dadm.2018.10.002.
Supplementary data
References
- 1.Brookmeyer R., Johnson E., Ziegler-Graham K., Arrighi H.M. Forecasting the global burden of Alzheimer's disease. Alzheimers Dement. 2007;3:186–191. doi: 10.1016/j.jalz.2007.04.381. [DOI] [PubMed] [Google Scholar]
- 2.Sperling R.A., Aisen P.S., Beckett L.A., Bennett D.A., Craft S., Fagan A.M. Toward defining the preclinical stages of Alzheimer's disease: Recommendations from the national institute on aging-Alzheimer's association workgroups on diagnostic guidelines for Alzheimer's disease. Alzheimers Dement. 2011;7:280–292. doi: 10.1016/j.jalz.2011.03.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Sperling R.A., Rentz D.M., Johnson K.A., Karlawish J., Donohue M., Salmon D.P. The A4 study: stopping AD before symptoms begin? Sci Transl Med. 2014;6:228fs13. doi: 10.1126/scitranslmed.3007941. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Seifan A., Schelke M., Obeng-Aduasare Y., Isaacson R. Early life epidemiology of Alzheimer's disease-a critical review. Neuroepidemiology. 2015;45:237–254. doi: 10.1159/000439568. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Schelke M.W., Hackett K., Chen J.L., Shish C., Shum J., Montgomery M.E. Nutritional interventions for Alzheimer's prevention: A clinical precision medicine approach. Ann N Y Acad Sci. 2016;1367:50–56. doi: 10.1111/nyas.13070. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Racine A.M., Clark L.R., Berman S.E., Koscik R.L., Mueller K.D., Norton D. Associations between performance on an abbreviated CogState battery, other measures of cognitive function, and biomarkers in people at risk for Alzheimer's disease. J Alzheimers Dis. 2016;54:1395–1408. doi: 10.3233/JAD-160528. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Gershon R.C., Wagster M.V., Hendrie H.C., Fox N.A., Cook K.F., Nowinski C.J. NIH toolbox for assessment of neurological and behavioral function. Neurology. 2013;80:S6. doi: 10.1212/WNL.0b013e3182872e5f. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Weintraub S., Dikmen S.S., Heaton R.K., Tulsky D.S., Zelazo P.D., Slotkin J. The cognition battery of the NIH toolbox for assessment of neurological and behavioral function: Validation in an adult sample. J Int Neuropsychol Soc. 2014;20:567–578. doi: 10.1017/S1355617714000320. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Casaletto K.B., Umlauf A., Beaumont J., Gershon R., Slotkin J., Akshoomoff N. Demographically corrected normative standards for the English version of the NIH toolbox cognition battery. J Int Neuropsychol Soc. 2015;21:378–391. doi: 10.1017/S1355617715000351. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Casaletto K.B., Umlauf A., Marquine M., Beaumont J.L., Mungas D., Gershon R. Demographically corrected normative standards for the Spanish language version of the NIH toolbox cognition battery. J Int Neuropsychol Soc. 2016;22:364–374. doi: 10.1017/S135561771500137X. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Tulsky D.S., Carlozzi N.E., Holdnack J., Heaton R.K., Wong A., Goldsmith A. Using the NIH toolbox cognition battery (NIHTB-CB) in individuals with traumatic brain injury. Rehabil Psychol. 2017;62:413. doi: 10.1037/rep0000174. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Tulsky D.S., Heinemann A.W. The clinical utility and construct validity of the NIH toolbox cognition battery (NIHTB-CB) in individuals with disabilities. Rehabil Psychol. 2017;62:409. doi: 10.1037/rep0000201. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Carlozzi N.E., Tulsky D.S., Wolf T.J., Goodnight S., Heaton R.K., Casaletto K.B. Construct validity of the NIH toolbox cognition battery in individuals with stroke. Rehabil Psychol. 2017;62:443. doi: 10.1037/rep0000195. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Holdnack J.A., Tulsky D.S., Brooks B.L., Slotkin J., Gershon R., Heinemann A.W. Interpreting patterns of low scores on the NIH toolbox cognition battery. Arch Clin Neuropsychol. 2017;32:574–584. doi: 10.1093/arclin/acx032. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Buckley R.F., Sparks K.P., Papp K.V., Dekhtyar M., Martin C., Burnham S. Computerized cognitive testing for use in clinical trials: A comparison of the NIH toolbox and cogstate C3 batteries. J Prev Alzheimers Dis. 2017;4:3–11. doi: 10.14283/jpad.2017.1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Folstein M.F., Folstein S.E., McHugh P.R. “Mini-mental state”: A practical method for grading the cognitive state of patients for the clinician. J Psychiatr Res. 1975;12:189–198. doi: 10.1016/0022-3956(75)90026-6. [DOI] [PubMed] [Google Scholar]
- 17.Jorm A.F., Christensen H., Korten A.E., Henderson A.S., Jacomb P.A., Mackinnon A. Do cognitive complaints either predict future cognitive decline or reflect past cognitive decline? A longitudinal study of an elderly community sample. Psychol Med. 1997;27:91–98. doi: 10.1017/s0033291796003923. [DOI] [PubMed] [Google Scholar]
- 18.Molinuevo J.L., Rabin L.A., Amariglio R., Buckley R., Dubois B., Ellis K.A. Implementation of subjective cognitive decline criteria in research studies. Alzheimers Dement. 2017;13:296–311. doi: 10.1016/j.jalz.2016.09.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Albert M.S., DeKosky S.T., Dickson D., Dubois B., Feldman H.H., Fox N.C. The diagnosis of mild cognitive impairment due to Alzheimer's disease: Recommendations from the national institute on aging-Alzheimer's association workgroups on diagnostic guidelines for Alzheimer's disease. Alzheimers Dement. 2011;7:270–279. doi: 10.1016/j.jalz.2011.03.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.McKhann G.M., Knopman D.S., Chertkow H., Hyman B.T., Jack C.R., Kawas C.H. The diagnosis of dementia due to Alzheimer's disease: recommendations from the national institute on aging-Alzheimer's association workgroups on diagnostic guidelines for Alzheimer's disease. Alzheimers Dement. 2011;7:263–269. doi: 10.1016/j.jalz.2011.03.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Petersen R.C. Mild cognitive impairment as a diagnostic entity. J Intern Med. 2004;256:183–194. doi: 10.1111/j.1365-2796.2004.01388.x. [DOI] [PubMed] [Google Scholar]
- 22.Slotkin J., Kallen M., Griffith J., Magasi S., Salsman H., Nowinski C. National Institutes of Health; Bethesda, MD: 2012. NIH toolbox technical manual. [Google Scholar]
- 23.Wechsler D. Psychological Corporation; New York: 1987. WMS-R: Wechsler memory scale-revised. [Google Scholar]
- 24.Rentz D.M., Amariglio R.E., Becker J.A., Frey M., Olson L.E., Frishe K. Face-name associative memory performance is related to amyloid burden in normal elderly. Neuropsychologia. 2011;49:2776–2783. doi: 10.1016/j.neuropsychologia.2011.06.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Benton A.L., Hamsher K.D., Varney N.R. Oxford UP; New York: 1983. Spreen 0. Contributions to neuropsychological assessment. [Google Scholar]
- 26.Parker D.M., Crawford J.R. Assessment of frontal lobe dysfunction. In: Crawford J.R., Parker D.M., McKinnley W.W., editors. A handbook of neuropsychological assessment. East Sussex, England; Erlbaum: 1992. pp. 267–291. [Google Scholar]
- 27.Reitan R.M. Reitan Neuropsychological Laboratory; Tucson, Arizona: 1979. Manual for administration of neuropsychological test batteries for adults and children. [Google Scholar]
- 28.Spreen O., Strauss E. Oxford; New York, NY: 1991. A compendium of neuropsychological tests: Administration, norms, and commentary. [Google Scholar]
- 29.Monsch A.U., Bondi M.W., Butters N., Salmon D.P., Katzman R., Thal L.J. Comparisons of verbal fluency tasks in the detection of dementia of the Alzheimer type. Arch Neurol. 1992;49:1253. doi: 10.1001/archneur.1992.00530360051017. [DOI] [PubMed] [Google Scholar]
- 30.Tombaugh T.N., Kozak J., Rees L. Normative data stratified by age and education for two measures of verbal fluency: FAS and animal naming. Arch Clin Neuropsychol. 1999;14:167–177. [PubMed] [Google Scholar]
- 31.Ala T.A., Hughes L.F., Kyrouac G.A., Ghobrial M.W., Elble R.J. The Mini Mental State Exam may help in the differentiation of dementia with lewy bodies and Alzheimer's disease. Int J Geriatr Psychiatry. 2002;17:503–509. doi: 10.1002/gps.550. [DOI] [PubMed] [Google Scholar]
- 32.Razani J., Wong J.T., Dafaeeboini N., Edwards-Lee T., Lu P., Alessi C. Predicting everyday functional abilities of dementia patients with the mini-mental state examination. J Geriatr Psychiatry Neurol. 2009;22:62–70. doi: 10.1177/0891988708328217. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Mentzer J.T., Flint D.J. Validity in logistics research. J Bus Logist. 1997;18:199. [Google Scholar]
- 34.Hutcheson C.D., Sofroniou N. Sage; London: 1999. The multivariate social scientist: Introductory statistics using generalized linear models. [Google Scholar]
- 35.Cohen J. A coefficient of agreement for nominal scales. Educ Psychol Meas. 1960;20:37–46. [Google Scholar]
- 36.Mungas D., Heaton R., Tulsky D., Zelazo P.D., Slotkin J., Blitz D. Factor structure, convergent validity, and discriminant validity of the NIH toolbox cognitive health battery (NIHTB-CHB) in adults. J Int Neuropsychol Soc. 2014;20:579–587. doi: 10.1017/S1355617714000307. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Manly J.J., Jacobs D.M., Sano M., Bell K., Merchant C.A., Small S. Cognitive test performance among nondemented elderly African Americans and whites. Neurology. 1998;50:1238–1245. doi: 10.1212/wnl.50.5.1238. [DOI] [PubMed] [Google Scholar]
- 38.Manly J.J., Jacobs D.M., Touradji P., Small S.A., Stern Y. Reading level attenuates differences in neuropsychological test performance between African American and White elders. J Int Neuropsychol Soc. 2002;8:341–348. doi: 10.1017/s1355617702813157. [DOI] [PubMed] [Google Scholar]
- 39.Karlawish J., Jack C.R., Jr., Rocca W.A., Snyder H.M., Carrillo M.C. Alzheimer's disease: The next frontier—Special Report 2017. Alzheimers Dement. 2017;13:374–380. doi: 10.1016/j.jalz.2017.02.006. [DOI] [PubMed] [Google Scholar]
- 40.Seifan A., Isaacson R. The Alzheimer's prevention clinic at Weill Cornell Medical College/New York-Presbyterian Hospital: risk stratification and personalized early intervention. J Prev Alzheimers Dis. 2015;2:254. doi: 10.14283/jpad.2015.81. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Grober E., Hall C.B., Lipton R.B., Zonderman A.B., Resnick S.M., Kawas C. Memory impairment, executive dysfunction, and intellectual decline in preclinical Alzheimer's disease. J Int Neuropsychol Soc. 2008;14:266–278. doi: 10.1017/S1355617708080302. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Bondi M.W., Edmonds E.C., Jak A.J., Clark L.R., Delano-Wood L., McDonald C.R. Neuropsychological criteria for mild cognitive impairment improves diagnostic precision, biomarker associations, and progression rates. J Alzheimers Dis. 2014;42:275–289. doi: 10.3233/JAD-140276. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Wild K., Howieson D., Webbe F., Seelye A., Kaye J. Status of computerized cognitive testing in aging: A systematic review. Alzheimers Dement. 2008;4:428–437. doi: 10.1016/j.jalz.2008.07.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Fredrickson J., Maruff P., Woodward M., Moore L., Fredrickson A., Sach J. Evaluation of the usability of a brief computerized cognitive screening test in older people for epidemiological studies. Neuroepidemiology. 2010;34:65–75. doi: 10.1159/000264823. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.