Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2021 Oct 12.
Published in final edited form as: J Alzheimers Dis. 2021;83(2):861–877. doi: 10.3233/JAD-210251

A comparison of cross-sectional and longitudinal methods of defining objective subtle cognitive decline in preclinical Alzheimer’s disease based on Cogstate One Card Learning accuracy performance

Shehroo B Pudumjee a, Emily S Lundt b, Sabrina M Albertson b, Mary M Machulda a, Walter K Kremers b, Clifford R Jack Jr c, David S Knopman d, Ronald C Petersen e, Michelle M Mielke d,e, Nikki H Stricker a
PMCID: PMC8506654  NIHMSID: NIHMS1733745  PMID: 34366338

Abstract

BACKGROUND:

Longitudinal, but not cross-sectional, cognitive testing is one option proposed to define transitional cognitive decline for individuals on the Alzheimer’s continuum.

OBJECTIVE:

Compare diagnostic accuracy of cross-sectional subtle objective cognitive impairment (sOBJ) and longitudinal objective decline (ΔOBJ) over 30 months for identifying (1) cognitively unimpaired participants with preclinical Alzheimer’s disease defined by elevated brain amyloid and tau (A+T+) and (2) incident mild cognitive impairment (MCI) based on Cogstate One Card Learning (OCL) accuracy performance.

METHOD:

Mayo Clinic Study of Aging cognitively unimpaired participants aged 50+ with amyloid and tau PET scans (n=311) comprised the biomarker-defined sample. A case-control sample of participants aged 65+ remaining cognitively unimpaired for at least 30 months included 64 who subsequently developed MCI (incident MCI cases) and 184 controls, risk-set matched by age, sex, education, and visit number. sOBJ was assessed by OCL z-scores. ΔOBJ was assessed using within subjects’ standard deviation and annualized change from linear regression or linear mixed effects (LME) models. Concordance measures Area Under the ROC Curve (AUC) or C-statistic and odds ratios (OR) from conditional logistic regression models were derived. sOBJ and ΔOBJ were modeled jointly to compare methods.

RESULTS:

sOBJ and ΔOBJ-LME methods differentiated A+T+ from A−T− (AUC=0.64, 0.69) and controls from incident MCI (C-statistic=0.59, 0.69) better than chance; other ΔOBJ methods did not. ΔOBJ-LME improved prediction of future MCI over baseline sOBJ (p=0.003) but not over 30-month sOBJ (p=0.09).

DISCUSSION:

Longitudinal decline did not offer substantial benefit over cross-sectional assessment in detecting preclinical Alzheimer’s disease or incident MCI.

Keywords: Sensitivity and Specificity, Neuropsychology, Memory, Validity, Amyloid, Tau, Cognigram, Reliable Change Index, Biomarker, Transitional Cognitive Decline

INTRODUCTION

Research criteria proposed by the National Institute on Aging and Alzheimer’s Association (NIA-AA) workgroup biologically defines Alzheimer’s disease as the presence of amyloidosis and atypical tau deposition [1]. Both amyloid and tau have been found to impact cognitive performance, particularly memory [25]. Although PiB-PET and tau-PET methods detect in-vivo amyloid and tau deposition and inform risk of future cognitive decline [68], these methods are expensive, precluding widespread use. Similarly, person-administered neuropsychological measures are sensitive to change up to 20 years prior to a diagnosis of MCI [9] but are time intensive and not suited for population-wide cognitive screening. Consequently, computerized cognitive assessment, capable of broader dissemination, shows promise for early detection of cognitive decline [10].

The Cogstate Brief Battery (CBB) [11, 12] is a brief computerized cognitive assessment that can be administered in clinic settings or remotely. The One Card Learning (OCL) subtest from the CBB has previously shown some promise for identifying individuals with amyloidosis. Performance on the OCL subtest of the CBB using a ≤5th or ≤10th percentile sample-specific slope was used to identify memory decliners in a community dwelling sample [13, 14]. A higher proportion of those meeting the definition of memory decline had amyloidosis relative to non-decliners [13]. Participant registries following a large cohort of volunteers may be able to use computerized cognitive measures to identify and selectively recruit individuals who are more likely to have amyloidosis to enrich clinical trials, and several participant registries use the CBB [1517]. Given that the success of enrichment strategies may depend, in part, on the ability of the CBB to accurately identify individuals with amyloidosis, additional validation of the ability of the CBB to differentiate individuals with and without AD imaging biomarkers is needed. Further, although intra-individual decline on longitudinal cognitive testing is one of several ways to meet criteria for “transitional cognitive decline” based on the numeric clinical staging scheme for individuals on the Alzheimer’s continuum (e.g., with amyloidosis) per the NIA-AA workgroup [1], subtle cognitive decline as determined from a single assessment is omitted from consideration.

This study compares different approaches to assessing objective subtle cognitive decline in cognitively unimpaired individuals based on OCL accuracy performance (Figure 1). We define subtle objective cognitive impairment (sOBJ) as performance on cognitive measures below normative expectations on a single assessment (cross-sectional) [18]. We define subtle objective longitudinal cognitive decline (∆OBJ) as a change over time using two or more assessments. Although these specific sOBJ and ∆OBJ terms are recently proposed by our group in an effort to differentiate objective subtle cognitive decline from the widely used subjective cognitive decline terminology [18, 19], these concepts and the importance of differentiating impairment from decline are not new. For example, Baker and colleagues [20] previously provided support for the presence of both cognitive impairment and cognitive decline in preclinical AD in their meta-analysis. We recently demonstrated that sOBJ defined as performance ≤ −1 standard deviation (SD) below age-specific Cogstate norms on the CBB Learning/Working Memory Composite did not differentiate individuals with elevated brain amyloid (A+) either alone or in combination with elevated tau (T+) better than chance [18]. It did predict incident mild cognitive impairment (MCI) better than chance with an AUC value of .70 [18]. However, longitudinal performance was not considered in that study.

Figure 1.

Figure 1.

Study definitions: Methods of defining subtle cognitive decline.

The primary aim of this study was to compare the diagnostic accuracy of cross-sectional (sOBJ) and longitudinal (ΔOBJ) performance on Cogstate OCL accuracy for identifying (1) cognitively unimpaired individuals with preclinical AD determined by amyloid and tau PET, and (2) cognitively unimpaired individuals who later developed incident MCI. We predicted that ΔOBJ methods (serial assessment) would have greater diagnostic and prognostic accuracy than sOBJ methods (single assessment). We also compared several methods of defining ΔOBJ.

METHOD

Participants were from the Mayo Clinic Study of Aging (MCSA), a prospective population-based study that uses age and sex-stratified random sampling methods and is conducted among residents of Olmsted County, MN [21]. MCSA participants are invited to undergo imaging studies, including PiB-PET (since 2008) and tau-PET (since 2016). MCSA study visits occurred approximately every 15 months. Briefly, clinic visits included an interview and completion of the Clinical Dementia Rating® scale [22] by a study coordinator, neurological evaluation and administration of the Short Test of Mental Status [23] by a physician, and neuropsychological testing [21]. Consensus diagnoses of cognitively unimpaired, MCI, or dementia were made blind to CBB performance, biomarker status, and prior clinical information including past performance on neuropsychological testing and prior diagnoses [21, 24]. Administration of the CBB began in 2012 for individuals 50–69 years old and in 2013 for individuals 70+. CBB was completed in clinic during MCSA study visits and at a midpoint between full study visits. Between clinic visits, home administration of the CBB was offered to all participants at 4-month intervals beginning in 2015, following initial piloting. Some participants chose to come to the clinic at the midpoint between MCSA study visits to complete Cogstate instead of completing it at home. See Table 1 for the approximate timing of CBB sessions across individuals selecting the 1) in-clinic only or 2) in-clinic and at home tracks.

Table 1.

Target Cogstate Brief Battery (CBB) session schedule depicted by in clinic and clinic/home tracks.

Participants who chose the clinic only track.
Time (months) 0 7.5 15 22.5 30
Location in-clinic in-clinic in-clinic in-clinic in-clinic
MCSA visit 1 2 3
Participants who chose the clinic/home track.
Time (months) 0 3–4 7–8 11–12 15 18–19 22–23 26–27 30
Location in-clinic home home home in-clinic home home home in-clinic
MCSA visit 1 2 3

Note. All participants completed the CBB during full MCSA clinic visits that are scheduled every 15 months. Participants chose whether to complete intervening CBB sessions in the clinic (less frequent) or at home (more frequent). CBB = Cogstate Brief Battery. MCSA = Mayo Clinic Study of Aging.

The study protocols were approved by the Mayo Clinic and Olmsted Medical Center Institutional Review Boards. All participants provided written informed consent.

Inclusion Criteria

Participants may have been enrolled in the MCSA prior to initiation of CBB testing and so baseline CBB is not always the first MCSA study visit. For the purposes of this manuscript, baseline will represent the participants’ visit that included the initial CBB session.

Participants with biomarker data:

Cognitively unimpaired individuals aged 50 years or older at the time of their baseline CBB were eligible. Participants were required to have two MCSA follow-up visits to be used for computations of change in CBB measures as described in Study Definitions. All participants had 30 months of follow-up CBB data on average so in remaining sections, the CBB sessions occurring from the MCSA visit where the CBB was first administered to approximately the second MCSA follow-up visit after that initial CBB session will be referred to as the 30-month follow-up period. Individuals were also required to have amyloid and tau PET scans within +/−3 years of their baseline; the most recent imaging was used if repeat imaging was available. Amyloid and tau positivity was determined using Pittsburgh Compound B PET (PiB-PET) and tau PET (AV1451) [2527]. Cutoffs to determine amyloid and tau positivity were SUVR ≥ 1.48 (centiloid 22) [28] and ≥ 1.25 [26] respectively. Participants, all cognitively unimpaired, were grouped based on amyloid and tau status to include those meeting criteria for biological AD [1] (A+T+ n=32), Alzheimer’s pathologic change (A+T− n=68), and those with normal AD biomarkers (A−T− n=211).

Participants for incident MCI analyses:

Cognitively unimpaired individuals aged 65 years or older at the time of their baseline CBB were eligible. In order to compute a rate of change in OCL accuracy scores that could fairly be compared against a single CBB session, participants were required to remain cognitively unimpaired through a run-in period of 2 follow-up visits (approximately 30-months from baseline to 2nd follow-up) and have usable accuracy data at baseline and 2nd follow-up (i.e., no missing data or data failing completion criteria; See Figure 2 ‘a’). Incident MCI cases were participants receiving a diagnosis of MCI after the run-in period. 184 stable controls were 3:1 matched to 64 incident MCI cases on age, sex, education, and year of initial CBB to account for potential temporal effects. To ensure equitable opportunity for development of MCI, cases were also risk-set matched to controls having at least as much time in the study subsequent to the run-in period (see Figure 2 ‘b’); a match was valid if that length of time for a control was greater than or equal to the case. Incident MCI cases without at least one control participant match identified (n=3) were excluded. There were five cases for whom three matches were not identified within our broader sample; 1 or 2 matches were used for those cases.

Figure 2. Sample timeline and matching for cognitively unimpaired, incident MCI cases and stable controls.

Figure 2.

Note: Visualize the timing of events in a risk-set matched design with a run-in period. Eligible subjects had a run-in period of 30 months of CBB assessments (see ‘a’) which was used to derive the metrics of interest sOBJ and ΔOBJ. Incident MCI cases were those who went on to receive a diagnosis of MCI (yellow dot) after the run in period and were matched to controls (blue X) by age, sex, education, and year of initial CBB to account for potential temporal effects. To ensure equitable opportunity for development of MCI, cases were also matched to controls having at least as much time in study subsequent to the run-in period, i.e. risk-set matching (see ‘b’). We used a 3-to-1 control-to-case matching ratio. To compare the relationship between each metric of interest and incidence of MCI while accounting for a matched-design, conditional logistic regression (CLR) models were fit and stratified by matched-sets.

Cogstate Brief Battery and One Card Learning accuracy

The CBB includes four tasks. To maintain parsimony, OCL was the focus of this manuscript. OCL is a continuous visual recognition memory task that was the focus of prior studies using the CBB to identify individuals with memory decline [13, 14]. OCL requires the examinee to recognize whether a playing card has been presented earlier in the subtest by providing a yes/no response. Both correct recognition and rejection are considered for accuracy [12]. OCL accuracy (arcsine transformed) was the primary outcome variable. Sessions with failed completion flags were excluded; see our prior work for additional details about CBB administration in the MCSA [29, 30]. Additional analyses were also completed for the remaining CBB subtests and composite scores; those results are available in Supplemental Online Materials. Most participants (99%) completed their baseline CBB assessment during a clinic visit on a PC or iPad. Across all CBB sessions, 21% were completed at home.

Study Definitions (see Figure 1)

Subtle Objective Cognitive Impairment (sOBJ):

Normative data in the form of means and SDs by age groups were provided by Cogstate and used to derive age-corrected z-scores [31]. sOBJ at baseline was based on the z-score at the CBB initial session. Since cases were required to remain cognitively unimpaired throughout the 30-month run-in period for the incident MCI analyses, sOBJ at 30 months represents the z-score at the CBB session associated with the 2nd MCSA follow-up visit (approximately 30-months after baseline CBB session) for the case-control cohort.

Subtle Objective Cognitive Decline (∆OBJ):

∆OBJ was defined using an intra-individual, longitudinal approach across baseline and follow-up assessments with three different methods, as described below. Data from the baseline CBB through two follow-up MCSA study visits were included in the window used to examine change in OCL accuracy performance (approximately 30 months).

Within-subjects standard deviation (∆OBJ-WSD):

WSD is the method recommended by Cogstate to determine intra-individual change from baseline to a follow-up session, defined in this study as the last CBB session in the 30-month period. A conventional cut-off of z ≤ −1 is recommended to identify individuals with a decline in test performance. WSD values obtained from Cogstate normative data were derived across an approximately 1-month interval for ages 18–99 [31]. This process is analogous to a reliable change index and can be used to provide a z-score for change in scores across time [32, 33]. Some have criticized this method of calculating reliable change [34, 35], thus we used two additional methods to obtain change in accuracy per year (annualized change) using all available Cogstate session data over the 30-months’ time-span. Models were fit on OCL accuracy with years since baseline.

∆OBJ-LR Method:

Using a 2-stage approach, first separate linear regression (LR) models were fit for each participant. Next, the per-person slopes were extracted from each model. With a single predictor variable of time in years, the slope is equivalent to the change in accuracy given 1 unit of time. More explicitly, if yi is predicted OCL accuracy and xi is time in years then slope = (y2y1)/(x2x1) where x2x1 is 1 year. This allows us to interpret the slope as annualized change in OCL accuracy.

∆OBJ-LME Method:

Linear mixed effects (LME) models were fit between accuracy and time since baseline including subject-specific intercepts and slopes. This approach considers data from both the group and the individual. Group data is used to stabilize those with fewer observations allowing for individual slopes to align with the overall group slope more closely. This allows for the estimation of subject specific slopes that have been shrunk toward the group slope, reducing noise. Subject-specific slopes were extracted from the model and, like above, can be interpreted as measures of annualized change in the response.

Reference group:

A reference group was required to identify annualized change cut points to be applied to our groups of interest. This reference group comprised 732 cognitively unimpaired individuals aged 50–65 at baseline CBB with at least two follow-up visits, who remained cognitively unimpaired over this period (mean age = 59; mean education = 15 years; 50% male). We used a younger cohort since the prevalence of biomarker positivity increases greatly with age [7], and because defining this reference group by biomarker status alone would result in circularity for primary analyses. Change in test performance was computed as described above for ∆OBJ-LR and ∆OBJ-LME methods. The annualized change equivalent to the slope at the 10th percentile within this reference group was used as the a priori conventional cut-off for ∆OBJ-LR and ∆OBJ-LME (e.g. change < 10th%ile), similar to methods previously reported [13, 14].

Statistical Methods

We compared OCL performance and demographic characteristics between (1) biomarker positive and negative groups and (2) stable controls and incident MCI cases using linear model ANOVA tests of mean differences and chi-squared tests for frequency differences. An alpha level of .05 was used to discern statistical significance for all analyses. We report effect size using Hedge’s g which expresses the difference of group means scaled by the pooled standard deviation.

We conducted Area Under the Receiver Operating Curve (AUROC) analyses on each measure of interest to provide an aggregate measure of performance across all possible classification thresholds for differentiating between biomarker groups. We then compared the AUC values obtained from each method (sOBJ, WSD, LR, LME) [36]. Examining results based on conventional cut-off values of ≤ −1 z (sOBJ and ∆OBJ-WSD) and the 10th %ile slope (∆OBJ-LR and ∆OBJ-LME) was our primary focus, as this allows for generalizability of study findings and facilitates clinical translation of results. In exploratory analyses, we also used the Youden index method to identify study specific “optimal” cut-points [37]. This selected cut-off is the value that maximized the Youden function, i.e., maximized the difference between the true positive and false positive rate out of all possible cut-off values.

As part of the incident MCI analyses, we examined concordance statistics (c-statistic) from conditional logistic regression (CLR) models to predict individuals who developed MCI with stratification by matched-sets. Due to the binary outcome (incident MCI or stable control) the c-statistic can be interpreted in the same way as the area under the Receiver Operating Characteristic curve. To test for differences in c-statistics between methods, jack-knife variance estimates of group differences were computed. Odds ratios (OR) from each CLR model were assessed to compare the magnitude of increasing odds of future MCI per unit of cognition decrement. ANOVA F-statistics were computed to test whether joint CLR models including both sOBJ and ∆OBJ measures provided a better fit to the data over a univariate CLR model with sOBJ, allowing us to assess if one method was better than the other. To facilitate comparison of ORs for the various measures of interest, we needed to have equitable scaling. Division by sample standard deviation was used to rescale the LR and LME annualized change estimation methods making a 1-unit decline equivalent to a 1 SD decrement which is comparable to a 1z decrement.

AUROC methods do not account for the matching in a case-control framework. However, use of consistent methods with previously published manuscripts facilitates comparisons of results. Thus, AUROC analyses for discriminating cases and controls were included in our supplemental materials, despite this being a lower-powered analysis as compared to c-statistics from CLR models.

RESULTS

Cognitively unimpaired participants with biomarker data

Group mean comparisons:

See Table 2 for participant demographics. Compared to the A−T− group, the A+T+ group showed significantly lower mean OCL accuracy performance at baseline (sOBJ absolute mean difference = 0.324 z, p=0.026) and less benefit from practice over time based on ∆OBJ-LME (absolute mean difference of annualized change 0.002; p<0.001), with medium effect sizes (see Table 2 and Figure 3). A−T− and A+T+ groups did not differ statistically on OCL accuracy performance when ∆OBJ-WSD and ∆OBJ-LR methods were employed, and resultant effect sizes were small. While a similar pattern of results was observed across A+T− and A−T− comparisons, the magnitude of group differences was smaller. In addition, only ∆OBJ-LME was significantly different between groups, with benefit from practice over time in A+T− individuals compared to A−T− individuals characterized by a small effect size.

Table 2.

Demographic characteristics and performance on One Card Learning accuracy by biomarker group. All participants are cognitively unimpaired. Statistics reported are of the form mean (standard deviation, SD) unless otherwise specified.

A−T−
(n = 211)
A+T+
(n = 32)
p-value
A−T− vs A+T+
Hedge’s g A+T−
(n = 68)
p-value
A−T− vs A+T−
Hedge’s g
Age, years 62.83 (9.26) 75.26 (7.58) < 0.001 71.60 (10.37) < 0.001
Education, years 15.32 (2.24) 14.47 (2.59) 0.051 15.10 (2.55) 0.507
Male count (%) 118 (56%) 17 (53%) 0.767 36 (53%) 0.667
Follow up duration in years 2.52 (0.21) 2.49 (0.25) 0.434 2.50 (0.27) 0.512
Total CBB administrations 6.08 (1.72) 5.59 (1.46) 0.134 5.49 (1.80) 0.016
OCL accuracy
 sOBJ, z-score 0.348 (0.746) 0.024 (0.884) 0.026 0.42 0.177 (0.985) 0.131 0.21
 ΔOBJ-WSD, z-score 0.504 (0.766) 0.338 (0.914) 0.267 0.21 0.489 (1.025) 0.894 0.02
 ΔOBJ-LR, change per year 0.021 (0.035) 0.012 (0.040) 0.186 0.25 0.022 (0.046) 0.803 −0.03
 ΔOBJ-LME, change per year 0.019 (0.003) 0.017 (0.004) < 0.001 0.72 0.018 (0.004) 0.03 0.30

Note. p-values represent linear model ANOVAs for mean comparisons or Pearson’s Chi-squared test for frequency comparisons. A = amyloid; T = tau; OCL= One Card Learning; sOBJ = subtle objective cognitive impairment (cross-sectional). ΔOBJ = subtle objective cognitive decline (longitudinal); WSD = within subjects standard deviation, positive value indicates higher score across repeated assessments (e.g., benefit from practice); LR = Linear regression model; LME = linear mixed effects model.

Figure 3. Boxplots of One Card Learning accuracy performance across methods of defining subtle cognitive decline and biomarker groups.

Figure 3.

Note: Box plots representing performances on OCL accuracy separately by biomarker defined groups show the interquartile range (box), median (horizontal solid line within each box) and the extent of outlier performances (vertical lines extending above and below each box) in each of the sOBJ and ∆OBJ methods. Dashed line (---) denotes the conventional cut-off used (≤ −1 z score for sOBJ and ∆OBJ-WSD; <10th percentile slope for ∆OBJ-LR and ∆OBJ-LME methods). Individuals in all biomarker based groups with performances below this line are identified as decliners based on these cut-offs. A = amyloid, T = tau, sOBJ = subtle objective cognitive impairment at baseline based on OCL accuracy age-corrected z-score derived from Cogstate norms. ∆OBJ = subtle objective cognitive decline since baseline assessment. WSD = within subjects standard deviation z-score based on Cogstate norms. LR = Linear regression model; <10%ile slope is equivalent to a < −0.0199 annualized change on OCL accuracy raw score (arcsine transformed). LME = linear mixed effects model; <10%ile slope is equivalent to a < 0.0162 annualized change on OCL accuracy raw score (arcsine transformed). <10%ile slope cutoffs were derived from the cognitively unimpaired reference sample.

Overall diagnostic accuracy

Differentiating A−T− and A+T+ groups:

Overall diagnostic accuracy for differentiating between the A−T− and A+T+ groups was better than chance for the single baseline sOBJ and ∆OBJ-LME methods, but not for the ∆OBJ-WSD and ∆OBJ-LR methods (Table 3). When comparing total AUC across methods, sOBJ performed comparably to all three longitudinal methods (p’s > .05). The total AUC value was marginally higher for the ∆OBJ-LME method (AUC = .69) relative to the sOBJ method (AUC = .64), but this did not represent a statistically significant difference.

Table 3.

Comparison of total area under the curve (AUC) with 95% Confidence Intervals (CI) from ROC analyses to distinguish between cognitively unimpaired biomarker groups based on subtle objective cognitive impairment (sOBJ) at baseline and longitudinal decline (∆OBJ) defined using three different methods.

A−T− and A+T+ comparisons

ΔOBJ Methods (p values)
AUC (CI) WSD LR LME
sOBJ 0.64 (0.53, 0.75) 0.31 0.43 0.34
ΔOBJ-WSD 0.53 (0.42, 0.65) -- 0.35 0.03
ΔOBJ-LR 0.56 (0.44, 0.68) -- -- 0.07
ΔOBJ-LME 0.69 (0.58, 0.80) -- -- --

A−T− and A+T− comparisons

ΔOBJ Methods (p values)
AUC (CI) WSD LR LME
sOBJ 0.55 (0.47, 0.63) 0.41 0.32 0.59
ΔOBJ-WSD 0.49 (0.41, 0.58) -- 0.61 0.10
ΔOBJ-LR 0.48 (0.40, 0.57) -- -- 0.06
ΔOBJ-LME 0.57 (0.49, 0.66) -- -- --

Note. AUC confidence intervals (CI) excluding .50 indicated differentiation of biomarker groups better than chance. Total AUC values were compared across methods [36]. sOBJ = subtle objective cognitive impairment (cross-sectional). ΔOBJ = subtle objective cognitive decline (longitudinal). WSD = within subjects’ standard deviation. LR = annualized change from linear regression model. LME = annualized change from linear mixed effects model.

Differentiating A−T− and A+T− groups:

Overall diagnostic accuracy for differentiating between the A−T− and A+T− groups was not better than chance (95% CI’s of total AUC values contained 0.5) for baseline or longitudinal change methods. All methods had comparable AUC values (p’s>.05; see Table 3).

Examining performance of differing cut-offs

Differentiating A−T− and A+T+ groups:

Using conventional cut-offs of z ≤ −1 based on Cogstate norms (sOBJ and ∆OBJ-WSD) and slope < 10th %ile of the reference group (∆OBJ-LR and ∆OBJ-LME), sensitivity values for identifying A+T+ individuals were low for all methods (9–38%; see Table 4 and Figure 3). ∆OBJ-LME showed the highest sensitivity relative to other methods. Specificity values were high across methods when using conventional cut-offs (85–98%). When exploring optimal, derived cutoffs, sensitivity values showed improvement, increasing to 62% sensitivity for ∆OBJ-LME, although this leads to a decrease in specificity (75%).

Table 4.

Sensitivity and specificity for One Card Learning accuracy performance based on derived, optimal cut-offs and conventional cut-offs for subtle objective cognitive impairment (sOBJ) at baseline and longitudinal decline (∆OBJ) defined using three different methods.

A−T− and A+T+ comparisons

Index Threshold type Threshold Sensitivity Specificity
sOBJ Optimal <−0.215 z 0.50 (0.34, 0.66) 0.82 (0.76, 0.87)
Conventional ≤ −1 z 0.09 (0.00, 0.22) 0.97 (0.94, 0.99)
ΔOBJ-WSD Optimal <−0.691 z 0.19 (0.06, 0.31) 0.97 (0.95, 0.99)
Conventional ≤ −1 z 0.12 (0.03, 0.25) 0.98 (0.95, 1.00)
∆OBJ-LR Optimal < 29.6%ile (0.0047) 0.50 (0.31, 0.69) 0.69 (0.63, 0.75)
Conventional < 10%ile (−0.0199) 0.22 (0.09, 0.38) 0.89 (0.84, 0.93)
∆OBJ-LME Optimal < 16.5%ile (0.0173) 0.62 (0.47, 0.78) 0.75 (0.70, 0.81)
Conventional < 10%ile (0.0162) 0.38 (0.22, 0.56) 0.85 (0.80, 0.90)
A−T− and A+T− comparisons

Index Threshold type Threshold Sensitivity Specificity

sOBJ Optimal < 0.293 z 0.56 (0.44, 0.68) 0.57 (0.50, 0.64)
Conventional ≤−1 z 0.06 (0.01, 0.12) 0.97 (0.94, 0.99)
ΔOBJ-WSD Optimal <−0.636 z 0.12 (0.04, 0.21) 0.96 (0.93, 0.98)
Conventional ≤ −1 z 0.06 (0.01, 0.12) 0.98 (0.95, 1.00)
∆OBJ-LR Optimal < 6.4%ile (−0.0257) 0.16 (0.07, 0.25) 0.92 (0.89, 0.96)
Conventional < 10%ile (−0.0199) 0.16 (0.09, 0.25) 0.89 (0.84, 0.93)
∆OBJ-LME Optimal < 17%ile (0.0172) 0.41 (0.29, 0.53) 0.79 (0.73, 0.84)
Conventional < 10%ile (0.0162) 0.31 (0.21, 0.43) 0.85 (0.80, 0.90)

Note: Conventional cut-offs for sOBJ and ∆OBJ-WSD are ≤ −1 SD, which corresponds to ≤ −1 z using age-corrected Cogstate norms and norms for change respectively. Annualized change percentiles used in ∆OBJ-LR and ∆OBJ-LME correspond to our cognitively unimpaired reference group (n=732 aged 50–65 at baseline); the corresponding cut-off for annualized change for OCL accuracy raw score (arcsine transformed) is provided in parentheses. Optimal cutoffs were derived using the Youden index [37], which maximizes the difference between the true positive and false positive rate across all possible cut-point values. A = amyloid; T = tau; OCL= One Card Learning; sOBJ = subtle objective cognitive impairment at baseline; ∆OBJ = objective longitudinal cognitive decline; WSD = within subjects standard deviation; LR = linear regression model; LME = linear mixed effects model; A−T− group N = 211; A+T+ group N = 32; A+T− group N = 68.

Differentiating A−T− and A+T− groups:

Using conventional cut-offs of z ≤ −1 and annualized change < 10th %ile of the reference group, sensitivity values for identifying A+T− individuals were low for all methods (6–31%; Table 4), with high specificity. Optimal cut-offs are not discussed given that AUC values suggested OCL accuracy performance was unable to differentiate A−T− and A+T− groups better than chance.

Additional Supplemental Results

Supplemental analyses for other available CBB subtests and composite scores from the CBB confirm that the OCL accuracy subtest best differentiated A+T+ and A+T− from A−T− groups relative to other available CBB measures, including composite scores (see Supplemental Table 1). Of note, ONB accuracy and the Learning/Working Memory Composite results showed greater improvement over time in both biomarker positive groups (A+T+ and A+T−) relative to the A−T− group for the ∆OBJ-LME method, which was unexpected.

A frequency cross-tabulation (2×2 Table) of OCL accuracy results by sOBJ and ∆OBJ methods is provided to allow future derivation of additional estimates of test accuracy, as recommended by the reporting standards for studies of diagnostic test accuracy in dementia [38] (see Supplemental Tables 2 and 3).

Cognitively unimpaired participants who progressed to MCI (incident MCI cases) and matched controls

Group Mean Comparisons:

See Table 5 for participant demographics. Group differences were statistically significant for sOBJ at baseline, sOBJ at 30 months, and ∆OBJ-LME methods (p’s < .05). The incident MCI group showed significantly lower OCL accuracy performance at baseline and 30-month follow-up, as well as less benefit from practice over time based on annualized change (LME method), compared to the stable group. Differences were characterized by medium effect sizes for sOBJ, sOBJ 30-month, and ∆OBJ-LME methods, and minimal effect sizes for ∆OBJ-WSD and ∆OBJ-LR methods.

Table 5.

Demographic characteristics and performance on One Card Learning accuracy by those who progressed to MCI after the 30-month run-in period (incident MCI) and those who remained cognitively unimpaired (stable controls). Statistics reported are of the form mean (standard deviation, SD) unless otherwise specified.

Stable controls
(n = 184)
Incident MCI
(n = 64)
p-value
Hedge’s g
Age, years 76.91 (6.02) 77.64 (6.45) 0.414
Education, years 14.58 (2.53) 14.34 (2.48) 0.516
Male count (%) 84 (46%) 28 (44%) 0.792
CBB sessions in run-in period 4.40 (1.23) 4.30 (1.09) 0.546
Follow-up in years1 3.20 (1.11) 2.25 (1.05) < 0.001
Time to conversion in years2 -- 4.81 (1.04) --
OCL accuracy
 sOBJ baseline, z-score 0.233 (0.821) −0.144 (0.773) 0.002 0.46
 sOBJ 30-month, z-score 0.560 (0.780) 0.031 (0.981) < 0.001 0.63
 ΔOBJ-WSD, z-score 0.312 (0.869) 0.160 (1.013) 0.251 0.17
 ΔOBJ-LR, change per year 0.010 (0.046) 0.007 (0.047) 0.655 0.06
 ΔOBJ-LME, change per year 0.016 (0.005) 0.013 (0.005) < 0.001 0.62

Note. p-values represent linear model ANOVAs for mean comparisons or Pearson’s Chi-squared test for frequency comparisons. OCL= One Card Learning; sOBJ = subtle objective cognitive impairment (cross-sectional). ΔOBJ = subtle objective cognitive decline (longitudinal); WSD = within subjects standard deviation, positive value indicates higher score across repeated assessments (e.g., benefit from practice); LR = annualized change from linear regression model. LME = annualized change from linear mixed effects model.

1

Follow-up time starts just after the run-in period until MCI diagnosis (cases) or last visit (controls).

2

Time to conversion represents time from Cogstate baseline to MCI diagnosis.

Overall diagnostic accuracy

Overall diagnostic accuracy for differentiating between the incident MCI and stable groups was better than chance for the sOBJ at baseline, sOBJ at 30 months and ∆OBJ-LME methods, but not for ∆OBJ-WSD and ∆OBJ-LR methods (Table 6). Although no change methods were significantly better than sOBJ at baseline at differentiating groups (p’s > .05), sOBJ at 30 months and ∆OBJ-LME methods had the highest concordance values (both .69). sOBJ at 30 months and ∆OBJ-LME better differentiated between groups as compared to ∆OBJ-WSD and ∆OBJ-LR methods (p’s < .05). AUROC values showed the same pattern of results (see Supplemental Table 4).

Table 6.

Concordance statistic (C-statistic) from conditional logistic regression models and comparison of predictive accuracy between cases and controls by method of defining subtle cognitive decline. P-values computed using jack-knifed variance estimates.

Comparing concordances (p values)
Concordance
(p value)
sOBJ at 30 months WSD LR LME
sOBJ
 baseline 0.59 (0.003) 0.12 0.38 0.21 0.09
 30 months 0.69 (<0.001) -- 0.004 <0.001 0.95
ΔOBJ
 WSD 0.50 (0.21) -- -- 0.33 0.02
 LR 0.46 (0.58) -- -- -- 0.004
 LME 0.69 (<0.001) -- -- -- --

Note. sOBJ = subtle objective cognitive impairment (cross-sectional). ΔOBJ = subtle objective cognitive decline (longitudinal). WSD = within subjects’ standard deviation. LR = annualized change from linear regression model. LME = annualized change from linear mixed effects model.

Conditional logistic regression models

There was a similar pattern of findings when examining odds ratios from independent conditional logistic regression models (see Table 7): sOBJ at 30 months had the largest effect; for every decrement in performance equivalent to 1 z score, the odds of progression to MCI was 2.09 times higher (p < .001). Odds ratios were also statistically significant for sOBJ at baseline and ∆OBJ-LME (OR of 1.70 and 2.02 respectively), but not ∆OBJ-WSD and ∆OBJ-LR. We performed joint conditional logistic regression models for select comparisons (see Table 8). Results show there is significant improvement to the model over sOBJ at baseline when either sOBJ at 30 months or ∆OBJ-LME is added to the model. ∆OBJ-LME and sOBJ at 30 months were highly correlated (r = 0.82), yielding concerns that multicollinearity may confound joint conditional logistic regression models. We thus performed ANOVAs for these analyses. A model with both sOBJ at 30 months and ∆OBJ-LME did not provide a significantly better fit to the data as compared to sOBJ at 30 months alone (p=0.09), nor ∆OBJ-LME alone (p=0.25).

Table 7.

Odds ratio with 95% confidence intervals (CI) from conditional logistic regression models for specified unit of decrement in cross-sectional or longitudinal One Card Learning performance with corresponding increases in likelihood of progression to MCI.

Decrement Odds Ratios (95% CI) p-value
sOBJ
 baseline1 1 z 1.70 (1.20, 2.42) 0.003
 30 months1 1 z 2.09 (1.44, 3.03) <0.001
ΔOBJ
 WSD1 1 z 1.20 (0.88, 1.64) 0.25
 LR1,2 0.046 SD 1.06 (0.80, 1.41) 0.67
 LME1,2 0.005 SD 2.02 (1.43, 2.86) <0.001

Note. Odds ratios represent the likelihood of progressing from cognitively unimpaired to MCI given a 1 SD lower performance for each individual method. sOBJ = subtle objective cognitive impairment at a single assessment. ΔOBJ = subtle objective cognitive decline (longitudinal). WSD = within subjects’ standard deviation. LR = annualized change from linear regression model. LME = annualized change from linear mixed effects model.

1

Reverse scaling applied so that the odds ratios are greater than one.

2

ΔOBJ-LR and ΔOBJ-LME were scaled by dividing by their standard deviations.

Table 8.

Joint conditional logistic regression models with odds ratios for planned model comparisons.

Model Term Decrement Odds Ratios (95% CI) p-value
1 sOBJ at baseline1 1 z 1.35 (0.92, 1.99) 0.13
sOBJ at 30 months1 1 z 1.89 (1.27, 2.81) 0.002
2 sOBJ at baseline 1 1 z 1.06 (0.66, 1.70) 0.82
ΔOBJ-LME1,2 0.0068 SD 1.96 (1.26, 3.03) 0.003

Note. Odds ratios represent the likelihood of progressing from cognitively unimpaired to MCI given a 1 SD lower performance for each individual method. sOBJ = subtle objective cognitive impairment (cross-sectional). ΔOBJ = subtle objective cognitive decline (longitudinal). WSD = within subjects’ standard deviation. LR = annualized change from linear regression model. LME = annualized change from linear mixed effects model.

1

Reverse scaling applied so that the odds ratios are greater than one.

2

ΔOBJ-LME was scaled by dividing by its standard deviation.

Comparing performance of different cut-offs

Using conventional cut-offs of z ≤ −1 based on Cogstate norms (sOBJ at baseline, sOBJ at 30 months, and ∆OBJ-WSD) and annualized change < 10th %ile of the reference group (∆OBJ-LR and ∆OBJ-LME), sensitivity values were low across methods (11–23%; Table 9) with the exception of the ∆OBJ-LME method (77% sensitivity). However, specificity for the ∆OBJ-LME method was poor (46%) suggesting that the 10th %ile cut-off for ∆OBJ-LME slope was overly liberal. Optimal derived cutoffs tended to provide a better balance of sensitivity and specificity, although specificity remained poor for several methods.

Table 9.

Comparison of diagnostic accuracy of Cogstate One Card Learning accuracy (arcsine transformed) to differentiate stable controls from incident MCI individuals for different subtle objective cognitive impairment (sOBJ) and longitudinal decline (∆OBJ) methods.

Threshold type Threshold Sensitivity Specificity
sOBJ
 Baseline Optimal 0.1699 z 0.70 (0.58, 0.81) 0.56 (0.49, 0.63)
Conventional ≤−1 z 0.11 (0.05, 0.19) 0.93 (0.89, 0.96)
 30 months Optimal 0.8624 z 0.86 (0.77, 0.94) 0.41 (0.34, 0.48)
Conventional ≤−1 z 0.16 (0.08, 0.25) 0.97 (0.95, 0.99)
ΔOBJ
 WSD Optimal −1.1714 z 0.14 (0.06, 0.23) 0.97 (0.94, 0.99)
Conventional ≤−1 z 0.14 (0.06, 0.23) 0.95 (0.91, 0.97)
 LR Optimal −0.0337 0.19 (0.09, 0.28) 0.88 (0.83, 0.92)
Conventional <10th %ile (−0.0199) 0.23 (0.14, 0.34) 0.78 (0.72, 0.84)
 LME Optimal 0.0140 0.64 (0.53, 0.77) 0.67 (0.60, 0.73)
Conventional <10th %ile (0.0162) 0.77 (0.66, 0.86) 0.46 (0.39, 0.53)

Note: Optimal cutoffs were derived using the Youden index [37], which maximizes the difference between the true positive and false positive rate across all possible cut-point values. sOBJ = subtle objective cognitive impairment (cross-sectional). ΔOBJ = subtle objective cognitive decline (longitudinal). WSD = within subjects standard deviation z-score based on Cogstate norms. LR = linear regression model; <10%ile slope is equivalent to a < −0.0199 annual change on OCL accuracy raw score (arcsine transformed). LME = linear mixed effects model; <10%ile slope is equivalent to a < 0.0162 annual change on OCL accuracy raw score (arcsine transformed). <10%ile slope cutoffs were derived from the cognitively unimpaired reference sample (N=732). Controls = cognitively unimpaired and stable (N = 184). Cases = incident MCI (N = 64).

DISCUSSION

This study compared different approaches to detecting subtle cognitive decline based on Cogstate OCL performance. We compared diagnostic accuracy of cross-sectional (sOBJ) and longitudinal (ΔOBJ) performance on OCL accuracy for (1) differentiating cognitively unimpaired biomarker negative individuals from cognitively unimpaired individuals with preclinical Alzheimer’s disease (A+T+) and Alzheimer’s pathologic change (A+T−) and (2) differentiating incident MCI cases from matched stable controls. Inconsistent with our study hypothesis, longitudinal cognitive decline over 30 months based on OCL accuracy was not substantially better at detecting transitional cognitive decline relative to a single cross-sectional assessment, with results varying significantly across different ∆OBJ methods.

Biomarker positive, cognitively unimpaired groups had lower baseline OCL accuracy performance (sOBJ) and less improvement over time relative to the biomarker negative group. sOBJ and ∆OBJ-LME differentiated between A−T− and A+T+ (but not A+T−) individuals better than chance. ∆OBJ-LME offered a non-significant incremental benefit over sOBJ for differentiating A−T− and A+T+ groups, increasing the AUC values from .64 to .69. Even with this slight increase, the maximum AUC value seen with the ∆OBJ-LME method was just below what is typically considered acceptable (AUC ≥ .70) [39, 40]. ∆OBJ-LME offered better sensitivity relative to sOBJ, particularly when using optimal cutoffs. However, the marginal benefits offered by ∆OBJ-LME required data over the course of 30 months, which must be weighed against the practicality and brevity of a single assessment.

There was a similar pattern of findings when differentiating between incident MCI cases and stable cognitively unimpaired controls. Notably, inclusion of an additional cross-sectional (sOBJ) comparison at the 30-month follow-up suggests that intra-individual change per se is not the most important aspect of the added benefit of repeated assessment for OCL accuracy. sOBJ at baseline, sOBJ at 30 months, and ∆OBJ-LME all differentiated groups better than chance, with comparable concordance values. Adding either sOBJ at 30 months or ∆OBJ-LME to a model with sOBJ at baseline resulted in significant improvement to the model. However, ∆OBJ-LME did not lead to a significant model improvement over sOBJ at 30 months. This suggests that a single assessment proximal to the incidence of MCI was an equally effective predictor as intraindividual change.

The slight advantage in sensitivity with the ∆OBJ-LME method in the biomarker defined groups was again seen to some degree when we applied the same < 10th%ile conventional cut-off to incident MCI and stable cognitively unimpaired groups. However, specificity was poor, with 54% of those who remained cognitively unimpaired after the run-in period also showing a decline in OCL. This raises concerns about the suitability of this cut-off for broad application.

∆OBJ-WSD and ∆OBJ-LR did not differentiate between biomarker groups or case-control groups better than chance. More importantly, the WSD method, the only method currently provided by Cogstate to assess longitudinal change over time using the CBB, was outperformed by single assessment results achieved with sOBJ at 30 months when distinguishing incident MCI from stable cognitively unimpaired participants. Further, sensitivity values of ∆OBJ-WSD and ∆OBJ-LR methods were very low when using conventional cut-offs (6–23%). Use of ∆OBJ-WSD and ∆OBJ-LR methods for detection of preclinical AD and incident MCI with OCL accuracy is not recommended in circumstances where high sensitivity to preclinical cognitive change is needed.

Practice Effects

Consideration of practice effects is necessary when comparing the utility of baseline and longitudinal assessments on OCL accuracy. Mean annualized change on OCL accuracy in each biomarker group was positive, suggesting improvement on successive sessions consistent with our prior results demonstrating that the OCL subtest is prone to practice effects [30, 41]. In addition, diminished practice effects are associated with cognitive decline in the context of incident MCI/dementia [4245]. We expected that the ∆OBJ-LME model would best capture these potentially variable effects given its utility in accounting for fixed and random effects. However, as noted above, ∆OBJ-LME provided relatively modest benefit in overall diagnostic accuracy. ∆OBJ-WSD only considers the baseline and terminal assessments without including intervening sessions and is based on typical change from baseline to a single follow-up session, so may not fully account for practice effects across numerous sessions. This may contribute to the poor performance of the ∆OBJ-WSD method, and others have also advised against using this method to determine a reliable change due to methodological weaknesses associated with this approach [34, 35]. We focused primarily on ∆OBJ methods provided for use by Cogstate (∆OBJ-WSD) and those previously shown to have potential utility for detecting amyloidosis (∆OBJ-LME) [13, 14]. However, these models alone do not directly account for practice effects. Instead, practice effects were accounted for in this manuscript through the study design that prioritized a similar amount of follow-up time (30 months) and sessions across primary groups of interest. Regression-based norms for determining cognitive decline provide a method of modeling practice effects more directly and often perform better than simpler methods of determining reliable change [45, 46]. Regression-based norms for change represent an important future direction.

Reliability

Variations in reliability are of particular relevance to longitudinal studies seeking to detect cognitive decline. Similar to current results, Ivnik et al. [47] compared diagnostic accuracy across four methods of interpreting neuropsychological test data and found that longitudinal data interpretation did not offer significant benefit in predicting cognitive decline. Increased test-retest variability among older cognitively unimpaired individuals limited the utility of longitudinal data interpretation. Low reliability may similarly be driving our findings, as ∆OBJ methods did not substantially outperform the single assessment sOBJ method in differentiating groups of interest. ∆OBJ-LME uses advanced statistical methods to account for intra-individual variability, thus diminishing the effect of low reliability, and yet was not substantially better than sOBJ at baseline and sOBJ at 30 months when predicting incident MCI. Reliability estimates of CBB subtests have varied across different cohorts and statistical methods [34, 48]. Lim and colleagues (2013) reported an average intraclass correlation coefficient of .87 across 4 CBB sessions when collapsing across 3 clinical groups, with slightly lower ICC values within subgroups (.77 in healthy adults, .82 for those with amnestic MCI, .68 for those AD dementia). We previously reported that reliability coefficients for several CBB tests fall below levels required for clinical decision making, and this was particularly relevant for OCL accuracy, with reliability coefficients ranging from 0.26 – 0.59 across 2 successive sessions [29]. Our prior work demonstrated that platform of administration may add an additional confound to reliability across serial assessments [29], and reporting average reliability values across several sessions may yield different reliability coefficients than when reported session by session.

Implications and application of these findings

The numeric clinical staging scheme described by the National Institute on Aging-Alzheimer’s Association (NIA-AA) Working Group proposed evidence of subtle decline on longitudinal cognitive testing as one option to define transitional cognitive decline for individuals on the Alzheimer’s continuum [1]. This definition does not include potential use of sOBJ, i.e., a single cross-sectional assessment. However, in the present study, the CBB which is marketed for both single and serial assessments, did not offer substantially better diagnostic accuracy when using longitudinal data over 30 months. Although these results are confined to the CBB and may be related to test-specific factors described above, this study highlights the importance of directly comparing cross-sectional and longitudinal approaches to objectively defining subtle cognitive decline. Considering only longitudinal definitions of decline may compound test-retest reliability problems and may miss an important subset of individuals at risk for future cognitive decline. These results support the FDA guidance on early AD trials that includes defining Stage 2 with a single time point measure of cognition (U.S. Department of Health and Human Services, 2018). Future studies should investigate whether this same pattern of results is observed when using traditional neuropsychological measures.

Cognitive decline preceding clinical diagnosis of dementia by several years is well established, particularly with an accelerated rate of decline in later years [4951]. Caselli et al. [9] showed that there is a discernible inflection point after which those who develop MCI decline at a more rapid rate than those who remain cognitively unimpaired. This inflection point precedes MCI diagnoses by up to 20 years. It is possible that our 30-month window was too short to fully capture transitional cognitive decline. However, this window is representative of many research and clinical contexts. Further, it is possible that the relative proximity of the sOBJ at 30 months assessment to the subsequent progression to MCI contributes to its stronger performance compared to sOBJ at baseline. However, this does not explain the advantage of sOBJ at 30 months over some ∆OBJ methods since all longitudinal methods included information from the 30-month session.

Sensitivity was low across most methods in the biomarker defined and incident MCI groups. Application of conventional cut-offs using Cogstate norms for a single cross-sectional assessment or their recommended method of determining change over time (∆OBJ-WSD) identified only 6–12% of cognitively unimpaired A+ individuals and 11–16% of incident MCI individuals. Given high specificity, any individuals identified as decliners based on these cut-offs have a high probability for biomarker positivity or future cognitive decline. However, these sensitivity values may be too low for optimal use in registry models, as only a small proportion of potentially eligible at-risk participants are identified. Although use of optimal cut-offs improved sensitivity, they were associated with 1) a notable decrement in specificity and 2) thresholds well within what is usually considered an unimpaired range. Conventional cut-offs are based on normative data that do not account for underlying neuropathology that has been shown to affect the detection of cognitive decline in older adults [52] which may explain the improved sensitivity when using optimal cut-offs relying on a much smaller change in performance. We previously reported lower than expected sensitivity of the CBB to prevalent MCI [53], suggesting that improvements in the CBB normative data may be needed to enhance its clinical utility. Ideally, higher sensitivity could be achieved while maintaining specificity through use of more sensitive cognitive measures. Recent results by Lim and colleagues demonstrate that with a sensitive measure of learning, differential practice effects across cognitively unimpaired amyloid positive and negative individuals can be elicited over a week with repeated daily assessments [54]. More work towards developing novel, sensitive, and efficient cognitive measures optimized for remote assessment is needed and represents an important future direction.

Limitations

This study has several limitations. We focused on OCL accuracy to maintain parsimony in this manuscript. Importantly, our definitions of sOBJ and ∆OBJ as used in this study were therefore limited to a single measure to maintain focus on our primary aim of comparing cross-sectional and longitudinal approaches, and are not intended to represent proposed optimal methods of defining subtle cognitive decline. This is an important limitation, as single test approaches are less stable and robust than multi-test definitions of cognitive decline [55]. Prior work has highlighted the importance and utility of defining objective subtle cognitive decline cross-sectionally using a battery of neuropsychological tests [5658]. Our group recently demonstrated that a measure of ∆OBJ based on composite z-scores of in-person neuropsychological measures of either memory or attention/executive domains captured more individuals meeting NIA-AA Stage 2 criteria than subjective cognitive decline or neurobehavioral change [19]. In that study, Petersen and colleagues applied cutoff points of ≤ −0.1 z units/year and ≤ −0.2 z units/year on these memory and attention/executive composites derived from linear regression models over 30 months. We want to emphasize that the definitions of ∆OBJ used in the current manuscript differ from that approach in the cut-offs applied and were optimized based on prior Cogstate manuscripts and our goal of directly comparing cross-sectional and longitudinal approaches. This brings up the question of whether use of composite scores may provide an alternative to the single test approach used in this study [59, 60]. However, our supplemental analyses show that the OCL accuracy subtest best differentiated A+T+ and A+T− from A−T− groups relative to CBB composite scores (see Supplemental Online Resources). This suggests that use of available CBB composite scores would be unlikely to alter the primary findings of this study. In addition, the Learning/Working Memory Composite showed unexpectedly greater improvement for A+T+ and A+T− groups compared to the biomarker negative group over 30-months when using the ∆OBJ-LME method. This pattern was driven by One Back accuracy. For OCL accuracy, there was a correlation of −.80 between the subject-specific slope and subject-specific intercept (i.e. baseline performance) terms with limited variability of individual trajectories over the course of several sessions. This implies that an individual’s performance at baseline significantly impacted subsequent slope-based interpretation in ∆OBJ-LME analyses.

Approximately 21% of all CBB sessions were completed at home. While there is some support for equivalent CBB performance across supervised and unsupervised settings [61], we previously showed subtly lower performances on OCL accuracy performance at home relative to in clinic, equivalent to approximately 5% of the raw accuracy value (out of 100%) [30]. It is possible that environmental distractions may impact performance during unsupervised sessions, reducing reliability and validity [62]. Finally, we did not consider neurodegeneration within our biomarker-defined samples because it is a relatively non-specific biomarker. However, it is possible that individuals in the reference and control groups have neurodegeneration that was unaccounted for [1].

Conclusions

In conclusion, our study addresses the critical and challenging need to directly compare cross-sectional and longitudinal methods for defining subtle cognitive decline. When investigating this question using the Cogstate OCL accuracy measure, none of the longitudinal ∆OBJ methods were substantially better than single cross-sectional assessment (sOBJ) methods at identifying cognitively unimpaired individuals at high risk of showing cognitive decline over time (biomarker positive or incident MCI). Our data suggest that the visual memory demands of OCL, tested with a yes/no recognition paradigm, do not provide the level of sensitivity needed for detecting subtle cognitive decline in preclinical AD. Results have important implications for patient registries using the CBB to help identify risk of AD biomarker positivity. OCL sensitivity based on conventional cut-offs is too low to be useful for this purpose. Either alternative, more lenient cut-offs or updated normative data are recommended. Future studies are needed to determine whether alternative measures with higher reliability and increased sensitivity to early memory changes associated with preclinical AD may show the expected benefits of longitudinal assessment.

Supplementary Material

Supplemental Material

ACKNOWLEDGEMENTS

The authors wish to thank the participants and staff at the Mayo Clinic Study of Aging, and Erin Twohy, M.S., who contributed to statistical analyses related to but not included in this manuscript. This work was supported by the Rochester Epidemiology Project (R01 AG034676), the National Institutes of Health (grant numbers P50 AG016574, P30 AG062677, U01 AG006786, R37 AG011378, R01 AG041851, RF1 AG55151), a grant from the Alzheimer’s Association (AARG-17–531322), Zenith Award from the Alzheimer’s Association, the Robert Wood Johnson Foundation, The Elsie and Marvin Dekelboum Family Foundation, Alexander Family Alzheimer’s Disease Research Professorship of the Mayo Clinic, Liston Award, Schuler Foundation, GHR Foundation, AVID Radiopharmaceuticals, and the Mayo Foundation for Education and Research. We would like to greatly thank AVID Radiopharmaceuticals, Inc., for their support in supplying AV-1451 precursor, chemistry production advice and oversight, and FDA regulatory cross-filing permission and documentation needed for this work. NHS and MMMi serve as consultants to Biogen and Lundbeck. DSK serves on a Data Safety Monitoring Board for the DIANTU study and is an investigator in clinical trials sponsored by Lilly Pharmaceuticals, Biogen, and the University of Southern California. RCP has served as a consultant for Hoffman-La Roche Inc., Merck Inc., Genentech Inc., Biogen Inc., Eisai, Inc., and GE Healthcare.

ABBREVIATIONS

AD

Alzheimer’s disease

AUROC

Area under the Receiver Operating Curve

CBB

Cogstate Brief Battery

CLR

Conditional logistic regression

A+

Elevated brain amyloid

T+

Elevated brain tau

Lrn/WM

Learning / Working Memory

LME

Linear mixed effects

LR

Linear regression

MCI

Mild Cognitive Impairment

OCL

One Card Learning

PiB

Pittsburgh compound B

SD

Standard deviation

sOBJ

Subtle objective cognitive impairment

ΔOBJ

Subtle objective cognitive decline

WSD

Within-subjects standard deviation

References

  • [1].Jack CR Jr., Bennett DA, Blennow K, Carrillo MC, Dunn B, Haeberlein SB, Holtzman DM, Jagust W, Jessen F, Karlawish J, Liu E, Molinuevo JL, Montine T, Phelps C, Rankin KP, Rowe CC, Scheltens P, Siemers E, Snyder HM, Sperling R, Contributors (2018) NIA-AA Research Framework: Toward a biological definition of Alzheimer’s disease. Alzheimers Dement 14, 535–562. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Lim YY, Ellis KA, Pietrzak RH, Ames D, Darby D, Harrington K, Martins RN, Masters CL, Rowe C, Savage G, Szoeke C, Villemagne VL, Maruff P, Aibl Research Group (2012) Stronger effect of amyloid load than APOE genotype on cognitive decline in healthy older adults. Neurology 79, 1645–1652. [DOI] [PubMed] [Google Scholar]
  • [3].Lim YY, Maruff P, Pietrzak RH, Ellis KA, Darby D, Ames D, Harrington K, Martins RN, Masters CL, Szoeke C, Savage G, Villemagne VL, Rowe CC, Group AR (2014) Abeta and cognitive change: examining the preclinical and prodromal stages of Alzheimer’s disease. Alzheimers Dement 10, 743–751 e741. [DOI] [PubMed] [Google Scholar]
  • [4].Lowe VJ, Bruinsma TJ, Wiste HJ, Min H-K, Weigand SD, Fang P, Senjem ML, Therneau TM, Boeve BF, Josephs KA, Pandey MK, Murray ME, Kantarci K, Jones DT, Vemuri P, Graff-Radford J, Schwarz CG, Machulda MM, Mielke MM, Roberts RO, Knopman DS, Petersen RC, Jack CR (2019) Cross-sectional associations of tau-PET signal with cognition in cognitively unimpaired adults. Neurology 93, e29–e39. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [5].Rowe CC, Bourgeat P, Ellis KA, Brown B, Lim YY, Mulligan R, Jones G, Maruff P, Woodward M, Price R, Robins P, Tochon-Danguy H, O’Keefe G, Pike KE, Yates P, Szoeke C, Salvado O, Macaulay SL, O’Meara T, Head R, Cobiac L, Savage G, Martins R, Masters CL, Ames D, Villemagne VL (2013) Predicting Alzheimer disease with β-amyloid imaging: Results from the Australian imaging, biomarkers, and lifestyle study of ageing. Ann Neurol 74, 905–913. [DOI] [PubMed] [Google Scholar]
  • [6].Doraiswamy PM, Sperling RA, Coleman RE, Johnson KA, Reiman EM, Davis MD, Grundman M, Sabbagh MN, Sadowsky CH, Fleisher AS, Carpenter A, Clark CM, Joshi AD, Mintun MA, Skovronsky DM, Pontecorvo MJ, Group FtA-AS (2012) Amyloid-β assessed by florbetapir F 18 PET and 18-month cognitive decline: A multicenter study. Neurology 79, 1636–1644. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [7].Jack CR Jr., Wiste HJ, Therneau TM, Weigand SD, Knopman DS, Mielke MM, Lowe VJ, Vemuri P, Machulda MM, Schwarz CG, Gunter JL, Senjem ML, Graff-Radford J, Jones DT, Roberts RO, Rocca WA, Petersen RC (2019) Associations of Amyloid, Tau, and Neurodegeneration Biomarker Profiles With Rates of Memory Decline Among Individuals Without Dementia. JAMA 321, 2316–2325. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [8].Knopman DS, Lundt ES, Therneau TM, Vemuri P, Lowe VJ, Kantarci K, Gunter JL, Senjem ML, Mielke MM, Machulda MM, Boeve BF, Jones DT, Graff-Radford J, Albertson SM, Schwarz CG, Petersen RC, Jack CR Jr. (2019) Entorhinal cortex tau, amyloid-β, cortical thickness and memory performance in non-demented subjects. Brain 142, 1148–1160. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [9].Caselli RJ, Langlais BT, Dueck AC, Chen Y, Su Y, Locke DEC, Woodruff BK, Reiman EM (2020) Neuropsychological decline up to 20 years before incident mild cognitive impairment. Alzheimers Dement 16, 512–523. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [10].Sabbagh M, Boada M, Borson P, Doraiswamy PM, Dubois B, Ingram J, Iwata A, Porsteinsson AP, Possin KL, Rabinovici GD, Vellas B, Chao S, Vergallo A, Hampel H (2020) Early Detection of Mild Cognitive Impairment MCI in an At Home Setting. J Prev Alzheimers Dis. [DOI] [PubMed]
  • [11].Maruff P, Lim YY, Darby D, Ellis KA, Pietrzak RH, Snyder PJ, Bush AI, Szoeke C, Schembri A, Ames D, Masters CL (2013) Clinical utility of the cogstate brief battery in identifying cognitive impairment in mild cognitive impairment and Alzheimer’s disease. BMC psychology 1, 30–30. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [12].Maruff P, Thomas E, Cysique L, Brew B, Collie A, Snyder P, Pietrzak RH (2009) Validity of the CogState brief battery: relationship to standardized tests and sensitivity to cognitive impairment in mild traumatic brain injury, schizophrenia, and AIDS dementia complex. Arch Clin Neuropsychol 24, 165–178. [DOI] [PubMed] [Google Scholar]
  • [13].Darby DG, Brodtmann A, Pietrzak RH, Fredrickson J, Woodward M, Villemagne VL, Fredrickson A, Maruff P, Rowe C (2011) Episodic memory decline predicts cortical amyloid status in community-dwelling older adults. J Alzheimers Dis 27, 627–637. [DOI] [PubMed] [Google Scholar]
  • [14].Darby DG, Pietrzak RH, Fredrickson J, Woodward M, Moore L, Fredrickson A, Sach J, Maruff P (2012) Intraindividual cognitive decline using a brief computerized cognitive screening test. Alzheimers Dement 8, 95–104. [DOI] [PubMed] [Google Scholar]
  • [15].Racine AM, Clark LR, Berman SE, Koscik RL, Mueller KD, Norton D, Nicholas CR, Blennow K, Zetterberg H, Jedynak B, Bilgel M, Carlsson CM, Christian BT, Asthana S, Johnson SC (2016) Associations between Performance on an Abbreviated CogState Battery, Other Measures of Cognitive Function, and Biomarkers in People at Risk for Alzheimer’s Disease. J Alzheimers Dis 54, 1395–1408. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [16].Walter S, Langford OG, Clanton TB, Jimenez-Maggiora GA, Raman R, Rafii MS, Shaffer EJ, Sperling RA, Cummings JL, Aisen PS (2020) The Trial-Ready Cohort for Preclinical and Prodromal Alzheimer’s Disease (TRC-PAD): Experience from the First 3 Years. J Prev Alzheimers Dis 7, 234–241. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [17].Weiner MW, Nosheny R, Camacho M, Truran-Sacrey D, Mackin RS, Flenniken D, Ulbricht A, Insel P, Finley S, Fockler J, Veitch D (2018) The Brain Health Registry: An internet-based platform for recruitment, assessment, and longitudinal monitoring of participants for neuroscience studies. Alzheimers Dement 14, 1063–1076. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [18].Stricker NH, Lundt ES, Albertson SM, Machulda MM, Pudumjee SB, Kremers WK, Jack CR, Knopman DS, Petersen RC, Mielke MM (2020) Diagnostic and Prognostic Accuracy of the Cogstate Brief Battery and Auditory Verbal Learning Test in Preclinical Alzheimer’s Disease and Incident Mild Cognitive Impairment: Implications for Defining Subtle Objective Cognitive Impairment J Alzheimers Dis 76, 261–274. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [19].Petersen RC, Wiste HJ, Weigand SD, Fields JA, Geda YE, Graff-Radford J, Knopman DS, Kremers WK, Lowe V, Machulda MM, Mielke MMP, Stricker NH, Therneau TM, Vemuri P, Jack CR Jr. (2021) NIA-AA Alzheimer’s Disease Framework: Clinical Characterization of Stages. Ann Neurol. [DOI] [PMC free article] [PubMed]
  • [20].Baker JE, Lim YY, Pietrzak RH, Hassenstab J, Snyder PJ, Masters CL, Maruff P (2017) Cognitive impairment and decline in cognitively normal older adults with high amyloid-beta: A meta-analysis. Alzheimers Dement (Amst) 6, 108–121. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [21].Roberts RO, Geda YE, Knopman DS, Cha RH, Pankratz VS, Boeve BF, Ivnik RJ, Tangalos EG, Petersen RC, Rocca WA (2008) The Mayo Clinic Study of Aging: Design and sampling, participation, baseline measures and sample characteristics. Neuroepidemiology 30, 58–69. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [22].Morris JC (1993) The Clinical Dementia Rating (CDR): Current version and scoring rules. Neurology 43, 2412–2414. [DOI] [PubMed] [Google Scholar]
  • [23].Kokmen E, Smith GE, Petersen RC, Tangalos E, Ivnik RC (1991) The short test of mental status: Correlations with standardized psychometric testing. Arch Neurol 48, 725–728. [DOI] [PubMed] [Google Scholar]
  • [24].Petersen RC, Roberts RO, Knopman DS, Geda YE, Cha RH, Pankratz VS, Boeve B, Tangalos EG, Ivnik RJ, Rocca WA (2010) Prevalence of mild cognitive impairment is higher in men: The Mayo Clinic Study of Aging. Neurology 75, 889–897. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [25].Jack CR Jr., Lowe VJ, Senjem ML, Weigand SD, Kemp BJ, Shiung MM, Knopman DS, Boeve BF, Klunk WE, Mathis CA, Petersen RC (2008) 11C PiB and structural MRI provide complementary information in imaging of Alzheimer’s disease and amnestic mild cognitive impairment. Brain 131, 665–680. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [26].Jack CR Jr., Wiste HJ, Weigand SD, Therneau TM, Lowe VJ, Knopman DS, Gunter JL, Senjem ML, Jones DT, Kantarci K, Machulda MM, Mielke MM, Roberts RO, Vemuri P, Reyes DA, Petersen RC (2017) Defining imaging biomarker cut points for brain aging and Alzheimer’s disease. Alzheimers Dement 13, 205–216. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [27].Vemuri P, Lowe VJ, Knopman DS, Senjem ML, Kemp BJ, Schwarz CG, Przybelski SA, Machulda MM, Petersen RC, Jack CR Jr. (2017) Tau-PET uptake: Regional variation in average SUVR and impact of amyloid deposition. Alzheimers Dement (Amst) 6, 21–30. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [28].Klunk WE, Koeppe RA, Price JC, Benzinger TL, Devous MD Sr., Jagust WJ, Johnson KA, Mathis CA, Minhas D, Pontecorvo MJ, Rowe CC, Skovronsky DM, Mintun MA (2015) The Centiloid Project: standardizing quantitative amyloid plaque estimation by PET. Alzheimers Dement 11, 1–15 e11–14. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [29].Stricker NH, Lundt ES, Edwards KK, Machulda MM, Kremers WK, Roberts RO, Knopman DS, Petersen RC, Mielke MM (2019) Comparison of PC and iPad administrations of the Cogstate Brief Battery in the Mayo Clinic Study of Aging: Assessing cross-modality equivalence of computerized neuropsychological tests. The Clinical neuropsychologist 33, 1102–1126. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [30].Stricker NH, Lundt ES, Alden EC, Albertson SM, Machulda MM, Kremers WK, Knopman DS, Petersen RC, Mielke MM (2020) Longitudinal Comparison of in Clinic and at Home Administration of the Cogstate Brief Battery and Demonstrated Practice Effects in the Mayo Clinic Study of Aging. The journal of prevention of Alzheimer’s disease 7, 21–28. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [31].Cogstate (2018) Cogstate Pediatric and Adult Normative Data, Cogstate, Inc., New Haven, CT. [Google Scholar]
  • [32].Hinton-Bayre AD (2010) Deriving reliable change statistics from test-retest normative data: comparison of models and mathematical expressions. Arch Clin Neuropsychol 25, 244–256. [DOI] [PubMed] [Google Scholar]
  • [33].Lewis MS, Maruff P, Silbert BS, Evered LA, Scott DA (2007) The influence of different error estimates in the detection of postoperative cognitive dysfunction using reliable change indices with correction for practice effects. Arch Clin Neuropsychol 22, 249–257. [DOI] [PubMed] [Google Scholar]
  • [34].Fratti S, Bowden SC, Cook MJ (2017) Reliability and validity of the CogState computerized battery in patients with seizure disorders and healthy young adults: comparison with standard neuropsychological tests. The Clinical Neuropsychologist 31, 569–586. [DOI] [PubMed] [Google Scholar]
  • [35].Hinton-Bayre AD (2011) Specificity of reliable change models and review of the within-subjects standard deviation as an error term. Arch Clin Neuropsychol 26, 67–75. [DOI] [PubMed] [Google Scholar]
  • [36].DeLong ER, DeLong DM, Clarke-Pearson DL (1988) Comparing the areas under two or more correlated receiver operating characteristic curves: a nonparametric approach. Biometrics 44, 837–845. [PubMed] [Google Scholar]
  • [37].Perkins NJ, Schisterman EF (2005) The Youden Index and the optimal cut-point corrected for measurement error. Biom J 47, 428–441. [DOI] [PubMed] [Google Scholar]
  • [38].Noel-Storr AH, McCleery JM, Richard E, Ritchie CW, Flicker L, Cullum SJ, Davis D, Quinn TJ, Hyde C, Rutjes AW, Smailagic N, Marcus S, Black S, Blennow K, Brayne C, Fiorivanti M, Johnson JK, Kopke S, Schneider LS, Simmons A, Mattsson N, Zetterberg H, Bossuyt PM, Wilcock G, McShane R (2014) Reporting standards for studies of diagnostic test accuracy in dementia: The STARDdem Initiative. Neurology 83, 364–373. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [39].Hosmer D, Lemeshow S (2000) In Applied Logistic Regression John Wiley and Sons, New York, NY, pp. 160–164. [Google Scholar]
  • [40].Mandrekar JN (2010) Receiver operating characteristic curve in diagnostic test assessment. J Thorac Oncol 5, 1315–1316. [DOI] [PubMed] [Google Scholar]
  • [41].Mielke MM, Machulda MM, Hagen CE, Christianson TJ, Roberts RO, Knopman DS, Vemuri P, Lowe VJ, Kremers WK, Jack CR Jr, Petersen RC (2016) Influence of amyloid and APOE on cognitive performance in a late middle-aged cohort. Alzheimer’s & Dementia 12, 281–291. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [42].Hassenstab J, Ruvolo D, Jasielec MS, Xiong C, Grant E, Morris JC (2015) Absence of practice effects in preclinical Alzheimer’s Disease. Neuropsychology 29, 940–948. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [43].Machulda MM, Hagen CE, Wiste HJ, Mielke MM, Knopman DS, Roberts RO, Vemuri P, Lowe VJ, Jack CR Jr., Petersen RC (2017) Practice effects and longitudinal cognitive change in clinically normal older adults differ by Alzheimer imaging biomarker status. Clin Neuropsychol 31, 99–117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [44].Machulda MM, Pankratz VS, Christianson TJ, Ivnik RC, Mielke MM, Roberts RO, Knopman DS, Boeve BF, Petersen RC (2013) Practice effects and longitudinal cognitive change in normal aging vs. incident mild cognitive impairment and dementia in the Mayo Clinic Study of Aging. The Clinical Neuropsychologist 27, 1247–1264. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [45].Nation DA, Ho JK, Dutt S, Han SD, Lai MHC (2019) Neuropsychological Decline Improves Prediction of Dementia Beyond Alzheimer’s Disease Biomarker and Mild Cognitive Impairment Diagnoses. J Alzheimers Dis 69, 1171–1182. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [46].Duff K, Suhrie KR, Dalley BCA, Anderson JS, Hoffman JM (2019) External validation of change formulae in neuropsychology with neuroimaging biomarkers: A methodological recommendation and preliminary clinical data. Clin Neuropsychol 33, 478–489. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [47].Ivnik RC, Smith GE, Petersen RC, Boeve BF, Kokmen E, Tangalos E (2000) Diagnostic accuracy of four approaches to interpreting neuropsychological test data. Neuropsychology 14, 163–177. [DOI] [PubMed] [Google Scholar]
  • [48].Lim YY, Jaeger J, Harrington K, Ashwood T, Ellis KA, Stoffler A, Szoeke C, Lachovitzki R, Martins RN, Villemagne VL, Bush AI, Masters CL, Rowe CC, Ames D, Darby D, Maruff P (2013) Three-month stability of the CogState brief battery in healthy older adults, mild cognitive impairment, and Alzheimer’s disease: Results from the Australian Imaging, Biomarkers, and Lifestyle-Rate of Change Substudy (AIBL-ROCS). Arch Clin Neuropsychol 28, 320–330. [DOI] [PubMed] [Google Scholar]
  • [49].Grober E, Hall CB, Lipton RB, Zonderman AB, Resnick SM, Kawas C (2008) Memory impairment, executive dysfunction, and intellectual decline in preclinical Alzheimer’s disease. J Int Neuropsychol Soc 14, 266–278. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [50].Wilson RS, Segawa E, Boyle PA, Anagnos SE, Hizel LP, Bennett DA (2012) The natural history of cognitive decline in Alzheimer’s disease. Psychol Aging 27, 1008–1017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [51].Wilson R, Leurgans S, Boyle P, Schneider J, Bennett D (2010) Neurodegenerative basis of age-related cognitive decline. Neurology 75, 1070–1078. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [52].Harrington KD, Aschenbrenner AJ, Maruff P, Masters CL, Fagan AM, Benzinger TL, Gordon BA, Cruchaga C, Morris JC, Hassenstab J (2021) Undetected Neurodegenerative Disease Biases Estimates of Cognitive Change in Older Adults. Psychol Sci, 0956797620985518. [DOI] [PMC free article] [PubMed]
  • [53].Alden EC, Pudumjee SB, Lundt ES, Albertson SM, Machulda MM, Kremers WK, Jack CR Jr, Knopman DS, Petersen RC, Mielke MM, Stricker NH (2021) Diagnostic accuracy of the Cogstate Brief Battery for prevalent MCI and prodromal AD (MCI A+T+) in a population-based sample. Alzheimer’s & Dementia 17, 584–594. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [54].Lim YY, Baker JE, Bruns L, Mills A, Fowler C, Fripp J, Rainey-Smith SR, Ames D, Masters CL, Maruff P (2020) Association of deficits in short-term learning and Aβ and hippoampal volume in cognitively normal adults. Neurology, 10.1212/WNL.0000000000010728. [DOI] [PubMed]
  • [55].Thomas KR, Edmonds EC, Eppig JS, Wong CG, Weigand AJ, Bangen KJ, Jak AJ, Delano-Wood L, Galasko DR, Salmon DP, Edland SD, Bondi MW, Alzheimer’s Disease Neuroimaging I (2019) MCI-to-normal reversion using neuropsychological criteria in the Alzheimer’s Disease Neuroimaging Initiative. Alzheimers Dement 15, 1322–1332. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [56].Machulda MM, Lundt ES, Albertson SM, Kremers WK, Mielke MM, Knopman DS, Bondi MW, Petersen RC (2019) Neuropsychological subtypes of incident mild cognitive impairment in the Mayo Clinic Study of Aging. Alzheimers Dement 15, 878–887. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [57].Thomas KR, Bangen KJ, Weigand AJ, Edmonds EC, Wong CG, Cooper S, Delano-Wood L, Bondi MW, Alzheimer’s Disease Neuroimaging I (2019) Objective subtle cognitive difficulties predict future amyloid accumulation and neurodegeneration. Neurology. [DOI] [PMC free article] [PubMed]
  • [58].Thomas KR, Edmonds EC, Eppig J, Salmon DP, Bondi MW, Alzheimer’s Disease Neuroimaging I (2018) Using Neuropsychological Process Scores to Identify Subtle Cognitive Decline and Predict Progression to Mild Cognitive Impairment. Journal of Alzheimer’s disease : JAD 64, 195–204. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [59].Weintraub S, Carrillo MC, Farias ST, Goldberg TE, Hendrix JA, Jaeger J, Knopman DS, Langbaum JB, Park DC, Ropacki MT, Sikkes SAM, Welsh-Bohmer KA, Bain LJ, Brashear R, Budur K, Graf A, Martenyi F, Storck MS, Randolph C (2018) Measuring cognition and function in the preclinical stage of Alzheimer’s disease. Alzheimers Dement (N Y) 4, 64–75. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [60].Gates TM, Kamminga J, Jayewardene A, Vincent T, Quan D, Brew BJ, Bloch M, Cysique LA (2021) An examination of reliable change methods for measuring cognitive change with the Cogstate Computerized Battery: Research and clinical implications. Arch Clin Neuropsychol 36, 597–612. [DOI] [PubMed] [Google Scholar]
  • [61].Cromer JA, Harel BT, Yu K, Valadka JS, Brunwin JW, Crawford CD, Mayes LC, Maruff P (2015) Comparison of Cognitive Performance on the Cogstate Brief Battery When Taken In-Clinic, In-Group, and Unsupervised. Clin Neuropsychol 29, 542–558. [DOI] [PubMed] [Google Scholar]
  • [62].Madero EN, Anderson J, Bott NT, Hall A, Newton D, Fuseya N, Harrison JE, Myers JR, Glenn JM (2021) Environmental Distractions during Unsupervised Remote Digital Cognitive Assessment. The Journal of Prevention of Alzheimer’s Disease, 1–4. [DOI] [PMC free article] [PubMed]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplemental Material

RESOURCES