Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2020 Jan 15;15(1):e0226784. doi: 10.1371/journal.pone.0226784

Selection of memory clinic patients for CSF biomarker assessment can be restricted to a quarter of cases by using computerized decision support, without compromising diagnostic accuracy

Hanneke F M Rhodius-Meester 1,2,*, Ingrid S van Maurik 1,3, Juha Koikkalainen 4, Antti Tolonen 5, Kristian S Frederiksen 6, Steen G Hasselbalch 6, Hilkka Soininen 7, Sanna-Kaisa Herukka 7, Anne M Remes 7,8,9, Charlotte E Teunissen 10, Frederik Barkhof 11,12, Yolande A L Pijnenburg 1, Philip Scheltens 1, Jyrki Lötjönen 4, Wiesje M van der Flier 1,3
Editor: Han Zhang13
PMCID: PMC6961870  PMID: 31940390

Abstract

Introduction

An accurate and timely diagnosis for Alzheimer’s disease (AD) is important, both for care and research. The current diagnostic criteria allow the use of CSF biomarkers to provide pathophysiological support for the diagnosis of AD. How these criteria should be operationalized by clinicians is unclear. Tools that guide in selecting patients in which CSF biomarkers have clinical utility are needed. We evaluated computerized decision support to select patients for CSF biomarker determination.

Methods

We included 535 subjects (139 controls, 286 Alzheimer’s disease dementia, 82 frontotemporal dementia and 28 vascular dementia) from three clinical cohorts. Positive (AD like) and negative (normal) CSF biomarker profiles were simulated to estimate whether knowledge of CSF biomarkers would impact (confidence in) diagnosis. We applied these simulated CSF values and combined them with demographic, neuropsychology and MRI data to initiate CSF testing (computerized decision support approach). We compared proportion of CSF measurements and patients diagnosed with sufficient confidence (probability of correct class ≥0.80) based on an algorithm with scenarios without CSF (only neuropsychology, MRI and APOE), CSF according to the appropriate use criteria (AUC) and CSF for all patients.

Results

The computerized decision support approach recommended CSF testing in 140 (26%) patients, which yielded a diagnosis with sufficient confidence in 379 (71%) of all patients. This approach was more efficient than CSF in none (0% CSF, 308 (58%) diagnosed), CSF selected based on AUC (295 (55%) CSF, 350 (65%) diagnosed) or CSF in all (100% CSF, 348 (65%) diagnosed).

Conclusions

We used a computerized decision support with simulated CSF results in controls and patients with different types of dementia. This approach can support clinicians in making a balanced decision in ordering additional biomarker testing. Computer-supported prediction restricts CSF testing to only 26% of cases, without compromising diagnostic accuracy.

Introduction

An accurate and timely clinical diagnosis of Alzheimer’s disease (AD) is important both for care and research [1]. Over the last decade the field has moved from a purely clinical concept to a more biological definition of AD [2]. The development of CSF biomarkers has played a pivotal role in this paradigm-shift, enabling accurate detection of AD pathology in vivo [3]. The current diagnostic criteria emphasize the use of these biomarkers in suspected AD [2, 4].

Diagnostic guidelines do not specify which patients should receive CSF testing however, and as a result there is a large practice variation [5]. Recently, a set of criteria for the use of CSF has been published, mirroring the appropriate use criteria (AUC) for amyloid PET, but these are still rather difficult to translate to clinical practice [6]. First, they are phrased in rather general terms, e.g. ‘CSF biomarkers can be ordered for all patients with suspected underlying AD pathology’, which carries the risk of performing CSF unnecessarily. In addition, the AUC for CSF state that there ‘should be a determination of how CSF biomarkers might contribute to the diagnosis and clinical decision making’, but it is not self-evident how this should be done. Furthermore, ordering of CSF should also depend on ‘the confidence of the clinician in the diagnosis’ [6].

Data-driven support tools could be helpful to answer the question ‘for which patient should CSF biomarkers be measured?’ before embarking on actual testing. By using simulated positive (AD like) or negative (normal) CSF biomarker profiles for a specific patient added to their priori clinical information, such a tool could inform the clinician on how the knowledge of CSF biomarkers would impact (confidence in a) diagnosis [7].

The Disease state index (DSI) classifier is a clinical decision support system (CDSS) which supports the differential diagnosis of several types of dementia by combining and weighing all available information, including cognitive tests, automated MRI features and CSF biomarkers [8]. This multi-class classifier has previously shown to be able to differentiate between several types of dementia [9, 10]. The DSI classifier can achieve high probability of correct diagnosis even without CSF biomarkers, by using an optimal combination of tests [11, 12].

In this work, we took the functionality of our CDSS as an input to predict who would benefit from CSF testing. More specifically, we tested whether the CDSS may help to answer the following question: if a clinician already has information on neuropsychological tests and MRI brain, would additional CSF testing contribute significantly to a more accurate diagnosis? We thus aimed to evaluate how a computer-based tool can help in selecting those patients by using a data-driven approach and simulated CSF outcome.

Methods

Subjects

We included 535 subjects from three memory clinics, as part of the PredictND project [13]: 463 subjects from the Amsterdam Dementia Cohort from the Alzheimer center of the Amsterdam UMC (the Netherlands) [14, 15]), 50 subjects from the Danish Dementia Research Centre at Copenhagen University Hospital, Rigshospitalet (Denmark) and 22 subjects from the Neurocenter at Kuopio University Hospital (Finland). The pooled cohort consisted of subjects with the following diagnosis: 286 (53%) AD, 82 (13%) frontotemporal dementia (FTD), 28 (5%) vascular dementia (VaD) and 139 (26%) controls with subjective cognitive decline (SCD). Subjects were eligible for inclusion if both CSF and brain MRI results were available.

All subjects had received a standardized work-up, including medical history, physical, neurological and neuropsychological assessment, MRI, laboratory tests and CSF. Subjects were diagnosed as SCD when the cognitive complaints could not be confirmed by cognitive testing and criteria for mild cognitive impairment (MCI) or dementia were not met. Probable AD was diagnosed using the core clinical criteria of the NIA-AA for AD dementia [4]. Probable FTD (including the behavioural variant of FTD, progressive non-fluent aphasia and semantic dementia) was diagnosed using the criteria from Rasckovsky and Gorno-Tempini, respectively [16, 17]. VaD was diagnosed using the NINDS-AIREN criteria [18]. The local Medical Ethical Committee reviewed and approved the study, including its consent procedure, in accordance with the declaration of Helsinki. Capacity to consent was determined by the clinician, who performed the work-up. All patients provided written informed consent for their data to be used for research purposes.

Neuropsychology

Cognitive functions were assessed with a brief standardized test battery including widely used tests. We used the Mini-Mental State Examination (MMSE) for global cognitive functioning [19]. For memory we included the Rey auditory verbal learning task (RAVLT) [20]. To measure mental speed and executive functioning we used Trail Making Test A and B (TMT-A, TMT-B) [21]. Language and executive functioning were tested by category fluency (animals) [22]. For behavioral symptoms we used the Neuropsychiatric Inventory (NPI) [23]. Missing data ranged from n = 3 (1%) (MMSE) to n = 126 (24%) (TMT-B).

Cerebrospinal fluid biomarkers

CSF beta amyloid 1–42 (AB42), total tau and tau phosphorylated at threonine 181 (p-tau) were measured with commercially available ELISA tests (Innotest, Fujirebio, Ghent, Belgium) locally according to standard procedures. Raw data values were used for analysis. Analysis showed comparable values between the centers and hence no further correction was performed.

APOE genotype

Apolipoprotein-E (APOE) genotyping were performed locally in each center. Patients were dichotomized into APOE e4 carriers (hetero- and homozygous) and non-carriers. APOE data were available in 472 (88%) subjects.

Imaging markers

MRI images were acquired using 1.5 T or 3 T scanners. The voxel size varied between 0.5–1.1 × 0.5–1.1 × 0.9–2.2 mm for T1 images and 0.4–1.3 × 0.4–1.2 x 0.6–6.5 mm for FLAIR images. We extracted five imaging markers using the cNeuro® cMRI quantification tool: 1) computed medial temporal lobe atrophy (cMTA), 2) computed global cortical atrophy (cGCA), 3) AD similarity scale, 4) Anterior Posterior index and 5) white matter hyperintensities (WMH). First, we automatically computed the MTA and GCA scores [24, 25]. To do so, we defined volumes of brain structures from T1 image segmentations produced by a multi-atlas segmentation algorithm [9, 26]. The cMTA was defined from the volumes of the hippocampi and inferior lateral ventricles separately for left and right. In addition to volumetry, voxel-based morphometry (VBM) [27] was applied to compute gray matter concentrations which were used to define the cGCA. For the AD similarity scale, the region of interest (ROI) in the patient image was represented as a linear combination of the corresponding ROIs from a database of previously diagnosed patients [28, 17]. A ROI-based grading imaging marker was defined as the share of the weights from the linear model having the diagnostic label AD while using the hippocampus area as a ROI. The Anterior Posterior index is a specific measure for characterizing fronto-temporal atrophy, a well-known hallmark in FTD. API measures the relative volume of the cortex in the fronto-temporal and parieto-occipital regions. [29]. In addition, we automatically extracted the volume of WMH from FLAIR images [9, 30]. All imaging markers are corrected for head size [31] and for age and sex [32].

Disease State Index classifier

The Disease State Index (DSI) classifier is a simple supervised and data-driven machine learning method that compares different diagnostic groups with each other; controls, AD, FTD and VaD in this work [8, 13]. DSI is a continuous value between zero and one, measuring the similarity of all patient data to diagnostic groups defined [8]. For each single test result from cognitive tests, APOE, imaging or CSF, a reference dataset determines a fitness function f(x) = FN(x)/(FN(x)+FP(x)) where FN and FP are false negative and positive rates, respectively, when using x as a cut-off to classify patients into two diagnostic groups. Then, this function is used to compute fitness value between zero and one to each test result of the patient. 2) The ‘relevance’ of each test result is defined as sensitivity+specificity-1 from the reference dataset. 3) DSI is calculated by averaging all fitness values by weighting them with their relevance. This process is repeated for each pairwise comparison (controls-AD, controls-FTD, controls-VaD, AD-FTD, AD-VaD, VaD-FTD). The total DSI for each diagnostic group is the average over all pairwise comparisons, e.g. AD-controls, AD-FTD and AD-VaD for AD. All variables are corrected for age and sex [32]. Due to the design of DSI, there is no need to impute data or exclude cases with incomplete data, as only available data are used [10].

Probability of correct class

A high total DSI value or a big difference in DSI values between the two most similar diagnostic groups (first and second DSI) provides more confidence for making a diagnosis than a low value or a small difference. Therefore, we define a new measure, probability of correct class (PCC), which estimates the probability that the suggested diagnosis for a patient is correct, when compared to the clinical diagnosis (ground truth). If the highest DSI value is 0.75 and the second highest is 0.67 for the patient being studied, PCC measures the share of correctly classified cases in the reference database having DSI close to 0.75 and difference close to 0.75–0.67 = 0.08. In other words, PCC estimates (personalized) confidence in classification, i.e., classification accuracy for this very patient. In clinical practice the clinician can adjust the applied PCC cutoff depending on the required accuracy. In this study, patients were considered as having a diagnosis with sufficient accuracy if PCC was at least 0.80. For comparison we performed a sensitivity analysis and repeated our analyses for different PCC cut-offs. Details on the definition and calculation of PCC can be found in the supplement file (S1 File).

Simulated CSF to guide clinical decision making

In our search for a scenario with the largest proportion of patients diagnosed with high confidence (PCC≥0.80), against the smallest proportion of CSF tested, we applied four scenarios:

  • Scenario A (“computerized decision support”): We performed CSF only when predicted to be useful (Fig 1A). This was modelled using a stepwise approach: first, patients diagnosed with PCC≥0.80 based on neuropsychology, MRI and APOE were regarded as clear cases and no CSF was ordered (step one). For the remaining patients with a PCC<0.80, we simulated adding CSF biomarkers; PCC was recomputed after adding positive (age- and sex-normalized median AB42, tau and p-tau values of AD group) and negative (age- and sex-normalized median AB42, tau and p-tau values of SCD group) CSF values (step two) [32]. If either of the resulting simulated PCC values reached the cut-off of≥0.80, actual CSF values were added, and DSI and PCC calculated using these real values (step three).

  • Scenario B (“No CSF”): We calculated DSI and PPC for each patient using neuropsychology, MRI and APOE, excluding CSF (Fig 1B).

  • Scenario C (“AUC”): We performed CSF based on the AUC criteria (Fig 1C), which state that CSF biomarkers can be applied to all patients with suspected underlying AD pathology [6]. We operationalized this as a DSI for AD >0.6. In all patients with a DSI for AD >0.6, we then calculated PCC adding CSF (step two). For patients with a DSI for AD ≤0.6, no CSF was added and DSI and PCC were calculated using only neuropsychology, MRI and APOE.

  • Scenario D (“All CSF”): We performed CSF in all patients (Fig 1D). We calculated DSI and PCC for each patient using neuropsychology, MRI, APOE and CSF.

Fig 1. Flow chart for the four diagnostic scenarios.

Fig 1

AUC: appropriate use criteria according to [33], operationalized as a DSI for AD >0.6, PCC: probability of correct class, NP: neuropsychology, MRI: magnetic resonance imaging, CSF: cerebrospinal fluid biomarkers, Sim.: simulate, FU: follow-up. Numbers in circles denote groups described in section 3.3 and Table 3.

Statistical analyses

We used a two-sample test of proportion to test differences in proportion of patients 1) with CSF tested, and 2) diagnosed with sufficient confidence (PCC≥0.80), between the computerized decision support and the other scenarios [34]. We compared the different groups derived from the computerized decision support (scenario A), using Analysis of Variance (ANOVA).

In an additional sensitivity analysis, we visualized the share of patients diagnosed (percentage of patients above PCC cutoff) and the share of patients with CSF measurement for a range of PCC cut-offs (0.5–1.0), and calculated the classification accuracy of the diagnosed patients. This was done for the four diagnostic scenarios: A) the computerized decision support, B) No CSF, C) AUC, and D) All CSF.

Statistical analyses were performed using STATA version 14.1 and R version 3.5.3. A MATLAB toolbox created by [35] was used in the DSI analyses. The analyses were performed in MATLAB version R2015b (MathWorks, Natick, MA, USA).

Results

Baseline characteristics

Overall mean age was 65±8 and 262 (49%) were females. Details on cognitive tests, MRI markers and CSF results and by group can be found in Table 1.

Table 1. Baseline characteristics according to baseline diagnosis.

n = Control n = 139 AD n = 286 FTD n = 82 VaD n = 28
Female, n (%) 535 60 (43) 158 (55) 36 (44) 8 (29)
Age, in years 535 62 ±8 67 ±8 63 ±6 70 ±8
APOE e4 carrier, n (%) 472 52 (42) 165 (65) 25 (34) 10 (48)
Neuropsychology
MMSE 532 28 ±1 21 ±5 24 ±4 24±4
RAVLT, learning 506 42 ±9 22 ±8 28 ±8 24 ±8
RAVLT, recall 506 9 ±3 2 ±2 4 ±3 4 ±3
TMT-A, in seconds 515 37 ±20 79 ±54 57 ±36 96 ±57
TMT-B, in seconds 409 85 ±39 193 ±80 155 ±77 220 ±86
Animal fluency 519 22 ±5 13 ±5 12 ±7 11 ±4
NPI, total score 445 7 ±9 11 ±11 22 ±17 16 ±10
MRI
cMTA right 535 0.3 ±0.5 1.4 ±0.8 1.6 ±1.2 1.5 ±1.0
cMTA left 535 0.3 ±0.5 1.5 ±0.9 2.0 ±1.4 1.5 ±1.2
cGCA 535 0.3 ±0.5 1.2 ±0.8 1.3 ±0.8 1.4 ±0.7
WMH, ml 535 2.9 ±4.7 6.6 ±9.9 3.6 ±7.7 44.5 ±29.6
AD similarity scale 535 0.4 ±0.1 0.6 ±0.1 0.4 ±0.1 0.6 ±0.1
Anterior Posterior index 535 0.1 ±0.7 0.1 ±1.2 -2.1 ±1.7 0.1 ±1.1
CSF
AB42, pg/ml 535 928 ±280 535 ±183 914 ±250 704 ±252
Total tau, pg/ml 535 322 ±208 693 ±405 337 ±140 308 ±162
P- tau, pg/ml 535 52 ±24 86 ±39 45 ±18 43 ±18
AUC+ (DSI for AD>0.6), n (%) 535 5 (4) 250 (87) 31 (38) 9 (32)

AD: Alzheimer´s disease, FTD: Frontotemporal dementia, VAD: Vascular dementia, MMSE: Mini-Mental state Examination, RAVLT: Rey Auditory Verbal Learning Test, TMT: Trail Making Test, NPI: Neuropsychiatric Inventory score, cMTA: computed medial temporal lobe atrophy scale (0–4), derived from volume of hippocampus and volume of inferor lateral ventricle, cGCA: computed global cortical atrophy scale (0–3), derived from concentration of cortical grey matter using voxel based morphometry, WMH: volume of white matter hyperintensities, AD similarity scale:based on hippocampus ROI, Anterior posterior index: weighted ratio of volumes of the frontal/temporal lobes and parietal/occipital lobes. MRI volumes are adjusted for head size, AB42: beta amyloid 1–42; p-tau: tau phosphorylated at threonine 181. AUC+: number of patients fulfilling appropriate use criteria according to [33], operationalized as a DSI for AD >0.6.

Data are presented as mean ± SD, unless otherwise specified.

Contribution of CSF to diagnosis

To evaluate in which patients the knowledge of CSF changed the confidence in the diagnosis, we examined four diagnostic scenarios, as shown in Fig 1.

In the computerized decision support (Fig 1A) 308 (58%) patients that had PCC≥0.80 in step one (similar to scenario B), did not receive CSF testing. For the remaining 227 (42%) patients, we added simulated CSF values to identify how such values would impact diagnosis. In 140 patients the cutoff of PCC 0.80 was reached (72 with only positive CSF biomarkers, 46 only negative CSF biomarkers and 22 conflicting CSF biomarkers). When we added the actual CSF results of these patients to calculate DSI and PCC (step three), an additional 71 patients reached our threshold to be diagnosed with sufficient confidence. As a result restricting measurement of CSF to only 26% (140/535), diagnosis with sufficient confidence was reached in 71% of the patients ((308+71)/535). In comparison with the other scenarios, the computerized decision support diagnosed a significantly higher proportion of patients diagnosed with sufficient confidence (Table 2). Of note, the computerized approach even outranks scenario D (all CSF), emphasizing that performing CSF in patients were it is not useful, causes confusing and thus lower confidence in the diagnosis. Furthermore, this approach had a significantly lower proportion of CSF tested compared to scenario C based on AUC.

Table 2. Comparison of the four diagnostic scenarios.

Scenario Diagnosis with sufficient confidence, PCC≥0.80, n (%) CSF performed, n (%)
A. Computerized decision support 379 (71%) *,, 140 (26%)§
B. No CSF 308 (58%) 0 (0%)
C. AUC 350 (65%) 295 (55%)
D. All CSF 348 (65%) 535 (100%)

AUC: appropriate use criteria according to [33], operationalized as a DSI for AD >0.6, PCC: probability of correct class.

Proportion diagnosed

*Scenario A vs Scenario B: difference 13%[719], z = 4.44, p<0.001

†Scenario A vs Scenario C: difference 6% [0.4–12], z = 2.104, p = 0.035

‡Scenario A vs Scenario D: difference 6%[0.4–12], z = 2.104, p = 0.035; Proportion CSF performed

§Scenario A vs Scenario C: difference -29% [–35 ––23], z = -9.662, p<0.001

Groups comparison computerized decision support

When we used the computerized decision support (scenario A, Fig 1A), we can discern different groups. Table 3 shows the baseline characteristics of these different groups. In the largest group 1, patients already had PCC≥0.80 using neuropsychology, MRI and NPO. This group contained mainly controls and patients with a clear clinical profile, i.e. large difference between the highest and second-highest DSI. In group 2, CSF was not helpful as PCC remained <0.80 also after adding simulated CSF values. This group consisted mainly of AD patients, yet with more WMH and a small difference to the second-highest DSI, indicating high probability of comorbid neuropathology. In group 3 CSF was helpful as PCC increased from <0.80 to ≥0.80 by adding simulated CSF, and was ≥0.80 also after adding the actual CSF values (a diagnosis could be made). This group contained mainly FTD and AD patients, where especially tau and p-tau were more abnormal. Finally, in group 4 PCC increased from <0.80 to ≥0.80 by adding simulated CSF, but remained <0.80 when adding the actual CSF values. Patients in group 4 seemed to have multiple underlying neuropathology, with higher tau and more WMH, and a smaller difference with the second-highest DSI.

Table 3. Patient groups based on computerized decision support (matching Fig 1A).

Directly diagnosed CSF not useful Diagnosis based on CSF Not diagnosed Group wise comparison
Group 1 Group 2 Group 3 Group 4 P value
  n = 308 n = 87 n = 71 n = 69
Female, n (%) 164 (53) 31 (36) 34 (48) 33 (48) p = 0.036
Age, in years 64 ± 8 66 ± 8 65 ± 8 67 ± 8 p = 0.0045
APOE e4 carrier, n (%) 150 (55) 34 (43) 30 (47) 38 (66) p = 0.046
MMSE 24 ± 5 25 ± 4 22 ± 4 22 ± 5 p<0.001
MRI
cMTA right 1.1 ± 1.0 1.1 ± 1.0 1.4 ± 0.8 1.4 ± 0.8 p<0.001
cMTA left 1.1 ± 1.1 1.2 ± 1.1 1.5 ± 1.1 1.6 ± 1.0 p = 0.0061
cGCA 0.9 ± 0.9 0.9 ± 0.8 1.4 ± 0.8 1.3 ± 0.8 p<0.001
WMH, ml 7.3 ± 16.5 9.1 ± 10.9 3.9 ± 4.4 7.2 ± 9.8 p = 0.127
AD similarity scale 0.5 ± 0.2 0.5 ± 0.1 0.6 ± 0.1 0.6 ± 0.1 p<0.001
Anterior Posterior index -0.1 ± 1.5 -0.4 ± 1.5 -0.3 ± 1.2 -0.6 ± 1.3 p = 0.066
CSF
AB42, pg/ml 737 ± 300 696 ± 296 653 ± 276 619 ± 251 p = 0.0074
Total tau, pg/ml 485 ± 365 485 ± 315 667 ± 399 554 ± 380 p = 0.0018
P- tau, pg/ml 65 ± 36 65 ± 32 87 ± 45 69 ± 34 p<0.001
Clinical diagnosis p<0.001
Control, n (%) 109 (78) 25 (18) 2 (1) 3 (2)
AD, n (%) 134 (47) 42 (15) 55 (19) 55 (19)
FTD, n (%) 44 (54) 15 (18) 13 (16) 10 (12)
VaD, n (%) 21 (75) 5 (18) 1 (4) 1 (4)
Difference with second DSI without CSF 0.32 ± 0.09 0.08 ± 0.05 0.11 ± 0.05 0.09 ± 0.04 p<0.001
AUC+ (DSI for AD>0.6), n (%) 138 (45) 22 (25) 67 (94) 68 (99) p<0.001

AD: Alzheimer´s disease, FTD: Frontotemporal dementia, VAD: Vascular dementia, MMSE: Mini-Mental state Examination, cMTA: computed medial temporal lobe atrophy scale (0–4), derived from volume of hippocampus and volume of inferor lateral ventricle, cGCA: computed global cortical atrophy scale (0–3), derived from concentration of cortical grey matter using voxel based morphometry, WMH: volume of white matter hyperintensities, AD similarity scale:based on hippocampus ROI, Anterior posterior index: weighted ratio of volumes of the frontal/temporal lobes and parietal/occipital lobes. MRI volumes are adjusted for head size. AB42: beta amyloid 1–42; p-tau: tau phosphorylated at threonine 181. Difference with second DSI without CSF: difference between the two most similar diagnostic groups (first and second DSI), AUC+: number of patients fulfilling appropriate use criteria according to [33], operationalized as a DSI for AD >0.6.

Data are presented as mean ± SD, unless otherwise specified

Nearly all patients in groups 3 and 4 were considered appropriate for CSF following the AUC, showing that our computer-supported decision does not result in a risk of missing patients who would actually have benefited from CSF. In addition, 45% of the patients in group 1 and 30% of the patients in group 2 were considered eligible for CSF testing according to AUC. Yet our computerized decision support predicted that additional CSF testing would not yield additional diagnostic accuracy.

Different PCC cut-offs

For the current manuscript, we arbitrarily defined an accurate diagnosis as PCC≥0.80. To circumvent this inherent arbitrariness, we repeated our analyses for different PCC cutoffs. Fig 2 shows the share of patients with CSF tested and the share of patients diagnosed for different PCC cutoffs, comparing the four diagnostic scenarios: the computerized decision support, no CSF, AUC and CSF for all patients. Overall, a higher PCC cutoff resulted in a lower number of diagnosed patients. Independent of PCC cutoff, applying no CSF resulted in the smallest share of patients diagnosed, while the stepwise approach gave the largest share of patients diagnosed with CSF tested in max 26.5% patients.

Fig 2. Share of patients diagnosed and share of patients with CSF measured for different probability of correct class cutoffs, comparing computerized decision support, no CSF, AUC and CSF for all patients.

Fig 2

Blue: proportion of patients diagnosed, Red: proportion of patients with CSF measured, PCC: probability of correct class. Solid lines show results for the computerized decision support (Fig 1A), dotted lines show results for using no CSF, but only neuropsychology, MRI and APOE (Fig 1B), dashed dotted lines show results for AUC (Fig 1C) and dashed lines using all data (Fig 1D).

Regardless of PCC cutoff, the accuracy of the four different approaches remained comparable (see S1 Fig) demonstrating that the stepwise approach does not compromise accuracy.

Visualization of computerized decision support

To enable clinicians to use this method in daily practice, we have visualized the computerized decision support in Fig 3. Using the DSI classifier, clinicians select the PCC threshold (default PCC≥0.80) and simulate positive and negative CSF biomarkers. The resulting change in PCC can then support the clinician to refrain from or embark on CSF testing. In Fig 3, two examples of the visualization for clinical use are shown. For patient A, the clinician chose the default PCC cutoff of 0.80. When using only neuropsychology, MRI and APOE data, the highest DSI was obtained for AD with a 72% probability of correct diagnosis. When adding simulated CSF, PCC increased to 0.89 (AD-like CSF), resulting also in a higher DSI for AD. When CSF would be negative, PCC also increased to 0.86 resulting in higher DSI for FTD. As the simulated CSF values resulted in a PCC above the cutoff of 0.80, the suggested advice by the tool is that “CSF measurement is considered potentially useful”. Following this advice, the results of the actually performed CSF testing are shown in the bottom panel. The actual CSF biomarker results were AD-like, resulting in a clinical AD diagnosis with 91% probability of correct diagnosis. For patient B, however, simulating only negative CSF increased the PCC above 0.80. And indeed, actual CSF ruled out AD and indicated FTD as the most probable diagnosis. This was also the clinical diagnosis.

Fig 3. Examples of visualization simulated CSF for clinical use.

Fig 3

PCC: probability of correct class, AD: Alzheimer´s disease, FTD: Frontotemporal dementia, VAD: Vascular dementia, CN: control. For patient A both simulated positive and negative CSF resulted in an increase of PCC. Adding actual CSF confirmed the AD diagnosis. For patient B simulated negative CSF increased the PCC. Adding actual CSF ruled out AD and indicated FTD as the most probable diagnosis.

Discussion

Although relevant for diagnosing and excluding AD, it remains unclear in which patients applying CSF biomarkers has most added value. This study showed that a data-driven, individualized approach, simulating CSF outcome before embarking on actual CSF testing, can help in identifying those patients in whom CSF biomarkers are most likely to contribute to a more accurate diagnosis. Recommending CSF testing in only one quarter of patients, our computerized decision support led to a diagnosis with sufficient confidence in 71% of the patients. By contrast, in the three other scenarios (CSF in none, based on AUC or CSF in all) less patients were diagnosed with sufficient confidence. By visualizing the effect of simulated CSF (Fig 3) on the probability of correct class (PCC), clinicians can be supported in making a conscious decision in ordering additional biomarker testing.

Neurodegenerative disorders in general and AD in specific, are increasingly considered as biologically defined diseases. Current guidelines and diagnostic criteria therefore emphasize the use of biomarkers [2, 4]. By using these biomarkers, the clinician can either confirm or rule out underlying AD. Yet, clinicians do not often apply CSF biomarkers [36]. The appropriate use criteria (AUC) which have been proposed as a guideline, are generally worded and challenging to translate to clinical practice [6]. In our study, more than half of the patients would be selected for CSF measurement based on fulfillment of the AUC. In addition to specifying criteria, the AUC further mention that the clinician should estimate ‘how CSF biomarkers might contribute to the diagnosis and clinical decision making’ [6]. Operationalization of this statement is difficult. Tools are therefore needed in search for an answer on the question ‘which patient benefits most from CSF?’. To select patients for amyloid measurement, previous studies have aimed to ‘predict’ amyloid positivity in controls and patients with AD [3740]. Yet a clinician requires information on how knowledge on amyloid predicts clinical outcome. We added to these studies by using a data-driven approach including not only controls and AD, but also patients with VaD and FTD, and simulated AB42, tau and p-tau values to guide decision making.

In the computerized decision support approach, PCC is first calculated using only neuropsychology, MRI and APOE data. If the clinician considers the PCC (= the confidence in the correct diagnosis) high enough, no additional testing is needed. This was frequently the case for controls and VaD [11]. When the PCC is too low, simulated normal or AD-like CSF values are added, enabling the clinicians to decide whether the change in PCC is enough for them to order additional testing. If the PCC remains low, the clinician could use this information to refrain from embarking on CSF testing, which would only yield unnecessary costs without improving diagnostic accuracy. Contrariwise the clinician could take other further steps to improve diagnostic certainty, e.g. genetic screening or FDG-PET. One could imagine that comparable approaches would be developed for these other diagnostic tests.

As an example, we used PCC≥0.80 as a cut-off. There is however some inherent arbitrariness to the cut-off of diagnostic certainty. It is for example unclear what level of certainty is acceptable in clinical practice. Also, clinicians may value the PCC differently, and may want to use less or more stringent cut-off values. This is why we repeated our analyses for a wide range of cut-offs, resulting in Fig 2. Independent of PCC cut-off, the results remained largely similar: using estimated CSF values to support the choice to embark on testing, helps to reduce costs (smaller proportion CSF testing) while improving diagnostic certainty.

An often mentioned, and important limitation of data-driven decision support systems is that they require multiple pieces of data that are not readily available to the clinician, or that the models are too complex which limit their clinical footprint. In order for clinical decision support tools to be implemented in clinical practice, they should have an understandable basis and be intuitive (i.e. not a complex black box), time efficient, and assist rather than replace the clinician [7]. Our stepwise approach meets all these requirements; well-known CSF values are simulated in a simple way and the tool provides a suggestion on the usefulness of CSF testing. We used the DSI classifier, an existing, validated machine learning algorithm [10, 12]. The DSI classifier has a graphical counterpart which makes interpretation of results to clinicians more transparent (an example is shown in Fig 3), tolerates missing data (no imputation needed), and gives information about the confidence of the classification.

Furthermore, we used data of three different memory clinic cohorts, warranting a heterogeneous sample. All patients came to the memory clinic seeking medical help; our data thus reflects real-life, clinical patients. In each clinic, patients were diagnosed using a standard clinical workup, and we were able to use measures that overlapped. CSF analyses were performed in three different laboratories, yet we found no significant differences in values between the three centers. MRI scans were acquired on systems with different field strengths, but the automatic analyses of these scans were all performed by the same software [9].

Some limitations also warrant discussion. First, we categorized patients as having single pathologies, while on average 20–40% patients have in fact multiple pathologies, contributing to their syndrome of dementia [41]. This is probably also the case in our sample, as can be seen in diagnostic group 4, described in Fig 1 and Table 3. Here we see a smaller difference with the second DSI (and subsequently a low PCC), suggesting mixed pathology. Second, a limitation could be the use of SCD as controls. However, if SCD is indeed an enriched population, then the ability of our computerized decision support approach to select the correct patient for CSF testing would be underestimated, and would even be better if ‘healthy controls’ were used. In fact, clinical practice is not about comparing AD patients with healthy controls, but on the differential diagnosis. As SCD patients do seek help for their complaints, they also reflect daily clinical practice, where clinical decision support systems should work.Third, we simulated positive or negative CSF biomarkers and did not include amyloid-PET. This would hamper use of our computerized approach, for example, in the United States, where amyloid-PET is more frequently used than CSF. However, we know CSF correlates strongly with amyloid-PET, especially when both AB42 and tau have been used (as we did here) [42]. Research replicating this study using PET instead, would be a useful addition to this work. Another interesting field for future work is MCI. CSF is a known predictor of progression from MCI to dementia; applying a stepwise approach might be helpful in guiding clinicians also in this challenge [43]. Finally, future studies should evaluate this approach in a prospective design.

The diagnostic routine for AD is very suitable for shared decision making as there is typically more than one reasonable option to choose, such as either perform or not perform additional CSF testing [44]. In shared decision making, the clinician and patient decide together which care plan best fits with individual preferences and needs [45]. An important aspect of shared decision making is that appropriate expectations are set, and recognize that the decision for or against biomarker testing is a highly individualized decision that warrants a shared decision making approach [46]. However, previous research showed that informing patients on diagnostic testing in general was not standard in the memory clinic [36, 44, 47]. The current paper helps clinicians to appreciate whether biomarkers are potentially useful for a specific patient. This information could then be the topic of a dialogue between clinician and patient/caregiver before actually ordering the test. This in turn aids the clinicians in appropriate expectation setting before embarking on biomarker testing.

In conclusion, we showed that computerized decision support can guide clinicians in who will benefit from CSF and who will not. This resulted in an optimized diagnosis that is at the same time cost-effective. Clinicians can thus order CSF applying a data-driven, individualized approach. By visualizing the effect of simulated CSF on diagnostic certainty, clinicians can be supported in making a weighted decision in ordering additional biomarker testing.

Supporting information

S1 File. Probability of correct class (PCC).

(PDF)

S1 Fig. Accuracy for different probability of correct class cutoff’s, comparing computerized decision support, no CSF, AUC and CSF for all patients.

PCC: probability of correct class

Solid lines show results for the computerized decision support (Fig 1A), dotted lines show results for using no CSF, but only neuropsychology, MRI and APOE (Fig 1B), dashed dotted lines show results for AUC (Fig 1C) and dashed lines using all data (Fig 1D).

(TIF)

Acknowledgments

Research of the Alzheimer center Amsterdam is part of the neurodegeneration research program of Amsterdam Neuroscience. The Alzheimer Center Amsterdam is supported by Stichting Alzheimer Nederland and Stichting VUmc fonds. The clinical database structure was developed with funding from Stichting Dioraphte. Wiesje M van der Flier holds the Pasman chair.

Data Availability

The data from this study are available upon request. For this study we requested data from the PredictND consortium. The data-sharing agreement of the PredictND consortium however allows us use of data for this specific project only. The data underlying the results presented in the study are available from the PredictND-board; team@predictnd.eu.

Funding Statement

This study is partly funded by Combinostics. The funder provided support in the form of salaries for authors [JK and JL], and had an additional role in the study as Juha Koikkalainen and Jyrki Lötjönen developed the method and quantitative raw data were generated using Combinostics’ tools. They also reviewed the manuscript. The specific roles of these authors are articulated in the ‘author contributions’ section. For development of the PredictAD tool, VTT Technical Research Centre of Finland has received funding from European Union’s Seventh Framework Programme for research, technological development and demonstration under grant agreements 601055 (VPH-DARE@IT), 224328 (PredictAD), and 611005 (PredictND). The latter had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.'

References

  • 1.2016 Alzheimer's disease facts and figures. Alzheimers Dement. 2016;12(4):459–509. Epub 2016/08/30. . [DOI] [PubMed] [Google Scholar]
  • 2.Jack CR Jr., Bennett DA, Blennow K, Carrillo MC, Dunn B, Haeberlein SB, et al. NIA-AA Research Framework: Toward a biological definition of Alzheimer's disease. Alzheimers Dement. 2018;14(4):535–62. Epub 2018/04/15. 10.1016/j.jalz.2018.02.018 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Mattsson N, Lonneborg A, Boccardi M, Blennow K, Hansson O. Clinical validity of cerebrospinal fluid Abeta42, tau, and phospho-tau as biomarkers for Alzheimer's disease in the context of a structured 5-phase development framework. Neurobiol Aging. 2017;52:196–213. Epub 2017/03/21. 10.1016/j.neurobiolaging.2016.02.034 . [DOI] [PubMed] [Google Scholar]
  • 4.McKhann GM, Knopman DS, Chertkow H, Hyman BT, Jack CR Jr., Kawas CH, et al. The diagnosis of dementia due to Alzheimer's disease: recommendations from the National Institute on Aging-Alzheimer's Association workgroups on diagnostic guidelines for Alzheimer's disease. Alzheimers Dement. 2011;7(3):263–9. S1552-5260(11)00101-4 [pii]; 10.1016/j.jalz.2011.03.005 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Frisoni GB, Boccardi M, Barkhof F, Blennow K, Cappa S, Chiotis K, et al. Strategic roadmap for an early diagnosis of Alzheimer's disease based on biomarkers. Lancet Neurol. 2017;16(8):661–76. Epub 2017/07/20. 10.1016/S1474-4422(17)30159-X . [DOI] [PubMed] [Google Scholar]
  • 6.Shaw LM, Arias J, Blennow K, Galasko D, Molinuevo JL, Salloway S, et al. Appropriate use criteria for lumbar puncture and cerebrospinal fluid testing in the diagnosis of Alzheimer's disease. Alzheimers Dement. 2018;14(11):1505–21. Epub 2018/10/15. 10.1016/j.jalz.2018.07.220 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Shortliffe EH, Sepulveda MJ. Clinical Decision Support in the Era of Artificial Intelligence. Jama. 2018;320(21):2199–200. Epub 2018/11/07. 10.1001/jama.2018.17163 . [DOI] [PubMed] [Google Scholar]
  • 8.Mattila J, Koikkalainen J, Virkki A, Simonsen A, van GM, Waldemar G, et al. A disease state fingerprint for evaluation of Alzheimer's disease. J Alzheimers Dis. 2011;27(1):163–76. KG54325631131N10 [pii]; 10.3233/JAD-2011-110365 [DOI] [PubMed] [Google Scholar]
  • 9.Koikkalainen J, Rhodius-Meester H, Tolonen A, Barkhof F, Tijms B, Lemstra AW, et al. Differential diagnosis of neurodegenerative diseases using structural MRI data. Neuroimage Clin. 2016;11:435–49. 10.1016/j.nicl.2016.02.019 S2213-1582(16)30040-7 [pii]. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Tolonen A, F. M. Rhodius-Meester H, Bruun M, Koikkalainen J, Barkhof F, Lemstra A, et al. Data-Driven Differential Diagnosis of Dementia Using Multiclass Disease State Index Classifier. Front Aging Neurosci. 2018;10:111 10.3389/fnagi.2018.00111 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Bruun M, Rhodius-Meester HFM, Koikkalainen J, Baroni M, Gjerum L, Lemstra AW, et al. Evaluating combinations of diagnostic tests to discriminate different dementia types. Alzheimers Dement (Amst). 2018;10:509–18. Epub 2018/10/16. 10.1016/j.dadm.2018.07.003 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Rhodius-Meester HFM, Liedes H, Koikkalainen J, Wolfsgruber S, Coll-Padros N, Kornhuber J, et al. Computer-assisted prediction of clinical progression in the earliest stages of AD. Alzheimers Dement (Amst). 2018;10:726–36. Epub 2019/01/09. 10.1016/j.dadm.2018.09.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Bruun M, Frederiksen KS, Rhodius-Meester HFM, Baroni M, Gjerum L, Koikkalainen J, et al. Impact of a Clinical Decision Support Tool on Dementia Diagnostics in Memory Clinics: The PredictND Validation Study. Current Alzheimer research. 2019;16(2):91–101. Epub 2019/01/04. 10.2174/1567205016666190103152425 . [DOI] [PubMed] [Google Scholar]
  • 14.van der Flier WM, Pijnenburg YA, Prins N, Lemstra AW, Bouwman FH, Teunissen CE, et al. Optimizing patient care and research: the Amsterdam Dementia Cohort. J Alzheimers Dis. 2014;41(1):313–27. LR61J34616453267 [pii]; 10.3233/JAD-132306 [DOI] [PubMed] [Google Scholar]
  • 15.van der Flier WM, Scheltens P. Amsterdam Dementia Cohort: Performing Research to Optimize Care. J Alzheimers Dis. 2018;62(3):1091–111. Epub 2018/03/23. 10.3233/JAD-170850 . [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Rascovsky K, Hodges JR, Knopman D, Mendez MF, Kramer JH, Neuhaus J, et al. Sensitivity of revised diagnostic criteria for the behavioural variant of frontotemporal dementia. Brain. 2011;134(Pt 9):2456–77. awr179 [pii]; 10.1093/brain/awr179 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Gorno-Tempini ML, Hillis AE, Weintraub S, Kertesz A, Mendez M, Cappa SF, et al. Classification of primary progressive aphasia and its variants. Neurology. 2011;76(11):1006–14. WNL.0b013e31821103e6 [pii]; 10.1212/WNL.0b013e31821103e6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Roman GC, Tatemichi TK, Erkinjuntti T, Cummings JL, Masdeu JC, Garcia JH, et al. Vascular dementia: diagnostic criteria for research studies. Report of the NINDS-AIREN International Workshop. Neurology. 1993;43(2):250–60. 10.1212/wnl.43.2.250 [DOI] [PubMed] [Google Scholar]
  • 19.Folstein MF, Folstein SE, McHugh PR. "Mini-mental state". A practical method for grading the cognitive state of patients for the clinician. J Psychiatr Res. 1975;12(3):189–98. 0022-3956(75)90026-6 [pii]. 10.1016/0022-3956(75)90026-6 [DOI] [PubMed] [Google Scholar]
  • 20.Schmand B, Houx P, de Koning I. Normen van psychologische tests voor gebruik in de klinische neuropsychologie (in Dutch). Nederlands Instituut van Psychologen; 2012;www.psynip.nl/website/sectoren-en-secties/sector-gezondheidszorg/neuropsychologie. [Google Scholar]
  • 21.Reitan R. Validity of the Trail Making Test as an indicator of organic brain damage. Percept Mot Skills. 1958;8:271–6. [Google Scholar]
  • 22.Van der Elst W, Van Boxtel MP, Van Breukelen GJ, Jolles J. Normative data for the Animal, Profession and Letter M Naming verbal fluency tests for Dutch speaking participants and the effects of age, education, and sex. J Int Neuropsychol Soc. 2006;12(1):80–9. S1355617706060115 [pii]; 10.1017/S1355617706060115 [DOI] [PubMed] [Google Scholar]
  • 23.Cummings JL, Mega M, Gray K, Rosenberg-Thompson S, Carusi DA, Gornbein J. The Neuropsychiatric Inventory: comprehensive assessment of psychopathology in dementia. Neurology. 1994;44(12):2308–14. 10.1212/wnl.44.12.2308 [DOI] [PubMed] [Google Scholar]
  • 24.Scheltens P, Leys D, Barkhof F, Huglo D, Weinstein HC, Vermersch P, et al. Atrophy of medial temporal lobes on MRI in "probable" Alzheimer's disease and normal ageing: diagnostic value and neuropsychological correlates. J Neurol Neurosurg Psychiatry. 1992;55(10):967–72. 10.1136/jnnp.55.10.967 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Koikkalainen JR, Rhodius-Meester HFM, Frederiksen KS, Bruun M, Hasselbalch SG, Baroni M, et al. Automatically computed rating scales from MRI for patients with cognitive disorders. European radiology. 2019. Epub 2019/02/24. 10.1007/s00330-019-06067-1 . [DOI] [PubMed] [Google Scholar]
  • 26.Lotjonen JM, Wolz R, Koikkalainen JR, Thurfjell L, Waldemar G, Soininen H, et al. Fast and robust multi-atlas segmentation of brain magnetic resonance images. Neuroimage. 2010;49(3):2352–65. Epub 2009/10/28. 10.1016/j.neuroimage.2009.10.026 . [DOI] [PubMed] [Google Scholar]
  • 27.Ashburner J, Friston KJ. Voxel-based morphometry—the methods. Neuroimage. 2000;11(6 Pt 1):805–21. Epub 2000/06/22. 10.1006/nimg.2000.0582 . [DOI] [PubMed] [Google Scholar]
  • 28.Vos SJ, Visser PJ, Verhey F, Aalten P, Knol D, Ramakers I, et al. Variability of CSF Alzheimer's disease biomarkers: implications for clinical practice. PLoS One. 2014;9(6):e100784 10.1371/journal.pone.0100784 PONE-D-14-16936 [pii]. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Bruun M, Koikkalainen J, Rhodius-Meester HFM, Baroni M, Gjerum L, van Gils M, et al. Detecting frontotemporal dementia syndromes using MRI biomarkers. NeuroImage Clinical. 2019;22:101711 Epub 2019/02/12. 10.1016/j.nicl.2019.101711 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Wang Y, Catindig JA, Hilal S, Soon HW, Ting E, Wong TY, et al. Multi-stage segmentation of white matter hyperintensity, cortical and lacunar infarcts. Neuroimage. 2012;60(4):2379–88. Epub 2012/03/06. 10.1016/j.neuroimage.2012.02.034 . [DOI] [PubMed] [Google Scholar]
  • 31.Buckner RL, Head D, Parker J, Fotenos AF, Marcus D, Morris JC, et al. A unified approach for morphometric and functional data analysis in young, old, and demented adults using automated atlas-based head size normalization: reliability and validation against manual measurement of total intracranial volume. Neuroimage. 2004;23(2):724–38. Epub 2004/10/19. 10.1016/j.neuroimage.2004.06.018 . [DOI] [PubMed] [Google Scholar]
  • 32.Cole TJ, Green PJ. Smoothing reference centile curves: the LMS method and penalized likelihood. Statistics in medicine. 1992;11(10):1305–19. Epub 1992/07/01. 10.1002/sim.4780111005 . [DOI] [PubMed] [Google Scholar]
  • 33.Girdler-Brown BV, Dzikiti LN. Hypothesis tests for the difference between two population proportions using Stata. Southern African Journal of Public Health. 2018;2(3):63–8. 10.7196/SHS.2018.v2.i3.71. [DOI] [Google Scholar]
  • 34.Cluitmans L, Mattila J, Runtti H, van Gils M, Lotjonen J. A MATLAB toolbox for classification and visualization of heterogenous multi-scale human data using the Disease State Fingerprint method. Stud Health Technol Inform. 2013;189:77–82. Epub 2013/06/07. . [PubMed] [Google Scholar]
  • 35.Da X, Toledo JB, Zee J, Wolk DA, Xie SX, Ou Y, et al. Integration and relative value of biomarkers for prediction of MCI to AD progression: spatial patterns of brain atrophy, cognitive scores, APOE genotype and CSF biomarkers. NeuroImage Clinical. 2014;4:164–73. Epub 2013/12/29. 10.1016/j.nicl.2013.11.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Visser LNC, Kunneman M, Murugesu L. Clinician-patient communication during the diagnostic work-up: the ABIDE project. Alzheimer's & Dementia: The Journal of the Alzheimer's Association. 2019;11:520–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Prestia A, Caroli A, Wade SK, van der Flier WM, Ossenkoppele R, Van BB, et al. Prediction of AD dementia by biomarkers following the NIA-AA and IWG diagnostic criteria in MCI patients from three European memory clinics. Alzheimers Dement. 2015. S1552-5260(14)02890-8 [pii]; 10.1016/j.jalz.2014.12.001 [DOI] [PubMed] [Google Scholar]
  • 38.Insel PS, Palmqvist S, Mackin RS, Nosheny RL, Hansson O, Weiner MW, et al. Assessing risk for preclinical beta-amyloid pathology with APOE, cognitive, and demographic information. Alzheimers Dement (Amst). 2016;4:76–84. Epub 2016/10/11. 10.1016/j.dadm.2016.07.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Verberk I, Slot RE, Verfaillie SC, Heijst H, Prins ND, Van Berckel B, et al. Plasma-amyloid as pre-screener for the earliest Alzheimer's pathological changes. Ann Neurol. 2018;in press. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Palmqvist S, Insel PS, Zetterberg H, Blennow K, Brix B, Stomrud E, et al. Accurate risk estimation of beta-amyloid positivity to identify prodromal Alzheimer's disease: Cross-validation study of practical algorithms. Alzheimers Dement. 2019;15(2):194–204. Epub 2018/10/27. 10.1016/j.jalz.2018.08.014 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Zekry D, Hauw JJ, Gold G. Mixed dementia: epidemiology, diagnosis, and treatment. J Am Geriatr Soc. 2002;50(8):1431–8. Epub 2002/08/08. 10.1046/j.1532-5415.2002.50367.x . [DOI] [PubMed] [Google Scholar]
  • 42.Zwan M, van Harten A, Ossenkoppele R, Bouwman F, Teunissen C, Adriaanse S, et al. Concordance between cerebrospinal fluid biomarkers and [11C]PIB PET in a memory clinic cohort. J Alzheimers Dis. 2014;41(3):801–7. Epub 2014/04/08. 10.3233/JAD-132561 . [DOI] [PubMed] [Google Scholar]
  • 43.van Maurik IS, Zwan MD, Tijms BM, Bouwman FH, Teunissen CE, Scheltens P, et al. Interpreting Biomarker Results in Individual Patients With Mild Cognitive Impairment in the Alzheimer's Biomarkers in Daily Practice (ABIDE) Project. JAMA Neurol. 2017;74(12):1481–91. Epub 2017/10/20. 10.1001/jamaneurol.2017.2712 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Van der Flier W, Kunneman M, Bouwman FH, Petersen R, Smets EMA. Diagnostic dilemmas in Alzheimer's disease: Room for shared decision making. Alzheimers Dement. 2017;3(3):301–4. 10.1016/j.trci.2017.03.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Stiggelbout AM, Pieterse AH, De Haes JC. Shared decision making: Concepts, evidence, and practice. Patient education and counseling. 2015;98(10):1172–9. Epub 2015/07/29. 10.1016/j.pec.2015.06.022 . [DOI] [PubMed] [Google Scholar]
  • 46.Grill JD, Apostolova LG, Bullain S, Burns JM, Cox CG, Dick M, et al. Communicating mild cognitive impairment diagnoses with and without amyloid imaging. Alzheimers Res Ther. 2017;9(1):35 Epub 2017/05/06. 10.1186/s13195-017-0261-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Kunneman M, Pel-Littel R, Bouwman FH, Gillissen F, Schoonenboom NSM, Claus JJ, et al. Patients' and caregivers' views on conversations and shared decision making in diagnostic testing for Alzheimer's disease: The ABIDE project. Alzheimer's & dementia (New York, N Y). 2017;3(3):314–22. Epub 2017/10/27. 10.1016/j.trci.2017.04.002 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Han Zhang

23 Oct 2019

PONE-D-19-25997

Selection of memory clinic patients for CSF biomarker assessment can be restricted to a quarter of cases by using computerized decision support, without compromising diagnostic accuracy

PLOS ONE

Dear Dr. Rhodius-Meester,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please do improve the readability and clarity during the revision to make it easier for general readership. 

We would appreciate receiving your revised manuscript by Dec 07 2019 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.

We look forward to receiving your revised manuscript.

Kind regards,

Han Zhang, Ph.D

Academic Editor

PLOS ONE

Journal Requirements:

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please describe in your methods section how capacity to provide consent was determined for the participants in this study. Please also state whether your ethics committee or IRB approved this consent procedure. If it was not necessary to assess the capacity to provide consent of your participants please outline why this was not required in this case

3. We note that you have indicated that data from this study are available upon request. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions.

In your revised cover letter, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially identifying or sensitive patient information) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings as either Supporting Information files or to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. Please see http://www.bmj.com/content/340/bmj.c181.long for guidelines on how to de-identify and prepare clinical data for publication. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories.

We will update your Data Availability statement on your behalf to reflect the information you provide.

4. Thank you for stating the following in the Financial Disclosure section:

'The author(s) received no specific funding for this work.'

We note that one or more of the authors are employed by a commercial company: Combinostics Ltd and VTT Technical Research Centre of Finland Ltd.

  1. Please provide an amended Funding Statement declaring this commercial affiliation, as well as a statement regarding the Role of Funders in your study. If the funding organization did not play a role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript and only provided financial support in the form of authors' salaries and/or research materials, please review your statements relating to the author contributions, and ensure you have specifically and accurately indicated the role(s) that these authors had in your study. You can update author roles in the Author Contributions section of the online submission form.

Please also include the following statement within your amended Funding Statement.

“The funder provided support in the form of salaries for authors [insert relevant initials], but did not have any additional role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. The specific roles of these authors are articulated in the ‘author contributions’ section.”

If your commercial affiliation did play a role in your study, please state and explain this role within your updated Funding Statement.

2. Please also provide an updated Competing Interests Statement declaring this commercial affiliation along with any other relevant declarations relating to employment, consultancy, patents, products in development, or marketed products, etc. 

Within your Competing Interests Statement, please confirm that this commercial affiliation does not alter your adherence to all PLOS ONE policies on sharing data and materials by including the following statement: "This does not alter our adherence to  PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests) . If this adherence statement is not accurate and  there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please know it is PLOS ONE policy for corresponding authors to declare, on behalf of all authors, all potential competing interests for the purposes of transparency. PLOS defines a competing interest as anything that interferes with, or could reasonably be perceived as interfering with, the full and objective presentation, peer review, editorial decision-making, or publication of research or non-research articles submitted to one of the journals. Competing interests can be financial or non-financial, professional, or personal. Competing interests can arise in relationship to an organization or another person. Please follow this link to our website for more details on competing interests: http://journals.plos.org/plosone/s/competing-interests

Please include both an updated Funding Statement and Competing Interests Statement in your cover letter. We will change the online submission form on your behalf.

5. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

Additional Editor Comments (if provided):

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The study showed that using a computerized decision support as a tool to guide in selecting patients who need to be detected CSF is helpful in clinical operations. The computerized decision support approach reduced the patients who need CSF testing and classified varied types of dementia with increased sufficient confidence. The work has the potential clinical utility value.

I still have some questions for this work.

1. In this study, I noticed the controls were individuals with subjective cognitive decline. As far as I know, SCD is defined as the population in higher risk for future cognitive impairment and may have already exist abnormal CSF alteration. Thus, I would like to ask why to chose SCD as controls.

2. The description of imaging markers is too general. In this study, T1 images were analysed to extracted quantitative indicators. However, the procedures were not clear. For example,

VBM was conducted by which software and whether to eliminate the effect of volume of cranial cavity. Please specify the processes which would help others to verified the results.

3. In the study, CSF biomarkers included beta amyloid 1-42 (AB42), total tau and tau phosphorylated at threonine 181 (p-tau), which were all AD-related markers. Thus, whether to test these CSF markers would help improve the accuracy of diagnosis in VaD and FTD?

Reviewer #2: 1) Line 131, the reference associated with RAVLT needs to be revised.

2) Line 134, in [21] the paper presents different verbal fluency tests, why the choice of animals ? and did the author try with other categories to see if this leads to different findings ?

3) Line 157, incorrect citation [9], please provide the original references, I suppose the author needs to cite this paper in his/her case; Lötjönen J., Wolz R., Koikkalainen J., Thurfjell L., Waldemar G., Soininen H., Rueckert D. Fast and robust multi-atlas segmentation of brain magnetic resonance images. NeuroImage. 2010;49(3):2352–2365

4) Line 247, in table 1, not all the characteristics are available for all patients, how did the author deal with missing information ? How confident is he/she about their effect on the results.

5) Line 265, it is unclear to me how the author has simulated the CSF values.

6) Line 276, table2 , I would appreciate if the author interprets and explains why scenario A had higher rate than scenario D.

7) According to the figure in the supplementary notes, the higher DSI/delta_DSI the closer PCC to 1. Hence, the higher PCC the better for the diagnosis, how does the author explain that for values larger than 0.8 the accuracy of diagnosis decreased (line 335).

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2020 Jan 15;15(1):e0226784. doi: 10.1371/journal.pone.0226784.r002

Author response to Decision Letter 0


8 Nov 2019

Amsterdam, 31 October 2019

Regarding PONE-D-19-25997

Dear dr Han Zhang,

Please find uploaded our revised manuscript for publication in PLOS ONE as part of the special collection on ‘Early diagnosis and treatment of Alzheimer’s disease’.

We thank the editor and reviewers for their careful reading, thoughtful comments, and recommendation for revision. Please find below our responses to the comments in a point-by-point fashion. We have highlighted changes (via track&change) in response to the reviewers’ comments in the manuscript and supplemental data.

We hope the rebuttal adequately addresses the points raised during the review process,

Kind regards, also on behalf of the co-authors,

Hanneke Rhodius- Meester

===========================================================================

Journal Requirements:

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

Reply: We have updated our manuscript and file naming according to your requirements.

2. Please describe in your methods section how capacity to provide consent was determined for the participants in this study. Please also state whether your ethics committee or IRB approved this consent procedure. If it was not necessary to assess the capacity to provide consent of your participants please outline why this was not required in this case

Reply: We added information on the consent procedure to the paragraph ‘subjects’ in the methods section (L124-126).

3. We note that you have indicated that data from this study are available upon request. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions.

In your revised cover letter, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially identifying or sensitive patient information) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

Reply: We address this in our revised cover letter by adding a paragraph on data access. For this study we requested data from the PredictND consortium. The data-sharing agreement of the PredictND consortium however allows us use of data for this specific project only. The data underlying the results presented in the study are available upon request from the PredictND-board; team@predictnd.eu. We hope you understand the restrictions we unfortunately have to deal with.

4. Thank you for stating the following in the Financial Disclosure section:

'The author(s) received no specific funding for this work.'

We note that one or more of the authors are employed by a commercial company: Combinostics Ltd and VTT Technical Research Centre of Finland Ltd.

1. Please provide an amended Funding Statement declaring this commercial affiliation, as well as a statement regarding the Role of Funders in your study. If the funding organization did not play a role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript and only provided financial support in the form of authors' salaries and/or research materials, please review your statements relating to the author contributions, and ensure you have specifically and accurately indicated the role(s) that these authors had in your study. You can update author roles in the Author Contributions section of the online submission form.

Please also include the following statement within your amended Funding Statement.

“The funder provided support in the form of salaries for authors [insert relevant initials], but did not have any additional role in the study design, data collection and analysis, decision to publish, or preparation of the manuscript. The specific roles of these authors are articulated in the ‘author contributions’ section.”

If your commercial affiliation did play a role in your study, please state and explain this role within your updated Funding Statement.

2. Please also provide an updated Competing Interests Statement declaring this commercial affiliation along with any other relevant declarations relating to employment, consultancy, patents, products in development, or marketed products, etc.

Within your Competing Interests Statement, please confirm that this commercial affiliation does not alter your adherence to all PLOS ONE policies on sharing data and materials by including the following statement: "This does not alter our adherence to PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests) . If this adherence statement is not accurate and there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please know it is PLOS ONE policy for corresponding authors to declare, on behalf of all authors, all potential competing interests for the purposes of transparency. PLOS defines a competing interest as anything that interferes with, or could reasonably be perceived as interfering with, the full and objective presentation, peer review, editorial decision-making, or publication of research or non-research articles submitted to one of the journals. Competing interests can be financial or non-financial, professional, or personal. Competing interests can arise in relationship to an organization or another person. Please follow this link to our website for more details on competing interests: http://journals.plos.org/plosone/s/competing-interests

Please include both an updated Funding Statement and Competing Interests Statement in your cover letter. We will change the online submission form on your behalf.

Reply: We have updated our Funding Statement and Competing Interest Statement, adding that Juha Koikkalainen and Jyrki Lötjönen, as employees of Combinostics, developed the method and quantitative raw data were generated using Combinostics’ tools. They also reviewed the manuscript. The specific roles of these authors are articulated in the ‘author contributions’ section. We confirm that this commercial affiliation does not alter our adherence to all PLOS ONE policies on sharing data and materials and included the following statement: "This does not alter our adherence to PLOS ONE policies on sharing data and materials.” Thank you for changing the online submission form.

5. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

Reply: We updated captions for the supporting information files and in-text citations as requested.

===========================================================================

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

Reviewer #1: Partly

Reviewer #2: Yes

________________________________________

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

________________________________________

3. Have the authors made all data underlying the findings in their manuscript fully available?

Reviewer #1: Yes

Reviewer #2: Yes

________________________________________

4. Is the manuscript presented in an intelligible fashion and written in standard English?

Reviewer #1: Yes

Reviewer #2: Yes

________________________________________

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The study showed that using a computerized decision support as a tool to guide in selecting patients who need to be detected CSF is helpful in clinical operations. The computerized decision support approach reduced the patients who need CSF testing and classified varied types of dementia with increased sufficient confidence. The work has the potential clinical utility value.

I still have some questions for this work.

1. In this study, I noticed the controls were individuals with subjective cognitive decline. As far as I know, SCD is defined as the population in higher risk for future cognitive impairment and may have already exist abnormal CSF alteration. Thus, I would like to ask why to choose SCD as controls.

Reply: Thank you for this observation; the use of SCD patients as controls can be seen as a topic of discussion, which we added (p19, L461-467). SCD might be an enriched population, and we think that the reviewers’ concern is that the ability of our computerized decision support approach to select the correct patient for CSF testing would be underestimated. If we used ‘healthy controls’ the tool might even perform better. However, we feel that using SCD patients in this computerized decision support approach is actually a strength of our study. In fact, clinical practice is not about comparing AD patients with healthy controls, but on the differential diagnosis. As SCD patients do seek help for their complaints, they also reflect daily clinical practice, where clinical decision support systems should work. Finally, biomarker status in healthy controls is often unknown, whereas a proportion of these ‘healthy’ controls will be biomarker positive.

2. The description of imaging markers is too general. In this study, T1 images were analyzed to extracted quantitative indicators. However, the procedures were not clear. For example,

VBM was conducted by which software and whether to eliminate the effect of volume of cranial cavity. Please specify the processes which would help others to verified the results.

Reply: We apologize that the description on imaging markers was unclear to you. As described in the second line of this section, the processes were conducted by the cNeuro® cMRI quantification tool. To enhance readability, and because our core message is not on the imaging markers, we have chosen to show limited details on these markers. These processes of deriving these markers have been described elaborately in several papers, to which we refer in the text. All markers were corrected for age, sex and head size, we added this information to the methods section (L169-170).

3. In the study, CSF biomarkers included beta amyloid 1-42 (AB42), total tau and tau phosphorylated at threonine 181 (p-tau), which were all AD-related markers. Thus, whether to test these CSF markers would help improve the accuracy of diagnosis in VaD and FTD?

Reply: You are of course correct in stating that the chosen CSF biomarkers are all AD-related markers. In clinical practice these are the CSF biomarkers that clinicians can use for detecting AD pathology during life; CSF biomarkers for FTD and VaD unfortunately have not been developed yet. Nonetheless, CSF biomarkers for AD may still be useful in the differential diagnosis of FTD and VaD. In clinical practice the clinician is faced with a differential diagnostic dilemma’s for all his/her patients. It is thus not about ‘controls versus AD’, but about ‘this type of dementia or that type of dementia’. A tool that supports clinicians in ordering CSF biomarkers in daily practice, should thus be able to cope with this differential diagnostic dilemma. Especially since there can be a broad overlap in clinical symptoms in AD, FTD and VaD. Negative AD-biomarkers increase the likelihood of another underlying pathology, and may in turn help the clinician with this differential diagnostic dilemma. Our computerized decision approach aids for all diagnostic groups, as can be seen in table 3, pointing out both patients in which CSF is not needed and patient in which CSF is useful. For this group with different types of dementia, reflecting clinical practice, we need CSF in only 1 out of 4 patients, without compromising diagnostic accuracy.

Reviewer #2: 1) Line 131, the reference associated with RAVLT needs to be revised.

Reply: We apologize for this mistake and added the correct reference (L132).

2) Line 134, in [21] the paper presents different verbal fluency tests, why the choice of animals ? and did the author try with other categories to see if this leads to different findings ?

Reply: Thank you for this interesting suggestion. One can definitely argue for example letter fluency, and the difference between letter and animal fluency, can be very indicative. In our dataset (which we used retrospectively) we only had information on animal fluency, not on other categories. However we were able to construct a brief standardized test battery that contains widely used tests, and is thus representative for clinical practice, where our tool should work.

3) Line 157, incorrect citation [9], please provide the original references, I suppose the author needs to cite this paper in his/her case; Lötjönen J., Wolz R., Koikkalainen J., Thurfjell L., Waldemar G., Soininen H., Rueckert D. Fast and robust multi-atlas segmentation of brain magnetic resonance images. NeuroImage. 2010;49(3):2352–2365

Reply: We added this reference to the revised manuscript (L158).

4) Line 247, in table 1, not all the characteristics are available for all patients, how did the author deal with missing information ? How confident is he/she about their effect on the results.

Reply: Indeed, information on APOEe4 status and several cognitive tests is missing in a small proportion of the patients. Due to the design of DSI, there is no need to impute data or exclude cases with incomplete data, as only available data are used. Furthermore, we repeated our analysis with whole cases only, and found no relevant change. We added a remark on this to the paragraph on the DSI in the Methods section and included a reference to Tolonen Front Aging Neurosciene 2018 (ref 10), were this is explained in detail (L187-188).

5) Line 265, it is unclear to me how the author has simulated the CSF values.

Reply: We simulated the CSF values by taking age- and sex-normalized median AB42, tau and p-tau values of the AD group (=positive CSF values) and by taking age- and sex-normalized median AB42, tau and p-tau values of the SCD group (=negative CSF values). This age- and sex normalization is done in the same way as for the MRI markers (see ref 31). We described this in the paragraph on ‘Simulated CSF to guide clinical decision making’ in the Methods section and added reference 31 to the text (L216). Hopefully this clarifies the used method.

6) Line 276, table2 , I would appreciate if the author interprets and explains why scenario A had higher rate than scenario D.

Reply: Thank you for the opportunity to elaborate on this interesting finding. Scenario A results in a higher proportion of patients diagnosed with sufficient probability as compared to the other scenario’s (including scenario D), while proportion of performed CSF is much lower. This shows exactly why tools like develop in our study are needed; CSF testing is clearly only useful when it increases the confidence in the diagnosis. One can imagine that borderline CSF values or in case of multiple pathologies, adding CSF values only confuses and leads to lower proportion of diagnosed patients. This can also be seen in f.e. group 2 and group 4 (as shown in Table 3) were CSF did not help. We added a sentence on this to the section above Table 2 (L276-278).

7) According to the figure in the supplementary notes, the higher DSI/delta_DSI the closer PCC to 1. Hence, the higher PCC the better for the diagnosis, how does the author explain that for values larger than 0.8 the accuracy of diagnosis decreased (line 335).

Reply: You are correct in assuming that the PCC correlates to the accuracy of the diagnosis. As can be seen in S1 Fig, the accuracy (as indicated by the black line) increases with increasing PCC, also for values larger than 0.8. Setting the cut-off for a ‘certain’ diagnosis higher, results in less patients diagnosed. This can be seen in Fig 2: the proportion of diagnosed patients decreases with increasing PCC, since fewer patients while have a very high PCC leading to a smaller proportion of patients. Please also note that in this line we refer to the number of patients diagnosed, and not to the accuracy. We hope this explains both Fig 2 and S1 Fig.

________________________________________

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Attachment

Submitted filename: Response to reviewers.docx

Decision Letter 1

Han Zhang

6 Dec 2019

Selection of memory clinic patients for CSF biomarker assessment can be restricted to a quarter of cases by using computerized decision support, without compromising diagnostic accuracy

PONE-D-19-25997R1

Dear Dr. Rhodius-Meester,

We are pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it complies with all outstanding technical requirements.

Within one week, you will receive an e-mail containing information on the amendments required prior to publication. When all required modifications have been addressed, you will receive a formal acceptance letter and your manuscript will proceed to our production department and be scheduled for publication.

Shortly after the formal acceptance letter is sent, an invoice for payment will follow. To ensure an efficient production and billing process, please log into Editorial Manager at https://www.editorialmanager.com/pone/, click the "Update My Information" link at the top of the page, and update your user information. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, you must inform our press team as soon as possible and no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

With kind regards,

Han Zhang, Ph.D

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The manuuscript has been revised thoughtfully.The application of this tool in CSF biomarkers in clinic is valuable. I agree to accept this version.

Reviewer #2: (No Response)

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

Acceptance letter

Han Zhang

10 Dec 2019

PONE-D-19-25997R1

Selection of memory clinic patients for CSF biomarker assessment can be restricted to a quarter of cases by using computerized decision support, without compromising diagnostic accuracy

Dear Dr. Rhodius-Meester:

I am pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

For any other questions or concerns, please email plosone@plos.org.

Thank you for submitting your work to PLOS ONE.

With kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Han Zhang

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 File. Probability of correct class (PCC).

    (PDF)

    S1 Fig. Accuracy for different probability of correct class cutoff’s, comparing computerized decision support, no CSF, AUC and CSF for all patients.

    PCC: probability of correct class

    Solid lines show results for the computerized decision support (Fig 1A), dotted lines show results for using no CSF, but only neuropsychology, MRI and APOE (Fig 1B), dashed dotted lines show results for AUC (Fig 1C) and dashed lines using all data (Fig 1D).

    (TIF)

    Attachment

    Submitted filename: Response to reviewers.docx

    Data Availability Statement

    The data from this study are available upon request. For this study we requested data from the PredictND consortium. The data-sharing agreement of the PredictND consortium however allows us use of data for this specific project only. The data underlying the results presented in the study are available from the PredictND-board; team@predictnd.eu.


    Articles from PLoS ONE are provided here courtesy of PLOS

    RESOURCES