Abstract
Cross-validation (CV) is a resampling approach to evaluate machine learning models when sample size is limited. The number of all possible combinations of folds for the training data, known as CV rounds, are often very small in leave-one-out CV. Alternatively, Monte Carlo cross-validation (MCCV) can be performed with a flexible number of simulations when computational resources are feasible for a study with limited sample size. We conduct extensive simulation studies to compare accuracy between MCCV and CV with the same number of simulations for a study with binary outcome (e.g., disease progression or not). Accuracy of MCCV is generally higher than CV although the gain is small. They have similar performance when sample size is large. Meanwhile, MCCV is going to provide reliable performance metrics as the number of simulations increases. Two real examples are used to illustrate the comparison between MCCV and CV.
Keywords: Alzheimer’s disease, Binary outcome, Cross-validation, Machine learning, Monte Carlo cross-validation
Introduction
Machine learning (ML) methods are increasingly applied to improve diagnostic classification in clinical research [5, 9]. For a study with categorical outcome whose classes are known beforehand, supervised ML methods can be used to predict outcomes for a separate data [18, 31]. Linear discriminant analysis (LDA) finds a linear combination of features that separates two or more classes. Logistic regression is a special case of LDA for a two-class classification problem. The k-nearest neighbors algorithm is a simple and easy method that assumes similar individuals being close to each other. Regression tree starts from root of a tree with all the features as nodes [28, 34, 36]. For every possible route from root to the end of a branch that does not split any further, a classification is made. Ensemble classification and boosting are two techniques to improve weak methods, such as stochastic gradient boosting [7]. Random forest achieves classification via a majority voting from all decision trees [20, 21, 26, 42]. Support vector machine (SVM) finds a decision function that maximizes the margin around the separating hyperplane by developing a mapping from features to classes as a combination of kernels [11]. SVMs are preferable in many researches due to their high accuracy in model prediction [38].
Cross-validation (CV) procedure is traditionally applied to build ML models. To perform CV, data are split into k small folds (e.g., ). The majority of these folds are used as the training data and the remaining folds are the testing data. For leave-one-out CV with 10 folds, there is a total of possible 10 CV rounds that is the number of all possible combinations of folds for the training data. When it is computationally intensive, researchers may only run one round in the model building, which could introduce a significant amount of bias in the model performance metrics. In light of this issue, researchers may run all CV rounds to reduce the estimate bias (e.g., 10 rounds in leave-one-out CV with 10 folds).
Alternatively, one may consider Monte Carlo cross-validation (MCCV) that splits data into two subsets by sampling without replacement. MCCV is a simple and effective approach in data science [37]. Xu et al. [40, 41] compared MCCV with leave-one-out CV for a linear regression model with continuous outcome, and found that MCCV has a better performance than leave-one-out CV with regards to mean squared error of prediction. Shao [37] pointed out that the traditional CV approach is asymptotically inconsistent which is a important statistical property of model selection. When the probability of selecting the model with the best predictive ability goes to 1 as the number of observations goes to . Shao [37] provided the asymptoticl consistent property of MCCV for linear models.
In many clinical studies, dichotomous outcome is the parameter of interest, such as disease progression or not in cancer trials [27, 33, 35], and amyloid- status (either positive or negative) in Alzheimer’s disease (AD) research [2, 22, 29]. Amyloid- is considered as one of the two pathologies for diagnosis of AD [14], and has been the target in many recent AD trials using disease modified therapies (DMTs) (e.g., the BAN2401 trial [15]). One important drug is Aducanumab that is an antibody drug to remove amyloid- plaques for individuals at early stages of AD [16]. In that trial, a positive amyloid Positron Emission Tomography (PET) scan was one of the inclusion criteria. However, PET scan is very expensive and it is often not covered by insurance. Effective ML methods have the potential to save costs and screen proper participants faster.
Methods
We are going to compare MCCV with CV for a study with binary outcome (e.g., disease progression or not). We use 10-fold leave-two-out CV to build a predictive model by using 8 subsets as the testing data and the remaining 2 subsets as the testing data. There is a total of 45 CV rounds which are the number of combinations choosing 8 folds from a total of 10 folds. To perform MCCV, data are split into a training set (80%) and a testing set (the remaining 20%) without replacement in each simulation. For a fair comparison between MCCV and CV, we use the same number of rounds as that in the CV: 45 simulations in MCCV.
The following 12 supervised ML methods for binary outcome are studied: (1) linear discriminant analysis (LDA); (2) generalized linear model (GLM); (3) logistic regression (LOG); (4) naive bayes (BAY); (5) bagged classification and regression tree (CART); (6) recursive partitioning and regression trees (TREE); (7) k-nearest neighbors (KNN); (8) random forest (RF); (9) learning vector quantization (LVQ); (10, 11) support vector machines with linear kernel (SVM-L) or polynomial kernel (SVM-P); and stochastic gradient boosting (SGB). We use the statistical package caret from R to implement these ML methods [10, 31], with the detailed function values in Table 1. In the statistical package caret, an inner CV with 10-fold is performed on the training set, also known as the nested CV.
Table 1.
Method | ML model | Method value in R |
---|---|---|
LDA | Linear discriminant analysis | lda |
GLM | Generalized linear model | glm |
LOG | Boosted logistic regression | LogitBoost |
BAY | Naive Bayes | Naive_bayes |
CART | Bagged CART | Treebag |
TREE | Recursive partitioning and regression trees | rpart |
KNN | k-Nearest neighbors | knn |
RF | Random forest | rf |
LVQ | Learning vector quantization | lvq |
SVM-L | Support vector machines with linear kernel | svmLinear |
SVM-P | Support vector machines with polynomial kernel | svmPoly |
SGB | Stochastic gradient boosting | gbm |
Accuracy is one of the most common performance metrics to evaluate ML models, and it is calculated as the proportion of all samples from a testing data that are correctly predicted by using the predictive ML model built from a training data [18]. It is defined as:
where TP, FN, TN, and FP are the numbers of true positive, false negative, true negative, and false positive, respectively. It is easy to show that the total sample size is TP+TN+FP+FN. When comparing different ML methods, the one having the highest accuracy is preferable. In this article, we focus on comparing accuracy between MCCV and CV instead of identifying optimal ML methods.
Results
We first apply the aforementioned ML methods to predict amyloid- positivity using two data sets: (1) Alzheimer’s Disease Neuroimaging Initiative (ADNI), and (2) Center for Neurodegeneration and Translational Neuroscience (CNTN).
ADNI data
We first use data from the ADNI [39] to illustrate the application of the considered ML methods to predict amyloid- positivity among individuals with significant memory concern (SMC). Individuals with SMC are at an early stage of dementia, and they become one of the target population in AD clinical trials to alter the disease progression by starting intervention earlier [16]. The SMC group was enrolled during the second phase and the third phase of the ADNI [1]. The ADNI study is a longitudinal study having one of the goals to accelerate the AD drug development by discovering new biomarkers.
In this example, let the outcome be the amyloid- status, which is defined by using a threshold of 1.11 from the computed standardized uptake value ratio (SUVR). The SUVR is an average of weighted four cortical retention means divided by the whole cerebellum SUVR. Four regions are: frontal, cingulate, parietal, and temporal regions [6, 12, 13]. The SUVR value is obtained from the baseline amyloid positron emission tomography (PET) scan.
The following 11 features are used in the ML models. APOE 4 gene is the well-known genetic risk factor for patients with AD. Additional one copy of APOE 4 gene would increase the risk of developing AD by 4-fold or more [8]. Six demographic features are: age, sex, race, years of eduction, hispanic ethnicity, and marital status. The neuropsychological scores from the following four tests are also included as features: (1) Clinical Dementia Rating-Sum of Boxes (CDR-SB), (2) Mini Mental State Exam (MMSE), (3) Montreal Cognitive Assessment (MoCA), and (4) the 13-item ADAS-cog (ADAS-cog13). Among these features, many of them are continuous measures, especially the cognitive tests.
The characteristics of the SMC individuals are presented in Table 2. Participates are elderly with the mean age of 71.85, and the majority of the participants are Whites (close to 90%). We also present the Pearson correlation coefficient of each feature with the amyloid- status. The genetic risk factor of AD (APOE 4) and age are two features having significant correlations with the amyloid- status. The other features do not have strong correlations with the outcome, while the marital status and sex have week correlations with the outcome.
Table 2.
N = 149 | (p value) | |
---|---|---|
Amyloid- positivity (%) | 59 (35%) | |
APOE | 0.2670 (0.0004) | |
0 Copy | 109 (64.50%) | |
1 Copy | 54 (31.95%) | |
2 Copies | 6 (3.55%) | |
Age | 71.85 (6.11) | 0.2255 (0.0032) |
Edu | 16.83 (2.54) | 0.0136 (0.8606) |
Sex | 99 (58.58%) | (0.0756) |
Hispanic | 6 (3.55%) | (0.9347) |
Race | (0.2306) | |
Whites | 152 (89.94%) | |
African American | 10 (5.92%) | |
Other | 7 (4.14%) | |
Marry status | 0.1435 (0.0627) | |
Married | 123 (72.78%) | |
Never married | 10 (5.92%) | |
Divorced | 18 (10.65 %) | |
Widowed | 18 (10.65%) | |
ADAS-cog13 | 10.40 (4.53) | 0.01239 (0.8734) |
MoCA | 25.93 (2.65) | (0.3370) |
CDR-SB | 0.06 (0.17) | 0.02516 (0.7454) |
MMSE | 29.07 (1.17) | (0.4750) |
The last column is the Pearson correlation coefficient between each feature and the outcome (Amyloid-)
We have 169 individuals with SMC in the database. We choose to randomly pick 150 individuals for 100 times. For each selected data, we follow the aforementioned procedure to randomly split data into 10 folds in leave-two-out CV, with a total of 45 CV rounds. Similarly, we run 45 simulations in MCCV, and compare its average accuracy with the average accuracy in CV. The value of d is the mean accuracy difference between MCCV and CV (MCCV-CV), t is the test statistics from a paired t-test, and p is the pvalue from the paired t-test to assess the difference between MCCV and CV. When d is positive, it suggests that MCCV has a higher average accuracy than CV. In Fig. 1, we find that MCCV always has a higher average accuracy than CV in all the considered ML methods. In addition to accuracy, we also compare these two approaches with F1 score which takes into account of possible mis-classifications. We present the model comparison using F1 score in the model building in Fig. 2. It can be seen that MCCV has a better performance than CV with all the d values are positive for all the ML methods.
In Table 3, we present the detailed performance metrics (accuracy and F1) for each ML method. For this data set, it can be seen that MCCV has better performance than CV. The accuracy and F1 values of MCCV are larger than those using CV. The p values for normality test based on the Shapiro Wilk’s test and equal variance test based on the Levene test are provided in this table. All these p values are above 0.05. We do not have sufficient evidence to the reject the normality assumption and the equal variance assumption. The paired t-test is used to compare the accuracy difference between MCCV and CV. All ML methods show that MCCV has a statistically significant higher accuracy than CV. Although not presented here for the paired t-test for F1 in this table, the results are similar to the findings using accuracy as seen in Fig. 2.
Table 3.
Method | Normality | Levene test | Paired t-test | Accuracy | F1 | |||
---|---|---|---|---|---|---|---|---|
Test statistic | p Value | MCCV | CV | MCCV | CV | |||
LDA | 0.102 | 0.962 | 0.009 | 0.000 | 0.649 | 0.639 | 0.654 | 0.648 |
GLM | 0.798 | 0.443 | 0.009 | 0.000 | 0.651 | 0.642 | 0.658 | 0.651 |
LOG | 0.898 | 0.129 | 0.005 | 0.002 | 0.606 | 0.600 | 0.612 | 0.605 |
BAY | 0.249 | 0.457 | 0.008 | 0.000 | 0.667 | 0.659 | 0.669 | 0.665 |
CART | 0.473 | 0.754 | 0.004 | 0.054 | 0.610 | 0.606 | 0.614 | 0.612 |
TREE | 0.632 | 0.193 | 0.010 | 0.000 | 0.641 | 0.630 | 0.643 | 0.635 |
KNN | 0.129 | 0.086 | 0.008 | 0.000 | 0.621 | 0.612 | 0.625 | 0.616 |
RF | 0.733 | 0.475 | 0.011 | 0.000 | 0.635 | 0.624 | 0.639 | 0.633 |
LVQ | 0.095 | 0.828 | 0.010 | 0.000 | 0.629 | 0.619 | 0.632 | 0.625 |
SVM-L | 0.591 | 0.288 | 0.009 | 0.000 | 0.654 | 0.646 | 0.661 | 0.652 |
SVM-P | 0.475 | 0.435 | 0.011 | 0.000 | 0.621 | 0.610 | 0.626 | 0.619 |
SGB | 0.609 | 0.946 | 0.006 | 0.000 | 0.651 | 0.645 | 0.656 | 0.648 |
CNTN data
We use another data from the CNTN study [3, 25] to compare CV and MCCV to predict amyloid- positivity. In this dataset, we have 53 amyloid- positivity cases from the total of 117 participants. The following features are studied: age, race, ethnicity, education, gender, MoCA, MMSE, and CDR. For the last cognitive measures, we only use their total scores. The ML procedure settings are the same as those used in the ADNI example. MCCV is shown to have better performance than CV when accuracy is used as the performance metric in Fig. 3, and F1 score is the performance metric in Fig. 4.
For this example, the time to run all 12 models is 12.6 s using MCCV as compared to 12.1 s with CV using a personal computer. If we run MCCV for 100 time, it will be 21 min. If we run CV only one time, the time would be much shorter, which is 12.1 s. When the total time is not too long, MCCV is recommended.
We conduct the variance analysis for the performance of MCCV and CV. These two methods have similar standard deviation (SD) of their accuracy values. In the ADNI data, the SD values are 0.292 and 0.301 for MCCV and CV respectively, and 0.312 and 0.291 for MCCV and CV in the CNTN data.
Numeric study
We conduct extensive simulation studies to compare the performance between MCCV and leave-two-out CV. We simulate data from normal distributions, multinomial distributions, and binomial distributions for 20 features, and a binominal distribution for the outcome (Y). The first feature () is simulated from , a normal distribution with mean of 0 and standard deviation (sd) of 2. The other 11 features follow , where is a random value from 0.5 to 20, and . The next four features follow a multinomial distribution with the maximum possible outcome randomly chosen from 3 to 8, and the probability randomly selected from 0.1 to 0.9. The last four features follow a binomial distribution with the probability randomly selected from 0.1 to 0.9. The outcome Y is simulated from a binomial distribution with the probability as a function of the first feature: , where and . The value of captures the correlation between Y and . A high value of represents a high correlation.
We present comparisons between MCCV and leave-two-out CV with 100 simulations when sample size N is 100 in Fig. 5. In each simulation, average accuracy is calculated from 45 CV runs as described above. In each plot, there are 100 dots representing average accuracy of MCCV and that of CV from each simulation. It can be seen from the figure that MCCV generally has a higher accuracy than CV. As sample size N is increased to 200 (Fig. 6), 500 (Fig. 7), and 1200 (Fig. 8), MCCV outperforms CV when N is not very large. In the case when , their accuracies are close to each other although MCCV is slightly better than CV with regards to accuracy. We observe similar results when data are simulated with in Fig. 9 when .
When sample size is large, CV and MCCV are almost the same. When a study’s sample size is 500 or below (e.g., Fig. 7), the performance gain using MCCV over CV is still substantial. When each simulation takes very short time, both approaches could be utilized, and MCCV often has better performance than CV.
In Fig. 10, we compare MCCV and CV as the number of features is increased from 6 to 20 for a study with . In all the presented cases, MCCV has a higher accuracy than CV with the calculated d values being positive. The paired t test statistics are relatively large that leads to a small p value showing the significant higher accuracy of MCCV as compared to CV. We do not find a clear trend of accuracy as the number of features goes up.
We also investigate the number of simulations, S, needed to have stable accuracy estimates in Fig. 11 with S from 200 to 12,000 when sample size N = 100, 300, and 800. When N is small (e.g., 100), the number of simulations has to be as large as S = 5000 in order to have a consistent accuracy. For a study with a large sample size (e.g., 800), it requires fewer simulations to have a stable estimate, such as S = 2000.
Discussion
We compare the performance between MCCV and CV based on popular ML methods when outcome is binary and sample size is limited. In simulation studies, we add a correlation between the outcome and one continuous feature. That pre-specified correlation has some effects on the model prediction. However, ML models are much more complicated than traditionally used statistical models (e.g., logistic regression model) with all features being used in the final predictive model through sophisticated mathematical algorithms (e.g., tree model, SVM). We do not observe a simple relationship between accuracy and that correlation. From simulation studies, we support the findings from this article that MCCV should be recommended for use in practice with a sufficient number of simulations: when and when .
When N is small (e.g., 100), the number of simulations has to be as large as in order to have a consistent accuracy. For a study with a large sample size (e.g., 800), it requires fewer simulations to have a stable estimate, such as .
Accuracy of ML models is used as the performance metric to compare MCCV with CV. Accuracy is a widely used performance metric in classification problems with known classes of the outcome. Other performance metrics may also be considered, such as sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and the Matthews Correlation Coefficient (MCC). The MCC is equivalent to the Pearson correlation coefficient between the actual outcome and the predicted outcome [19, 23, 24, 32].
For the paired t-test to compare the performance of MCCV and CV, we check the normality assumption of the paired data by using their difference based on Shapiro Wilk’s test [30] and the D’Agostino test [4], where the D’Agostino test is a goodness-of-fit measure based on the sample skewness and kurtosis. In a few configurations, the normality assumptions are not met. In such cases, the non-parametric Wilcoxon signed-ranked test may be used to calculate p value [17]. For the ADNI example using accuracy as performance metric, the average p value is 0.491 with the range from 0.095 to 0.898 for the SW test. The D’Agostino test has the average p value of 0.496 with the range from 0.084 to 0.904. The p values from the SW test and the D’Agostino test are often close to each other. For the CNTN data, the SW test has the mean p value of 0.460 with the range from 0.014 to 0.927. The p value of the Wilcoxon test could be slightly larger than that of the t-test, but their difference is often very small.
The number of features in ML methods is an important research topic. Xu et al. [41] provided tables for the frequencies of all the possible selected features. When the most relevant features are included in the available features, the performance of MCCV and CV should be similar to their performances with all the features included in the model.
In this article, we split data into 10 folds with 8 folds as the testing data in leave-two-out CV. In the traditional leave-one-out CV approach, we only need to run 10 rounds. With the leave-two-out approach, the number of runs is increased in order to reduce the variance of the model accuracy estimates. When the ML models are not that complicated and data are not extremely unbalanced, leave-t-out approach can be performed, where . It is also true that leave-t-out CV could be performed when more computational resources are available and sample sizes are large enough.
Acknowledgements
The authors are very grateful to Editor, Associate Editor, and three reviewers for their insightful comments that help improve the manuscript significantly.
Author contributions
The idea for the paper was originally developed by GS. GS drafted the manuscript and approved the final version. The author read and approved the final manuscript.
Funding
Shan’s research is partially supported by Grants from the National Institutes of Health: P20GM109025, R03CA248006 and R01AG070849.
Availability of data and materials
Data used in preparation of this article were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu), and the Center for Neurodegeneration and Translational Neuroscience (CNTN) database (nevadacntn.org). As such, the investigators within the ADNI and the CNTN contributed to the design and implementation of ADNI and/or provided data but did not participate in analysis or writing of this report.
Declarations
Ethics approval and consent to participate
The authors confirm that all methods were performed in accordance with the relevant guidelines and regulations.
Competing interests
The authors declare that they have no competing interests.
Footnotes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Aisen PS, Petersen RC, Donohue MC, Gamst A, Raman R, Thomas RG, Walter S, Trojanowski JQ, Shaw LM, Beckett LA, Jack CR, Jagust W, Toga AW, Saykin AJ, Morris JC, Green RC, Weiner MW. Clinical core of the Alzheimer’s disease neuroimaging initiative: progress and plans. Alzheimers Dement. 2010;6(3):239–246. doi: 10.1016/j.jalz.2010.03.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Cummings J, Zhong K. Alzheimer’s disease drug develoment & emerging therapies. Pract Neurol, 2019:95–99.
- 3.Cummings JL, Fulkerson N. Neurodegeneration research: advances in clinical translational neuroscience infrastructure and methods. Alzheimer’s Dement Transl Res Clin Interv. 2018;4:326–329. doi: 10.1016/j.trci.2018.03.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.D’Agostino R, Pearson ES. Tests for departure from normality. Biometrika. 1973;60(3):613–622. [Google Scholar]
- 5.Ezzati A, Lipton RB. Machine learning predictive models can improve efficacy of clinical trials for Alzheimer’s disease. J Alzheimers Dis. 2020;74(1):55–63. doi: 10.3233/JAD-190822. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Jagust WJ, Bandy D, Chen K, Foster NL, Landau SM, Mathis CA, Price JC, Reiman EM, Skovronsky D, Koeppe RA. The Alzheimer’s disease neuroimaging initiative positron emission tomography core. Alzheimers Dement. 2010;6(3):221–229. doi: 10.1016/j.jalz.2010.03.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Khanna S, Domingo-Fernández D, Iyappan A, Emon MA, Hofmann-Apitius M, Fröhlich H. Using multi-scale genetic, neuroimaging and clinical data for predicting Alzheimer’s disease and reconstruction of relevant biological mechanisms. Sci Rep. 2018;8(1):1–13. doi: 10.1038/s41598-018-29433-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Kim J, Basak JM, Holtzman DM. The role of apolipoprotein E in Alzheimer’s disease. Neuron. 2009;63:287–303. doi: 10.1016/j.neuron.2009.06.026. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Kourou K, Exarchos TP, Exarchos KP, Karamouzis MV, Fotiadis DI. Machine learning applications in cancer prognosis and prediction. Comput Struct Biotechnol J. 2015;13:8–17. doi: 10.1016/j.csbj.2014.11.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Kuhn M. Building predictive models in R using the caret package. J Stat Softw. 2008;28(5):1–26. doi: 10.18637/jss.v028.i05. [DOI] [Google Scholar]
- 11.Lahmiri S, Boukadoum M. New approach for automatic classification of Alzheimer’s disease, mild cognitive impairment and healthy brain magnetic resonance images. Healthc Technol Lett. 2014;1(1):32–36. doi: 10.1049/htl.2013.0022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Landau SM, Breault C, Joshi AD, Pontecorvo M, Mathis CA, Jagust WJ, Mintun MA. Amyloid- imaging with Pittsburgh compound B and florbetapir: comparing radiotracers and quantification methods. J Nucl Med. 2013;54(1):70–77. doi: 10.2967/jnumed.112.109009. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Landau SM, Fero A, Baker SL, Koeppe R, Mintun M, Chen K, Reiman EM, Jagust WJ. Measurement of longitudinal -amyloid change with 18F-florbetapir PET and standardized uptake value ratios. J Nucl Med. 2015;56(4):567–574. doi: 10.2967/jnumed.114.148981. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Paul Murphy M. Amyloid-beta solubility in the treatment of Alzheimer’s disease. N Engl J Med. 2018;378(4):391–392. doi: 10.1056/NEJMe1714638. [DOI] [PubMed] [Google Scholar]
- 15.Panza F, Lozupone M, Logroscino G, Imbimbo BP. A critical appraisal of amyloid--targeting therapies for Alzheimer disease. Nat Rev Neurol. 2019;15(2):73–88. doi: 10.1038/s41582-018-0116-6. [DOI] [PubMed] [Google Scholar]
- 16.Sevigny J, Chiao P, Bussière T, Weinreb PH, Williams L, Maier M, Dunstan R, Salloway S, Chen T, Ling Y, O’Gorman J, Qian F, Arastu M, Li M, Chollate S, Brennan MS, Quintero-Monzon O, Scannevin RH, Moore Arnold H, Engber T, Rhodes K, Ferrero J, Hang Y, Mikulskis A, Grimm J, Hock C, Nitsch RM, Sandrock A. The antibody aducanumab reduces A plaques in Alzheimer’s disease. Nature. 2016;537(7618):50–56. doi: 10.1038/nature19323. [DOI] [PubMed] [Google Scholar]
- 17.Shan G. New nonparametric rank-based tests for paired data. Open J Stat. 2014;04(07):495–503. doi: 10.4236/ojs.2014.47047. [DOI] [Google Scholar]
- 18.Shan G. Exact statistical inference for categorical data. 1. San Diego: Academic Press; 2015. [Google Scholar]
- 19.Shan G. Improved confidence intervals for the Youden index. PLoS ONE. 2015;10(7):e0127272. doi: 10.1371/journal.pone.0127272. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Shan G. Comments on ‘Two-sample binary phase 2 trials with low type I error and low sample size’. Stat Med. 2017;36(21):3437–3438. doi: 10.1002/sim.7359. [DOI] [PubMed] [Google Scholar]
- 21.Shan G. Accurate confidence intervals for proportion in studies with clustered binary outcome. Stat Methods Med Res. 2020;29(10):3006–3018. doi: 10.1177/0962280220913971. [DOI] [PubMed] [Google Scholar]
- 22.Shan G. Optimal two-stage designs based on restricted mean survival time for a single-arm study. Contemp Clin Trials Commun. 2021;21:100732. doi: 10.1016/j.conctc.2021.100732. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Shan G. Conservative confidence intervals for the intraclass correlation coefficient for clustered binary data. J Appl Stat. 2022;49(10):2535–2549. doi: 10.1080/02664763.2021.1910939. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Shan G. Randomized two-stage optimal design for interval-censored data. J Biopharm Stat. 2022;32(2):298–307. doi: 10.1080/10543406.2021.2009499. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Shan G, Banks S, Miller JB, Ritter A, Bernick C, Lombardo J, Cummings JL. Statistical advances in clinical trials and clinical research. Alzheimer’s Dement Transl Res Clin Interv. 2018;4:366–371. doi: 10.1016/j.trci.2018.04.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Shan G, Bernick C, Caldwell JZK, Ritter A. Machine learning methods to predict amyloid positivity using domain scores from cognitive tests. Sci Rep. 2021;11(1):4822. doi: 10.1038/s41598-021-83911-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Shan G, Dodge-Francis C, Wilding GE. Exact unconditional tests for dichotomous data when comparing multiple treatments with a single control. Ther Innov Regul Sci. 2020;54(2):411–417. doi: 10.1007/s43441-019-00070-w. [DOI] [PubMed] [Google Scholar]
- 28.Shan G, Ma C, Hutson AD, Wilding GE. An efficient and exact approach for detecting trends with binary endpoints. Stat Med. 2012;31(2):155–164. doi: 10.1002/sim.4411. [DOI] [PubMed] [Google Scholar]
- 29.Shan G, Ritter A, Miller J, Bernick C. Effects of dose change on the success of clinical trials. Contemp Clin Trials Commun. 2022;30:100988. doi: 10.1016/j.conctc.2022.100988. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Shan G, Vexler A, Wilding GE, Hutson AD. Simple and exact empirical likelihood ratio tests for normality based on moment relations. Commun Stat Simul Comput. 2011;40(1):129–146. doi: 10.1080/03610918.2010.532896. [DOI] [Google Scholar]
- 31.Shan G, Wang W. ExactCIdiff: an R package for computing exact confidence intervals for the difference of two proportions. R J. 2013;5(2):62–71. doi: 10.32614/RJ-2013-026. [DOI] [Google Scholar]
- 32.Shan G, Wilding G. Unconditional tests for association in 2 * 2 contingency tables in the total sum fixed design. Stat Neerl. 2015;69(1):67–83. doi: 10.1111/stan.12047. [DOI] [Google Scholar]
- 33.Shan G, Wilding GE, Hutson AD, Gerstenberger S. Optimal adaptive two-stage designs for early phase II clinical trials. Stat Med. 2016;35(8):1257–1266. doi: 10.1002/sim.6794. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Shan G, Zhang H. Two-stage optimal designs with survival endpoint when the follow-up time is restricted. BMC Med Res Methodol. 2019;19:74. doi: 10.1186/s12874-019-0696-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Shan G, Zhang H, Jiang T. Adaptive two-stage optimal designs for phase II clinical studies that allow early futility stopping. Seq Anal. 2019;38(2):199–213. doi: 10.1080/07474946.2019.1611307. [DOI] [Google Scholar]
- 36.Shan G, Zhang H, Jiang T, Peterson H, Young D, Ma C. Exact p-values for Simon’s two-stage designs in clinical trials. Stat Biosci. 2016;8(2):351–357. doi: 10.1007/s12561-016-9152-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Shao J. Linear model selection by cross-validation. J Am Stat Assoc. 1993;88:486–494. doi: 10.1080/01621459.1993.10476299. [DOI] [Google Scholar]
- 38.Uddin S, Khan A, Hossain ME, Moni MA. Comparing different supervised machine learning algorithms for disease prediction. BMC Med Inform Decis Mak. 2019;19(1):281. doi: 10.1186/s12911-019-1004-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Weiner MW, Veitch DP, Aisen PS, Beckett LA, Cairns NJ, Cedarbaum J, Donohue MC, Green RC, Harvey D, Jack CR, Jagust W, Morris JC, Petersen RC, Saykin AJ, Shaw L, Thompson PM, Toga AW, Trojanowski JQ. Impact of the Alzheimer’s disease neuroimaging initiative, 2004 to 2014. Alzheimers Dement. 2015;11(7):865–884. doi: 10.1016/j.jalz.2015.04.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Xu QS, Liang YZ. Monte Carlo cross validation. Chemom Intell Lab Syst. 2001;56(1):1–11. doi: 10.1016/S0169-7439(00)00122-2. [DOI] [Google Scholar]
- 41.Xu QS, Liang YZ, Du YP. Monte Carlo cross-validation for selecting a model and estimating the prediction error in multivariate calibration. J Chemom. 2004;18(2):112–120. doi: 10.1002/cem.858. [DOI] [Google Scholar]
- 42.Zhang H, Song Y, Jiang B, Chen B, Shan G. Two-stage bagging pruning for reducing the ensemble size and improving the classification performance. Math Probl Eng 2019:8906034.
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Data used in preparation of this article were obtained from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database (adni.loni.usc.edu), and the Center for Neurodegeneration and Translational Neuroscience (CNTN) database (nevadacntn.org). As such, the investigators within the ADNI and the CNTN contributed to the design and implementation of ADNI and/or provided data but did not participate in analysis or writing of this report.