Abstract
Aims
We developed new machine learning (ML) models and externally validated existing statistical models [ischaemic stroke predictive risk score (iScore) and totalled health risks in vascular events (THRIVE) scores] for predicting the composite of recurrent stroke or all-cause mortality at 90 days and at 3 years after hospitalization for first acute ischaemic stroke (AIS).
Methods and results
In adults hospitalized with AIS from January 2005 to November 2016, with follow-up until November 2019, we developed three ML models [random forest (RF), support vector machine (SVM), and extreme gradient boosting (XGBOOST)] and externally validated the iScore and THRIVE scores for predicting the composite outcomes after AIS hospitalization, using data from 721 patients and 90 potential predictor variables. At 90 days and 3 years, 11 and 34% of patients, respectively, reached the composite outcome. For the 90-day prediction, the area under the receiver operating characteristic curve (AUC) was 0.779 for RF, 0.771 for SVM, 0.772 for XGBOOST, 0.720 for iScore, and 0.664 for THRIVE. For 3-year prediction, the AUC was 0.743 for RF, 0.777 for SVM, 0.773 for XGBOOST, 0.710 for iScore, and 0.675 for THRIVE.
Conclusion
The study provided three ML-based predictive models that achieved good discrimination and clinical usefulness in outcome prediction after AIS and broadened the application of the iScore and THRIVE scoring system for long-term outcome prediction. Our findings warrant comparative analyses of ML and existing statistical method–based risk prediction tools for outcome prediction after AIS in new data sets.
Keywords: Stroke, Mortality, Prediction models, Machine-based learning, Statistical
Graphical Abstract
Graphical Abstract.
Introduction
In the USA, stroke is one of the leading causes of hospitalization1 with high rates of stroke recurrence (15–18%) and mortality (29–48%) after 3 years, despite considerable advances in drug and device interventions.2–4 Accurate risk stratification of inpatients with an acute ischaemic stroke (AIS) is crucial for guiding targeted interventions and clinical decision-making.5 Evidence suggests that even stroke experts achieve <50% accuracy in predicting the clinical outcome after incident stroke.6 In previous studies largely based on conventional likelihood–based statistical (CS) methods, the investigators have reported individual predictors of recurrent stroke or mortality following initial stroke.7
Numerous risk prediction models, largely based on CS methods, have been developed for predicting short-term clinical outcomes to improve inpatient care and offer limited clinical utility.8–12 Acute ischaemic stroke is a complex and heterogenous condition with a wide range of variables that potentially determine the eventual outcome, especially in the long term. Currently available risk prediction models for AIS use a limited number of candidate predictors to compute the risk of outcome.8–12 These risk prediction models fail to account for a large number of variables13 and, therefore, performed poorly for outcome prediction when compared with machine learning (ML) models that were trained on a larger number of variables.14 However, artificial intelligence–based ML models provide unique analytic opportunities to integrate large-scale information and condense and select the most relevant predictors from a broader high-dimensional data set for predicting outcome but have been under-investigated in AIS.15,16 Currently, none of the prediction models for AIS outcome have demonstrated usefulness in informing clinical decision-making using statistical methods such as decision curve analysis (DCA).17 Therefore, novel risk prediction models with robust methodology that account for diverse predictor variables with improved performance relative to existing models are needed for the risk stratification of patients with AIS. Furthermore, the present ML models and statistical method–based scoring systems for outcome prediction after hospitalization for AIS have been limited by short-term follow-up and uncertain clinical applicability. Their utility in predicting recurrent stroke and mortality beyond 1 year is unknown.
Accordingly, we aimed, primarily, to develop, calibrate, and validate novel ML models, trained on a broader array of predictor variables than previously reported,18 to predict the composite of recurrent stroke or all-cause mortality at 90 days or 3 years in AIS survivors after hospital discharge. Our secondary aims were to externally validate the ischaemic stroke predictive risk score (iScore)19 and totalled health risks in vascular events (THRIVE) scores,20 two well-known and extensively validated prognostic models as exemplars for identifying patients at high risk for composite outcomes after AIS, and to assess the performance of each ML model and risk score for clinical utility across a range of thresholds.
Methods
Study design and population
In this retrospective study, we screened 1024 patients for eligibility and identified a final cohort of 721 consecutive patients who survived their first-ever AIS after index hospitalization at the Mayo Clinic, Rochester, MN, from 1 January 2005 to 30 November 2016. The International Classification of Disease Clinical Modification Tenth Revision (ICD-10-CM) codes were used to identify primary discharge diagnoses (see Supplementary material online, Panel S1). The patients were followed up for recurrent stroke or death, whichever occurred first, until 30 November 2019; if neither occurred, the patients were censored on that day. The following patients were excluded: 45 patients (4%) with non-stroke conditions, 99 patients (10%) with previous strokes not categorized as AIS, and 159 patients (16%) aged ≥ 90 years (the Institutional Review Board restricts data sharing to patients aged below 90 years). We followed the transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD) guidelines21 for study design, conduct, and reporting. We also adhered to the guidelines for strengthening the reporting of observational studies in epidemiology (STROBE). The STROBE flow diagram for selection of steady heart is provided in Supplementary material online, Figure S1. Data were extracted and deidentified prior to the analysis. The details of the data extraction process have been published elsewhere.22 The Mayo Clinic Institutional Review Board approved the study and waived the requirement for informed consent. The deidentified data that support the results of the present study are available upon specific request from the corresponding author.
Baseline characteristics
Stroke was diagnosed through a physical examination supported by brain and vascular imaging. The first-ever stroke was defined as a stroke occurring in a patient who had not experienced any prior stroke within a minimal look-back period of 9 years. For each patient, the severity of AIS on admission was assessed using the National Institutes of Health Stroke Scale (NIHSS) score.23 The NIHSS score ranges from 0 to 42, with higher scores indicating greater severity. Acute ischaemic stroke subtypes were classified into five categories24: (i) large artery atherosclerosis; (ii) cardioembolism; (iii) small vessel occlusion; (iv) other determined aetiology; and (v) undetermined aetiology. On the day of dismissal, basic activities of daily living, including bathing, dressing, grooming, toileting, hygienic measures, mobility, and eating, were assessed, and individually awarded scores of 1 (independent), 2 (requiring assistance), and 0 (dependent). Potential predictors were selected from multiple domains for model training based on a review of the literature, previous work, and expert opinion.22,25–28 Potential predictors were selected from multiple domains (see Supplementary material online, Table S1). The STROBE flow diagram shows that 14 patients (1.4%) had one or more missing data points. The missing values in the development and validation sets were imputed to prevent data leakage using the missForest R package based on the random forest (RF) method.29 Body mass index (BMI) was calculated by dividing weight in kilograms by the square of the height in metres. Categorical variables were encoded in numerically to facilitate the input data in the development data set.
Outcomes
We chose composite events as the outcome measure for the following reasons: We reviewed published literature and found considerable debate over the use of individual vs. combined clinical endpoints.30,31 Recent improvements in the treatment of AIS have led to a decline in the incidence of relevant outcomes that can be used as clinical endpoints, especially when challenged with low event rates.32 It is generally conceived that the higher the gradient in importance and frequency across the components of the combined outcome, the greater the inherent challenges in interpreting the results.33 Previous studies have clearly demonstrated that the risk of death is substantially greater after a recurrent stroke than after the first stroke,34 despite a decrease in the risk of stroke recurrence.35 These findings may argue for using aggregated recurrent stroke and mortality as the primary endpoint in stroke research.33 Combining recurrent stroke and mortality allowed for an evaluation of the overall impact and consequences of the event or intervention on patient health. Therefore, we choose a composite of recurrent stroke or death from any cause (both outcomes are of equal importance to the patient) within 90 days or 3 years in patients who survived to discharge after hospitalization for first-ever AIS as the outcome.
Machine learning–based approach
We followed four crucial steps to identify the best-performing risk prediction algorithm: (i) development and calibration of ML-based risk prediction models on a randomly generated development (n = 505; 70%) data set; (ii) external validation of the iScore and THRIVE scoring systems for the cumulative cohort (n = 721); and (iii) employment of DCA to assess the net benefit across wide-ranging thresholds for each model.
We built three independent ML-based models after assessing different ML methods. These models were trained in a development data set to predict binary composite outcomes. The models were based on RF,36 support vector machine (SVM),37 and extreme gradient boosting (XGBOOST)36 algorithms. The ML models were carefully selected according to their performance in predicting AIS outcomes in prior modelling studies.38,39 The relative advantages and disadvantages of the ML models are listed in Supplementary material online, Table S2. All ML models were constructed using 90 variables (see Supplementary material online, Table S1) and trained using 10-fold cross-validation, which was repeated 10 times to improve the prediction of composite outcomes. We constructed separate prediction models in the development data set using the top 10 variables that contributed the most to model prediction.
Further details on the model development, hyperparameter evaluation, and class imbalance are provided in the Supplementary material online, Panel S2. A schematic representation of the development process of ML algorithms is shown in Supplementary material online, Figure S2.
Ascertainment of ischaemic stroke predictive risk score and totalled health risks in vascular events scores
External validation refers to the evaluation of a prediction model’s performance outside the original model development setting to determine its reproducibility and generalizability.40 The Supplementary material online, Table S3 describes and Supplementary material online, Table S4 scores the components of the iScore and THRIVE scoring systems.
The iScore is a point-based prognostic score that was originally developed using a conventional statistical approach to predict the mortality rates within 30 days and 1 year following AIS in patients hospitalized from 2003 to 2008.19 We used the 30-day and 1-year iScore versions to validate the 90-day and 3-year composite outcomes in our study population. The score was calculated for each patient using a risk score algorithm. The score was generated based on age, sex, stroke severity, risk factors (atrial fibrillation, heart failure, coronary artery disease, and current smoking status), comorbid conditions (cancer and renal dialysis), and blood glucose levels on admission. Because of data limitations, comorbid diabetes mellitus was substituted for an admission blood glucose level of >130 mg/dL. The NIHSS score was converted to the Canadian Neurological Society score to validate the iScore in our study population.
The THRIVE score is also a point-based risk prediction tool derived from logistic regression.20 The THRIVE scoring system was initially developed to determine outcomes after endovascular treatment for AIS and was subsequently externally validated in patients with general stroke.
Both iScore and THRIVE models were evaluated using the area under the receiver operating characteristic curve (AUC), calibrated using the estimated slope and intercept. The goodness of fit was assessed using the Hosmer–Lemeshow test. The Youden index was calculated based on sensitivity and specificity. The maximum Youden index value was identified, and the corresponding sensitivity and specificity values were reported. Additionally, the positive predictive value (PPV) and negative predictive value (NPV) were calculated using the prevalence of adverse outcomes in the study sample. Additionally, the iScore and THRIVE scores were directly compared to predict the 3-year composite outcome.
Statistical analysis
Analyses were conducted using RStudio (R 4.0.516) for Windows, Stata (SE version 17.0; Stata Corp LLC, College Station, TX) and SAS statistical programs. We report the mean and standard deviation (SD) for normally distributed variables, the median and interquartile range (IQR) for non-normally distributed variables, and the number and proportion of categorical variables. Normally distributed data, non-normally distributed data, and categorical variables were compared using Student’s t-test, the Kruskal–Wallis test, and the Pearson χ2 test, respectively.
Model performance metrics
The performance metrics of all ML models for binary outcomes were measured as AUC for discriminative ability, Brier score for overall model performance, goodness of fit for calibration, and DCA for net benefit decision-making using a recently recommended framework.41 The Brier score is very familiar to the research community for evaluating and comparing ML models.42 It is an accepted metric for evaluating probabilistic prediction, particularly in the context of clinical prediction models. The Brier score effectively assesses the calibration and accuracy of predicted probabilities, which is crucial in various medical applications. Furthermore, the Brier score facilitates direct model comparisons by quantifying the average scored difference between predicted probabilities and observed outcomes. We aimed to optimize hyperparameters based on the Brier score to enhance model’s abilities to provide an accurate and well-calibrated probability estimate. A lower Brier score and higher AUC represent better performance of ML model.
The discriminatory performance of each model was assessed using receiver operating characteristic (ROC) curve analysis in both the development and internal validation cohorts with all 90 variables and the top 10 most predictive variables. Differences in the AUC across the models were assessed using the DeLong test. Calibration plots were constructed for each model in both the development and validation cohorts with all 90 variables and the top 10 most predictive variables in each data set. The significance of the calibration plots was assessed using the Brier score [which ranges from 0 (best) to 1 (worst)]43 and Hosmer–Lemeshow44 goodness-of-fit test statistics (well calibrated if P > 0.05). Separate ROC curves, calibration plots, and decision analytic curves were generated in the development and validation data sets for the ML-based models and in the cumulative data set for the iScore and THRIVE scores.
The DCA is an approach to quantifying the utility of a prediction model for clinical decision-making.17 Decision curve analysis measures a ‘net benefit’ as a function of default strategies of treating all or treating none across a range of threshold probabilities.45 The net benefit complements the accuracy metrics of the model by incorporating the consequences of decisions made based on the prediction model. We assessed the net benefit of clinical utility across a range of thresholds for each model and the existing scoring system to ascertain whether the model should be used for clinical decision-making. In our analysis, we utilized the dcurves package in R to conduct the DCA. The net benefit derived from the DCA for all three ML models was computed using the 30% validation cohort. In contrast, the net benefit for the iScore and THRIVE was computed for the cumulative cohort.
First, we compared the results of the RF, SVM, and XGBOOST models in the development and internal validation data sets using the AUC, accuracy, Brier score, and Hosmer–Lemeshow test. Second, we compared the performance metrics for outcome prediction between study models to those derived from the validation and recalibration of the iScore and THRIVE scores in our study cohort.
We acknowledge that class imbalance is an important consideration in ML. However, in our study, we chose not to specifically address class imbalance because we wanted to evaluate the performance of our models in a real-world scenario where class imbalance is often present and not always addressed. By not explicitly handling the class imbalance, we aimed to have the models learn from the natural data distribution since it would be encountered in practice. This approach allows us to assess the models’ performance under real-world conditions and provides insights into their generalizability.
Selection of the top 10 variables for the construction of parsimonious machine learning models
We selected the top 10 variables for the construction of parsimonious models for predicting outcomes from the following clinical and modelling perspectives: Stroke is a complex clinical syndrome with variability in outcome. Multivariable risk-adjusted models were superior to expert opinion for predicting stroke outcome.6 Variables in multiple domains improve the performance of a model derived from six clinical variables.46 The NIHSS score, which is the current gold standard for the assessment of disease severity, utilizes 13 features.23 In this study, we used features such as measures of basic activities of daily living, which are very important in the assessment of stroke recovery. Five variables may not be able to cover all the important domains of a condition as complex as AIS, whereas >10 variables become burdensome for bedside use; therefore, we choose to use 10 features for parsimonious model. From a modelling perspective, we chose to focus on the top 10 features to strike a balance between model complexity and interpretability. While other choices, such as selecting 5 or 15 features, could have been made, the selection of 10 features allowed us to capture a meaningful subset of highly important variables without overwhelming the model or sacrificing its interpretability.
Complementary analysis
The THRIVE score was originally developed for predicting 90-day mortality in patients with AIS receiving thrombolytic therapy; subsequently, it exhibited broader utility and validation in other stroke types and for predicting in-hospital or 1-year mortality.47–49 Similarly, the iScore was initially developed to predict 3-month and 1-year mortality after AIS and subsequently validated at different time points.50 Although our primary objective was to assess whether the iScore or THRIVE scoring system can predict outcomes at 90 days and beyond 1 year (3 years in the present study) after index AIS, we constructed separate ROC curves to predict composite outcomes and mortality within 1 year of AIS to assess potential differences in discriminatory function. Additionally, we (i) validated and compared the iScore and THRIVE scores in a sensitivity analysis in a restricted 30% validation cohort for fairer evaluation against ML models and (ii) compared iScore and THRIVE scores for predicting a 90-day composite outcome.
The STROBE checklist is provided in Supplementary material online, Table S5.
Results
Study population
The median age in the study population was 75 years (IQR: 65–83 years); 49% of the patients were men, and 8% were non-white. During the follow-up, 80 (11%) and 246 (34%) patients experienced 90-day and 3-year outcomes, respectively. Baseline patient characteristics according to the study outcomes are presented in Table 1.
Table 1.
Baseline characteristics of the study population by outcome
| Variables | Cumulative cohort, n = 721 | 90-day composite of recurrent stroke and mortality | 3-year composite of recurrent stroke or mortality | ||||
|---|---|---|---|---|---|---|---|
| No (n = 641) | Yes (n = 80) | P value | No (n = 482) | Yes (n = 239) | P value | ||
| Age, median (IQR) | 75.3 (65.0, 82.5) | 74.8 (63.5, 82.2) | 80.5 (74.4, 85.4) | <0.001 | 73.0 (61.6, 80.3) | 81.0 (72.9, 85.3) | <0.001 |
| Sex at birth, female | 355 (49.2%) | 328 (51.2%) | 27 (33.8%) | 0.003 | 249 (51.7%) | 106 (44.4%) | 0.065 |
| Race, non-white | 55 (7.6%) | 47 (7.3%) | 8 (10.0%) | 0.397 | 35 (7.3%) | 20 (8.4%) | 0.598 |
| BMI, median (IQR) | 27 (24, 37) | 27 (24, 32) | 25.9 (22, 30) | 0.061 | 28 (24, 32) | 26 (23, 30) | <0.001 |
| Current smoker | 130 (18.0%) | 124 (19.3%) | 6 (7.5%) | 0.009 | 94 (19.5%) | 36 (15.1%) | 0.144 |
| LOS, median (IQR) | 3.0 (2.0, 6.0) | 3.0 (2.0, 5.0) | 5.0 (2.8, 8.0) | <0.001 | 3.0 (2.0, 5.0) | 4.0 (3.0, 7.0) | <0.001 |
| Hyperlipidaemia | 464 (64.4%) | 424 (66.1%) | 40 (50.0%) | 0.004 | 336 (69.7%) | 128 (53.6%) | <0.001 |
| Hypertension | 580 (80.4%) | 517 (80.7%) | 63 (78.8%) | 0.685 | 379 (78.6%) | 201 (84.1%) | 0.081 |
| Depression | 138 (19.1%) | 118 (18.4%) | 20 (25.0%) | 0.158 | 77 (16.0%) | 61 (25.5%) | 0.002 |
| Diabetes mellitus | 196 (27.2%) | 179 (27.9%) | 17 (21.2%) | 0.206 | 128 (26.6%) | 68 (28.5%) | 0.59 |
| Arthritis | 190 (26.4%) | 168 (26.2%) | 22 (27.5%) | 0.805 | 118 (24.5%) | 72 (30.1%) | 0.105 |
| Cancer | 47 (6.5%) | 37 (5.8%) | 10 (12.5%) | 0.022 | 14 (2.9%) | 33 (13.8%) | <0.001 |
| Atrial fibrillation | 212 (29.4%) | 180 (28.1%) | 32 (40.0%) | 0.027 | 123(25.5%) | 89 (37.2%) | 0.001 |
| Asthma | 34 (4.7%) | 31 (4.8%) | 3 (3.8%) | 0.666 | 25 (5.2%) | 9 (3.8%) | 0.397 |
| CAD | 205 (28.4%) | 177 (27.6%) | 28 (35.0%) | 0.167 | 121 (25.1%) | 84 (35.1%) | 0.005 |
| PAD | 59 (8.2%) | 51 (8.0%) | 8 (10.0%) | 0.529 | 27 (5.6%) | 32 (13.4%) | <0.001 |
| Substance use | 49 (6.8%) | 43 (6.7%) | 6 (7.5%) | 0.791 | 33 (6.8%) | 16 (6.7%) | 0.939 |
| COPD | 64 (8.9%) | 53 (8.3%) | 11 (13.8%) | 0.104 | 42 (8.7%) | 22 (9.2%) | 0.827 |
| Osteoporosis | 74 (10.3%) | 60 (9.4%) | 14 (17.5%) | 0.024 | 36 (7.5%) | 38 (15.9%) | <0.001 |
| CKD | 87 (12.1%) | 70 (10.9%) | 17 (21.2%) | 0.007 | 44 (9.1%) | 43 (18.0%) | <0.001 |
| Heart failure | 80 (11.1%) | 62 (9.7%) | 18 (22.5%) | <0.001 | 33 (6.8%) | 47 (19.7%) | <0.001 |
| Valve disease | 53 (7.4%) | 45 (7.0%) | 8 (10.0%) | 0.336 | 34 (7.1%) | 19 (7.9%) | 0.664 |
| Dementia | 69 (9.6%) | 50 (7.8%) | 19 (23.8%) | <0.001 | 25 (5.2%) | 44 (18.4%) | <0.001 |
| Seizure | 6 (0.8%) | 5 (0.8%) | 1 (1.2%) | 0.663 | 4 (0.8%) | 2 (0.8%) | 0.992 |
| Anaemia | 76 (10.5%) | 64 (10.0%) | 12 (15.0%) | 0.168 | 35 (7.3%) | 41 (17.2%) | <0.001 |
| Migraine | 26 (3.6%) | 24 (3.8%) | 2 (2.5%) | 0.572 | 20 (4.2%) | 6 (2.5%) | 0.264 |
| OSA | 49 (6.8%) | 44 (6.9%) | 5 (6.2%) | 0.837 | 34 (7.1%) | 15 (6.3%) | 0.696 |
| Parkinson’s disease | 11 (1.5%) | 8 (1.2%) | 3 (3.8%) | 0.085 | 3 (0.6%) | 8 (3.3%) | 0.005 |
| Liver disease | 9 (1.2%) | 8 (1.2%) | 1 (1.2%) | 0.999 | 6 (1.2%) | 3 (1.3%) | 0.991 |
| Hypothyroidism | 63 (8.7%) | 53 (8.3%) | 10 (12.5%) | 0.206 | 40 (8.3%) | 23 (9.6%) | 0.553 |
| Other psychiatric conditions | 1 (0.1%) | 0 (0.0%) | 1 (1.2%) | 0.005 | 0 (0.0%) | 1 (0.4%) | 0.155 |
| Unilateral symptoms | 684 (94.9%) | 608 (94.9%) | 76 (95.0%) | 0.955 | 458 (95.0%) | 226 (94.6%) | 0.792 |
| MCA | 489 (67.8%) | 432 (67.4%) | 57 (71.2%) | 0.486 | 311 (64.5%) | 178 (74.5%) | 0.007 |
| Anterior circulation | 536 (74.3%) | 475 (74.15) | 61 (76.2%) | 0.678 | 341 (70.7%) | 195 (81.6%) | 0.002 |
| Aetiology of stroke | <0.001 | <0.001 | |||||
| LAA | 83 (11.5%) | 75 (11.7%) | 8 (10.0%) | 56 (11.6%) | 27 (11.3%) | ||
| Cardioembolism | 216 (30.0%) | 182 (28.4%) | 34 (42.5%) | 127 (26.3%) | 89 (37.2%) | ||
| SVO | 158 | 154 | 4 (5.0%) | 127 (26.3%) | 31 (13.0%) | ||
| Undetermined | 226 (31.3%) | 194 (30.3%) | 32 (40.0%) | 143 (29.7%) | 83 (34.7%) | ||
| Determined | 38 (5.3%) | 36 (5.6%) | 2 (2.5%) | 29 (6.0%) | 9 (3.8%) | ||
| NIHSS score, median (IQR) | 4.0 (2.0,8.0) | 4.0 (2.0,7.0) | 8.0 (4.0,16.0) | <0.001 | 4.0 (2.0,6.0) | 6.0 (3.0,10.5) | <0.001 |
| SBP, median (IQR) | 148 (134, 166) | 149 (135, 165) | 140 (132, 170) | 0.286 | 149 (135, 166) | 146 (134, 164) | 0.203 |
| DBP, median (IQR) | 77.3 (69, 86) | 77 (70, 86) | 76.0 (66, 86) | 0.261 | 77 (7 087) | 76 (67, 85) | 0.219 |
| Thrombolytics | 71 (9.8%) | 63 (9.8%) | 8 (10.0%) | 0.961 | 45 (9.3%) | 26 (10.9%) | 0.513 |
| Embolectomy | 3 (0.4%) | 3 (0.5%) | 0 (0.0%) | 0.54 | 3 (0.6%) | 0 (0.0%) | 0.222 |
| Aspirin | 487 (67.5%) | 449(70%) | 38 (47.5%) | <0.001 | 340(70.5%) | 147 (61.5%) | 0.015 |
| Clopidogrel | 132(18.3%) | 119(18.6%) | 13 (16.2%) | 0.614 | 95 (19.7%) | 37 (15.5%) | 0.167 |
| Aggrenox | 49 (6.8%) | 48 (7.5%) | 1 (1.2%) | 0.037 | 37 (7.7%) | 12 (5.0%) | 0.182 |
| Warfarin | 218 (30.2%) | 199 (31%) | 19 (23.8%) | 0.18 | 149 (30.9%) | 69 (28.9%) | 0.574 |
| DOACs | 16 (2.2%) | 15 (2.3%) | 1 (1.2%) | 0.533 | 6 (1.2%) | 10 (4.2%) | 0.012 |
| Statin | 473 (65.6%) | 438 (68.3%) | 35 (43.8%) | <0.001 | 338(70.1%) | 135 (56.5%) | <0.001 |
| High-intensity statin | 21 (2.9%) | 20 (3.1%) | 1 (1.2%) | 0.348 | 15 (3.1%) | 6 (2.5%) | 0.651 |
| ACEI/ARBs | 301(41.7%) | 274 (42.7%) | 27 (33.8%) | 0.124 | 200 (41.5%) | 101 (42.3%) | 0.844 |
| Beta-blockers | 357 (49.5%) | 327 (51%) | 30 (37.5%) | 0.023 | 227 (47.1%) | 130 (54.4%) | 0.065 |
| Calcium channel blocker | 127(17.6%) | 105(16.4%) | 22 (27.5%) | 0.014 | 73 (15.1%) | 54 (22.6%) | 0.013 |
| HCTZ | 226 (31.3%) | 204 (31.8%) | 22 (27.5%) | 0.432 | 155 (32.2%) | 71 (29.7%) | 0.504 |
| Oral diabetic drugs | 83 (11.5%) | 82 (12.8%) | 1 (1.2%) | 0.002 | 66 (13.7%) | 17 (7.1%) | 0.009 |
| Insulin | 82 (11.4%) | 74 (11.5%) | 8 (10.0%) | 0.682 | 46 (9.5%) | 36 (15.1%) | 0.028 |
| SSRI | 97 (13.5%) | 88 (13.7%) | 9 (11.2%) | 0.54 | 57 (11.8%) | 40 (16.7%) | 0.069 |
| Antiepileptics | 24 (3.3%) | 23 (3.6%) | 1 (1.3%) | 0.29 | 17 (3.5%) | 7 (3.0%) | 0.702 |
| Antiplatelets | 579 (80.3%) | 534 (83.3%) | 45 (56.2%) | <0.001 | 403 (83.6%) | 176 (73.6%) | 0.002 |
| All anticoagulant | 234 (32.5%) | 214 (33.4%) | 20 (25.0%) | 0.131 | 155 (32.2%) | 79 (33.1%) | 0.809 |
| Antiplatelet or anticoagulant | 684 (94.9%) | 629 (98.1%) | 55 (68.8%) | <0.001 | 473 (98.1%) | 211 (88.3%) | <0.001 |
| Any antihypertensive | 547 (75.9%) | 497 (77.5%) | 50 (62.5%) | 0.003 | 363 (75.3%) | 184 (77%) | 0.62 |
| Hemiparesis | 325 (45.1%) | 270 (42.1%) | 55 (68.8%) | <0.001 | 188 (39%) | 137 (57.3%) | <0.001 |
| hemisensory defect | 76 (10.5%) | 63 (9.8%) | 13 (16.2%) | 0.078 | 43 (8.9%) | 33 (13.8%) | 0.044 |
| Hemineglect | 42 (5.8%) | 36 (5.6%) | 6 (7.5%) | 0.498 | 25 (5.2%) | 17 (7.1%) | 0.299 |
| Impaired speech | 327 (45.4%) | 285 (44.5%) | 42 (52.5%) | 0.173 | 205 (42.5%) | 122 (51%) | 0.031 |
| Hemianopia | 82 (11.4%) | 70 (10.9%) | 12 (15.0%) | 0.279 | 53 (11.0%) | 29 (12.1%) | 0.65 |
| Gaze palsy | 64 (8.9%) | 51 (8.0%) | 13 (16.2%) | 0.014 | 33 (6.8%) | 31 (13.0%) | 0.006 |
| Facial palsy | 322 (44.7%) | 281 (43.8%) | 41 (51.2%) | 0.209 | 206 (42.7%) | 116 (48.5%) | 0.141 |
| Arm weakness | 418 (58%) | 355 (55.4%) | 63 (78.8%) | <0.001 | 252 (52.3%) | 166 (69.5%) | <0.001 |
| Lower extremity weakness | 360 (49.9%) | 308 (48%) | 52 (65.0%) | 0.004 | 216 (44.8%) | 144 (60.3%) | <0.001 |
| Arm and lower extremity weakness | 330 (45.8%) | 280 (43.7%) | 50 (62.5%) | 0.001 | 196 (40.7%) | 134 (56.1%) | <0.001 |
| Dysarthria | 238(33%) | 209 (32.6%) | 29 (36.2%) | 0.513 | 163 (33.8%) | 75 (31.4%) | 0.512 |
| Ataxia | 107 (14.8%) | 101 (15.8%) | 6 (7.5%) | 0.05 | 84 (17.4%) | 23 (9.6%) | 0.006 |
| Gait imbalance | 74 (10.3%) | 73 (11.4%) | 1 (1.2%) | 0.005 | 61 (12.7%) | 13 (5.4%) | 0.003 |
| Vertigo | 28 (3.9%) | 27 (4.2%) | 1 (1.2%) | 0.196 | 24 (5.0%) | 4 (1.7%) | 0.031 |
| Coma | 11 (1.5%) | 5 (0.8%) | 6 (7.5%) | <0.001 | 3 (0.6%) | 8 (3.3%) | 0.005 |
| Visual symptoms | 41 (5.7%) | 39 (6.1%) | 2 (2.5%) | 0.199 | 33 (6.8%) | 8 (3.4%) | 0.058 |
| Incoordination | 44 (6.1%) | 42 (6.6%) | 2 (2.5%) | 0.153 | 35 (7.3%) | 9 (3.8%) | 0.065 |
| Headache | 22 (3.1%) | 21 (3.3%) | 1 (1.2%) | 0.32 | 20 (4.1%) | 2 (0.8%) | 0.015 |
| Confusion | 11 (1.5%) | 10 (1.6%) | 1 (1.2%) | 0.831 | 6 (1.2%) | 5 (2.1%) | 0.382 |
| Dressing | <0.001 | <0.001 | |||||
| 0 | 90 (12.5%) | 49 (7.6%) | 41 (51.2%) | 24 (5.0%) | 66 (27.6%) | ||
| 1 | 374 (51.9%) | 360 (56.2%) | 14 (17.5%) | 310 (64.3%) | 64 (26.8%) | ||
| 2 | 257 (35.6%) | 232 (36.2%) | 25 (31.2%) | 148 (30.7%) | 109 (45.6%) | ||
| Grooming | <0.001 | <0.001 | |||||
| 0 | 79 (11.0%) | 42 (6.6%) | 37 (46.2%) | 18 (3.7%) | 61 (25.5%) | ||
| 1 | 389 (54%) | 372 (58%) | 17 (21.2%) | 318 (66%) | 71 (29.7%) | ||
| 2 | 253 (35.1%) | 227 (35.4%) | 26 (32.5%) | 146 (30.3%) | 107 (44.8%) | ||
| Toileting | <0.001 | <0.001 | |||||
| 0 | 87 (12.1%) | 50 (7.8%) | 37 (46.2%) | 26 (5.4%) | 61 (25.5%) | ||
| 1 | 377 (52.3%) | 363 (56.6%) | 14 (17.5%) | 311 (64.5%) | 66 (27.6%) | ||
| 2 | 257 (35.6%) | 228 (35.6%) | 29 (36.2%) | 145 (30.1%) | 112 (46.9%) | ||
| Eating | <0.001 | <0.001 | |||||
| 0 | 61 (8.5%) | 30 (4.7%) | 31 (38.8%) | 14 (2.9%) | 47 (19.7%) | ||
| 1 | 519 (72%) | 489 (76.3%) | 30 (37.5%) | 396 (82.2%) | 123 (51.5%) | ||
| 2 | 141 (19.6%) | 122 (19%) | 19 (23.8%) | 72 (14.9%) | 69 (28.9%) | ||
| Hygiene maintenance | <0.001 | <0.001 | |||||
| 0 | 84 (11.7%) | 46 (7.2%) | 38 (47.5%) | 21 (4.4%) | 63 (26.4%) | ||
| 1 | 395 (54.8%) | 381 (59.4%) | 14 (17.5%) | 323 (67%) | 72 (30.1%) | ||
| 2 | 242 (33.6%) | 214 (33.4%) | 28 (35.0%) | 138 (28.6%) | 104 (43.5%) | ||
| Bathing | <0.001 | <0.001 | |||||
| 0 | 91 (12.6%) | 52 (8.1%) | 39 (48.8%) | 27 (5.6%) | 64 (26.8%) | ||
| 1 | 364 (50.5%) | 352 (54.9%) | 12 (15.0%) | 302 (62.7%) | 62 (25.9%) | ||
| 2 | 266 (36.9%) | 237 (37%) | 29 (36.2%) | 153 (31.7%) | 113 (47.3%) | ||
| Mobility | <0.001 | <0.001 | |||||
| 0 | 96 (13.3%) | 58 (9.0%) | 38 (47.5%) | 32 (6.6%) | 64 (26.8%) | ||
| 1 | 372 (51.6%) | 358 (55.9%) | 14 (17.5%) | 305 (63.3%) | 67 (28.0%) | ||
| 2 | 253 (35.1%) | 225 (35.1%) | 28 (35.0%) | 145 (30.1%) | 108 (45.2%) | ||
0, dependent; 1, independent; 2, required assistance; ACEI/ARBs, angiotensin-converting enzyme inhibitor/angiotensin type II receptor blocker; BMI, body mass index; CAD, coronary artery disease; CKD, chronic kidney disease; COPD, chronic obstructive pulmonary disease; DOACs, direct-acting oral anticoagulants; DBP, diastolic blood pressure; HCTZ, hydrochlorothiazide; IQR, interquartile range, LAA, large artery atherosclerosis; LOS, length of stay; MCA, middle cerebral artery territory; n, number of patients; NIHSS, National Institutes of Health Stroke Scale; OSA, obstructive sleep apnoea; PAD, peripheral artery disease; SBP, systolic blood pressure; SVO, small vessel occlusion; SSRI, selective serotonin receptor inhibition.
Machine learning models
The ROC curves and calibration plots for each model for the 90-day and 3-year outcomes are shown in Figures 1 and 2, respectively. The full set of performance metrics based on the timescales of the outcomes is presented in Tables 2 and 3. Notably, all models achieved good discrimination and low Brier scores (indicators of the models’ performance). Additional details of the models are provided in Supplementary material online, Table S6. The TRIPOD checklist is provided in Supplementary material online, Table S7.
Figure 1.
Receiver operating characteristic curves and calibration plots for 90-day outcome. Receiver operating characteristic curves for predicting composite or recurrent stroke or mortality at 90 days after hospitalization for acute ischaemic stroke are shown in the development (A) and internal validation (B) data sets stratified according to individual models using all the predictors. The corresponding calibration plots are depicted for the development (C) and internal validation (D) data sets. The lower panel illustrates receiver operating characteristic curves for each model [(E) development, (F) internal validation data sets] and calibration plots [(G) development, (H) internal validation data sets] for predicting 90-day outcome using the 10 most important predictors. AUC, area under the receiver operating characteristic curve; CI, confidence interval; RF, random forest; SVM, support vector machine; XGBOOST, extreme gradient boosting.
Figure 2.
Receiver operating characteristic curves and calibration plots for 3-year outcome. Receiver operating characteristic curves for predicting composite or recurrent stroke or mortality at 3 years after hospitalization for acute ischaemic stroke are shown in the development (A) and internal validation (B) data sets stratified according to individual models using all the predictors. The corresponding calibration plots are depicted for the development (C) and internal validation (D) data sets. The lower panel illustrates receiver operating characteristic curves for each model [(E) development, (F) internal validation data sets] and calibration plots [(G) development, (H) internal validation data sets] for predicting 3-year outcome using the 10 most important predictors. AUC, area under the receiver operating characteristic curve; CI, confidence interval; RF, random forest; SVM, support vector machine; XGBOOST, extreme gradient boosting.
Table 2.
Performance metrics of machine learning models for predicting composite of recurrent stroke or mortality at 90 days and 3 years after ischaemic stroke
| Prediction | Features | Model | Development set, n = 505 (70%) | Validation set, n = 216 (30%) | ||||
|---|---|---|---|---|---|---|---|---|
| AUC (95% CI) | Brier score | Hosmer–Lemeshow, P | AUC (95% CI) | Brier score | Hosmer–Lemeshow, P | |||
| 90-day composite outcome | All 90 features | RF | 0.782 (0.760–0.803) | 0.079 | 0.915 | 0.779 (0.671–0.887) | 0.072 | 0.692 |
| SVM | 0.784 (0.762–0.806) | 0.077 | 0.999 | 0.771 (0.660–0.882) | 0.074 | 0.929 | ||
| XGBOOST | 0.789 (0.768–0.810) | 0.077 | 0.999 | 0.772 (0.674–0.869) | 0.072 | 0.999 | ||
| Top 10 identified features | RF | 0.828 (0.810–0.847) | 0.076 | 0.944 | 0.777 (0.765–0.878) | 0.072 | 0.999 | |
| SVM | 0.799 (0.778–0.819) | 0.077 | 0.999 | 0.682 (0.547–0.817) | 0.074 | 0.997 | ||
| XGBOOST | 0.819 (0.800–0.838) | 0.075 | 0.999 | 0.799 (0.704–0.894) | 0.072 | 0.999 | ||
| 3-year composite outcome | All 90 features | RF | 0.768 (0.754–0.782) | 0.181 | 0.989 | 0.743 (0.674–0.811) | 0.176 | 0.999 |
| SVM | 0.779 (0.766–0.793) | 0.173 | 0.999 | 0.777 (0.711–0.843) | 0.171 | 0.958 | ||
| XGBOOST | 0.765 (0.751–0.779) | 0.177 | 0.999 | 0.773 (0.702–0.844) | 0.161 | 0.999 | ||
| Top 10 identified features | RF | 0.770 (0.756–0.784) | 0.178 | 0.944 | 0.736 (0.663–0.839) | 0.180 | 0.754 | |
| SVM | 0.773 (0.759–0.789) | 0.175 | 0.999 | 0.746 (0.670–0.823) | 0.171 | 0.999 | ||
| XGBOOST | 0.776 (0.763–0.790) | 0.174 | 0.999 | 0.763 (0.692–0.835) | 0.166 | 0.999 | ||
AUC, area under the receiver operating characteristic curve; CI, confidence interval; RF, random forest; SVM, support vector machine; XGBOOST, extreme gradient boost.
Table 3.
Performance metrics of ischaemic stroke predictive risk score and totalled health risks in vascular events score for predicting composite outcome at 90 days and 3 years after ischaemic stroke in cumulative cohort
| Prediction | Model | Features | AUC (95% CI) | Cut-off | Sensitivity | Specificity | Youden’s index | PPV | NPV | Hosmer–Lemeshow, P |
|---|---|---|---|---|---|---|---|---|---|---|
| 90-day composite outcome | iScore | 8 | 0.715 (0.654–0.777) | 70 | 0.700 | 0.647 | 0.347 | 0.199 | 0.945 | 0.081 |
| THRIVE | 10 | 0.664 (0.595–0.731) | 5 | 0.350 | 0.913 | 0.263 | 0.333 | 0.918 | 0.117 | |
| 3-year composite outcome | iScore | 10 | 0.671 (0.636–0.721) | 40 | 0.657 | 0.622 | 0.279 | 0.463 | 0.785 | 0.124 |
| THRIVE | 10 | 0.675 (0.663–0.716) | 4 | 0.460 | 0.790 | 0.251 | 0.521 | 0.747 | 0.282 |
AUC, area under the receiver operating characteristic curve; CI, confidence interval; iScore, ischaemic stroke predictive risk score; THRIVE, totalled health risks in vascular events.
Models with full set of predictors
Ninety-day prediction
The differences between the discriminatory performance of the RF [0.779, 95% confidence interval (CI) 0.671–0.887], SVM (0.771, 95% CI 0.660–0.882), and XGBOOST (0.772, 95% CI 0.674–0.869) models were not statistically significant in the validation data set.
Three-year prediction
The discriminatory performances of the RF (0.743, 95% CI 0.674–0.811), SVM (0.777, 95% CI 0.711–0.843), and XGBOOST (0.773, 95% CI 0.702–0.844) models were comparable in the validation data set.
Top 10 most important variables
The top 10 most important variables selected by each model from a total of 90 variables for predicting the composite outcome and the average of the two models are shown in Figure 3. The 10 important variables were derived solely from the training data set during the cross-validation process. All models consistently selected the variables ‘no antithrombotic drugs at discharge’ and ‘advanced age’ as important for predicting the 90-day and 3-year outcomes, respectively.
Figure 3.
Top 10 important predictors for the composite outcome identified by random forest, extreme gradient boosting, and the average of all models. The bars represent the relative contribution of variables in predicting the clinical outcome. All models are consistent in identifying exposure/no exposure to antithrombotic drugs as the most important predictor for 90-day clinical outcome. Similarly, all models are consistent in identifying age as the most important predictor for 3-year clinical outcome. ACEI, angiotensin-converting enzyme inhibitor or angiotensin type II receptor blocker; BMI, body mass index; CCB, calcium channel blocker; DBP, diastolic blood pressure; DM, diabetes mellitus; LOS, length of stay; NIHSS, National Institutes of Health Stroke Scale; SBP, systolic blood pressure; XGBOOST, extreme gradient boosting.
Ninety-day prediction
All models showed comparable performances; the RF, SVM, and XGBOOST models exhibited AUCs of 0.777 (95% CI 0.765–0.878), 0.682 (95% CI 0.547–0.817), and 0.799 (95% CI 0.704–0.894), respectively, in the validation set. The AUC achieved using 90 variables for the 90-day composite outcome was not greater than that achieved using the top 10 variables. Of the top 10 variables, no antithrombotic drugs at discharge, dependency on basic activities, NIHSS score, or prescription of calcium channel blockers at discharge were shared across the ML models. All three models selected antithrombotic medications at discharge as the most important variable.
Three-year prediction
The three ML models exhibited comparable performances: RF, SVM, and XGBOOST models showed AUC of 0.736 (95% CI 0.547–0.818), 0.746 (95% CI 0.670–0.823), and 0.763 (95% CI 0.692–0.835), respectively, in the validation set. The AUC achieved using 90 variables for the 3-year composite outcome was not greater than that achieved using the top 10 variables. Advanced age, dependency on basic activities, NIHSS score, and calcium channel blocker use at discharge were shared between the RF and XGBOOST models. Both the RF and XGBOOST models selected age as the most important factor.
Validation of the ischaemic stroke predictive risk score and totalled health risks in vascular events scores
Ninety-day prediction
The iScore achieved an AUC of 0.715 (95% CI 0.654–0.777) and good calibration (Hosmer–Lemeshow test, P = 0.081; Figure 4). At a cut-off point of 70, the iScore exhibited 70% sensitivity, 65% specificity, 20% PPV, and 95% NPV for the 90-day outcome prediction. The THRIVE score achieved a modest AUC of 0.664 (95% CI 0.595–0.731) and good calibration (Hosmer–Lemeshow test, P = 0.117). At a cut-off point of 5, the THRIVE score exhibited 35% sensitivity, 91% specificity, 33% PPV, and 92% NPV for the 90-day outcome prediction.
Figure 4.
External validation with generation of receiver operating characteristic curves, calibration, and decision analysis curves of the ischaemic stroke predictive risk score and totalled health risks in vascular events scoring systems. The left two panels represent external validation of the ischaemic stroke predictive risk score with construction of receiver operating characteristic curves, recalibration plots, and decision curve analysis for 90-day and 3-year outcome prediction. Similarly, the right two panels represent the external validation of the totalled health risks in vascular events score with the generation of receiver operating characteristic curves, recalibration plots, and decision curve analysis for 90-day and 3-year outcome prediction.
Three-year prediction
The iScore achieved an AUC of 0.671 (95% CI 0.636–0.721) and good calibration (Hosmer–Lemeshow test, P = 0.124). At a cut-off point of 40, the iScore exhibited 66% sensitivity, 62% specificity, 46% PPV, and 78% NPV for the 3-year outcome prediction. Similarly, the THRIVE score achieved an AUC of 0.675 (95% CI 0.663–0.716), with good calibration (Hosmer–Lemeshow test, P = 0.282). At a cut-off point of 4, the THRIVE score exhibited 46% sensitivity, 79% specificity, 52% PPV, and 75% NPV for the 3-year outcome prediction. The results of the direct comparison of the iScore (AUC 0.679; 95% CI 0.636–0.722) and THRIVE score (AUC 0.676; 95% CI 0.634–0.717) for predicting the 3-year outcome demonstrated no significant difference (DeLong test 0.886), as shown in Supplementary material online, Figure S3 and Table S8.
Decision curve analysis
We plotted the net benefit on the y-axis and threshold probability (preferences) on the x-axis.45 Both the iScore and THRIVE scores provided comparable net benefits for clinical utility across a range of thresholds in the cumulative cohort for both the 90-day and 3-year outcome predictions (Figure 4). Random forest, SVM, and XGBOOST consistently showed net benefits across a wide range of thresholds for outcome prediction in both the development and validation data sets regardless of whether all 90 variables or the 10 most predictive variables were used to construct the plots (Figure 5).
Figure 5.
Decision curve analysis plots for each study model with all predictors are shown at 90 days in the development (A) and the internal validation sets (B) and at 3 years in the development (C) and internal validation sets (D). Similarly, decision curve analysis plots for each study model with the top 10 most important predictors are shown at 90 days in the development set (E) and internal validation set (F) and at 3 years in the development (G) and internal validation sets (H). RF, random forest; SVM, support vector machine; XGBOOST, extreme gradient boosting.
Comparative performances
The RF, SVM, and XGBOOST models and the iScore and THRIVE scoring systems demonstrated a good fit according to the Hosmer–Lemeshow test, with a P value consistently higher than 0.05. All ML models achieved numerically higher AUCs, greater accuracy, and lower Brier scores in predicting the 90-day than the 3-year composite outcomes after AIS. The AUCs for predicting both the 90-day and 3-year composite outcomes after AIS with all 90 or the top 10 most important variables were comparable across ML models in both the development and validation sets. The results of ML models were not statistically compared with the results of iScore and THRIVE scores because the validation of the ML models in 30% of the data set with the external validation of existing models in the cumulative cohort post challenges.
Complementary analysis
The discriminatory performance of the iScore and THRIVE scoring systems for predicting 1-year composite events (iScore: AUC 0.680, 95% CI 0.631–0.730; THRIVE: AUC 0.663, 95% CI 0.609–0.717; DeLong P value 0.494) or 1-year mortality alone (iScore: AUC 0.733, 95% CI 0.682–0.785; THRIVE: AUC 0.713, 95% CI 0.656–0.770; DeLong 0.432) was not different from AUC for the 3-year prediction of composite events (see Supplementary material online, Figure S4 and Table S9). Additional validation of the iScore and THRIVE scores in the 30% validation cohort yielded findings comparable to those of the primary analysis (iScore: AUC 0.679, 95% CI 0.636–0.721; THRIVE: AUC 0.676, 95% CI 0.634–0.717; DeLong 0.886), as shown in Supplementary material online, Figure S5 and Table S10. We presented the sensitivity and specificity measures of ML models for outcome protection in Supplementary material online, Table S11. Finally, for 90-day composite outcome prediction, iScore outperformed THRIVE score (AUC 0.715 vs. 0.655; DeLong test P = 0.023) as shown in Supplementary material online, Figure S6 and Table S12.
Discussion
In this retrospective study, we assessed the performance and clinical usefulness of three ML models and validated the iScore and THRIVE score in predicting the 90-day or 3-year composite outcome of recurrent stroke or all-cause mortality after AIS. The important contributions of this study are as follows. In hospitalized AIS patients who survived to discharge, each of the three ML models, which were carefully developed using 90 predictor variables of high diversity, achieved similar performance in predicting the composite outcome at both 90 days and 3 years after AIS and similar usefulness for informing clinical decision-making. Additionally, the top 10 predictor-based parsimonious ML models retained the performance and clinical utility of the initial 90-variable-based models. Each study model demonstrated good discrimination and calibration in predicting the outcome over 90 days and 3 years. In this study, the utility of the iScore and THRIVE scoring systems was expanded for predicting recurrent stroke, predictive performance was extended beyond 1 year of follow-up, and clinical decision-making was improved through the construction of DCA curves.
The existing ML algorithms were modestly successful with a focus on functional outcomes and short-term follow-up (30 days to 1 year); they are insufficient for guiding routine clinical decision-making.51,52 None of the previous ML models examined the outcome beyond 1 year or systematically assessed their clinical usefulness using DCA.51,52 Conversely, the models built in this study have several advantages: The patients were well characterized with near-complete data collection and follow-up with only 0.1% missing values (which were imputed), the models were trained on manually abstracted large-scale granular data (total of 90 variables) from multiple domains to avoid artificial intelligence bias,53 and predictor variables such as those related to daily living activities not often used in previous studies were employed in model development.54
To avoid overfitting and to enhance generalizability, we developed and validated a separate set of pragmatic models for outcome prediction using the top 10 predictors from the original set of 90 variables with no loss of performance or clinical utility. We observed a predominance of stroke-associated variables contributing to the 90-day prediction (no antithrombotic medication use, stroke aetiology, arm weakness, dependency on daily activity, NIHSS score, and facial palsy), whereas demographics and comorbidities (age, dyslipidaemia, diabetes mellitus, cancer, BMI, and systolic blood pressure) prevailed in the 3-year prediction. Additionally, dependency in daily activities and NIHSS score were predictors for both the short- and long-term outcomes. Assessment of the directionality of variables indicated that small vessel occlusion stroke subtype (lacunar),55 increasing mean systolic blood pressure,56 hyperlipidaemia,57 and increasing BMI58 were inversely related to the 3-year composite outcome, consistent with previous reports.7,22,59–63
The performance characteristics of the ML models for risk prediction were similar in our study, consistent with several recent reports, including a meta-analysis of 71 studies.18,51 One potential explanation of these findings is the use of a priori variables as input features.64 The results might be different when variables are directly captured from large electronic databases. The study provides valuable insights into the performance metrics between the iScore and THRIVE score in predicting the composite of recurrent stroke or all-cause mortality. Although both the iScore and THRIVE scores were comparable in discriminatory performance, the THRIVE score demonstrated a lower sensitivity, allowing physicians to identify only 35 and 46% of the patients with AIS with composite outcomes at 90 days and 3 years, respectively, as opposed to the iScore, which allowed the identification of 70 and 66% of patients with composite outcomes at 90 days and 3 years, respectively. These findings suggest that using the THRIVE scores is likely to miss a portion of patients with unfavourable outcomes, supporting the findings of a previous report.65 Notably, the iScore and THRIVE scores consistently showed a net benefit across a range of thresholds in DCA, a finding not explored in previous validation studies.
The study was not intended to generate a point-based scoring system for risk calculation at the bedside, but, instead, to assess the performance of ML models in comparison to existing validated risk prediction tools. We conceived that ML models are complimentary to statistical-based models in providing meaningful insights and may not be mutually exclusive.66
Clinical implications
Physicians are often faced with the difficult task of estimating a prognosis early after AIS to assist in decision-making, patient and family counselling, and resource allocation, but clinical expertise and knowledge of individual prognostic indicators are often insufficient for predicting clinical outcomes after AIS.6 Although both the iScore and THRIVE scores offer effective risk stratification after AIS, they have some limitations. Our advanced artificial intelligence–based ML models, equipped with improved decision curve analytics, provided enhanced accuracy and potentially scalable integration into workflows for informing clinical decision-making for short- and long-term follow-up in patients with AIS. Compared with statistical method–based risk prediction tools, ML models can be self-trained with additional data in the future to improve the current results and are not necessarily simple given the context of increasingly wider use of built-in automation of electronic medical records
Research implications
There exists considerable uncertainty and confusion regarding artificial intelligence algorithms for prognostication, particularly among practising physicians, as opposed to the relatively small percentage of academics or physicians with interest in research. The overarching goal of our artificial intelligence research is making these algorithms more accessible to practising physicians. Therefore, in this study, we opted for iScore and THRIVE scoring systems, given their broad external validity as baseline models that are more familiar to practising physicians. We identified many opportunities and challenges for future research into the use of ML in outcome prediction, including the following: (i) advancing the design and innovation of the ML models to further improve their predictive performance; (ii) assessing the interoperability and generalizability of the models; (iii) testing the ability of the models to integrate into electronic health records for routine clinical use; and (iv) determining whether the models improve the clinical outcomes of individualized care and decision-making and reduce costs.
Strengths and limitations
Our study had some limitations. Comparing the internal evaluation of ML models to the external validation of existing scores (iScore and THRIVE scores) presents a challenge in directly establishing the superiority of one approach over the other. Therefore, we did not compare ML models with the existing models. We applied novel ML strategies to data obtained from a single academic centre in predominantly Caucasian population involving adults aged <90 years; this design limited the generalizability of the study, and thus, external validation in an independent data set, especially one more inclusive of minority populations, is warranted. However, the study design minimizes the spectrum bias associated with large heterogeneous pooled data from multiple nations used in some studies, which limits model calibration and potential utility.54 Furthermore, the modest sample size precluded subgroup and sensitivity analyses. Although we included 90 variables, we cannot rule out the possibility of unmeasured confounders or misclassification because our data were extracted from electronic medical records. We evaluated the composite of recurrent stroke or mortality as the composite outcome because of low event rates and other reasons described above. The results may be different if components of the composite outcome are evaluated separately. This study also had several strengths. Comparing the internal evaluation of ML models to the external validation of existing scores (iScore and THRIVE scores) presents a challenge in directly establishing the superiority of one approach over the other. To enable equitable comparison, we generated new ROC curves for both the iScore and THRIVE scoring systems to predict 3-year composite outcomes in 30% validation data set. Notably, we found no discernable variations in predictive performance when compared with the results from the overall cohort. The study adhered to the TRIPOD reporting guidelines. We externally validated the previously identified risk prediction models, the iScore and the THRIVE scores, for reference to develop and validate ML models, extending their prognostic value to 3 years. Receiver operating characteristic curves, calibration plots, and decision analysis curves were created for each model for visual verification.
Conclusions
In this retrospective study, three artificial intelligence–based ML models were developed to predict the short- and long-term combined risk of recurrent stroke or all-cause death after the first AIS. The ML models, iScore, and THRIVE score systems consistently demonstrated good discrimination and clinical utility in predicting the outcome across 90 days and 3 years after AIS. The utility of the iScore and THRIVE score has expanded to include recurrent stroke and extended prediction outcomes from 1 to 3 years after AIS. The ML models have the advantages of further improvement in the results by incorporating additional complex data in the future, self-training, automation, and ease of implementation in the workflow. Before wider clinical implementation, further research is needed to establish their generalizability in a larger and more heterogeneous patient population to ensure that predictive performance and clinical utility are preserved.
Supplementary Material
Contributor Information
Daniel Axford, Department of Information Technology, Mathematics and Statistics, College of Science, Health, Engineering and Education, Murdoch University, Murdoch, Australia.
Ferdous Sohel, Department of Information Technology, Mathematics and Statistics, College of Science, Health, Engineering and Education, Murdoch University, Murdoch, Australia.
Vida Abedi, Department of Public Health Science, Penn State College of Medicine, Hershey, PA, USA.
Ye Zhu, Robert D. and Patricia E. Kern Centre for the Science of Healthcare Delivery, Mayo Clinic, Rochester, MN, USA.
Ramin Zand, Neuroscience Institute, Geisinger Health System, 100 North Academy Ave, Danville, PA 17822, USA; Neuroscience Institute, The Pennsylvania State University, Hershey, PA 17033, USA.
Ebrahim Barkoudah, Internal Medicine/Hospital Medicine, Brigham and Women’s Hospital, Harvard University, Boston, MA, USA.
Troy Krupica, Internal Medicine/Hospital Medicine, West Virginial University, Morgantown, WV, USA.
Kingsley Iheasirim, Internal Medicine/Hospital Internal Medicine, Mayo Clinic Health System, Mankato, MN, USA.
Umesh M Sharma, Hospital Internal Medicine, Mayo Clinic, Phoenix, AZ, USA.
Sagar B Dugani, Hospital Internal Medicine, Mayo Clinic, Rochester, MN, USA.
Paul Y Takahashi, Community Internal Medicine, Mayo Clinic, Rochester, MN, USA.
Sumit Bhagra, Endocrinology, Diabetes and Metabolism, Mayo Clinic Health System, Austin, MN, USA.
Mohammad H Murad, Division of Public Health, Infectious Diseases, and Occupational Medicine, Mayo Clinic, Rochester, MN, USA.
Gustavo Saposnik, Stroke Outcomes and Decision Neuroscience Research Unit, Division of Neurology, Department of Medicine and Li Ka Shing Knowledge Institute, St.Michael’s Hospital, University of Toronto, Toronto, Ontario, Canada.
Mohammed Yousufuddin, Hospital Internal Medicine, Mayo Clinic Health System, 1000 1st Drive NW, Austin, MN 55912, USA.
Author contributions
D.A., F.S., V.A., Y.Z., E.B., T.K., R.Z., P.Y.T., S.B., M.H.M., G.S., and M.Y. conceived and designed the study. D.A., F.S., V.A., Y.Z., K.I., U.M.S., S.B.D., M.H.M., G.S., and M.Y. obtained the data, performed the analyses and interpretation, and drafted the manuscript. D.A., F.S., V.A., R.Z., E.B., T.K., K.I., U.M.S., S.B.D., P.Y.T., G.S., and M.Y. provided professional and statistical support and made critical revisions. All authors read and approved the final manuscript.
Supplementary material
Supplementary material is available at European Heart Journal – Digital Health.
Funding
None declared.
Ethical approval and consent to participate
The Mayo Clinic Institutional Review Board approved the study protocol. Written consent was waived because of minimal risk and retrospective nature of the study.
Data availability
Anonymized data to support the results of this study will be made available upon request of qualified investigators.
References
- 1. Virani SS, Alonso A, Aparicio HJ, Benjamin EJ, Bittencourt MS, Callaway CW, et al. Heart disease and stroke statistics—2021 update: a report from the American Heart Association. Circulation 2021;143:e254–e743. [DOI] [PubMed] [Google Scholar]
- 2. Feng W, Hendry RM, Adams RJ. Risk of recurrent stroke, myocardial infarction, or death in hospitalized stroke patients. Neurology 2010;74:588–593. [DOI] [PubMed] [Google Scholar]
- 3. Dhamoon MS, Sciacca RR, Rundek T, Sacco RL, Elkind MS. Recurrent stroke and cardiac risks after first ischemic stroke: the Northern Manhattan Study. Neurology 2006;66:641–646. [DOI] [PubMed] [Google Scholar]
- 4. Hankey GJ, Jamrozik K, Broadhurst RJ, Forbes S, Burvill PW, Anderson CS, et al. Long-term risk of first recurrent stroke in the Perth Community Stroke Study. Stroke 1998;29:2491–2500. [DOI] [PubMed] [Google Scholar]
- 5. Saposnik G, Cote R, Mamdani M, Raptis S, Thorpe KE, Fang J, et al. JURaSSiC: accuracy of clinician vs risk score prediction of ischemic stroke outcomes. Neurology 2013;81:448–455. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Ntaios G, Gioulekas F, Papavasileiou V, Strbian D, Michel P. ASTRAL, DRAGON and SEDAN scores predict stroke outcome more accurately than physicians. Eur J Neurol 2016;23:1651–1657. [DOI] [PubMed] [Google Scholar]
- 7. Kauw F, Takx RAP, de Jong HWAM, Velthuis BK, Kappelle LJ, Dankbaar JW. Clinical and imaging predictors of recurrent ischemic stroke: a systematic review and meta-analysis. Cerebrovasc Dis 2018;45:279–287. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Chaudhary D, Abedi V, Li J, Schirmer CM, Griessenauer CJ, Zand R. Clinical risk score for predicting recurrence following a cerebral ischemic event. Front Neurol 2019;10:1106. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9. Kernan WN, Viscoli CM, Brass LM, Makuch RW, Sarrel PM, Roberts RS, et al. The stroke prognosis instrument II (SPI-II): a clinical prediction instrument for patients with transient ischemia and nondisabling ischemic stroke. Stroke 2000;31:456–462. [DOI] [PubMed] [Google Scholar]
- 10. Thompson DD, Murray GD, Dennis M, Sudlow CL, Whiteley WN. Formal and informal prediction of recurrent stroke and myocardial infarction after stroke: a systematic review and evaluation of clinical prediction models in a new cohort. BMC Med 2014;12:58. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Weimar C, Diener HC, Alberts MJ, Steg PG, Bhatt DL, Wilson PWF, et al. The Essen stroke risk score predicts recurrent cardiovascular events: a validation within the REduction of Atherothrombosis for Continued Health (REACH) registry. Stroke 2009;40:350–354. [DOI] [PubMed] [Google Scholar]
- 12. Weimar C, Goertler M, Röther J, Ringelstein EB, Darius H, Nabavi DG, et al. Systemic risk score evaluation in ischemic stroke patients (SCALA): a prospective cross sectional study in 85 German stroke units. J Neurol 2007;254:1562–1568. [DOI] [PubMed] [Google Scholar]
- 13. Goldstein BA, Navar AM, Pencina MJ, Ioannidis JP. Opportunities and challenges in developing risk prediction models with electronic health records data: a systematic review. J Am Med Inform Assoc 2017;24:198–208. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Heo J, Yoon JG, Park H, Kim Young Dae, Nam HS, Heo JH. Machine learning-based model for prediction of outcomes in acute stroke. Stroke 2019;50:1263–1265. [DOI] [PubMed] [Google Scholar]
- 15. Patel UK, Anwar A, Saleem S, Malik P, Rasul B, Patel K, et al. Artificial intelligence as an emerging technology in the current care of neurological disorders. J Neurol 2021;268:1623–1642. [DOI] [PubMed] [Google Scholar]
- 16. Sun X, Douiri A, Gulliford M. Applying machine learning algorithms to electronic health records to predict pneumonia after respiratory tract infection. J Clin Epidemiol 2022;145:154–163. [DOI] [PubMed] [Google Scholar]
- 17. Vickers AJ, Elkin EB. Decision curve analysis: a novel method for evaluating prediction models. Med Decis Making 2006;26:565–574. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Wang W, Rudd AG, Wang Y, Curcin V, Wolfe CD, Peek N, et al. Risk prediction of 30-day mortality after stroke using machine learning: a nationwide registry-based cohort study. BMC Neurol 2022;22:195. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Saposnik G, Kapral MK, Liu Y, Hall R, O'Donnell M, Raptis S, et al. IScore: a risk score to predict death early after hospitalization for an acute ischemic stroke. Circulation 2011;123:739–749. [DOI] [PubMed] [Google Scholar]
- 20. Flint AC, Cullen SP, Faigeles BS, Rao VA. Predicting long-term outcome after endovascular stroke treatment: the totaled health risks in vascular events score. AJNR Am J Neuroradiol 2010;31:1192–1196. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Collins GS, Reitsma JB, Altman DG, Moons KG. Transparent reporting of a multivariable prediction model for individual prognosis or diagnosis (TRIPOD): the TRIPOD statement. BMJ 2015;350:g7594. [DOI] [PubMed] [Google Scholar]
- 22. Yousufuddin M, Bartley AC, Alsawas M, Sheely HL, Shultz J, Takahashi PY, et al. Impact of multiple chronic conditions in patients hospitalized with stroke and transient ischemic attack. J Stroke Cerebrovasc Dis 2017;26:1239–1248. [DOI] [PubMed] [Google Scholar]
- 23. Stroke National Institute of Neurological Disorders and Stroke . NIH Stroke Scale. 2019.
- 24. Adams HP Jr, Bendixen BH, Kappelle LJ, Biller J, Love BB, Gordon DL, et al. Classification of subtype of acute ischemic stroke. Definitions for use in a multicenter clinical trial. TOAST. Trial of Org 10172 in Acute Stroke Treatment. Stroke 1993;24:35–41. [DOI] [PubMed] [Google Scholar]
- 25. Yousufuddin M, Sharma UM, Bhagra S, Murad MH. Hyperlipidaemia and mortality among patients hospitalised with pneumonia: retrospective cohort and propensity score matched study. BMJ Open Respir Res 2021;8:e000757. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Yousufuddin M, Young N, Keenan L, Olson T, Shultz J, Doyle T, et al. Five-year mortality after transient ischemic attack: focus on cardiometabolic comorbidity and hospital readmission. Stroke 2018;49:730–733. [DOI] [PubMed] [Google Scholar]
- 27. Yousufuddin M, Young N, Shultz J, Doyle T, Fuerstenberg KM, Jensen K, et al. Predictors of recurrent hospitalizations and the importance of these hospitalizations for subsequent mortality after incident transient ischemic attack. J Stroke Cerebrovasc Dis 2019;28:167–174. [DOI] [PubMed] [Google Scholar]
- 28. Yousufuddin M, Zhu Y, Al Ward R, Peters J, Doyle T, Jensen KL, et al. Association of hyperlipidaemia with 5-year survival after hospitalisation for acute myocardial infarction: a propensity score matched analysis. Open Heart 2020;7:e001163. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Stekhoven DJ, Bühlmann P. MissForest—non-parametric missing value imputation for mixed-type data. Bioinformatics 2012;28:112–118. [DOI] [PubMed] [Google Scholar]
- 30. Freemantle N, Calvert M, Wood J, Eastaugh J, Griffin C. Composite outcomes in randomized trials: greater precision but with greater uncertainty? JAMA 2003;289:2554–2559. [DOI] [PubMed] [Google Scholar]
- 31. Goldberg R, Gore JM, Barton B, Gurwitz J. Individual and composite study endpoints: separating the wheat from the chaff. Am J Med 2014;127:379–384. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32. Ananth CV, Brandt JS, Keyes KM, Graham HL, Kostis JB, Kostis WJ. Epidemiology and trends in stroke mortality in the USA, 1975–2019. Int J Epidemiol 2023;52:858–866. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33. McCoy CE. Understanding the use of composite endpoints in clinical trials. West J Emerg Med 2018;19:631–634. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Skajaa N, Adelborg K, Horváth-Puhó E, Rothman KJ, Henderson VW, Thygesen LC, et al. Risks of stroke recurrence and mortality after first and recurrent strokes in Denmark: a nationwide registry study. Neurology 2021;98:e329–e342. [DOI] [PubMed] [Google Scholar]
- 35. Eriksson M, Grundberg A, Inge E, von Euler M. Stroke recurrence following 28 days after first stroke in men and women 2012 to 2020: observations from the Swedish Stroke Register. J Am Heart Assoc 2023;12:e028222. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36. Wright MN WS . A fast implementation of random forest. https://cran.r-project.org/web/packages/ranger/index.html. (2022).
- 37. RDocumentation, Kuhn M. caret/RegressionTests/Code/svmRadial.R. 2017. https://github.com/topepo/caret/blob/master/RegressionTests/Code/svmRadial.R. (2022).
- 38. Abedi V, Avula V, Chaudhary D, Shahjouei S, Khan A, Griessenauer CJ, et al. Prediction of long-term stroke recurrence using machine learning models. J Clin Med 2021;10:1286. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39. Abedi V, Avula V, Razavi SM, Bavishi S, Chaudhary D, Shahjouei S, et al. Predicting short and long-term mortality after acute ischemic stroke using EHR. J Neurol Sci 2021;427:117560. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Altman DG, Royston P. What do we mean by validating a prognostic model? Stat Med 2000;19:453–473. [DOI] [PubMed] [Google Scholar]
- 41. Steyerberg EW, Vickers AJ, Cook NR, Gerds T, Gonen M, Obuchowski N, et al. Assessing the performance of prediction models: a framework for traditional and novel measures. Epidemiology 2010;21:128–138. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42. Huang C, Li SX, Caraballo C, Masoudi FA, Rumsfeld JS, Spertus JA, et al. Performance metrics for the comparative analysis of clinical risk prediction models employing machine learning. Circ Cardiovasc Qual Outcomes 2021;14:e007526. [DOI] [PubMed] [Google Scholar]
- 43. Gerds TA, Cai T, Schumacher M. The performance of risk prediction models. Biom J 2008;50:457–479. [DOI] [PubMed] [Google Scholar]
- 44. Hosmer DW, Hosmer T, Le Cessie S, Lemeshow S. A comparison of goodness-of-fit tests for the logistic regression model. Stat Med 1997;16:965–980. [DOI] [PubMed] [Google Scholar]
- 45. Vickers AJ, van Calster B, Steyerberg EW. A simple, step-by-step guide to interpreting decision curve analysis. Diagn Progn Res 2019;3:18. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46. Johnston KC, Connors AF Jr, Wagner DP, Knaus WA, Wang X-Q, Haley EC. A predictive risk model for outcomes of ischemic stroke. Stroke 2000;31:448–455. [DOI] [PubMed] [Google Scholar]
- 47. Chen W, Liu G, Fang J, Wang Y, Song Y, Pan Y, et al. External validation of the totaled health risks in vascular events score to predict functional outcome and mortality in patients entered into the China National Stroke Registry. J Stroke Cerebrovasc Dis 2016;25:2331–2337. [DOI] [PubMed] [Google Scholar]
- 48. Shen B, Yang X, Sui RB, Yang B. The prognostic value of the THRIVE score, the iScore score and the ASTRAL score in Chinese patients with acute ischemic stroke. J Stroke Cerebrovasc Dis 2018;27:2877–2886. [DOI] [PubMed] [Google Scholar]
- 49. Wang Y, Liu J, Wu Q, Cheng Y, Liu M. Validation and comparison of multiple risk scores for prediction of symptomatic intracerebral hemorrhage after intravenous thrombolysis in VISTA. Int J Stroke 2023;18:338–345. [DOI] [PubMed] [Google Scholar]
- 50. Wang WY, Sang WW, Jin D, Yan S-M, Hong Y, Zhang H, et al. The prognostic value of the iScore, the PLAN score, and the ASTRAL score in acute ischemic stroke. J Stroke Cerebrovasc Dis 2017;26:1233–1238. [DOI] [PubMed] [Google Scholar]
- 51. Christodoulou E, Ma J, Collins GS, Steyerberg EW, Verbakel JY, Van Calster B. A systematic review shows no performance benefit of machine learning over logistic regression for clinical prediction models. J Clin Epidemiol 2019;110:12–22. [DOI] [PubMed] [Google Scholar]
- 52. Wang W, Kiik M, Peek N, Curcin V, Marshall IJ, Rudd AG, et al. A systematic review of machine learning models for predicting outcomes of stroke with structured data. PLoS One 2020;15:e0234722. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53. Burlina P, Paul W, Mathew P, Joshi N, Pacheco KD, Bressler NM. Low-shot deep learning of diabetic retinopathy with potential applications to address artificial intelligence bias in retinal diagnostics and rare ophthalmic diseases. JAMA Ophthalmol 2020;138:1070–1077. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54. Usher-Smith JA, Sharp SJ, Griffin SJ. The spectrum effect in tests for risk prediction, screening, and diagnosis. Bmj 2016;353:i3139. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 55. Norrving B. Long-term prognosis after lacunar infarction. Lancet Neurol 2003;2:238–245. [DOI] [PubMed] [Google Scholar]
- 56. Lin MP, Ovbiagele B, Markovic D, Towfighi A. Systolic blood pressure and mortality after stroke: too low, no go? Stroke 2015;46:1307–1313. [DOI] [PubMed] [Google Scholar]
- 57. Yeramaneni S, Kleindorfer DO, Sucharew H, Alwell K, Moomaw CJ, Flaherty ML, et al. Hyperlipidemia is associated with lower risk of poststroke mortality independent of statin use: a population-based study. Int J Stroke 2017;12:152–160. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58. Vemmos K, Ntaios G, Spengos K, Savvari P, Vemmou A, Pappa T, et al. Association between obesity and mortality after acute first-ever stroke: the obesity-stroke paradox. Stroke 2011;42:30–36. [DOI] [PubMed] [Google Scholar]
- 59. Bamford J, Dennis M, Sandercock P, Burn J, Warlow C. The frequency, causes and timing of death within 30 days of a first stroke: the Oxfordshire Community Stroke Project. J Neurol Neurosurg Psychiatry 1990;53:824–829. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60. Easton JD, Saver JL, Albers GW, Alberts MJ, Chaturvedi S, Feldmann E, et al. Definition and evaluation of transient ischemic attack: a scientific statement for healthcare professionals from the American Heart Association/American Stroke Association Stroke Council; Council on Cardiovascular Surgery and Anesthesia; Council on Cardiovascular Radiology and Intervention; Council on Cardiovascular Nursing; and the Interdisciplinary Council on Peripheral Vascular Disease. The American Academy of Neurology affirms the value of this statement as an educational tool for neurologists. Stroke 2009;40:2276–2293. [DOI] [PubMed] [Google Scholar]
- 61. Fonarow GC, Pan W, Saver JL, Smith EE, Reeves MJ, Broderick JP, et al. Comparison of 30-day mortality models for profiling hospital performance in acute ischemic stroke with vs without adjustment for stroke severity. JAMA 2012;308:257–264. [DOI] [PubMed] [Google Scholar]
- 62. Kleindorfer DO, Towfighi A, Chaturvedi S, Cockroft KM, Gutierrez J, Lombardi-Hill D, et al. 2021 guideline for the prevention of stroke in patients with stroke and transient ischemic attack: a guideline from the American Heart Association/American Stroke Association. Stroke 2021;52:e364–e467. [DOI] [PubMed] [Google Scholar]
- 63. Lau KK, Wong YK, Teo KC, Chang RS, Chan KH, Hon SF, et al. Epidemiology of stroke long-term prognostic implications of visit-to-visit blood pressure variability in patients with ischemic stroke. Int J Stroke 2014;3:162. [DOI] [PubMed] [Google Scholar]
- 64. Mortazavi BJ, Bucholz EM, Desai NR, Huang C, Curtis JP, Masoudi FA, et al. Comparison of machine learning methods with national cardiovascular data registry models for prediction of risk of bleeding after percutaneous coronary intervention. JAMA Netw Open 2019;2:e196835. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65. Matsumoto K, Nohara Y, Soejima H, Yonehara T, Nakashima N, Kamouchi M. Stroke prognostic scores and data-driven prediction of clinical outcomes after acute ischemic stroke. Stroke 2020;51:1477–1483. [DOI] [PubMed] [Google Scholar]
- 66. Rajula HSR, Verlato G, Manchia M, Antonucci N, Fanos V. Comparison of conventional statistical methods with machine learning in medicine: diagnosis, drug development, and treatment. Medicina (Kaunas) 2020;56:455. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
Anonymized data to support the results of this study will be made available upon request of qualified investigators.






