Skip to main content
PLOS One logoLink to PLOS One
. 2025 Sep 18;20(9):e0332371. doi: 10.1371/journal.pone.0332371

Identifying determinants of readmission and death post-stroke using explainable machine learning

Emir Veledar 1,*, Lili Zhou 1, Omar Veledar 2, Hannah Gardener 1, Carolina M Gutierrez 1, Scott C Brown 1, Farya Fakoori 1, Karlon H Johnson 1, Victor J Del Brutto 1, Ayham Alkhachroum 1, David Z Rose 3, Gillian Gordon Perue 1, Negar Asdaghi 1, Jose G Romano 1, Tatjana Rundek 1
Editor: Noah Hammarlund4
PMCID: PMC12445469  PMID: 40966264

Abstract

Background

Stroke remains a global health challenge with high rates of mortality and rehospitalization placing significant demands on healthcare systems. Identifying factors that determine outcomes of post-hospitalization improves resource allocation. Traditional statistical prediction models are suboptimal for the analysis of complex, multi-dimensional datasets. The objective of our study is to define the extended list of clinical and non-clinical predictors, which we believe can be achieved using Explainable Machine Learning (XML) models as an expansion of conventional methods.

Methods

We evaluated 11 established XML models that represent key ML methodologies to predict 90-day outcomes, namely mortality and rehospitalization among stroke survivors. The study population are 1,300 post-stroke individuals enrolled in the Transitions of Care Stroke Disparities Study (TCSD-S) (NIH/NIMH, NCT03452813) between June 2018 – October 2022. The care after transition data is sourced from participating comprehensive stroke centers and from the Florida Stroke Registry. The analysis incorporated clinical (e.g., age, stroke severity, comorbidities) and non-clinical factors including Social Drivers of Health (SDOH). A combined ranking approach, using Weighted Importance Scores and Frequency Counts, identified significant predictors across models.

Results

The resulting list of selected predictors included both established clinical factors and non-clinical factors, which enhanced prediction accuracy. Out of 38 identified predictors, 20 are non-clinical variables reflecting the importance of SDOH, environmental factors, and behavioral modifications beyond traditional clinical predictors of death/readmission. A secondary analysis restricted to ischemic stroke patients (n = 1,038) yielded virtually identical predictive performance, indicating robustness of the model within this subgroup.

Conclusions

Integrating SDOH, environmental factors, and behavioral modifications alongside traditional clinical predictors enhances the predictive accuracy of post-stroke outcome models. This underscores the critical role of addressing socioeconomic disparities during post-stroke transitions of care. Moreover, XML models’ ability to identify predictors spanning clinical and non-clinical domains suggests their potential to guide recovery. The resulting predictors are crucial for post-hospital care and hold strong potential for identifying individuals at risk of stroke, making them potentially significant across pre-stroke and hospitalization stages.

1. Introduction

Stroke challenges healthcare systems across the entire continuum of care. As the third leading cause of mortality and a major source of disability worldwide [1], stroke is a major chronic non-communicable condition associated with reduced population well-being [2]. Assessing stroke risk factors is a crucial initial step in decreasing its burden. Predictive models can be applied at three stages: primary prevention, response to acute treatment, and post-stroke outcomes including post-hospitalization recurrence and readmissions. The complexity of stroke adjudication, diagnosis, and treatment algorithms often restricts model effectiveness, demanding phase-specific approaches. Current meta-analyses reveal that no high-quality predictive or explanatory models exclusively addressing stroke risk currently exist [3,4]. Existing models often rely on non-modifiable risk factors and borrow heavily from frameworks developed for composite cardiovascular diseases. For the general population, all measures of model quality remain low, highlighting significant gaps in predictive accuracy and reliability. While recent studies have not substantially improved prediction intervals, they have introduced greater inclusivity by incorporating non-traditional biomarkers, such as genetic and polygenic scores, and leveraging novel ML methods, marking incremental progress in methodology and scope. Our focus is on developing and improving models specifically for the rehospitalization and death.

In the first (primary prevention) phase, predictive models aim to assess stroke risk within the general population but face substantial limitations, with most models achieving ROC scores below 0.7. These models rely on cardiovascular markers such as the AHA Life’s Essential 8 [5] which lack stroke specificity. Adding stroke-specific indicators is typically used for high-risk individuals to enhance prediction accuracy but is impractical for general screening [6].

The second phase focuses on patients receiving hospital care for acute stroke. Here, predictive power increases with detailed clinical data from advanced diagnostics like, brain and head/neck vessel imaging enabling the use of ML models to refine treatment. However, the wide variation in stroke subtypes and specific predictors for each subtype limits the development of a universal model, necessitating predictor-specific models for different stroke etiologies.

In the post-hospitalization stroke period, stroke survivors face heightened risks of complications, recurrence, and rehospitalization. Previous studies have the prevalence of stroke survivors discharged home ranging between 43–92% [712], often requiring ongoing care, rehabilitation, and adherence to treatment plans. A vast number of studies have focused on analyzing individual medical characteristics of stroke patients. However, our research takes a different approach, concentrating on additional variables that capture broader, non-individual characteristics. This study aims to predict 90-day post-discharge outcomes from acute stroke hospitalization by identifying key predictors of mortality and readmission. By leveraging data from multiple sources, the study employs XML models, which integrate clinical factors, SDOH, and health behaviors. Using multiple XML models, we aim to isolate robust, modifiable predictors of 90-day outcomes of death or readmission post-stroke. Our approach underscores XML’s ability to capture complex, non-linear relationships, enhancing interpretability and trust in predictive insights to inform post-stroke care. Therefore, our main research question targets to determine how XML models improve the prediction of 90-day mortality and readmission outcomes following acute stroke hospitalization even when the sample size is small and the design is unbalanced.

2. Materials and methods

We adopt the Weighted Importance Score and Frequency Count (WISFC) framework for aggregating feature‐importance across multiple models, as recently detailed [13]. This approach systematically combines magnitude and consistency information from diverse explainers to produce a robust, consensus‐based ranking of predictors.

2.1. Study population

Our study population is comprised of 1,300 post-stroke individuals enrolled in the Transitions of Care Stroke Disparities Study (TCSD-S) (NIH/NIMH, NCT03452813) between June 2018-October 2022 combined by information from participating comprehensive stroke centers from the Florida Stroke Registry [14]. The TCSD-S is an observational study of stroke survivors investigating factors that influence successful transitions post-stroke. Eligible participants were adults aged 18 years and older, diagnosed with either acute ischemic stroke or intracerebral hemorrhage, and discharged to either a rehabilitation facility or directly home. Although ischemic and intracerebral hemorrhage (ICH) strokes differ pathophysiologically, in our cohort, both groups had relatively low Modified Rankin Scale scores at discharge and showed no statistically significant difference in 90-day outcome rates. Given this similarity, and to preserve sample size and generalisability, both stroke types were retained in the analysis. Hospital care coordinators conducted interviews at discharge to assess SDOH. Follow-up structured interviews at 30 and 90 days post-discharge tracked readmissions, emergency room visits, discharge education, and behavioral modifications. The TCSD-S enrollee data from 10 collaborating comprehensive stroke centers were linked to the American Heart Association’s Get With The Guidelines–Stroke (GWTG-S) Database, providing additional clinical and demographic information such as race/ethnicity, sex, age, insurance status, stroke severity, and pre-stroke health conditions. Patients who died within 30 days or had specific post-discharge dispositions (e.g., transferred to hospice care) were excluded from the study.

Only those patients who provided written informed consent participated in the TCSD-S. The study protocol was approved by the Institutional Review Board of the University of Miami Protocol ID20170892.

2.2. Study variables

The primary outcome variable was a composite of hospital readmission or mortality within 90 days after discharge from the index hospitalization. Since approximately half of all readmissions within one year after a stroke occur within the first 90 days [1517], predicting 90-day outcomes is useful.

Independent variables were extracted from three sources: the TCSD-S which provides insight into individual SDOH, the GWTG-Stroke database provides clinical and index stroke characteristics, and the Social Contextual Indicators for Research and Analysis (SCIERA) database [18]. The database provides neighborhood-level SDOH factors (including detailed zip code characteristics), offering a broader context for individual patient data.

We included nine comprehensive arrays of potential predictors encompassing neighborhood socio-economic, and demographic characteristics, individual social determinants of health, health characteristics before the stroke, index stroke characteristics, acute care variables, hospital characteristics, discharge status, and measures of Adequate Transition of Care [19]:

  1. Patients’ sociodemographic Characteristics: Age, sex, race/ethnicity, and type of insurance.

  2. Individual SDOH: Language spoken at home, education status, prior work status, difficulty paying for medical care, difficulty paying for necessities (e.g., food, electricity), living arrangements, and social support [20].

  3. Health Characteristics Before the Stroke: Prior ambulation status, diabetes, hypertension, dyslipidemia, atrial fibrillation, peripheral vascular disease, coronary artery disease, previous stroke/Transient Ischemic Attack (TIA), carotid stenosis, chronic renal insufficiency, sleep apnea, depression, smoking status, drug/alcohol use, overweight/obesity, and history of deep vein thrombosis/pulmonary embolism (DVT/PE).

  4. Index Stroke Characteristics: Stroke etiology, National Institutes of Health Stroke Scale (NIHSS) score, final clinical diagnosis related to stroke, presence of weakness/paresis, altered level of consciousness, aphasia/language disturbance, other neurological signs/symptoms, mode of arrival (e.g., ambulance, self-transport), and length of hospital stay.

  5. Acute Care Variables: intravenous thrombolysis, endovascular therapy, discharge medications including antiplatelet agents, anticoagulants, and statins, provision of defect-free care, and presence of active bacterial or viral infection at admission or during hospitalization.

  6. Hospital Characteristics: Stroke center type (e.g., primary stroke center, comprehensive stroke center), teaching hospital status, and the number of beds.

  7. Discharge Status: Modified Rankin Scale (mRS) score at discharge, discharge disposition (e.g., home, rehabilitation facility), and ambulation status at discharge.

  8. Neighborhood (Zip-Code) Characteristics: Percentages of Hispanic, Non- Hispanic Black, and Non-Hispanic White residents; percentage of residents with a bachelor’s degree; median household income; percentage of high school completion; rural-urban commuting area codes; total housing population; the ratio of owner-occupied housing; housing density [21]; percentage below the poverty line; unemployment rate; densities of tobacco, alcohol, restaurant, fast food, grocery, pharmacy, and gym businesses; and counts of hospitals, clinics, and rehabilitation centers in the area. Crowding, in our context, is a measure of population density within a zip code. It is calculated by dividing the total housing population by the adjusted number of housing units, which is determined by dividing the count of housing units by the median number of rooms per unit. This metric provides insight into how densely populated a given area is concerning its housing capacity. Total Housing Population is defined as the total number of people living in owner or renter-occupied housing in a zip code. Social Support Size is the measure of how many persons a patient knows that they feel close to (i.e., persons they can talk to or reach out to if needed with 3 categories: “None”, “1-2”, “3 or more”).

  9. Adequate Transition of Care (ATOC): Adherence to at least 75% of applicable transition of care behavior modifications, including filled medications and taken as prescribed 90–100% of the time; attending outpatient therapy or attended and completed therapy if prescribed; has seen a medical provider after discharge; stopped using tobacco, alcohol, marihuana, and other drugs; exercising by regular walking on a treadmill or outside, or regular exercise other than walking; modified diet per recommendation after stroke.

The methods for obtaining outcome data and adjudicating endpoints have been detailed in the prior publication [19].

2.3. XML techniques and implementation

To predict post-discharge stroke outcomes, we applied a comprehensive array of XML techniques aimed at identifying the most important predictive variables from a pool of diverse clinical, socioeconomic, and behavioral data detailed above. This section outlines the methodology used to handle the 73 potential variables, the application of 11 different XML models, and the process of ranking variable importance.

Step 1: Identifying and Selecting Variables: The 73 study variables were selected as they represent strong candidates based on prior evidence [14,19,22] and their availability in the dataset. As detailed in the previous section, they were drawn directly from multiple data sources and span key domains such as sociodemographic characteristics, social determinants of health, pre-stroke health status, stroke-specific and acute care data, hospital and discharge characteristics, neighborhood factors, and transition of care adequacy. By incorporating these diverse variables, we ensured a comprehensive basis for modeling post-stroke recovery and readmission risk.

Step 2: Applying XML Methods: To process these 73 variables, we employed 11 different XML models. We use logistic regression as a universal benchmark, alongside 10 additional models encompassing regression-based, tree-based, and distance-based algorithms. These 10 models represent state-of-the-art approaches widely accepted in ML, ensuring comprehensive coverage of predictive methodologies. This method uses one data set randomly divided into 10 parts. Nine of those parts are used for training and a tenth for testing. This procedure is repeated 10 times reserving a different tenth for testing [23]. Each XML model training was carried out under 10-fold cross-validation to optimize hyperparameters and estimate out-of-sample performance. All continuous predictors in the training set were then mean-centered and scaled to unit variance; the same centering and scaling parameters were subsequently applied to the test set to avoid data leakage. No additional feature-engineering transformations were performed beyond this normalization step.

All candidate predictors were assessed for completeness prior to model development. Patients with missing values in any of the predictors listed in Table 1 were excluded from the analysis; no imputation methods were applied. Of the original cohort of 1,200 stroke patients, 73 (6.1%) were removed due to incomplete records, yielding a final sample of 1,127 patients with fully observed data.

Table 1. Participants’ demographic and stroke characteristics – selected noteworthy variables.

Selected Variables Total (N = 1300) Death or readmission in 90 days P Value
NO (N = 1094) YES (N = 206)
Age, mean (SD) 63.8 (13.9) 64.0 (13.9) 62.7 (14.3) 0.23
Male sex 732 (56.3%) 621 (56.8%) 111 (53.9%) 0.44
Ischemic stroke 1196 (92.0%) 1011 (92.4%) 185 (89.8%) 0.21
ICH 104 (8.0%) 83 (7.6%) 21 (10.2%)
NIH Stroke Scale, median (range) 3 (0–33) 2 (0–33) 4 (0–28) <0.001
mRS baseline, median (range) 1 (0–5) 1 (0–5) 2 (0–5) <0.001
Length of Stay, median (range) 4.1 (0.7–66.3) 4.0 (0.7–41.8) 4.5 (1.1–66.3) <0.001
Diabetes Mellitus 411 (31.6%) 332 (30.4%) 79 (38.4%) 0.023
CAD or prior MI 242 (18.6%) 192 (17.6%) 50 (24.3%) 0.023
Previous Stroke or TIA 270 (20.8%) 207 (18.9%) 63 (30.6%) <0.001
Chronic Renal Insufficiency 98 (7.5%) 65 (5.9%) 33 (16.0%) <0.001
ATOC 832 (64.0%) 728 (66.5%) 104 (50.5%) <0.001
Carotid Stenosis 56 (4.3%) 44 (4.0%) 12 (5.8%) 0.24
Non-Hispanic White race 665 (51.2%) 564 (51.6%) 101 (49.0%) 0.2
Non-Hispanic Black race 296 (22.8%) 244 (22.3%) 52 (25.2%)
Hispanic 286 (22.0%) 246 (22.5%) 40 (19.4%)
Other 53 (4.1%) 40 (3.7%) 13 (6.3%)
Private Insurance 291 (22.4%) 256 (23.4%) 35 (17.0%) 0.046
Medicare 580 (44.6%) 493 (45.1%) 87 (42.2%)
Medicaid 63 (4.9%) 51 (4.7%) 12 (5.8%)
Self/No Insurance 366 (28.2%) 294 (26.9%) 72 (35.0%)
Full-time Employment 472 (36.3%) 409 (37.4%) 63 (30.6%) 0.11
Part-time Employment 116 (8.9%) 91 (8.3%) 25 (12.1%)
Retired 556 (42.8%) 467 (42.7%) 89 (43.2%)
Unemployed 156 (12.0%) 127 (11.6%) 29 (14.1%)
Housing population, median (range) 29875 (569–75180) 29875 (569–75180) 29824 (766–73331) 0.85
Crowding, mean (SD) 0.48 (0.12) 0.48 (0.12) 0.49 (0.12) 0.26

To assess whether model performance differed when restricted to ischemic stroke, we performed the secondary analysis in ischemic stroke subset. We excluded the 89 hemorrhagic cases from our final cohort of 1,127 patients, yielding 1,038 ischemic strokes (92.1%). Predictors specific to stroke subtype (variables 16 and 38) were omitted. Discrimination (AUC) and calibration metrics remained virtually unchanged compared to the primary analysis.

All models were tuned or set under a uniform 10-fold cross-validation framework. For the penalized regressions (LASSO, Ridge, and Elastic Net with α = 0.5), we selected the smallest λ that minimized cross-validated error in each fold. Principal Component Regression was run with a fixed 10 components (k = 10). The k-Nearest Neighbors model evaluated 20 candidate k values via cross-validation. Support Vector Machines employed the radial basis kernel with default cost and γ settings. Random Forests were grown with 500 trees and the default mtry. Gradient Boosting was configured with 500 trees, interaction depth = 3, and shrinkage = 0.1. Finally, XGBoost models were trained for up to 100 rounds (with early stopping after 10 rounds), max_depth = 3, η = 0.3, subsample = 0.8, and colsample_bytree = 0.7.

These models were selected to capture both linear and non-linear relationships within the data and to explore varying perspectives on how the variables influence patient outcomes. The following methods were used:

  • Regression-based algorithms: Logistic regression, LASSO, ridge regression, elastic net [23].

  • Distance-based algorithms: Support Vector Machines (SVM), K-Nearest Neighbors (KNN) [24].

  • Tree-based algorithms: Random Forest, gradient boosting, XGBoost [25].

Each of these models assessed the data from different angles, providing a broad spectrum of variable importance evaluations. However, each of our 11 XML models carries its own assumptions and potential weaknesses, e.g., penalized regressions presume linear relationships and may miss nonlinear effects, PCR depends on variance‐based dimensionality reduction that can overlook low‐variance but predictive features, k-NN is sensitive to feature scaling and local density, SVMs hinge on kernel choice and can struggle with large datasets, and tree-based learners (RF, GBM, XGBoost) may overemphasize variables with many split points or categorical levels. By design, however, our framework remains agnostic to any single method’s limitation: every algorithm contributes its 12 most influential predictors under the same 10-fold CV regime, and we synthesize these into a unified importance profile. Even if one model misses interactions, overfits, or is skewed by sparse data, its particular weaknesses get smoothed out when we combine results from all models.

For regression-based algorithms, variable importance is determined by the magnitude of the coefficients, where larger absolute values of the t-statistics or standardized coefficients indicate greater importance. For tree-based algorithms, importance is assessed using model-specific criteria. In the case of Random Forest, variable importance is measured by the Mean Decrease in the Gini Index, which quantifies how much each variable reduces Gini impurity across all trees. In Gradient Boosting, importance is calculated based on the relative influence of each variable, determined by the reduction in the loss function whenever the variable is used for splitting. In XGBoost, importance is assessed based on the improvement in accuracy contributed by each variable when used for splitting branches. For KNN, variable importance is estimated using a proxy method that examines the impact of variable scaling. For SVM, importance is derived from the coefficients of the hyperplane, with larger absolute values indicating higher importance.

Step 3: Target Outcomes: The goal of applying these methods was to determine the importance of each variable in predicting the two main outcomes: death and hospital readmission.

Step 4: Quantifying Model Performance: We assessed the performance of each XML model using standard evaluation metrics, including accuracy, area under the ROC curve (AUC), and logistic loss. This quantifiable performance estimate ensured that each model’s predictive capacity was considered when interpreting the variable importance rankings.

Step 5: Variable Ranking for Each Method: For each XML method, we produced a ranked list of 73 variables and selected the top 12 variables based on their importance in predicting the target outcomes. The decision to include 12 variables accounts for the traditional “events per variable” rule of logistic regression, which typically supports one significant variable per 20 outcomes. To accommodate variations across different model types and ensure comprehensive coverage of the strongest predictors, we extended the list to include two additional variables. Since each method approaches the data differently, the top 12 variables differed across models. To capture these differences, a comparison of the variable rankings across all methods was created, showing how each method prioritized different variables.

Step 6: Aggregating and Filtering Variables: Based on the rankings generated by each method, some variables appeared in the top 12 across multiple methods, while others were highly ranked by only a few methods. Variables that did not appear in the top 12 for any method were excluded from further analysis. This process left us with 38 key variables for continued evaluation.

Step 7: Weighting Variable Importance: To ensure fair comparison and robust ranking, we ranked variables in 2 ways. The first way reflects the number of times the variable appears among the top 12 variables in each model. The second way assigns point weights ranging from 12 to 1 for each variable, according to their rank order within each model, ensuring that variables consistently ranked highly across multiple methods are given more importance.

Step 8: Final Ranking of Variables: The outcome of this process was a ranked list of variables that were considered most important in predicting post-discharge stroke outcomes. These top variables provide insights for healthcare providers to focus on during the 90-day post-stroke recovery period.

2.4. Evaluation metrics and statistical analysis

To assess the performance of the predictive models, several evaluation metrics were calculated: Accuracy [26], C-statistic (Area Under the ROC Curve) [27], Squared-Error Loss (Mean Squared Error, MSE) [28], Logistic Loss (Log Loss or Cross-Entropy Loss) [29] and Misclassification Rate [26]. These metrics provide a comprehensive assessment of each model’s strengths and weaknesses in terms of discrimination and calibration.

All statistical methods (including ML) suffer from class imbalance [30], which occurs when the distribution of classes in a dataset is highly skewed, leading to challenges in model training and performance as the algorithm may favor the majority class while neglecting the minority class. Applying multiple ML models and the creation of the combined predictor list offers a solution to class imbalance.

Continuous variables were summarized using means and standard deviations or medians with interquartile ranges, depending on their distribution. Categorical variables were presented as counts and percentages. Comparisons between patient groups were conducted using Chi-square tests for categorical variables and t-tests or Mann–Whitney U tests for continuous variables, as appropriate. Paired t-tests were used to compare the averages of estimates from different dataset pairs. All statistical analyses were performed using R version 4.4.

To avoid relying on any one method’s own biases, we employed 11 distinct XML models (regression-based models, tree-based models, and distance-based models), each of which independently identified its top 12 predictors and evaluated performance across multiple metrics via 10-fold cross-validation. This consensus‐driven approach ensures that our findings are not dominated by one method’s assumptions or parameter settings, but instead reflect variables that consistently emerge as influential across all algorithms. The result is a more stable and broadly applicable set of drivers for the outcome.

2.5. Creation of the combined predictor list

The list creation starts with the extraction of top predictors: from each model, we extracted the top 12 predictors. The aggregation combines the predictors from all models into a single list producing two ranking methods: Weighted Importance Scores, which assigned weights based on predictor rank and summed them across models, and Frequency Counts, which tracked the number of appearances in the top 12 predictors across all models. Final rankings are based on these scores, highlighting factors linked to 90-day readmission or mortality representing a merged outcome derived from two distinct methods.

The created and sorted list of the 38 strongest predictors combines the top 12 variables from each 73 sorted variables of the 11 models. The list captures insights from diverse modeling perspectives. This approach mitigates the lack of established consensus on the optimal method for combining predictors from different models to generate a comprehensive list [31,32]. The resulting aggregation method combines Weighted Importance Scores and Frequency Counts to ensure a balanced and robust selection process.

3. Results

A total of 1,300 stroke survivors were included in the analysis. The cohort had a mean age of 63.8 years (SD = 13.9), and 56% of the participants were male. The ethnic composition included 22% Hispanic, 23% Non-Hispanic Black, and 51% Non-Hispanic White individuals. Ischemic strokes accounted for 92% of cases, while 8% were intracerebral hemorrhages (ICH). The overall 90-day readmission or mortality rate was 15.8%, affecting 206 of the 1,300 patients. Table 1 summarises the demographic and stroke-related characteristics of the study population, stratified by the presence of the 90-day outcome (readmission or mortality).

Out of the 1,300 patients included in the study, 206 experienced an adverse 90-day outcome (either readmission or death) while 1,094 remained event-free. Specifically, 22 patients died and 187 were readmitted to the hospital within 90 days post-discharge.

To ensure the integrity of our analyses, we excluded the 73 patients with missing values and retained only the 1,227 individuals with complete data for all covariates. All subsequent analyses were therefore based on these 1,227 patients.

Table 2 summarizes the 10-fold cross-validation model fit statistics, including c-statistic, squared-error loss, logistic loss, misclassification rate, precision, recall and F1 across all models. The best values for each metric are highlighted in bold. Ridge Regression demonstrated the highest discriminative ability with a c-statistic of 0.660, while LASSO and Elastic Net excelled in capturing outcome probabilities, achieving a logistic loss of 0.414. Principal Component Regression (PCR) showed its effectiveness in minimizing classification errors with the lowest misclassification rate of 0.156.

Table 2. Model Fits for Algorithms with 10-fold cross-validation.

Algorithms c-Statistic Squared
error Loss
Logistic
Loss
Misclassification
Rate
Precision Recall F1 Score
Logistic Regression 0.615 0.144 0.533 0.181 0.578 0.541 0.603
Forward Selection 0.616 0.134 0.531 0.160 0.747 0.509 0.683
LASSO 0.655 0.126 0.414 0.158 0.725 0.502 0.841
Ridge 0.660 0.126 0.427 0.157 0.745 0.503 0.842
PCR 0.595 0.130 0.434 0.156 0.858 0.503 0.843
Elastic Net 0.651 0.126 0.414 0.159 0.725 0.502 0.841
KNN 0.612 0.131 0.507 0.157 0.767 0.500 0.915
SVM 0.622 0.145 0.557 0.157 0.843 0.500 0.915
Random Forest 0.642 0.128 0.423 0.158 0.804 0.506 0.784
Gradient Boosting 0.607 0.146 0.555 0.173 0.594 0.545 0.546
XGBoost 0.619 0.143 0.536 0.174 0.556 0.515 0.564

3.1. Variable importance

Table 3 presents the top 12 variables for each XML model, while Table 4 aggregates the importance of these variables across all models. These results highlight a combination of clinical and non-clinical factors that significantly impact patient outcomes after a stroke. Each model evaluates the contribution of different variables to predicting 90-day outcomes, resulting in rankings for each algorithm.

Table 3. Participants’ demographic and stroke characteristics – selected noteworthy variables.

Rank Regression-Based Algorithms
Logistic Regression Forward Selection LASSO Ridge Regression PCR Elastic Net
1 Chronic renal insufficiency 75% ATOC Chronic renal insufficiency Chronic renal insufficiency Length of stay Chronic renal insufficiency
2 75% ATOC Chronic renal insufficiency Discharge location Discharge location 75% ATOC Carotid stenosis
3 mRS at discharge Insurance 75% ATOC Active infection NIHSS Prior ambulation
4 Carotid stenosis mRS at discharge mRS at discharge Carotid stenosis Insurance Discharge location
5 Discharge location Employment status Insurance Prior ambulation Chronic renal insufficiency Race/Ethnicity
6 Race/Ethnicity Carotid stenosis Length of stay Race/Ethnicity mRS at discharge mRS at discharge
7 Insurance Crowding NIHSS mRS at discharge Rehab count Active infection
8 Prior ambulation Total housing population Age Ambulation at discharge Crowding 75% ATOC
9 Altered level of consciousness Discharge location Sex 75% ATOC Ambulation at discharge Ambulation at discharge
10 Ambulation at discharge %Black Race/Ethnicity Altered level of consciousness Median house income Altered level of consciousness
11 Drugs/Alcohol Race/Ethnicity Etiology Insurance Social support Insurance
12 Difficulty paying Medicare Social support Stroke type Difficulty paying Medicare Age Difficulty paying Medicare
Rank Distance-Based Algorithms Tree-Based Algorithms
KNN SVM Random Forest Gradient Boosting XGBoost
1 75% ATOC 75% ATOC Length of stay Length of stay Length of stay
2 mRS at discharge mRS at discharge Age mRS at discharge Total housing population
3 Length of stay Length of stay mRS at discharge NIHSS Age
4 Ambulation at discharge Ambulation at discharge NIHSS Chronic renal insufficiency NIHSS
5 Insurance Insurance Etiology Age mRS at discharge
6 NIHSS NIHSS Median house income Discharge location Etiology
7 Chronic renal insufficiency Chronic renal insufficiency Total housing population Insurance Housing density
8 Mode of arrival Mode of arrival %Bachelors Median house income Crowding
9 Age Age %High school Crowding %Hispanic
10 %White %White Housing density Gym business density Alcohol business density
11 Rehab count Rehab count Crowding %Bachelors %Below poverty
12 Diabetes Diabetes %Unemployment %White Restaurant business density

Table 4. Ranked the strongest variables across all 11 models.

ID Variable Count Sum R1 R2 ID Variable Count Sum R1 R2
1 mRS at discharge 11 99 1 1 20 Total housing population 3 22 20 14
2 Chronic renal insufficiency 9 88 2 2 21 %Bachelors degree 2 7 21 24
3 Insurance 9 59 3 5 22 Active infection 2 16 22 16
4 75% ATOC 8 77 4 3 23 Diabetes 2 2 23 34
5 Age 7 43 5 8 24 Housing density 2 9 24 21
6 Length of stay 7 75 6 4 25 Mode of arrival 2 10 25 20
7 NIHSS 7 58 7 6 26 Social support size 2 3 26 30
8 Ambulation at discharge 6 34 8 10 27 %Below poverty 1 2 27 35
9 Discharge location 6 50 9 7 28 %Black 1 3 28 29
10 Crowding 5 22 10 13 29 %High school 1 4 29 26
11 Race/Ethnicity 5 27 11 11 30 %Hispanic 1 4 30 27
12 Carotid stenosis 4 36 12 9 31 %Unemployment 1 1 31 37
13 %White 3 7 13 23 32 Alcohol business density 1 3 32 32
14 Altered level of consciousness 3 10 14 18 33 Drugs/Alcohol 1 2 33 33
15 Difficulty paying Medicare 3 3 15 28 34 Employment status 1 8 34 22
16 Etiology 3 17 16 15 35 Gym business density 1 3 35 31
17 Median household income 3 15 17 17 36 Restaurant business density 1 1 36 38
18 Prior ambulation 3 23 18 12 37 Sex 1 4 37 25
19 Rehab count 3 10 19 19 38 Stroke type 1 1 38 36

3.2. Summary of results

Table 4 provides a summary of the top variables across all 11 models. It ranks these 38 variables based on the number of models in which they appeared (count) among the top 12 predictors (R1). The cumulative ranking approach (sum) identifies variables consistently significant across models (R2), emphasizing their critical role in predicting 90-day outcomes for stroke survivors. The emerging socio-contextual risk determinants are shown in bold.

3.3. Sensitivity analysis

To assess the robustness of our findings, we repeated all analyses under three alternative scenarios, i.e., predicting 90-day readmission only; repeating the combined endpoint analysis in the ischaemic stroke subgroup; and predicting readmission only in the ischaemic subgroup. In all cases, the top predictors remained consistent with those reported above, and model discrimination and calibration measures show minimal deviation from the primary results (data not shown, but further discussed).

4. Discussion

Our study provides a unique look at estimating death and readmission risks in the first 90 days post-stroke, going beyond the typical clinical risk factors, and integrating 3 large data sets to provide a comprehensive view of the impact of clinical, individual social, and community-level risk factors on transitions of care. When only regression analysis models are used to model risks, they are limited to using a small number of predictors that operate in the same way on everyone, and uniformly throughout their range [33].

For studies where the goal is to predict the occurrence of an outcome and not measure the association between specific risk factors and an event in a clinically interpretable way, traditional regression models can be modified or abandoned in favour of models that

produce a more flexible relationship among the predictor variables and the outcome [33]. These methods have similar goals to regression-based approaches but different motivating philosophies (Fig 1). They do not require pre-specification of a model structure but instead search for the optimal fit within certain constraints (specific to the individual algorithm). This can result in a better final prediction model at the sacrifice of interpretability of how risk factors relate to the outcome of interest.

Predictive models can serve both explanatory and predictive purposes [34]. In our work, we focus on the former: identifying a concise set of explanatory variables without delving into the precise functional relationships among them. While this strategy does not eliminate multicollinearity, it has only a marginal effect on the robustness of our variable selection and on the overall interpretability of the resulting model.

By adding 10 XML models, our study advances the understanding of 90-day post-stroke outcomes by integrating clinical and non-clinical factors, particularly SDOH, into predictive models. By focusing on stroke-specific and non-classic predictors, we deliver a unified, systematic, and targeted perspective, advancing the precision and relevance of predictive modeling in stroke research. The stroke-specific list and findings underscore the value of Explainable XML models in identifying and ranking predictors that consistently influence readmission and mortality, thereby bridging gaps in traditional risk prediction frameworks. These XML models also assist medical practitioners in evaluating the validity of a diagnosis while ensuring the output is interpretable and comprehensible, even for patients [35]. By focusing on modifiable factors during the critical post-hospitalization period, this study contributes actionable insights for improving patient outcomes and resource allocation.

4.1. Clinical and non-clinical predictors

Key predictors of post-stroke outcomes, such as stroke severity, age, comorbidities (e.g., coronary artery disease) [36], active infection [37,38] and length of hospital stay, reinforce established clinical knowledge [39]. SDOH variables, such as socioeconomic status, education level, and neighborhood characteristics, emerged as significant contributors to model accuracy, aligning with prior findings [40], which demonstrated that incorporating SDOH improved mortality prediction for non-Hispanic Black patients with heart failure (HF). Similarly, our study found that integrating non-clinical factors enhanced predictive power, supporting the need for tailored, equitable healthcare strategies.

We also confirm the previously indicated importance of social determinants, such as (Medicare) insurance [17,41], housing status, social support, educational level, and employment status [42]. In particular, the role of an individual’s social support network has been shown to significantly impact functional recovery after stroke, especially within the first three months post-discharge. Future research should explore whether interventions outside the medical context, such as transitional care resources, community health workers, or rehabilitation strategies that strengthen social networks and incorporate group activities with family or caregivers, could mitigate the effects of limited social networks and enhance recovery outcomes for stroke survivors [42].

4.2. Machine learning methodologies

We employ 11 XML models, including tree-based algorithms like Random Forest and XGBoost, which excelled in capturing non-linear relationships within the data. The use of interpretable models provides transparency, fostering clinician trust and aligning with the recommendations by [43] and [44] for adopting trustworthy and explainable AI in healthcare. ML methods were also applied during patients’ initial hospitalization to identify those at high risk for readmission or mortality [45]. By synthesizing results through Weighted Importance Scores and Frequency Counts, this study provides a robust cumulative ranking of variable importance, extending beyond the limitations of traditional regression methods. This approach is novel, as the prevailing practice involves using multiple ML exploratory models in Phase 2 to predict clinical outcomes [46], without combining or synthesizing findings across different models. Our approach creates two distinct ranks, addressing the common challenge of interpreting results when multiple models are employed, a problem that remains unresolved in existing methodologies, some of which are shown in [47].

Over the past few years, AI has become an industry disruptor: as we demonstrate, it can more accurately refine and streamline data prediction; future AI processes may create patient-specific, guideline-based treatment plans as well, however, these projects must occur securely and ethically [48].

While we did not apply explicit resampling or class‐weighting to rebalance the training data, our assessment framework inherently counteracts imbalance in two ways. First, by integrating predictions from eleven diverse XML models—each with different inductive biases and error‐minimization strategies—we avoid overreliance on any single algorithm’s tendency to favor the majority class. Second, we evaluated every model using multiple performance metrics that capture different aspects of predictive quality: the c-Statistic (AUC) for discrimination, squared-error loss and logistic loss for probabilistic calibration, and misclassification rate for raw accuracy. By insisting that strong performance be sustained across all four metrics, we ensure that a model cannot succeed simply by predicting the majority class. Together, this approach provides a robust safeguard against the distortions introduced by class imbalance.

4.3. The role of SDOH and disparities in stroke outcomes

In the absence of relevant socio-environmental variables, XML models show only a modest improvement in prediction performance and explainability compared to traditional modeling techniques, even in cases where complete relevant electronic records are available (e.g., Sweden) [49]. Our findings emphasize the impact of SDOH on stroke outcomes, echoing evidence from the [40] study, which highlighted disparities in SDOH-related predictive improvements between Black and non-Black patients. While clinical predictors remain essential, the integration of neighborhood-level SDOH variables, such as access to healthcare and socioeconomic stability, offers a more holistic perspective on patient risk, paving the way for tailored interventions to address health disparities and community-level interventions. This represents a significant advancement in the field. While many studies focus on identifying disparities and social needs, the critical next step is to develop and implement solutions to address these needs effectively.

Future public health policies need to address the heightened mortality and readmission rates among stroke survivors from vulnerable areas highlighting the need to enhance care transitions and support for underserved patient populations.

4.4. Clinical implications

Prior analyses evaluating long-term stroke outcomes were based on electronic medical records [50] that do not include detailed SDOH data. Our study considers socio-environmental variables that significantly impact stroke survivors’ lives. By identifying high-impact variables, healthcare providers can prioritize high-risk patients, personalize preventive and rehabilitation strategies, and optimize resource allocation. The integration of XML models into clinical workflows may offer real-time risk assessment capabilities, enabling more proactive care management. These models also empower clinicians with interpretable outputs, bridging the gap between advanced analytics and practical decision-making, as recommended by other experts [44].

4.5. Strengths and limitations

Our use of a diverse stroke cohort from the state of Florida and a two-layer nested cross-validation approach ensures methodological robustness. However, the reliance on regional data may limit generalizability, and less than 10% of the cohort had acute intracerebral hemorrhage so results are not generalizable to this population. Although our cohort is geographically confined to Florida and predominantly ischemic stroke (92.1%), the secondary ischemic‐only analysis demonstrated equivalent performance, supporting the model’s robustness in this common stroke subtype. Limited individual-level SDOH data constrains the depth of insights into disparities. The Florida-specific dataset limits generalizability and challenges such as class imbalance and missing data persist. Validation in diverse populations and integration of real-time data could enhance predictive accuracy and enable dynamic care adjustments.

We conducted sensitivity analyses restricting to readmission only and to the ischaemic stroke subgroup, which confirmed that our main results, which focus on the prominence of SDOH factors, were robust to endpoint definition and stroke subtype. We have not included all of these additional tables here, but we plan to include them in a dedicated follow-up study.

Future research should aim to validate these models across diverse populations and incorporate real-time, individual-level data to enhance prediction accuracy and equity in care delivery.

While there are signs that similar predictors are vital across geographies, indicators of stroke outcomes and care may vary significantly between high-income and low-income countries due to differences in acute stroke management, post-stroke care, rehabilitation practices, and methodological approaches, making direct comparisons challenging [51].

Despite XML models’ methodological robustness, the study limitations are drawn from the sample size (n = 1,300), which is moderate relative to the number of potential predictors. This may constrain the detection of weaker associations and limit model generalizability. Second, the data are derived from stroke centers within a single U.S. state (Florida), which may introduce regional biases related to healthcare delivery, socioeconomic conditions, and demographic composition. As a result, the findings may not fully generalize to populations with different healthcare systems, geographic contexts, or stroke care practices. Future research should aim to validate these results using larger, multi-regional datasets that reflect broader clinical and social diversity.

Our cohort’s adverse outcome rate of 15.8% reflects the typical imbalance seen in post‐stroke prognostic studies. All stroke‐prediction models face this challenge, and few have achieved a ROC‐AUC substantially above 0.7 without risking overfitting. Since our goal was to evaluate an array of modeling methods rather than to fine‐tune a single algorithm, we did not apply formal imbalance‐correction techniques such as SMOTE or differential class‐weighting. Consequently, precision and recall (particularly for the minority class) are inherently constrained by the low event rate. Future work may explore resampling strategies or threshold adjustment to optimize these metrics, but such approaches must be balanced against potential bias and reduced generalizability.

4.6. May modeling help post-acute stroke care

The 15.8% adverse‐outcome rate in our cohort reflects the low‐prevalence challenge faced by all stroke‐prognosis models. Historically, no stroke‐prediction model has achieved an ROC‐AUC substantially above 0.7 without risking overfitting. Because our primary aim was to benchmark multiple modeling frameworks rather than to fine‐tune one classifier, we did not implement SMOTE, differential class weighting, or other resampling strategies. As a result, precision and recall (particularly for the minority adverse‐outcome class) remain limited, with recall values around 0.50. While these metrics are critical for clinical decision‐making, their potential improvement may necessitate imbalance‐focused methods that carry their own risk of bias and reduced generalizability.

In this paper, our focus is on answering three fundamental questions for post‑acute stroke care coordination: what interventions to implement, who should carry them out, and in which settings. Once a patient returns to their changed pre‑stroke environment, most medical and clinical factors are already addressed through established protocols; what remains poorly understood is the broader exposome, the totality of exposures and conditions a patient experienced prior to and after their stroke. We approximate these influences using variables commonly grouped under SDOH. Notably, 19 of the 38 top‑ranked predictors in our analysis represent modifiable or preventable socioeconomic factors. We do not claim to establish causal hierarchies or precise effect sizes for these variables, but we believe they highlight critical opportunities for improving post‑acute care coordination, an aspect often neglected in current practice.

We calculated precision, recall, and F1-score for each model and reported them in Table 3. Across models, recall values hover around 0.50, indicating that approximately half of the true adverse outcomes are identified (a characteristic consequence of the 15.8% event rate and consistent with prior stroke-prediction studies). Although these threshold-dependent metrics are critical for understanding the performance of the minority class, the low prevalence constrains improvements without risking overfitting. As our goal was to compare modeling frameworks rather than to implement imbalance-specific corrections, we have not pursued additional resampling or class-weighting strategies here.

4.7. Value of social determinants of health in post‐discharge prediction

As shown in Table 4, we find that 19 (marked in bold) of the 38 candidate predictors are SDOH variables. While clinical measures such as NIHSS score and comorbidity burden are paramount during the acute and early recovery phases, their prognostic influence attenuates once patients leave the hospital environment. In contrast, modifiable SDOH factors (housing security, access to care, social support networks, and neighborhood socioeconomic status) emerge as increasingly salient drivers of long‐term outcomes. In our permutation‐importance analysis, multiple SDOH predictors placed within the top ten features for several modeling approaches, underscoring their empirical contribution to discrimination and calibration. This finding aligns with prior studies from our group demonstrating SDOH’s role in shaping post‐stroke functional recovery and readmission risk [52,53]. Collecting SDOH data may impose additional burden, but these variables capture patient exposome dimensions essential for accurate, real‐world prognostication.

5. Conclusion

This study identifies several predictors of 90-day readmission or mortality in stroke survivors using explainable Machine Learning (XML) models, which outperform conventional regression techniques. These findings enable more accurate identification of high-risk patients and support tailored post-discharge care. With over seven million U.S. stroke survivors, leveraging predictive tools and expanding data sources can make a significant impact on the national stroke burden by refining interventions, improving outcomes, and inspiring innovation in care strategies. Machine learning models hold the potential to better harness large datasets across all phases, paving the way for unified, phase-spanning models that guide stroke risk reduction and improved recovery from onset to long-term care.

By applying eleven different modeling frameworks, we aim to present multiple perspectives on post-acute stroke care rather than to rank any one algorithm as inherently superior. Instead, these models shed light on diverse aspects of patient life and treatment pathways, particularly environmental exposures and Social Determinants Of Health (SDOH), that remain poorly understood. In a field urgently in need of better predictive tools and interpretability, our work demonstrates how ML methods can offer valuable preliminary insights to guide future research and intervention design when no prevailing explanatory frameworks yet exist.

With this knowledge, we plan to create a solution that can be integrated into Electronic Health Record (EHR) systems to harness diverse data sources (e.g., hospital EHRs, the Florida Stroke Registry, and socioeconomic datasets) and build a dynamic risk scoring system and Digital Twin model that predicts stroke outcomes and guides treatment decisions in real-time. By integrating established prognostic scores with guideline-driven management trees and leveraging real-world data, the future solution will deliver personalized, evidence-based care.

Acknowledgments

No acknowledgments are applicable to this manuscript.

Data Availability

The FSR uses data from Get with The Guidelines-Stroke® (GWTG-S). Due to data-sharing agreements, researchers must apply for access at http://www.heart.org/qualityresearch, with proposals reviewed by GWTG-S and FSR committees upon reasonable request.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Institute for Health Metrics and Evaluation (IHME). Global Burden of Disease 2021: Findings from the GBD 2021 Study. Seattle: Institute for Health Metrics and Evaluation (IHME). 2024. [Google Scholar]
  • 2.Norrving B, Davis SM, Feigin VL, Mensah GA, Sacco RL, Varghese C. Stroke prevention worldwide–what could make it work?. Neuroepidemiology. 2015;45:215–20. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Saxena A, McGranaghan P, Das S, Rubens M, Salami J, Veledar E. Abstract WP242: Modeling Prevention, Prediction or Explanation of Stroke Risk: Insights From Meta-Analysis of Studies With Models Predicting Stroke or Composite Outcomes. Stroke. 2019;50:AWP242–AWP242. [Google Scholar]
  • 4.Saxena A, Zhang Z, Ahmed MA, McGranaghan P, Rubens M, Veledar E. Abstract TMP98: Insights From Meta-analysis Of Studies With Models Predicting Stroke Or Composite Outcomes: A 2021 Study Update. Stroke. 2022;53(Suppl_1). doi: 10.1161/str.53.suppl_1.tmp98 [DOI] [Google Scholar]
  • 5.Lloyd-Jones DM, Allen NB, Anderson CAM, Black T, Brewer LC, Foraker RE, et al. Life’s Essential 8: Updating and Enhancing the American Heart Association’s Construct of Cardiovascular Health: A Presidential Advisory From the American Heart Association. Circulation. 2022;146(5):e18–43. doi: 10.1161/CIR.0000000000001078 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Islam U, Mehmood G, Al-Atawi AA, Khan F, Alwageed HS, Cascone L. NeuroHealth guardian: A novel hybrid approach for precision brain stroke prediction and healthcare analytics. J Neurosci Methods. 2024;409:110210. doi: 10.1016/j.jneumeth.2024.110210 [DOI] [PubMed] [Google Scholar]
  • 7.Cho J, Place K, Salstrand R, Rahmat M, Mansouri M, Fell N, et al. Developing a Predictive Tool for Hospital Discharge Disposition of Patients Poststroke with 30-Day Readmission Validation. Stroke Res Treat. 2021;2021:5546766. doi: 10.1155/2021/5546766 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Albert GP, McHugh DC, Roberts DE, Kelly AG, Okwechime R, Holloway RG, et al. Hospital Discharge and Readmissions Before and During the COVID-19 Pandemic for California Acute Stroke Inpatients. J Stroke Cerebrovasc Dis. 2023;32(8):107233. doi: 10.1016/j.jstrokecerebrovasdis.2023.107233 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Bettger JP, Thomas L, Liang L. Comparing recovery options for stroke patients. 2019. [PubMed]
  • 10.Schrage T, Thomalla G, Härter M, Lebherz L, Appelbohm H, Rimmele DL, et al. Predictors of Discharge Destination After Stroke. Neurorehabil Neural Repair. 2023;37(5):307–15. doi: 10.1177/15459683231166935 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Stein J, Borg-Jensen P, Sicklick A, Rodstein BM, Hedeman R, Bettger JP, et al. Are Stroke Survivors Discharged to the Recommended Postacute Setting?. Arch Phys Med Rehabil. 2020;101(7):1190–8. doi: 10.1016/j.apmr.2020.03.006 [DOI] [PubMed] [Google Scholar]
  • 12.Clery A, Bhalla A, Bisquera A, Skolarus LE, Marshall I, McKevitt C, et al. Long-Term Trends in Stroke Survivors Discharged to Care Homes: The South London Stroke Register. Stroke. 2020;51(1):179–85. doi: 10.1161/STROKEAHA.119.026618 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Veledar E, Zhou L, Veledar O, Gardener H, Gutierrez CM, Romano JG, et al. Synthesizing Explainability Across Multiple ML Models for Structured Data. Algorithms. 2025;18(6):368. doi: 10.3390/a18060368 [DOI] [Google Scholar]
  • 14.Johnson KH, Gardener H, Gutierrez C, Marulanda E, Campo-Bustillo I, Gordon Perue G, et al. Disparities in transitions of acute stroke care: The transitions of care stroke disparities study methodological report. J Stroke Cerebrovasc Dis. 2023;32(9):107251. doi: 10.1016/j.jstrokecerebrovasdis.2023.107251 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.National Institutes of Health. MyRIAD: Atrial Cardiopathy and Antithrombotic Drugs in Prevention After Cryptogenic Stroke.
  • 16.Benbassat J, Taragin M. Hospital readmissions as a measure of quality of health care: advantages and limitations. Arch Intern Med. 2000;160(8):1074–81. doi: 10.1001/archinte.160.8.1074 [DOI] [PubMed] [Google Scholar]
  • 17.Zhou LW, Lansberg MG, de Havenon A. Rates and reasons for hospital readmission after acute ischemic stroke in a US population-based cohort. PLoS One. 2023;18(8):e0289640. doi: 10.1371/journal.pone.0289640 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.SCIERA Inc. Your smart marketing intelligence partner. https://www.sciera.com. 2021.
  • 19.Dong C, Wang K, Di Tullio MR, Gutierrez C, Koch S, García EJ, et al. Disparities and temporal trends in stroke care outcomes in patients with atrial fibrillation: The FLiPER-AF stroke study. Int J Cerebrovasc Dis Stroke. 2019;2:19. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Griffith DM, Towfighi A, Manson SM, Littlejohn EL, Skolarus LE. Determinants of Inequities in Neurologic Disease, Health, and Well-being: The NINDS Social Determinants of Health Framework. Neurology. 2023;101(7 Suppl 1):S75–81. doi: 10.1212/WNL.0000000000207566 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Badland H, Foster S, Bentley R, Higgs C, Roberts R, Pettit C, et al. Examining associations between area-level spatial measures of housing with selected health and wellbeing behaviours and outcomes in an urban context. Health Place. 2017;43:17–24. doi: 10.1016/j.healthplace.2016.11.003 [DOI] [PubMed] [Google Scholar]
  • 22.Kapral MK, Wang H, Mamdani M, Tu JV. Effect of socioeconomic status on treatment and mortality after stroke. Stroke. 2002;33(1):268–73. doi: 10.1161/hs0102.101169 [DOI] [PubMed] [Google Scholar]
  • 23.Hastie T, Tibshirani R, Friedman J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction. 2nd ed. New: Springer; 2009. [Google Scholar]
  • 24.Murphy KP. Machine learning: a probabilistic perspective: MIT press; 2012. [Google Scholar]
  • 25.Molnar C. Interpretable machine learning. Lulu.com. 2020. [Google Scholar]
  • 26.Stehman SV. Selecting and interpreting measures of thematic classification accuracy. Remote Sensing of Environment. 1997;62(1):77–89. doi: 10.1016/s0034-4257(97)00083-7 [DOI] [Google Scholar]
  • 27.Harrell FE Jr, Lee KL, Califf RM, Pryor DB, Rosati RA. Regression modelling strategies for improved prognostic prediction. Stat Med. 1984;3(2):143–52. doi: 10.1002/sim.4780030207 [DOI] [PubMed] [Google Scholar]
  • 28.Zhou Wang, Bovik AC. Mean squared error: Love it or leave it? A new look at Signal Fidelity Measures. IEEE Signal Process Mag. 2009;26(1):98–117. doi: 10.1109/msp.2008.930649 [DOI] [Google Scholar]
  • 29.Bishop CM. Pattern Recognition and Machine Learning. New York: Springer. 2006. [Google Scholar]
  • 30.Leevy JL, Khoshgoftaar TM, Bauder RA, Seliya N. A survey on addressing high-class imbalance in big data. J Big Data. 2018;5(1). doi: 10.1186/s40537-018-0151-6 [DOI] [Google Scholar]
  • 31.Brankovic A, Cook D, Rahman J, Huang W, Khanna S. Evaluation of popular XAI applied to clinical prediction models: can they be trusted?. arXiv preprint. 2023. [Google Scholar]
  • 32.Krishna S, Han T, Gu A, Wu S, Jabbari S, Lakkaraju H. The disagreement problem in explainable machine learning: A practitioner’s perspective. arXiv preprint. 2022. doi: arXiv:2202.01602 [Google Scholar]
  • 33.Goldstein BA, Navar AM, Carter RE. Moving beyond regression techniques in cardiovascular risk prediction: applying machine learning to address analytic challenges. Eur Heart J. 2017;38(23):1805–14. doi: 10.1093/eurheartj/ehw302 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Veledar E, Veledar O, Gardener H, Rundek T, Garelnabi M. Harnessing Statistical and Machine Learning Approaches to Analyze Oxidized LDL in Clinical Research. Cell Biochem Biophys. 2025;:10.1007/s12013-025-01837–9. doi: 10.1007/s12013-025-01837-9 [DOI] [PubMed] [Google Scholar]
  • 35.Tahirovic E, Krivic S. Interpretability and explainability of logistic regression model for breast cancer detection. In: ICAART. 2023. 161–8.
  • 36.Rao A, Barrow E, Vuik S, Darzi A, Aylin P. Systematic Review of Hospital Readmissions in Stroke Patients. Stroke Res Treat. 2016;2016:9325368. doi: 10.1155/2016/9325368 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Boehme AK, Kulick ER, Canning M, Alvord T, Khaksari B, Omran S, et al. Infections Increase the Risk of 30-Day Readmissions Among Stroke Survivors. Stroke. 2018;49(12):2999–3005. doi: 10.1161/STROKEAHA.118.022837 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Boehme AK, Oka M, Cohen B, Elkind MSV, Larson E, Mathema B. Readmission Rates in Stroke Patients with and without Infections: Incidence and Risk Factors. J Stroke Cerebrovasc Dis. 2022;31(1):106172. doi: 10.1016/j.jstrokecerebrovasdis.2021.106172 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Chiou L-J, Lang H-C. Potentially preventable hospital readmissions after patients’ first stroke in Taiwan. Sci Rep. 2022;12(1):3743. doi: 10.1038/s41598-022-07791-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Segar MW, Hall JL, Jhund PS, Powell-Wiley TM, Morris AA, Kao D, et al. Machine Learning-Based Models Incorporating Social Determinants of Health vs Traditional Models for Predicting In-Hospital Mortality in Patients With Heart Failure. JAMA Cardiol. 2022;7(8):844–54. doi: 10.1001/jamacardio.2022.1900 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Spiegler KM, Irvine H, Torres J, Cardiel M, Ishida K, Lewis A, et al. Characteristics associated with 30-day post-stroke readmission within an academic urban hospital network. J Stroke Cerebrovasc Dis. 2024;33(11):107984. doi: 10.1016/j.jstrokecerebrovasdis.2024.107984 [DOI] [PubMed] [Google Scholar]
  • 42.Bishop L, Brown SC, Gardener HE, Bustillo AJ, George DA, Gordon Perue G, et al. The association between social networks and functional recovery after stroke. Int J Stroke. 2025;20(1):95–104. doi: 10.1177/17474930241283167 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Linardatos P, Papastefanopoulos V, Kotsiantis S. Explainable AI: A Review of Machine Learning Interpretability Methods. Entropy (Basel). 2020;23(1):18. doi: 10.3390/e23010018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Albahri AS, Duhaim AM, Fadhel MA, Alnoor A, Baqer NS, Alzubaidi L, et al. A systematic review of trustworthy and explainable artificial intelligence in healthcare: Assessment of quality, bias risk, and data fusion. Information Fusion. 2023;96:156–91. doi: 10.1016/j.inffus.2023.03.008 [DOI] [Google Scholar]
  • 45.Hung L-C, Sung S-F, Hu Y-H. A Machine Learning Approach to Predicting Readmission or Mortality in Patients Hospitalized for Stroke or Transient Ischemic Attack. Applied Sci. 2020;10(18):6337. doi: 10.3390/app10186337 [DOI] [Google Scholar]
  • 46.Matsumoto K, Nohara Y, Soejima H, Yonehara T, Nakashima N, Kamouchi M. Stroke Prognostic Scores and Data-Driven Prediction of Clinical Outcomes After Acute Ischemic Stroke. Stroke. 2020;51(5):1477–83. doi: 10.1161/STROKEAHA.119.027300 [DOI] [PubMed] [Google Scholar]
  • 47.Voigtlaender S, Pawelczyk J, Geiger M, Vaios EJ, Karschnia P, Cudkowicz M, et al. Artificial intelligence in neurology: opportunities, challenges, and policy implications. J Neurol. 2024;271(5):2258–73. doi: 10.1007/s00415-024-12220-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Saini H, Rose DZ. The Ghost in the Machine: Artificial Intelligence in Neurocardiology Will Advance Stroke Care. 2024. [DOI] [PMC free article] [PubMed]
  • 49.Otieno JA, Häggström J, Darehed D, Eriksson M. Developing machine learning models to predict multi-class functional outcomes and death three months after stroke in Sweden. PLoS One. 2024;19(5):e0303287. doi: 10.1371/journal.pone.0303287 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Bopche R, Gustad LT, Afset JE, Ehrnström B, Damås JK, Nytrø Ø. In-hospital mortality, readmission, and prolonged length of stay risk prediction leveraging historical electronic patient records. JAMIA Open. 2024;7(3):ooae074. doi: 10.1093/jamiaopen/ooae074 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Nkoke C, Jingi AM, Noubiap JJ, Nkouonlack C, Njume D, Dzudie A. Readmission and mortality during the first year after an acute stroke: A prospective cohort study from Cameroon. PLoS One. 2024;19(10):e0311893. doi: 10.1371/journal.pone.0311893 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Schoon BA, Hansen D, Roozenbeek B, Oude Groeniger J, van der Steen W, van der Lugt A, et al. Neighborhood Socioeconomic Status and the Functional Outcome of Patients Treated With Endovascular Thrombectomy for Ischemic Stroke. Neurology. 2025;105(1):e213615. doi: 10.1212/WNL.0000000000213615 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Voura EB, Abdul-Malak Y, Jorgensen TM, Abdul-Malak S. A retrospective analysis of the social determinants of health affecting stroke outcomes in a small hospital situated in a health professional shortage area (HPSA). PLOS Glob Public Health. 2024;4(1):e0001933. doi: 10.1371/journal.pgph.0001933 [DOI] [PMC free article] [PubMed] [Google Scholar]
PLoS One. 2025 Sep 18;20(9):e0332371. doi: 10.1371/journal.pone.0332371.r001

Author response to Decision Letter 0


Transfer Alert

This paper was transferred from another journal. As a result, its full editorial history (including decision letters, peer reviews and author responses) may not be present.

22 Mar 2025

Decision Letter 0

Noah Hammarlund

13 May 2025

Dear Dr. Veledar,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

The reviewers found the study important and methodologically strong but identified key issues that must be addressed for acceptance. Required changes include clarifying the outcome definition, detailing data preprocessing, addressing class imbalance with appropriate metrics, and adding supplemental materials defining predictors and reporting full model outputs. The inclusion of hemorrhagic strokes should be justified or evaluated through sensitivity analysis. While the use of SDOH is a strength, the discussion should acknowledge their limited impact on model performance and clarify their value.

Please submit your revised manuscript by Jun 27 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols .

We look forward to receiving your revised manuscript.

Kind regards,

Noah Hammarlund

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, we expect all author-generated code to be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. For studies involving third-party data, we encourage authors to share any data specific to their analyses that they can legally distribute. PLOS recognizes, however, that authors may be using third-party data they do not have the rights to share. When third-party data cannot be publicly shared, authors must provide all information necessary for interested researchers to apply to gain access to the data. (https://journals.plos.org/plosone/s/data-availability#loc-acceptable-data-access-restrictions) 

For any third-party data that the authors cannot legally distribute, they should include the following information in their Data Availability Statement upon submission:

1) A description of the data set and the third-party source

2) If applicable, verification of permission to use the data set

3) Confirmation of whether the authors received any special privileges in accessing the data that other researchers would not have

4) All necessary contact information others would need to apply to gain access to the data

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

Reviewer #1: Yes

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously? -->?>

Reviewer #1: Yes

Reviewer #2: No

**********

3. Have the authors made all data underlying the findings in their manuscript fully available??>

The PLOS Data policy

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English??>

Reviewer #1: Yes

Reviewer #2: Yes

**********

Reviewer #1: See attachment for additional information. Three areas are outline that should be addressed by the authors before the publication criteria are fully satisfied.

This study uses a rigorous analytic approach of multiple XML models, including regression-based, tree-based, and distance-based algorithms. This diversity in modeling approaches enhances the robustness of the findings. Using 10-fold cross-validation ensures that the models are tested on different subsets of the data, reducing the risk of overfitting. The study uses a combined ranking approach, incorporating Weighted Importance Scores and Frequency Counts to identify significant predictors across models. The performance of each model is assessed using standard evaluation metrics, including accuracy, area under the ROC curve (AUC), squared-error loss, logistic loss, and misclassification rate.

However, I strongly encourage the inclusion of additional information in the study to increase the interpretability and understanding of the findings. By incorporating additional information, robustness tests, and replicability measures, the study can further enhance its credibility and impact in the field of stroke research.

1. Detailed Description of Data Preprocessing:

• The manuscript should include a more detailed description of the data preprocessing steps, such as handling missing values, normalization, and applying feature engineering techniques.

2. Model Hyperparameters:

• Providing information on the hyperparameters used for each XML model would enhance the study's transparency and replicability.

3. Sensitivity Analysis:

• Including a sensitivity analysis to show how changes in model parameters or the inclusion/exclusion of certain variables affect the results would strengthen the robustness of the findings.

I would also encourage the authors to discuss the weaknesses and limitations of their modeling approach and their data sources. The following areas should be addresses:

1. Sample Size and Generalizability:

• Acknowledge the limitations related to sample size and the regional focus of the study. Discuss how these factors might affect the generalizability of the findings.

2. Class Imbalance:

• Address the issue of class imbalance in the dataset and how it was managed. Discuss any potential impacts on model performance and the interpretation of results.

3. Model Limitations:

• Highlight any limitations of the XML models used, such as their reliance on specific types of data or potential biases in variable selection.

4. Data Availability:

• Mention any restrictions on data availability and how they might limit the ability of other researchers to replicate the study.

Finally, the only thing missing from this study is a complete outline of the potential implications of these findings. I would like to know how these conclusions can be used without the following areas:

1. Clinical Practice:

• Discuss how the findings can be applied in clinical practice to improve post-stroke care, resource allocation, and patient outcomes. Highlight any specific recommendations for healthcare providers.

2. Policy and Public Health:

• Explore the implications for public health policy, particularly in addressing socioeconomic disparities and improving care transitions for stroke survivors.

3. Future Research:

• Identify areas for future research, such as developing interventions based on the identified predictors, validating the models in different populations, and exploring additional variables.

By including these additional details, the manuscript can provide a more comprehensive and nuanced understanding of the study's findings, their implications, and the context in which they were derived. This will also help address potential concerns and enhance the research's replicability and applicability.

Reviewer #2: This is a well written, important study that attempts to identify important clinical and non-clinical risk factors that help predict mortality and rehospitalization for patients post-stroke using a consolidated approach of multiple explainable ML techniques. I think the research has merit but there are a number of statistical problems that should be addressed before a satisfactory evaluation of this manuscript.

1. The introduction is well written and clear, there are a number of statements that would benefit from citations (e.g. while recent studies have not substantially improved prediction intervals...), Furthermore, the introduction could present more substantial information regarding performance of classifiers from previous studies as well as problems and limitations that makes this study unique (such as the inclusion of SDOH variables into these types of models)

2. It is unclear how the outcomes were combined into a single variable, what is the frequency for each outcome in the cohort?

3. Class Imbalance: The manuscript suggests that applying multiple machine learning (ML) methods and combining predictors helps address class imbalance. However, I am not convinced this alone mitigates the issue, especially in the absence of intrinsic strategies such as resampling, class weighting, or appropriate performance metrics. Furthermore, this assertion lacks citations, which weakens the credibility of the claim. I recommend supporting this statement with relevant literature or clarifying the methodological justification.

4. I would strongly recommend restricting the analysis to ischemic strokes. The dataset contains only ~100 cases of intracerebral hemorrhage, which likely have distinct clinical and demographic predictors compared to ischemic strokes. Including them may introduce noise that could impair the model's discriminatory performance and interpretability.

5. Table 1 Statistical Analysis: The manuscript does not clearly indicate what statistical tests were applied to generate Table 1. Given the presence of varied data types (e.g., means, medians with ranges, and percentages), it is important to specify whether the appropriate tests were used for each variable type. Additionally, it is unclear if corrections for multiple comparisons were performed.

6. Lasso, ridge and elastic nets are not distinct types of ML classifiers from logistic regression, rather there are regularization techniques aimed to reduce the important of coefficients by either shrinking their values or removing them altogether. Including all four variants (standard logistic regression, LASSO, ridge, and elastic net) in an ensemble may introduce redundancy, as they share the same underlying structure and linear decision boundaries. While the ensemble includes other diverse model types such as SVM, KNN, and decision tree-based methods, it may be more effective to select one or two representative logistic regression variants to avoid overemphasizing a single modeling framework and to maintain better balance across model types.

7. ML Model Performance: It would be valuable to include a supplemental table showing the full results for each machine learning model (Coefficients, predicted probabilities, odds ratios, etc). Was there an evaluation of the performance using the selected either 12 or 38 variables? is there an improvement in the prediction when using the most important variables?

8. Imbalanced Data: The dataset is highly imbalanced, yet the performance metrics presented do not adequately reflect this. Metrics such as precision, recall, and F1-score—particularly for the minority class—are essential in this context. Including confusion matrices would also help readers assess the classifiers' performance across both classes.

9. On Multicollinearity: The issue of multicollinearity among predictors is not addressed. Given the number of variables and the potential for correlated features, I recommend performing and reporting multicollinearity diagnostics (e.g., variance inflation factors), particularly for models like logistic regression where interpretation of coefficients is important.

10. On Clarity and Definition of Predictors:The manuscript lacks a clear definition and description of the predictors used. Many predictor acronyms are not defined in the text, and categorical variables are not described in terms of their levels or categories. I suggest including a supplemental table that clearly lists all predictors, defines each acronym, and indicates the data type (continuous, categorical), as well as levels for categorical variables.

11. Unfortunately, this study did not improve upon the performance metrics reported in previous analyses, despite increasing the number of variables by incorporating several non-clinical factors, such as social determinants of health (SDOH). Since the addition of these variables did not enhance model performance, it raises the question of their practical value—particularly given that SDOH data are often more difficult to obtain than standard clinical variables.

12. There are a few grammatical errors: last word of introduction unballanced and in the sentence... The TCSD-S enrollee data from 10 collaborating comprehensive stroke canters (is it centers)

**********

what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org . Please note that Supporting Information files do not need this step.

Attachment

Submitted filename: Review 20250420.docx

pone.0332371.s001.docx (17.9KB, docx)
PLoS One. 2025 Sep 18;20(9):e0332371. doi: 10.1371/journal.pone.0332371.r003

Author response to Decision Letter 1


9 Jun 2025

All our comments are already included in attached answer to reviewers. We tried to answer all questions and suggestions one by one.

Attachment

Submitted filename: Detailed_responses2revs_1and2.docx

pone.0332371.s002.docx (35.8KB, docx)

Decision Letter 1

Noah Hammarlund

17 Jun 2025

Dear Dr. Veledar,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Aug 01 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols .

We look forward to receiving your revised manuscript.

Kind regards,

Noah Hammarlund

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments:

Thank you for your revised manuscript. The study presents a thoughtful and innovative approach, and both reviewers noted major improvements. However, they also raise several remaining concerns that must be addressed before the manuscript can be considered for publication.

Where the reviewers raise methodological gaps (e.g., around class imbalance, model justification, or variable inclusion), you should either (a) conduct and report targeted additional analyses where feasible, or (b) clearly explain and justify your approach, and acknowledge resulting limitations where appropriate.

Please address the following points:

Composite Outcome: Justify the combination of readmission and mortality into a single outcome, especially given their differing frequencies. If separate models are not feasible, acknowledge this and discuss the implications as a limitation.

Stroke Type and Generalizability: Either conduct a sensitivity check restricted to ischemic stroke or justify your inclusion of ICH cases. Generalizability limitations due to geography and stroke subtype distribution should be made more explicit in the abstract and discussion.

Class Imbalance: The use of multiple models and evaluation metrics does not directly address class imbalance. Metrics like AUC and log loss can obscure performance issues on the minority class. Please provide a clearer discussion of the implications of imbalance and justify the decision not to apply standard techniques (e.g., weighting or resampling). Highlight how this may affect interpretation of metrics such as recall and variable rankings.

Multicollinearity: For regression-based models, address concerns about multicollinearity. Either assess and report relevant diagnostics (e.g., VIF, correlation checks), or justify your approach and acknowledge interpretability limitations.

Value of SDOH Variables: You make strong claims about the contribution of SDOH to model performance. While these variables appear in your ranked lists, the added predictive value is not directly assessed. We recommend a simple sensitivity analysis comparing model performance with and without SDOH predictors. If you do not include this, revise your language to more cautiously reflect what is shown, and note this limitation.

Explainability and Interpretation: Since the study emphasizes explainable ML, include a brief narrative interpretation of top predictors (e.g., mRS, ATOC) and how they could inform clinical decision-making.

Scope of Claims: Please revisit some of the manuscript’s stronger claims — such as statements that XML methods “enhance prediction” in small, imbalanced samples. Without comparative analyses or performance improvement tests, such claims should be tempered to reflect the more exploratory and descriptive nature of the work. Strengthening the limitations section will help appropriately frame the contribution.

We look forward to your resubmission.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

Reviewer #1: (No Response)

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions??>

Reviewer #1: Yes

Reviewer #2: Partly

**********

3. Has the statistical analysis been performed appropriately and rigorously? -->?>

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available??>

The PLOS Data policy

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English??>

Reviewer #1: Yes

Reviewer #2: Yes

**********

Reviewer #1: This study uses explainable machine learning (XML) to identify clinical and non-clinical predictors of 90-day readmission and mortality among stroke survivors. Drawing on data from the Transitions of Care Stroke Disparities Study (TCSD-S), the authors analyzed outcomes for 1,300 patients using 11 machine learning models. The study integrates information from clinical registries, neighborhood-level social determinants of health (SDOH), and transition-of-care metrics. The key contribution is a composite list of 38 variables, derived from model consensus, that predict adverse post-stroke outcomes. The authors argue for the superiority of XML over traditional regression methods, highlighting the added value of SDOH in predictive modeling and the potential application in digital health systems.

Contribution of the Methodology, Insights, Conclusions:

1) The manuscript extends current clinical studies by integrating socioeconomic, neighborhood, and behavioral characteristics into the machine learning framework to enhance model relevance.

2) The authors clearly describe model selection, performance metrics, and variable importance aggregation (frequency count and weighted score), providing transparency and reproducibility.

3) The application of 11 XML models to identify consensus predictors is a novel and methodologically rigorous approach to multidimensional health outcome modeling that is strengthened by the 10-fold cross-validation and inclusion of multiple performance metrics (C-statistic, log loss, precision, etc.).

4) These findings could inform risk-stratification tools and targeted post-discharge interventions.

Additional Clarity Needed Before Publication

1) Outcome Construction: The composite outcome of “90-day readmission or mortality” is not adequately justified. The frequency of each outcome (readmission vs. death) is provided late in the paper (n=22 deaths, n=187 readmissions), suggesting imbalanced outcome events. Provide a more explicit rationale for combining these outcomes or present separate models for each. At minimum, sensitivity analyses using each outcome separately would enhance interpretability.

2) Additional Preprocessing Details: While the authors briefly mention normalization and the exclusion of 73 individuals due to missing data, there is insufficient detail on how missing values were handled or imputed. It would be beneficial if the authors explicitly described missing data patterns, imputation methods (if used), and any variable selection criteria. A flowchart could help illustrate data preprocessing.

3) Limited Generalizability: The sample is geographically confined to Florida, and 92% of strokes are ischemic. Only ~100 ICH cases are included, limiting broader applicability. Please consider performing a secondary analysis that is restricted to ischemic stroke only or provide stratified results. Discuss generalizability limitations more prominently in the abstract and conclusion.

4) Imbalance: The adverse outcome rate is relatively low (15.8%), yet the authors do not apply formal balancing techniques (e.g., SMOTE, class weights). It would be beneficial if the authors included an additional discussion on how class imbalance may or may not influence metrics such as precision and recall.

5) Explainability: The manuscript emphasizes “explainable” ML but does not attempt to interpret the most influential variables beyond summary tables. I suggest including a short narrative interpretation of how top predictors could inform clinical decision-making.

Reviewer #2: The manuscript has improved dramatically from the first submission, I believe it is a stronger work either analytically as well as conceptually. The changes to the text and analysis have improved the credibility to the work produced. However, While the authors have attempted to address the reviewer’s concerns, the revisions provided fall short in several critical areas:

The authors emphasize that their goal is to produce a robust list of factors that are the most relevant for guiding post-discharge care decisions for stroke patients. They also claim that their ensemble approach help mitigate biases that individual classification methods bring to the analysis. These emphasis are repeated throughout the responses as a justification for not including performance metrics that would help the reader understand the real power of their final model. While in theory these justifications can be valid, due to the nature of their dataset (highly imbalanced, skewed population) it is extremely important to understand how these models address these biases and what are the statistical limitations. The analysis suggested: Sensitivity analysis, multicollinearity evaluation, model performance metrics, are not difficult to carry out and would give the reader the necessary tools to be assess the validity and limitations of the analysis.

Sensitivity Analysis: Even though this was not part of my comments, I agree that a sensitivity analysis would be important to understand the relationship of variables in the model. While using multiple modeling approaches helps mitigate algorithm-specific biases, the authors did not perform any sensitivity analysis (e.g., testing the impact of excluding variables, varying key parameters, or assessing model robustness under different conditions), as originally requested. This directly touches the value of adding SDOH to the analysis if there is minimal gain in the performance, as understanding the importance of these variables in the model against the outcome provide justification to their inclusion.

Class Imbalance: The authors acknowledge the imbalance and justify their focus on threshold-independent metrics like AUC and loss functions. However, they do not sufficiently address how the models perform on the minority class. Minority-class-specific metrics (e.g., precision, recall, F1) are dismissed due to instability rather than being transparently reported in aggregate. This leaves important aspects of model performance unassessed. Importantly, their recall values revolve around 0.50, meaning they only catch about half of the true positives, this is a classic symptom of class imbalance: high precision (fewer false positives), but lower recall (many false negatives).

Multicollinearity: The response deflects the issue by appealing to model diversity, without directly addressing the multicollinearity concern in models like logistic regression where coefficient interpretation is meaningful.

Social Determinants of Health (SDOH): The authors justify the inclusion of SDOH variables based on conceptual and ethical grounds but do not provide empirical evidence that these variables improved model performance. Without analyses demonstrating added value (e.g., feature importance, subgroup effects, or calibration improvement), it is unclear whether the additional complexity and burden of collecting SDOH data is warranted in this context.

**********

what does this mean? ). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org

Attachment

Submitted filename: Review 20250612.docx

pone.0332371.s003.docx (17.5KB, docx)
PLoS One. 2025 Sep 18;20(9):e0332371. doi: 10.1371/journal.pone.0332371.r005

Author response to Decision Letter 2


14 Jul 2025

Response to reviewers/editor(s)

All new responses are provided below in colours other than black.

Editor:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Reference list updated (in line with responses to reviewer comments – further below):

• Veledar E, Veledar O, Gardener H, Rundek T, Garelnabi M. Harnessing Statistical and Machine Learning Approaches to Analyze Oxidized LDL in Clinical Research. Cell Biochemistry and Biophysics. 2025; In press.

• Schoon BA, Hansen D, Roozenbeek B, et al. Neighborhood Socioeconomic Status and the Functional Outcome of Patients Treated With Endovascular Thrombectomy for Ischemic Stroke. Neurology. 2025;105(1):e213615. doi:10.1212/WNL.0000000000213615

• Voura EB, Abdul-Malak Y, Jorgensen TM, Abdul-Malak S. A retrospective analysis of the social determinants of health affecting stroke outcomes in a small hospital situated in a health professional shortage area (HPSA). PLOS Glob Public Health. 2024;4(1):e0001933. doi:10.1371/journal.pgph.0001933

Also, we have enhanced the Methods section by explicitly acknowledging the source of our algorithm (recently published). We now also state: “We adopt the Weighted Importance Score and Frequency Count (WISFC) framework for aggregating feature‐importance across multiple models, as originally detailed by Veledar et al. in Algorithms (2025). This approach systematically combines magnitude and consistency information from diverse explainers to produce a robust, consensus‐based ranking of predictors .” This change ensures that readers recognise the algorithm’s foundation in the literature and understand its motivation (i.e., to synthesise insights from an array of explainers) before proceeding to our implementation specifics of this manuscript. We trust this clarification strengthens the manuscript by situating our methodological choices within the broader scholarly context. The added reference is:

• Veledar, E.; Zhou, L.; Veledar, O.; Gardener, H.; Gutierrez, C.M.; Romano, J.G.; Rundek, T. Synthesizing Explainability Across Multiple ML Models for Structured Data. Algorithms 2025, 18, 368. https://doi.org/10.3390/a18060368

Additional Editor Comments:

Thank you for your revised manuscript. The study presents a thoughtful and innovative approach, and both reviewers noted major improvements. However, they also raise several remaining concerns that must be addressed before the manuscript can be considered for publication.

Where the reviewers raise methodological gaps (e.g., around class imbalance, model justification, or variable inclusion), you should either (a) conduct and report targeted additional analyses where feasible, or (b) clearly explain and justify your approach, and acknowledge resulting limitations where appropriate.

In addition to the detailed responses to individual points below, we feel that an additional clarification is needed in relation to the overal approach presented in this manuscript. Considering reviewer’s concern regarding our composite endpoint of 90-day readmission or mortality, we have clarified in the manuscript that these events (defined respectively as inpatient admission via the emergency department and death within 90 days of discharge) are not mutually exclusive and are commonly combined in stroke research to capture the overall burden of adverse post-discharge outcomes. We have also added a new “Sensitivity Analysis” subsection and a paragraph in “Strengths and Limitations” summarising three additional analyses (readmission only; ischaemic stroke with composite endpoint; and ischaemic stroke readmission only); all of which yielded consistent top predictors and similar discrimination and calibration metrics. Although we have not included the full tables for these additional runs in the main text due to space constraints, they are available upon request by the editor and will form part of a forthcoming manuscript in progress, which is focused on the interplay of correlated variables and predictors across different stroke populations. If you require these tables for the present review, we would be able to provide them.

Please address the following points:

Composite Outcome: Justify the combination of readmission and mortality into a single outcome, especially given their differing frequencies. If separate models are not feasible, acknowledge this and discuss the implications as a limitation.

Stroke Type and Generalizability: Either conduct a sensitivity check restricted to ischemic stroke or justify your inclusion of ICH cases. Generalizability limitations due to geography and stroke subtype distribution should be made more explicit in the abstract and discussion.

Class Imbalance: The use of multiple models and evaluation metrics does not directly address class imbalance. Metrics like AUC and log loss can obscure performance issues on the minority class. Please provide a clearer discussion of the implications of imbalance and justify the decision not to apply standard techniques (e.g., weighting or resampling). Highlight how this may affect interpretation of metrics such as recall and variable rankings.

Multicollinearity: For regression-based models, address concerns about multicollinearity. Either assess and report relevant diagnostics (e.g., VIF, correlation checks), or justify your approach and acknowledge interpretability limitations.

Value of SDOH Variables: You make strong claims about the contribution of SDOH to model performance. While these variables appear in your ranked lists, the added predictive value is not directly assessed. We recommend a simple sensitivity analysis comparing model performance with and without SDOH predictors. If you do not include this, revise your language to more cautiously reflect what is shown, and note this limitation.

Explainability and Interpretation: Since the study emphasizes explainable ML, include a brief narrative interpretation of top predictors (e.g., mRS, ATOC) and how they could inform clinical decision-making.

Scope of Claims: Please revisit some of the manuscript’s stronger claims — such as statements that XML methods “enhance prediction” in small, imbalanced samples. Without comparative analyses or performance improvement tests, such claims should be tempered to reflect the more exploratory and descriptive nature of the work. Strengthening the limitations section will help appropriately frame the contribution.

The above points are addressed individually as described below (one point at a time)

Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer 1

Reviewer #1: This study uses explainable machine learning (XML) to identify clinical and non-clinical predictors of 90-day readmission and mortality among stroke survivors. Drawing on data from the Transitions of Care Stroke Disparities Study (TCSD-S), the authors analyzed outcomes for 1,300 patients using 11 machine learning models. The study integrates information from clinical registries, neighborhood-level social determinants of health (SDOH), and transition-of-care metrics. The key contribution is a composite list of 38 variables, derived from model consensus, that predict adverse post-stroke outcomes. The authors argue for the superiority of XML over traditional regression methods, highlighting the added value of SDOH in predictive modeling and the potential application in digital health systems.

Contribution of the Methodology, Insights, Conclusions:

1) The manuscript extends current clinical studies by integrating socioeconomic, neighborhood, and behavioral characteristics into the machine learning framework to enhance model relevance.

2) The authors clearly describe model selection, performance metrics, and variable importance aggregation (frequency count and weighted score), providing transparency and reproducibility.

3) The application of 11 XML models to identify consensus predictors is a novel and methodologically rigorous approach to multidimensional health outcome modeling that is strengthened by the 10-fold cross-validation and inclusion of multiple performance metrics (C-statistic, log loss, precision, etc.).

4) These findings could inform risk-stratification tools and targeted post-discharge interventions.

Additional Clarity Needed Before Publication

1) Outcome Construction: The composite outcome of “90-day readmission or mortality” is not adequately justified. The frequency of each outcome (readmission vs. death) is provided late in the paper (n=22 deaths, n=187 readmissions), suggesting imbalanced outcome events. Provide a more explicit rationale for combining these outcomes or present separate models for each. At minimum, sensitivity analyses using each outcome separately would enhance interpretability.

Work/clarification: To clarify, the integrated outcome combines two events observed within 90 days after discharge:

• 90-day readmission, defined as return to the hospital requiring inpatient admission through the emergency department.

• 90-day mortality, defined as death within 90 days of discharge.

It is important to note that these two events are not mutually exclusive. A subset of patients experienced both readmission and subsequent death within the 90-day follow-up period.

Regarding the rationale for combining these outcomes, it is common in stroke and other acute care research to use composite endpoints to capture the overall burden of severe post-discharge adverse events (Bravata, et. al, 2007). Readmissions and mortality are related, clinically meaningful indicators of health system utilization, patient instability, and poor prognosis. Using a composite outcome allows us to assess the cumulative incidence of significant events that reflect failure to recover or deterioration after the index hospitalization, while increasing statistical power when individual event rates are relatively low - particularly for mortality in the 90-day timeframe.

That said, in light of your suggestion, we provide clearer justification in the manuscript and consider presenting separate models for readmission only as secondary analyses to enhance transparency and interpretability.

Response to Reviewer: We carried out three additional sets of sensitivity analyses on our patient cohort with complete data, mirroring the original Tables 2–4 under the following scenarios:

1. Readmission-only outcome (i.e. excluding deaths): we re-ran all logistic-regression and ML models predicting 90-day readmission alone.

2. Ischemic stroke subset with composite outcome: we restricted the cohort to the patients with ischaemic stroke and re-fitted all models for the combined endpoint of readmission or mortality.

3. Ischaemic stroke subset with readmission-only: we combined the above two restrictions.

Across all three scenarios, the same SDOH and clinical variables emerged as top predictors, and model discrimination and calibration metrics were essentially unchanged compared with the primary analysis. These findings confirm that our composite endpoint and mixed‐stroke cohort do not drive the key insights on SDOH predictors.

Given space constraints, we have not included full tables of these additional runs in the main text. Our intention is to further explore this point in a follow-up manuscript focused specifically on stroke subtype and individual endpoint models (subject to further research and data analysis).

Refined text:

1. Added section 3.3 Sensitivity Analysis: “To assess the robustness of our findings, we repeated all analyses under three alternative scenarios, i.e., predicting 90-day readmission only; repeating the combined endpoint analysis in the ischaemic stroke subgroup; and predicting readmission only in the ischaemic subgroup. In all cases, the top predictors remained consistent with those reported above, and model discrimination and calibration measures show minimal deviation from the primary results (data not shown, but further discussed).”

2. Added paragraph in setion 4.5 “Strengths and limitations” at the end of limitations and just before furter work “We conducted sensitivity analyses restricting to readmission only and to the ischaemic stroke subgroup, which confirmed that our main results, which focus on the prominence of SDOH factors, were robust to endpoint definition and stroke subtype. We have not included all of these additional tables here, but we plan to include them in a dedicated follow-up study.”

2) Additional Preprocessing Details: While the authors briefly mention normalization and the exclusion of 73 individuals due to missing data, there is insufficient detail on how missing values were handled or imputed. It would be beneficial if the authors explicitly described missing data patterns, imputation methods (if used), and any variable selection criteria. A flowchart could help illustrate data preprocessing.

Work: patients with missing records were excluded from the analysis. No imputations were made. Only patients with full records were used in the analysis.

Refined text: inserted the paragraph within step 2 of section 2.3, immediately following the mention of normalization: “All candidate predictors were assessed for completeness prior to model development. Patients with missing values in any of the predictors listed in Table 1 were excluded from the analysis; no imputation methods were applied. Of the original cohort of 1,200 stroke patients, 73 (6.1 %) were removed due to incomplete records, yielding a final sample of 1,127 patients with fully observed data.”

Response to Reviewer: We thank the reviewer for highlighting the need for greater clarity around our handling of missing data. In response, we have explicitly stated that no imputation was performed and that only patients with fully observed predictor records were included in our analysis. We have also updated the Methods section to describe the exclusion of subjects with any missing values and added the exact number of exclusions.

In our analysis, we elected to use a complete-case approach, retaining only records with complete information across all variables required for modeling. Specifically, 73 records were excluded due to missing values in one or more covariates. No imputation methods were applied.

Regarding the pattern of missingness, the majority of missing values were due to incomplete documentation in the source dataset. Thus, we assumed the data were missing at random (MAR) conditional on observed covariates.

3) Limited Generalizability: The sample is geographically confined to Florida, and 92% of strokes are ischemic. Only ~100 ICH cases are included, limiting broader applicability. Please consider performing a secondary analysis that is restricted to ischemic stroke only or provide stratified results. Discuss generalizability limitations more prominently in the abstract and conclusion.

Work to be described: variables related to ischemic stroke are listed as numbers 16 and 38 in order, so we do not really expect a big difference in our models once we exclude variables separating ischemic from hemorrhagic stroke. Hence, we do our analysis for ischemic stroke only.

Refined text:

1. Methods - within step 2 of section 2.3., added the following text:

To assess whether model performance differed when restricted to ischemic stroke, we performed the secondary analysis in ischemic stroke subset. We excluded the 89 hemorrhagic cases from our final cohort of 1,127 patients, yielding 1,038 ischemic strokes (92.1 %). Predictors specific to stroke subtype (variables 16 and 38) were omitted. Discrimination (AUC

Attachment

Submitted filename: PLOS_response2d.docx

pone.0332371.s004.docx (33.9KB, docx)

Decision Letter 2

Noah Hammarlund

21 Jul 2025

Dear Dr. Veledar,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

Thank you for your revised manuscript and detailed responses. The paper presents a thoughtful and well-structured modeling approach with clear improvements from the previous version. Most of the reviewer concerns have been adequately addressed, including those related to class imbalance, generalizability, and multicollinearity. We especially appreciate your attention to strengthening the limitations and clarifying analytic decisions.

However, two points still require further attention before the manuscript can be accepted:

  1. Explainability and Narrative Interpretation : In your response, you indicated that you had added a paragraph linking top predictors to specific clinical actions to address the concerns of both reviewers. However, this paragraph does not appear in the revised manuscript. Instead, there is a relatively repetitive statement about class imbalance. Please replace that with a brief narrative that interprets how the highest-ranked predictors might inform clinical decisions such as discharge planning or post-acute care coordination to tie the machine learning models to better match the framed contribution of explainability. This addition is especially important because the paper is framed as a contribution to explainable machine learning, but the methods primarily involve aggregating variable importance across multiple models and do not include standard explainability techniques. If you do not plan to add such techniques, we recommend revising the framing of the manuscript to more accurately reflect the scope of the contribution.

  2. Strength of Claims about Explainability and Model Superiority : Editor comments specifically asked for further empirical justification of claims about the superiority of XML models and the added value of SDOH, since the paper does not provide evidence that XML methods outperform traditional approaches or that SDOH variables meaningfully improve predictive performance. These concerns were not addressed in the revised manuscript. Additionally, claims that the models are “explainable” appear overstated given that no standard interpretability techniques are applied. If you do not plan to conduct additional comparative or sensitivity analyses (e.g., excluding SDOH variables or comparing against simpler models), we ask that you revise or soften language around these claims to reflect the nature of the work and to clearly acknowledge the limitations of the evidence provided. In particular, statements like “Our key contribution to academic knowledge lies in showing that XML models can enhance the prediction of 90-day mortality and readmission outcomes after acute stroke hospitalization, even with small, imbalanced samples” overstate what the current analysis supports.

Please submit your revised manuscript by Sep 04 2025 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org . When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols .

We look forward to receiving your revised manuscript.

Kind regards,

Noah Hammarlund

Academic Editor

PLOS ONE

Journal Requirements:

If the reviewer comments include a recommendation to cite specific previously published works, please review and evaluate these publications to determine whether they are relevant and should be cited. There is no requirement to cite these works unless the editor has indicated otherwise. 

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/ . PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org . Please note that Supporting Information files do not need this step.

PLoS One. 2025 Sep 18;20(9):e0332371. doi: 10.1371/journal.pone.0332371.r007

Author response to Decision Letter 3


6 Aug 2025

Dear Editor,

Thank you for coordinating the thoughtful reviews and for providing us with the opportunity to strengthen our manuscript. We have carefully considered each comment and revised the text—most notably renaming Section 4.6, adding a targeted narrative on socioeconomic predictors, and tempering claims around model superiority—to better reflect the scope and limitations of our work.

Response to comment 1:

Thank you for this suggestion. We have updated the manuscript accordingly:

a. Section Renaming: Section 4.6 has been renamed to “May Modeling Help Post-Acute Stroke Care” Because that question is really still very open.

b. New Narrative Paragraph: We added a concise narrative in Section 4.6 that links our highest-ranked predictors (particularly the 19 socioeconomic factors) to specific clinical actions, such as tailoring discharge planning and coordinating home-based or community services. This highlights how explainable insights can inform post-acute care interventions rather than merely ranking variable importance.

We trust this will strengthen the paper’s focus on real-world applicability of our explainable modeling framework.

Response to comment 2:

We have softened our original claim about model superiority and reframed the sentence to highlight the exploratory nature of our multi-model approach and the critical role of environmental variables (including SDOH). The previous statement has been replaced with a new paragraph, which emphasizes that our goal is to leverage machine learning for preliminary insights in a domain that currently lacks robust theories or interpretability methods for post-acute stroke outcomes, rather than to assert the unequivocal superiority of any single model.

Best Regards,

Emir Veledar

Attachment

Submitted filename: ResponseToEditor and Reviewers.docx

pone.0332371.s005.docx (15.7KB, docx)

Decision Letter 3

Noah Hammarlund

31 Aug 2025

Identifying Determinants of Readmission and Death Post-Stroke Using Explainable Machine Learning

PONE-D-25-14703R3

Dear Dr. Veledar,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager®  and clicking the ‘Update My Information' link at the top of the page. For questions related to billing, please contact billing support .

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Noah Hammarlund

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Acceptance letter

Noah Hammarlund

PONE-D-25-14703R3

PLOS ONE

Dear Dr. Veledar,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

You will receive further instructions from the production team, including instructions on how to review your proof when it is ready. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few days to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

You will receive an invoice from PLOS for your publication fee after your manuscript has reached the completed accept phase. If you receive an email requesting payment before acceptance or for any other service, this may be a phishing scheme. Learn how to identify phishing emails and protect your accounts at https://explore.plos.org/phishing.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Noah Hammarlund

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    Attachment

    Submitted filename: Review 20250420.docx

    pone.0332371.s001.docx (17.9KB, docx)
    Attachment

    Submitted filename: Detailed_responses2revs_1and2.docx

    pone.0332371.s002.docx (35.8KB, docx)
    Attachment

    Submitted filename: Review 20250612.docx

    pone.0332371.s003.docx (17.5KB, docx)
    Attachment

    Submitted filename: PLOS_response2d.docx

    pone.0332371.s004.docx (33.9KB, docx)
    Attachment

    Submitted filename: ResponseToEditor and Reviewers.docx

    pone.0332371.s005.docx (15.7KB, docx)

    Data Availability Statement

    The FSR uses data from Get with The Guidelines-Stroke® (GWTG-S). Due to data-sharing agreements, researchers must apply for access at http://www.heart.org/qualityresearch, with proposals reviewed by GWTG-S and FSR committees upon reasonable request.


    Articles from PLOS One are provided here courtesy of PLOS

    RESOURCES