Skip to main content
PLOS Digital Health logoLink to PLOS Digital Health
. 2024 Mar 15;3(3):e0000459. doi: 10.1371/journal.pdig.0000459

Identifying prognostic factors for survival in intensive care unit patients with SIRS or sepsis by machine learning analysis on electronic health records

Maximiliano Mollura 1, Davide Chicco 2,3,*, Alessia Paglialonga 4, Riccardo Barbieri 1
Editor: Nan Liu5
PMCID: PMC10942078  PMID: 38489347

Abstract

Background

Systemic inflammatory response syndrome (SIRS) and sepsis are the most common causes of in-hospital death. However, the characteristics associated with the improvement in the patient conditions during the ICU stay were not fully elucidated for each population as well as the possible differences between the two.

Goal

The aim of this study is to highlight the differences between the prognostic clinical features for the survival of patients diagnosed with SIRS and those of patients diagnosed with sepsis by using a multi-variable predictive modeling approach with a reduced set of easily available measurements collected at the admission to the intensive care unit (ICU).

Methods

Data were collected from 1,257 patients (816 non-sepsis SIRS and 441 sepsis) admitted to the ICU. We compared the performance of five machine learning models in predicting patient survival. Matthews correlation coefficient (MCC) was used to evaluate model performances and feature importance, and by applying Monte Carlo stratified Cross-Validation.

Results

Extreme Gradient Boosting (MCC = 0.489) and Logistic Regression (MCC = 0.533) achieved the highest results for SIRS and sepsis cohorts, respectively. In order of importance, APACHE II, mean platelet volume (MPV), eosinophil counts (EoC), and C-reactive protein (CRP) showed higher importance for predicting sepsis patient survival, whereas, SOFA, APACHE II, platelet counts (PLTC), and CRP obtained higher importance in the SIRS cohort.

Conclusion

By using complete blood count parameters as predictors of ICU patient survival, machine learning models can accurately predict the survival of SIRS and sepsis ICU patients. Interestingly, feature importance highlights the role of CRP and APACHE II in both SIRS and sepsis populations. In addition, MPV and EoC are shown to be important features for the sepsis population only, whereas SOFA and PLTC have higher importance for SIRS patients.

Author summary

Sepsis is defined as the dysregulated host response to infection causing a significant increase in patients’ mortality, thus resulting in an important global health problem. Systemic inflammatory response syndrome (SIRS), an exaggerated response of the body to a noxious stressor, and sepsis, an organ dysfunction caused by a dysregulated host response to infection, are two of the most critical conditions in the intensive care unit showing high patients’ mortality and resulting in an important global health problem. However, the major differences leading to an improvement in the patient conditions during SIRS and sepsis are not fully elucidated.

In this study, we assess the role of simple and easily available routinely collected blood count parameters as predictors for patients’ prognosis in 1,257 patients with SIRS or sepsis. We applied and compared the performance of five distinct machine-learning models when predicting ICU patient survival. Furthermore, we investigated the feature importance of the best-performing models for each population to highlight the major differences between the two populations. Results highlight the role of C-reactive protein and APACHE II score in both populations, whereas mean platelet volume and eosinophil counts show higher importance in sepsis patients and SOFA score and platelet count show higher importance for SIRS patients.

Introduction

Patient’s outcome has long been used as primary endpoint for trials in critical care as well as for determining the patient’s prognosis after treatments. Patient mortality and survival are indeed the major clinical outcomes, and they are main targets for assessing prognostic factors driving the patient conditions and the effectiveness of clinical interventions [1, 2], especially in the intensive care unit (ICU) where admitted patients are usually in very critical conditions and require constant monitoring and treatment.

In this context, sepsis represents an important global health problem accounting for about one-third of ICU deaths and its reported incidence is still increasing [36] and a proper and precise description of sepsis is still not available. Indeed, the definition of sepsis was subjected to several revisions during the years [79], and according to the Third International Consensus Definitions for Sepsis and Septic Shock [9] it is currently defined as a life-threatening organ dysfunction caused by a dysregulated host response to infection. This last update of the sepsis definition abandons the use of Systemic Inflammatory Response Syndrome (SIRS) criteria, which were recognized as presenting a lack of specificity, whereas it focuses on the life-threatening condition and the presence and progression of organ failure. However, the major differences between SIRS and sepsis leading to a positive or negative patient outcome are still not fully elucidated in the medical literature.

In particular, Gucyetmez et al. [10] evaluated the ability of hemogram parameters, a set of medical laboratory tests providing information about the cells in a person’s blood, and C-reactive protein (CRP) to distinguish non-sepsis SIRS from sepsis patients. The authors found that the combinations of CRP, lymphocytes count (LymC), and platelet count (PLTC) can be used to determine the likelihood of sepsis, however without exploring the association of these parameters with the patient survival for each population. This information can provide significant indications about the most important prognostic factors specifically for non-sepsis SIRS and sepsis patients. Also, the authors did not investigate the predictive power of the observed variables.

Especially for this last task, machine learning approaches have shown a good ability in the early identification of sepsis with data collected both from electronic health records [1113] and from physiological vital signs monitoring [14], also providing insights about the role of each feature in a multi-variable setting. Several studies focused on predicting ICU patient outcomes focusing on mortality or survival prediction task [1522], but multi-variable prognostic models estimating and comparing SIRS and sepsis outcomes are still lacking. The stratification of patients’ risk in particular for patients undergoing infections and with sepsis is important. In fact, these patients often require prompt management and interventions, like the initiation of antibiotic therapy and the administration of fluid and vasopressors for maintaining adequate tissue perfusion and hemodynamic stability [23]. These aspects led to an increasing interest toward the prediction of the patient outcome specifically for sepsis patients, in the last few years [2429].

In this context, simple and easily available laboratory measurements of blood cell counts (for example platelet, eosinophil, neutrophil, and lymphocyte counts) can be useful tools for patients’ risk stratification.

The goal of this study is to further explore the ability of hemogram parameters in estimating the survival of ICU patients with non-sepsis SIRS and with sepsis, by applying machine learning techniques in order to estimate the survival probability of ICU patients and to investigate the role of the different parameters in a multi-variable prediction setting. Specifically, our study makes further use of the features proposed by Gucyetmez and colleagues [10] to explore the predictive power of hemogram parameters in estimating the survival probability of patients with non-sepsis SIRS and with sepsis, by comparing different machine learning approaches.

Multi-variable feature importance of the best performing models is applied to further assess the role of each feature and to highlight differences between non-sepsis SIRS and sepsis cohorts.

Dataset

In this study, we use data retrospectively collected from 1,257 eligible medical and surgical patients admitted to the ICU’s of Acibadem International Hospital and Atasehir Memorial Hospital between 1 January 2006 and 31 December 2013, Istanbul, Turkey, and made available by Gucyetmez et al. [10]. The considered cohort includes 816 (64.9%) non-sepsis SIRS and 441 (35.1%) sepsis patients.

The dataset contains the following features for each patient: Age, sex, APACHE II and SOFA scores, diagnosis (medical, elective, and emergency surgery), length of ICU stay (LOS-ICU), mortality, CRP, WBCC, NeuC, LymC, NLCR, EoC, PLTC, MPV. A detailed description of the data here used is provided in Table 1. The target variable for our analysis was survival, indicating whether the patient survived (1) or died (0) in the ICU. A quantitative description of the distribution of each numeric and categorical feature for the non-sepsis SIRS (SIRS) and sepsis (SEPSIS) cohorts are reported in Tables 2 and 3. From the analysis of the target variable (Survival) it is possible to observe that both cohorts are unbalanced, with stronger unbalance in the SIRS (3.07% not survived) than in the SEPSIS (23.64% not survived) cohort.

Table 1. Description, unit of measure and range of values of each available feature in the dataset.

EC: Elective, AC: Emergency, and M: Medical. E: Male and K: Female.

feature description unit of measure values
Age Patient’s age at ICU admission years integer>0
APACHE II Illness severity score ordinal integer [0–71]
CRP Acute phase reactant produced in liver mg/dL continuous
Diagnosis Reason for ICU admission categorical [EC, AC, M]
EoC Eosinophils (cells) count 103/μL continuous
cohort Indication of sepsis (1) or non-sepsis SIRS (0) - binary
Sex Patient’s sex categorical [E,K]
LOS-ICU Patient’s length of stay in ICU days continuous
LymC Lymphocytes (cells) count 103/μL continuous
MPV Mean platelet volume fL continuous
NeuC Neutrofil (cells) count 103/μL continuous
NLCR Neutrophil-lymphocyte count ratio ratio continuous
SOFA Illness severity score ordinal integer [0–24]
PLTC Platelets count 103/μL continuous
WBCC White blood cells count 103/μL continuous
Mortality Patient’s outcome: dead or survived - binary

Table 2. Median and interquartile range (IQR) for each numeric variable of the dataset, stratified by Survival (S: Survived, NS: Not survived, T: Total cohort), and for the SIRS and SEPSIS cohorts.

feature SIRS median (IQR) SEPSIS median (IQR)
APACHE II—S 9 (5–12) 16 (11–20)
APACHE II—NS 25 (20–31) 27 (23–31.25)
APACHE II—T 9 (6–13) 18 (14–25)
Age—S 55 (36–69) 62 (51–75)
Age—NS 61 (51–74) 65.5 (55–78.25)
Age—T 55 (37–69) 63 (51.75–76)
CRP—S 2 (0.5–6.08) 5.42 (1.38–12.66)
CRP—NS 2.1 (0.5–5.66) 6.65 (2.3–16.21)
CRP—T 2 (0.5–6.07) 5.6 (1.6–13.97)
EOC—S 10 (0–40) 10 (0–30)
EOC—NS 50 (20–70) 15 (0–40)
EOC—T 10 (0–40) 10 (0–30)
LOS-ICU—S 1 (1–2) 4 (1–9)
LOS-ICU—NS 1 (1–5) 6 (3–17.25)
LOS-ICU—T 1 (1–2) 4 (2–10)
LymC—S 0.93 (0.61–1.34) 0.71 (0.45–1.17)
LymC—NS 1.22 (0.77–1.55) 0.72 (0.42–1.12)
LymC—T 0.93 (0.62–1.36) 0.71 (0.44–1.15)
MPV—S 10.1 (9.4–10.7) 10 (9.4–10.7)
MPV—NS 10 (9.1–10.4) 10 (9.1–11)
MPV—T 10.1 (9.4–10.7) 10 (9.3–10.8)
NLCR—S 10.02 (6.75–14.5) 11.41 (7.33–17.98)
NLCR—NS 8.94 (4.25–14.34) 11.57 (6.65–21.51)
NLCR—T 10.02 (6.7–14.5) 11.48 (7.19–18.57)
NeuC—S 9.22 (6.53–12.7) 8.2 (5.5–12.7)
NeuC—NS 10.17 (7.2–14.73) 8.76 (5.7–13.24)
NeuC—T 9.28 (6.56–12.73) 8.25 (5.57–12.72)
PLTC—S 191 (133–241) 174 (105.75–256.25)
PLTC—NS 172 (115–255) 150 (82–241.5)
PLTC—T 190 (132.25–241.75) 170.5 (101–255.25)
SOFA—S 1 (0–2) 2 (0–6)
SOFA—NS 9 (4–11) 8 (7–10)
SOFA—T 1 (0–2) 4 (1–7)
WBCC—S 11.23 (8.11–14.95) 9.91 (6.87–14.38)
WBCC—NS 12.14 (9.12–19.56) 10.56 (7.38–15.5)
WBCC—T 11.26 (8.17–15.05) 10.02 (7.08–14.6)

Table 3. Values, counts and percentages for each categorical variable of the dataset, stratified by Survival and for the full non-sepsis SIRS cohort.

DIAG.: Diagnosis, S: Survived, NS: Not Survived, EC: Elective, AC: Emergency, and M: Medical. E: Male, and K: Female.

SIRS SEPSIS
feature value counts % counts %
DIAG. S-AC 32 3.92 10 2.26
DIAG. NS-AC 2 0.25 1 0.23
DIAG. S-EC 537 65.81 75 16.97
DIAG. NS-EC 3 0.37 0 0
DIAG. S-M 220 26.96 251 56.79
DIAG. NS-M 20 2.45 103 23.30
SEX S-E 464 56.86 209 47.29
SEX NS-E 17 2.08 53 11.99
SEX S-K 325 39.83 127 28.73
SEX NS-K 8 0.98 51 11.54
TOTAL - 814/816 99.75 440/440 100
Survived - 789 96.93 336 76.36
Not Survived - 25 3.07 104 23.64

Methods

In this retrospective study, we developed predictive models of patient survival using machine learning algorithms and we evaluated the importance of features associated with patient survival using machine learning and biostatistics approaches for both the SIRS and SEPSIS populations separately (Fig 1). All the analyses were performed with the Python 3.8.3 programming language, and scikit-learn 1.0 and SciPy 1.7.1 software packages. Observations with missing information (three patients) were removed.

Fig 1. Schematized representation of the proposed data processing flow.

Fig 1

MCCV: Monte Carlo Cross-Validation; SMOTE: Synthetic Minority Oversampling Technique.

Associations between features and survival

The association between the input features and patient survival was also explored with classical statistical approaches. Specifically, differences in numeric features between survived and deceased groups in each cohort were tested with the Mann-Whitney U test, whereas differences in categorical features were assessed with χ2-test [30]. Statistical significance was defined for p < 0.005 as advocated by Benjamin et al. [31], which also accounts for multiple testing adjustments.

Survival prediction models

We trained five machine learning models with the goal of predicting patient’s survival considering the following features: Age, Sex, SOFA, APACHE II, CRP, WBCC, NeuC, LymC, EOC, NLCR, PLTC, and MPV, for both SIRS and SEPSIS cohorts. The approach consisted of 100 runs of Monte Carlo stratified Cross-Validation with 80%-20% train-test split as already proposed by Chicco et al. [32]. At each iteration, 80% of the data were used as a training set and 20% as a test set keeping constant the ratio between survived and dead patients. In order to limit the effect of class imbalance, we applied Synthetic Minority Oversampling Technique (SMOTE) [33] to the training set. Features were rescaled before feeding them to the classifier by removing the median and dividing by the interquartile range, as estimated on the training set [37], and five machine learning classifiers were used to develop patient survival prediction models. We considered the following classifiers: Logistic Regression (LR), Support Vector Machine (SVM) [34], Decision Tree (Tree), Random Forest (RF) [35] and XGBoost (XGB) [36] (evalution metric: logloss and objective function: binary/logistic). To evaluate the performance of the classifiers, Matthews correlation coefficients (MCC) [38] on the cross-validated test sets were considered because its proven ability to summarize results from contingency tables and its invariance to class swapping [3941, 60]. Specifically, the MCC can take values ranging from –1 to +1, where –1 represents the misclassification of all observations, 0 represents the random association, and 1 perfect classification. Average Receiver Operating Characteristic (ROC) curves and Precision-Recall curves (PRC) are also used to quantitatively assess the average model performances. Further details are reported in the supplementary material in Text A of S1 Appendix where additional sensitivity analyses summarizing model calibration on the test set (Text D of S1 Appendix) and the model performance with hyperparameters optimization (Text F of S1 Appendix) are also reported.

Feature importance

The best-performing model for each cohort was selected and feature importance was estimated through single feature elimination (SFE) approach, that is by evaluating the MCC obtained after removing one variable at a time. In this case, the smaller the resulting MCC, the higher the importance of the variable which generated that observed drop in performances. Feature importance analysis was executed for 100 runs of Monte Carlo stratified cross-validation partitions with 80%/20% train-test split [42]. The resulting MCCs for each run are obtained from the test set observations. Finally, we used the Spearman correlation coefficient and the Kendall coefficient [43] to quantify the correlation between the obtained ranks. Both coefficients range from –1 to 1 (from anticorrelation to perfect matching) whereas the absence of correlation is given by a 0 coefficient.

Results

Associations between features and survival

Results of the statistical analysis are reported in Table 4. It can be observed that APACHE II and SOFA scores showed significant (p<0.0001) association with survival in both SIRS and SEPSIS cohorts. EoC resulted significantly associated (p<0.0001) with survival in SIRS cohort only.

Table 4. p-values obtained from the statistical analysis when testing associations with patient survival in the SEPSIS and SIRS cohorts.

Differences in numeric features between survived and deceased groups in the two cohorts were tested with the Mann-Whitney U test [44], whereas differences in categorical features were assessed with the χ2-test.

SIRS SEPSIS
Age 0.0345 0.1460
APACHE II <0.0001 <0.0001
SOFA <0.0001 <0.0001
CRP 0.8693 0.0651
WBCC 0.1705 0.2607
NeuC 0.3202 0.4994
LymC 0.0255 0.9010
EoC <0.0001 0.5651
NLCR 0.5184 0.8086
PLTC 0.4588 0.0744
MPV 0.2796 0.6240
Sex 0.4754 0.0540

Survival prediction

Survival prediction performances for SEPSIS and SIRS cohorts are graphically summarized in Fig 2. Median MCCs, accuracy, sensitivity, specificity, F1-scores, positive predictive value, negative predictive value, areas under precision-recall and receiver operating characteristic curves, and the respective interquartile ranges are reported in the supplementary material (Text C of S1 Appendix). SVM and LR obtained the highest MCCs in predicting sepsis patient survival, that is 0.533 and 0.533, respectively. Random Forest performed as second best model in this cohort with MCC equal to 0.516 whereas XGBoost and Tree obtained the lowest results 0.459 and 0.368, respectively. LR was chosen as the best performing because of the highest third quartile.

Fig 2. Matthews correlation coefficients (MCC) of the different families of machine learning models in predicting Survival for the SIRS and SEPSIS cohorts.

Fig 2

Each violin plot shows the distribution of the data, whereas the small boxplot inside each violin plot shows the median and the first and third quartiles, the whiskers indicate 0.05 and 0.95 quantiles.

The best score when predicting survival on the SIRS cohort was achieved with the XGB method that reached 0.489. The second and third best-performing models in the SIRS cohort were RF with MCC equal to 0.39 and LR showing MCC equal to 0.379. SVM and Tree showed scores equal to 0.378 and 0.289, respectively.

Fig 3 shows the median ROC and PRC curves for SIRS and SEPSIS cohorts as an overall summary of models’ performances across all Monte Carlo runs.

Fig 3.

Fig 3

Receiver Operating Characteristic (ROC) curves for SIRS (panel (a)) and SEPSIS (panel (b)) cohorts showing the median performances of each model on the test sets generated during Monte Carlo cross-validation. Panels (c) and (d) depict Precision-Recall Curves (PRC) for SIRS and SEPSIS cohorts, respectively, showing the median performances of each model on the test sets generated during Monte Carlo cross-validation.

Feature ranking

This section describes the results obtained after the SFE approach performed on the models with the highest performance in the prediction task on each of the two cohorts. Median values and interquartile ranges for the resulting MCCs are reported in the supplementary material (Text E of S1 Appendix). A graphical representation of feature importance is shown in Fig 4a for the SEPSIS cohort and in Fig 4b for the SIRS cohort where features were ordered from lowest to the highest importance.

Fig 4. Matthews correlation coefficients after single feature elimination for the Survival prediction task performed with (a) the Logistic Regression model in the SEPSIS cohort and (b) with the XGBoost model in the SIRS cohort.

Fig 4

Features were ranked according to importance from left to right. Each violin plot shows the distribution of the data, whereas the small boxplot inside each violin plot shows the median and the first and third quartiles, the whiskers indicate 0.05 and 0.95 quantiles.

Specifically, APACHE II showed the highest importance, that is the lowest resulting median MCC equal to 0.436 (–0.097), in predicting SEPSIS patient survival with a Logistic Regression model. MPV ranked second in terms of feature importance for this specific cohort with MCC equal to 0.484. The other features did not induce a notable decrease in the model’s performance. Feature ranking for the survival prediction of SIRS patients with XGB algorithm showed that SOFA has the highest importance with resulting MCCs equal to 0.381 (–0.108) when the feature is removed.

Results with Spearman coefficient and Kendall distance did not show a significant correlation between the two series, with correlation equal to –0.091 (p = 0.737) and –0.007 (p = 0.983), respectively.

Discussion

Gucyetmez et al. [10] collected the data used in this study for exploring the ability of hemogram and CRP in discriminating between SIRS and SEPSIS cohorts. However, the authors did not investigate the prognostic role of the selected features within each cohort, therefore, our study aimed to investigate more in detail the importance of these features and the possible differences between the considered cohorts. Specifically, we performed the evaluation of the ability of hemogram parameters in predicting the survival of ICU patients diagnosed with SIRS or SEPSIS, using a set of parameters usually available in the patient clinical records. We used widely available features like patient sex, illness severity scores commonly measured and recorded at admission in the ICU, C-reactive protein, and blood cell count measurements. Patients’ comorbidities were not available in the patient’s records shared by Gucyetmez et al. despite they are commonly available in an ICU setting, which represents a significant lack of information. The developed models would have certainly benefited from more information about the patient’s history and this could have led to a more precise identification of differences in the prognostic factors. Therefore, future studies will focus on the extension of these analyses on more complete data including patients’ comorbidities.

The survival prediction models were developed and tested on SIRS and SEPSIS cohorts, with better performances observed in SEPSIS cohort.

Specifically, among all trained ML models, linear-based models like LR and SVM showed higher performances in the SEPSIS cohort, whereas RF and XGB performed better on the SIRS cohort.

Although average calibration curves are sub-optimal, which is likely due to the reduced sample size, the best-performing models show improved calibration with respect to the worst ones as expected.

This behavior suggests that a bigger population would allow for a proper calibration adjustment and translation of the model’s output score to an even more precise individualized patient survival probability. Also, this approach might account for a possible covariate shift due to changes in patient characteristics, without the need for the development of a new model. Therefore, we do consider that the relationships between the available variables both intra- and inter-population can be considered a reliable multivariable comparison of the major factor predicting survival for both SIRS and sepsis patients. As it can be observed in Text F of S1 Appendix, the sensitivity analysis implementing the hyperparameter optimization shows results very close to those observed without hyperparameter optimization, thus highlighting the robustness of the proposed framework.

The feature importance analysis attributed the highest importance to APACHE II and SOFA scores for SEPSIS and SIRS cohorts, respectively, thus confirming the importance of a preliminary assessment of patients’ risk at the admission in the ICU [45]. This result is also confirmed by statistical analysis, as shown in Table 4.

SEPSIS cohort

Results on the SEPSIS cohort showed that MPV was the second most important variable in predicting survival. This result is in line with the observed association of a higher MPV with an increased mortality risk as well as its predictive role [4648]. Our analysis ranked EoC and CRP as third and fourth most important features. In the literature, a lower EoC has been associated with mortality in critically ill medical patients [49], in patients admitted with an exacerbation of chronic obstructive pulmonary disease [50] and in patients with pneumonia [51]. Interestingly, although non-significant, our cohort showed an increase in EoC in deceased SEPSIS patients. An epidemiological study [52] pointed out that eosinophilia is a predictor of all-cause mortality and that an increased number of peripheral blood eosinophils may reflect an increased inflammatory response, resulting in tissue injury, a condition that may reflect our cohort. CRP was the fourth most important variable in our ML model. Of note, CRP had already shown the potential of being a predictor of survival of ICU patients [53], and more in general a predictor of mortality in ML frameworks [54].

SIRS cohort

Interestingly, the third most important variable in predicting survival for the SIRS cohort was platelet counts, with a smaller median value for the non-survived patients than for the survived group. Indeed, Vanderschueren et al. [55] found that Thrombocytopenia was associated with a higher risk of death in a septic cohort, in line with our results considering the definition of sepsis (sepsis-1) used in 2000 which only required two SIRS criteria. CRP ranked fourth in predicting patient survival with non-sepsis SIRS and similar considerations as for the SEPSIS cohort can be done, moreover, its importance in predicting survival of a non-sepsis SIRS cohort was already observed in animal studies [56]. The fifth and sixth-ranked features were lymphocytes and eosinophils. In literature, Lymphocytes counts were found to be associated with increased mortality risk in general ICU patients [57], heart failure [58], and COVID-19 patients [59]. Eosinophils count significantly differed between survived and deceased groups with an increase in the deceased one. Similar considerations as for the SEPSIS cohort can be done also for eosinophil counts, where we already pointed out that this apparently opposite behavior with respect to literature might be due to the specific cohort of our study with undergoing inflammatory response [52].

General considerations and applicability

The developed models show the ability to predict patient survival and specifically, this study can be considered as an important integration of the study performed by Gucyetmez et al. [10] so that once a patient with inflammatory response has been identified as septic or not the corresponding model can give us the possibility to immediately assess the likelihood of survival. Also, the feature importance analysis proposed in our study gives a clue on the main features that contributed to the developed cohort-specific score, and it suggests to clinicians which of the considered variables is more informative for a patient falling in the SIRS or SEPSIS cohort. It is important to notice that the SFE method presents some limitations when features are highly interdependent, since the contribution of a feature that is very important may still be underestimated due to the effect of other covariates that depend on it.

Finally, it is worth mentioning that we are not aware of whether these data were collected for administrative health reasons or whether they are commonly used for clinical practice, which might limit the general applicability of the results. However, the employment of data like these for scientific analyses based on computational intelligence can allow new scientific discoveries that otherwise would be impossible with traditional hospital technologies.

This study presents an original application of a statistical framework aimed at predicting patient survival. As the approach is mainly limited by the reduced sample size of the cohort, it is expected that a larger collection of data would allow for a more effective model calibration and optimization that would further improve the model generalizability, thus providing a more precise estimate of patient survival probability.

Conclusions

The proposed study applies an original machine learning paradigm for processing clinical information at admission in the ICU to predict patient survival. The proposed approach relies on a multi-variable predictive modeling approach based on information gathered at the ICU admission, and aimed at predicting the likelihood of patient survival for patients with SIRS and with SEPSIS. Results provide insights into the differences of the most relevant variables between the two groups. A Monte Carlo Cross-Validation procedure was further applied to have robust estimates of the obtained scores. The performed sensitivity analysis showed that results did not notably vary with hyperparameter tuning thus confirming the need for a larger cohort to advance to a fully calibrated deployable model.

In this context, Logisitic Regression and XGBoost algorithms are the best-performing models for SEPSIS and SIRS cohorts, respectively. Moreover, feature importance analysis revealed a high importance of APACHE II score and a comparable important role of C-reactive protein in both cohorts. Also, MPV and EoC were revealed to be important predictors of survival mainly in the SEPSIS cohort, whereas they showed a secondary role in the SIRS cohort.

SIRS cohort showed greater importance of SOFA and platelets count features which instead ranked last in SEPSIS.

Importantly, beyond Gucyetmez et al. [10] findings, the proposed framework addresses the question of whether a patient has sepsis or not, and our models give clinicians the possibility to estimate patient’s survival, as well as to identify the most important features involved in the stratification of patients’ risk with SIRS or SEPSIS, and that also led to the proposed survival estimates.

Of note, to our knowledge, this is the first study where the ability of hemogram parameters in predicting patient survival at the admission in the ICU and the role of the considered features are investigated to highlight differences between SIRS and SEPSIS patients.

Supporting information

S1 Appendix

Text A: Formulas of the confusion matrix rates. Text B: Model Hyperparameters. Text C: Model performances. Text D: Model Calibration. Text E: Single Feature Elimination. Text F: Sensitivity Analysis: Hyperparameter Optimization.

(PDF)

pdig.0000459.s001.pdf (227.7KB, pdf)
S1 Fig. Average calibration curves across the MCCV runs for the best (left column) and worst (right column) for SEPSIS (upper row) and SIRS (lower row).

(EPS)

pdig.0000459.s002.eps (1.5MB, eps)

Data Availability

The dataset is publically available at the following URL: https://figshare.com/articles/dataset/_C_Reactive_Protein_and_Hemogram_Parameters_for_the_Non_Sepsis_Systemic_Inflammatory_Response_Syndrome_and_Sepsis_What_Do_They_Mean_/1644426.

Funding Statement

The work of DC was funded by the European Union – Next Generation EU program, in the context of The National Recovery and Resilience Plan, Investment Partenariato Esteso PE8 “Conseguenze e sfide dell’invecchiamento”, Project Age-It (Ageing Well in an Ageing Society) and partially supported by Ministero dell’Università e della Ricerca of Italy under the “Dipartimenti di Eccellenza 2023-2027” ReGAInS grant assigned to Dipartimento di Informatica Sistemistica e Comunicazione at Università di Milano-Bicocca. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1. Veldhoen R.A., Howes D., Maslove D.M.: Is mortality a useful primary end point for critical care trials? Chest 158(1), 206–211 (2020). doi: 10.1016/j.chest.2019.11.019 [DOI] [PubMed] [Google Scholar]
  • 2. REMAP-CAP Writing Committee for the REMAP-CAP Investigators, Effect of antiplatelet therapy on survival and organ support–free days in critically ill patients with COVID-19: a randomized clinical trial JAMA 327, 1247–1259 (2022). doi: 10.1001/jama.2022.2910 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. The EPISEPSIS Study Group: EPISEPSIS: a reappraisal of the epidemiology and outcome of severe sepsis in French intensive care units. Intensive Care Medicine 30(4), 580–588 (2004). doi: 10.1007/s00134-003-2121-4 [DOI] [PubMed] [Google Scholar]
  • 4. Gaieski D.F., Edwards J.M., Kallan M.J., Carr B.G.: Benchmarking the incidence and mortality of severe sepsis in the United States. Critical Care Medicine 41(5), 1167–1174 (2013). doi: 10.1097/CCM.0b013e31827c09f8 [DOI] [PubMed] [Google Scholar]
  • 5. Yébenes J.C., Ruiz-Rodriguez J.C., Ferrer R., Clèries M., Bosch A., Lorencio C., Rodriguez A., Nuvials X., Martin-Loeches I., Artigas A.: Epidemiology of sepsis in Catalonia: analysis of incidence and outcomes in an European setting. Annals of Intensive Care 7(1) (2017). doi: 10.1186/s13613-017-0241-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Kaukonen K.-M., Bailey M., Suzuki S., Pilcher D., Bellomo R.: Mortality related to severe sepsis and septic shock among critically ill patients in Australia and New Zealand, 2000-2012. JAMA 311(13), 1308 (2014). doi: 10.1001/jama.2014.2637 [DOI] [PubMed] [Google Scholar]
  • 7. Bone R.C., Balk R.A., Cerra F.B., Dellinger R.P., Fein A.M., Knaus W.A., Schein R.M.H., Sibbald W.J.: Definitions for sepsis and organ failure and guidelines for the use of innovative therapies in sepsis. Chest 101(6), 1644–1655 (1992). doi: 10.1378/chest.101.6.1644 [DOI] [PubMed] [Google Scholar]
  • 8. Levy M.M., Fink M.P., Marshall J.C., Abraham E., Angus D., Cook D., Cohen J., Opal S.M., Vincent J.-L., Ramsay G.: 2001 SCCM/ESICM/ACCP/ATS/SIS international sepsis definitions conference. Critical Care Medicine 31(4), 1250–1256 (2003). doi: 10.1097/01.CCM.0000050454.01978.3B [DOI] [PubMed] [Google Scholar]
  • 9. Singer M., Deutschman C.S., Seymour C.W., Shankar-Hari M., Annane D., Bauer M., Bellomo R., Bernard G.R., Chiche J.-D., Coopersmith C.M., Hotchkiss R.S., Levy M.M., Marshall J.C., Martin G.S., Opal S.M., Rubenfeld G.D., van der Poll T., Vincent J.-L., Angus D.C.: The third international consensus definitions for sepsis and septic shock (SEPSIS-3). JAMA 315(8), 801–810 (2016). doi: 10.1001/jama.2016.0287 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10. Gucyetmez B., Atalan H.K.: C-reactive protein and hemogram parameters for the non-sepsis systemic inflammatory response syndrome and sepsis: what do they mean? PLOS One 11(2), 1–9 (2016). doi: 10.1371/journal.pone.0148699 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11. Desautels T., Calvert J., Hoffman J., Jay M., Kerem Y., Shieh L., Shimabukuro D., Chettipally U., Feldman M.D., Barton C., Wales D.J., Das R.: Prediction of sepsis in the intensive care unit with minimal electronic health record data: a machine learning approach. JMIR Medical Informatics 4(3), 28 (2016). doi: 10.2196/medinform.5909 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Fleuren L.M., Klausch T.L.T., Zwager C.L., Schoonmade L.J., Guo T., Roggeveen L.F., Swart E.L., Girbes A.R.J., Thoral P., Ercole A., Hoogendoorn M., Elbers P.W.G.: Machine learning for the prediction of sepsis: a systematic review and meta-analysis of diagnostic test accuracy. Intensive Care Medicine 46(3), 383–400 (2020). doi: 10.1007/s00134-019-05872-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Chicco D., Oneto L.: Data analytics and clinical feature ranking of medical records of patients with sepsis. BioData Mining 14(1) (2021). doi: 10.1186/s13040-021-00235-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Mollura M., Lehman L.-W.H., Mark R.G., Barbieri R.: A novel artificial intelligence based intensive care unit monitoring system: using physiological waveforms to identify sepsis. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences 379 (2212) (2021). doi: 10.1098/rsta.2020.0252 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Kim Sujin, P. R.W. Kim Woojae: A comparison of intensive care unit mortality prediction models through the use of data mining techniques. Healthcare Informatics Research 17(4), 232–243 (2011). doi: 10.4258/hir.2011.17.4.232 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16. Silva I., Moody G., Scott D.J., Celi L.A., Mark R.G.: Predicting in-hospital mortality of ICU patients: The PhysioNet/Computing in cardiology challenge 2012. In: 2012 Computing in Cardiology, pp. 245–248 (2012) [PMC free article] [PubMed] [Google Scholar]
  • 17. Pirracchio R., Petersen M.L., Carone M., Rigon M.R., Chevret S., van der Laan M.J.: Mortality prediction in intensive care units with the super ICU learner algorithm (SICULA): a population-based study. The Lancet Respiratory Medicine 3(1), 42–52 (2015). doi: 10.1016/S2213-2600(14)70239-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Awad A., Bader-El-Den M., McNicholas J., Briggs J.: Early hospital mortality prediction of intensive care unit patients using an ensemble learning approach. International Journal of Medical Informatics 108, 185–195 (2017). doi: 10.1016/j.ijmedinf.2017.10.002 [DOI] [PubMed] [Google Scholar]
  • 19.Johnson, A.E.W., Pollard, T.J., Mark, R.G.: Reproducibility in critical care: a mortality prediction case study. In: Doshi-Velez, F., Fackler, J., Kale, D., Ranganath, R., Wallace, B., Wiens, J. (eds.) Proceedings of MLHC 2017—the 2nd Machine Learning for Healthcare Conference. Proceedings of Machine Learning Research, vol. 68, pp. 361–376. PMLR, Boston, Massachusetts, USA (2017)
  • 20. Purushotham S., Meng C., Che Z., Liu Y.: Benchmarking deep learning models on large healthcare datasets. Journal of Biomedical Informatics 83, 112–134 (2018). doi: 10.1016/j.jbi.2018.04.007 [DOI] [PubMed] [Google Scholar]
  • 21. Nistal-Nuño B.: Developing machine learning models for prediction of mortality in the medical intensive care unit. Computer Methods and Programs in Biomedicine 216, 106663 (2022). doi: 10.1016/j.cmpb.2022.106663 [DOI] [PubMed] [Google Scholar]
  • 22. Hong C., Chen J., Yi F., Hao Y., Meng F., Dong Z., Lin H., Huang Z.: CD-Survdr: a contrastive-based model for dynamic survival analysis. Health Information Science and Systems 10(1) (2022). doi: 10.1007/s13755-022-00173-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Levy M.M., Evans L.E., Rhodes A.: The surviving sepsis campaign bundle: 2018 update. Intensive care medicine 44(6), 925–928 (2018) doi: 10.1007/s00134-018-5085-0 [DOI] [PubMed] [Google Scholar]
  • 24. Taylor R.A., Pare J.R., Venkatesh A.K., Mowafi H., Melnick E.R., Fleischman W., Hall M.K.: Prediction of in-hospital mortality in emergency department patients with sepsis: a local big data–driven, machine learning approach. Academic Emergency Medicine 23(3), 269–278 (2016). doi: 10.1111/acem.12876 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Tansakul V., Li X., Koszalinski R., Paiva W., Khojandi A.: Prediction of sepsis and in-hospital mortality using electronic health records. Methods of Information in Medicine 57, 185–193 (2018). doi: 10.3414/ME18-01-0014 [DOI] [PubMed] [Google Scholar]
  • 26. Chicco D., Jurman G.: Survival prediction of patients with sepsis from age, sex, and septic episode number alone. Scientific Reports 10(1) (2020). doi: 10.1038/s41598-020-73558-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Zhang K., Zhang S., Cui W., Hong Y., Zhang G., Zhang Z.: Development and validation of a sepsis mortality risk score for sepsis-3 patients in intensive care unit. Frontiers in Medicine 7 (2021). doi: 10.3389/fmed.2020.609769 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. van Doorn W.P.T.M., Stassen P.M., Borggreve H.F., Schalkwijk M.J., Stoffers J., Bekers O., Meex S.J.R.: A comparison of machine learning models versus clinical evaluation for mortality prediction in patients with sepsis. PLOS One 16(1), 0245157 (2021). doi: 10.1371/journal.pone.0245157 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29. Selcuk M., Koc O., Kestel A.S.: The prediction power of machine learning on estimating the sepsis mortality in the intensive care unit. Informatics in Medicine Unlocked 28, 100861 (2022). doi: 10.1016/j.imu.2022.100861 [DOI] [Google Scholar]
  • 30. Glantz S.A.: Primer of Biostatistics. The Mc Graw-Hill Companies Inc., California: (2012). [Google Scholar]
  • 31. Benjamin D.J., Berger J.O., Johannesson M., Nosek B.A., Wagenmakers E.-J., Berk R., Bollen K.A., Brembs B., Brown L., Camerer C., Cesarini D., Chambers C.D., Clyde M., Cook T.D., De Boeck P., Dienes Z., Dreber A., Easwaran K., Efferson C., Fehr E., Fidler F., Field A.P., Forster M., George E.I., Gonzalez R., Goodman S., Green E., Green D.P., Greenwald A.G., Hadfield J.D., Hedges L.V., Held L., Hua Ho T., Hoijtink H., Hruschka D.J., Imai K., Imbens G., Ioannidis J.P.A., Jeon M., Jones J.H., Kirchler M., Laibson D., List J., Little R., Lupia A., Machery E., Maxwell S.E., McCarthy M., Moore D.A., Morgan S.L., Munafó M., Nakagawa S., Nyhan B., Parker T.H., Pericchi L., Perugini M., Rouder J., Rousseau J., Savalei V., Schönbrodt F.D., Sellke T., Sinclair B., Tingley D., Van Zandt T., Vazire S., Watts D.J., Winship C., Wolpert R.L., Xie Y., Young C., Zinman J., Johnson V.E.: Redefine statistical significance. Nature Human Behaviour 2(1), 6–10 (2017). doi: 10.1038/s41562-017-0189-z [DOI] [PubMed] [Google Scholar]
  • 32. Chicco D., Jurman G.: Arterial disease computational prediction and health record feature ranking among patients diagnosed with inflammatory bowel disease. IEEE Access 9, 78648–78657 (2021). doi: 10.1109/ACCESS.2021.3084063 [DOI] [Google Scholar]
  • 33. Chawla N.V., Bowyer K.W., Hall L.O., Kegelmeyer W.P.: SMOTE: synthetic minority over-sampling technique. Journal of Artificial Intelligence Research 16, 321–357 (2002) doi: 10.1613/jair.953 [DOI] [Google Scholar]
  • 34. Cortes C., Vapnik V.: Support-vector networks. Machine Learning 20(3), 273–297 (1995). doi: 10.1007/BF00994018 [DOI] [Google Scholar]
  • 35. Breiman L.: Random forests. Machine Learning 45(1), 5–32 (2001). doi: 10.1023/A:1010933404324 [DOI] [Google Scholar]
  • 36.Chen, T., Guestrin, C.: XGBoost. Proceedings of KDD’16—the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (2016).
  • 37. Géron A., Hands-on machine learning with scikit-learn and TensorFlow: Concepts, Tools, and Techniques to build intelligent systems. O’Reilly Media; (2017). [Google Scholar]
  • 38. Matthews B.W.: Comparison of the predicted and observed secondary structure of T4 phage lysozyme. Biochimica et Biophysica Acta (BBA)—Protein Structure 405(2), 442–451 (1975). doi: 10.1016/0005-2795(75)90109-9 [DOI] [PubMed] [Google Scholar]
  • 39. Chicco D., Jurman G.: The advantages of the Matthews correlation coefficient (MCC) over F1 score and accuracy in binary classification evaluation. BMC Genomics 21(1) (2020). doi: 10.1186/s12864-019-6413-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40. Chicco D., Tötsch N., Jurman G.: The Matthews correlation coefficient (MCC) is more reliable than balanced accuracy, bookmaker informedness, and markedness in two-class confusion matrix evaluation. BioData Mining 14(1) (2021). doi: 10.1186/s13040-021-00244-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41. Chicco D., Warrens M.J., Jurman G.: The Matthews correlation coefficient (MCC) is more informative than Cohen’s Kappa and Brier score in binary classification assessment. IEEE Access 9, 78368–78381 (2021). doi: 10.1109/ACCESS.2021.3084050 [DOI] [Google Scholar]
  • 42. Chicco D.: Ten quick tips for machine learning in computational biology. BioData Mining 10(1), 1–17 (2017). doi: 10.1186/s13040-017-0155-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Chicco, D., Ciceri, E., Masseroli, M.: Extended Spearman and Kendall coefficients for gene annotation list correlation. In: Di Serio, C., Liò, P., Nonis, A., Tagliaferri, R. (eds.) Computational Intelligence Methods for Bioinformatics and Biostatistics, CIBB 2014 Revised Selected Papers, vol. 8623, pp. 19–32. Springer, Cambridge, England, United Kingdom (2015).
  • 44. MacFarland T.W., Yates J.M.: Mann–Whitney U test. In: Introduction to Nonparametric Statistics for the Biological Sciences Using R, pp. 103–132. Springer, Cham, Switzerland: (2016). [Google Scholar]
  • 45. Bertsimas D., Lukin G., Mingardi L., Nohadani O., Orfanoudaki A., Stellato B., Wiberg H., Gonzalez-Garcia S., Parra-Calderón C.L., Robinson K., Schneider M., Stein B., Estirado A., a Beccara L., Canino R., Dal Bello M., Pezzetti F., Pan A.: COVID-19 mortality risk assessment: an international multi-center study. PLOS One 15(12), 0243262 (2020). doi: 10.1371/journal.pone.0243262 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46. Tajarernmuang P., Phrommintikul A., Limsukon A., Pothirat C., Chittawatanarat K.: The role of mean platelet volume as a predictor of mortality in critically ill patients: a systematic review and meta-analysis. Critical Care Research and Practice 2016, 1–8 (2016). doi: 10.1155/2016/4370834 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Vardon-Bounes F., Gratacap M.-P., Groyer S., Ruiz S., Georges B., Seguin T., Garcia C., Payrastre B., Conil J.-M., Minville V.: Kinetics of mean platelet volume predicts mortality in patients with septic shock. PLOS One 14(10), 0223553 (2019). doi: 10.1371/journal.pone.0223553 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48. Vélez-Páez J.L., Legua P., Vélez-Páez P., Irigoyen E., Andrade H., Jara A., López F., Pérez-Galarza J., Baldeón L.: Mean platelet volume and mean platelet volume to platelet count ratio as predictors of severity and mortality in sepsis. PLOS One 17(1), 0262356 (2022). doi: 10.1371/journal.pone.0262356 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49. Abidi K., Belayachi J., Derras Y., Khayari M.E., Dendane T., Madani N., Khoudri I., Zeggwagh A.A., Abouqal R.: Eosinopenia, an early marker of increased mortality in critically ill medical patients. Intensive Care Medicine 37(7), 1136–1142 (2011). doi: 10.1007/s00134-011-2170-z [DOI] [PubMed] [Google Scholar]
  • 50. Holland M., Alkhalil M., Chandromouli S., Janjua A., Babores M.: Eosinopenia as a marker of mortality and length of stay in patients admitted with exacerbations of chronic obstructive pulmonary disease. Respirology 15(1), 165–167 (2010). doi: 10.1111/j.1440-1843.2009.01651.x [DOI] [PubMed] [Google Scholar]
  • 51. Echevarria C., Hartley T., Nagarajan T., Tedd H., Steer J., Gibson G.J., Bourke S.C.: 30 day mortality and eosinopenia in patients with pneumonia. European Respiratory Journal 44(Suppl 58) (2014) [Google Scholar]
  • 52. Hospers J.J., Schouten J.P., Weiss S.T., Postma D.S., Rijcken B.: Eosinophilia is associated with increased all-cause mortality after a follow-up of 30 years in a general population sample. Epidemiology 11(3), 261–268 (2000). doi: 10.1097/00001648-200005000-00006 [DOI] [PubMed] [Google Scholar]
  • 53. Karagoz I.: Does hemogram biomarkers predict mortality in intensive care population. Experimental Biomedical Research 2(4), 163–168 (2019). doi: 10.30714/j-ebr.2019454854 [DOI] [Google Scholar]
  • 54. Gebhardt C., Hirschberger J., Rau S., Arndt G., Krainer K., Schweigert F.J., Brunnberg L., Kaspers B., Kohn B.: Use of C-reactive protein to predict outcome in dogs with systemic inflammatory response syndrome or sepsis. Journal of Veterinary Emergency and Critical Care 19(5), 450–458 (2009). doi: 10.1111/j.1476-4431.2009.00462.x [DOI] [PubMed] [Google Scholar]
  • 55. Vanderschueren S., De Weerdt A., Malbrain M., Vankersschaever D., Frans E., Wilmer A., Bobbaers H.: Thrombocytopenia and prognosis in intensive care. Critical Care Medicine 28(6), 1871–1876 (2000). doi: 10.1097/00003246-200006000-00031 [DOI] [PubMed] [Google Scholar]
  • 56. Gebhardt C., Hirschberger J., Rau S., Arndt G., Krainer K., Schweigert F.J., Brunnberg L., Kaspers B., Kohn B.: Use of C-reactive protein to predict outcome in dogs with systemic inflammatory response syndrome or sepsis. Journal of Veterinary Emergency and Critical Care 19(5), 450–458 (2009). doi: 10.1111/j.1476-4431.2009.00462.x [DOI] [PubMed] [Google Scholar]
  • 57. Gutierrez-Rodriguez R., Aguilar-Alonso E., Arias-Verdu M., Castillo-Lorente E., Lopez-Caler C., Rojas-Amezcua M., De la Fuente-Martos C., Rivera-Fernandez R., Quesada-Garcia G.: Lymphopenia assessment in icu patients and relationship with mortality. Intensive Care Medicine Experimental 3(S1) (2015). doi: 10.1186/2197-425X-3-S1-A340 [DOI] [Google Scholar]
  • 58. Terpos E., Ntanasis-Stathopoulos I., Elalamy I., Kastritis E., Sergentanis T.N., Politou M., Psaltopoulou T., Gerotziafas G., Dimopoulos M.A.: Hematological findings and complications of COVID-19. American Journal of Hematology 95(7), 834–847 (2020). doi: 10.1002/ajh.25829 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59. Majmundar M., Kansara T., Park H., Ibarra G., Marta Lenik J., Shah P., Kumar A., Doshi R., Zala H., Chaudhari S., Kalra A.: Absolute lymphocyte count as a predictor of mortality and readmission in heart failure hospitalization. IJC Heart & Vasculature 39, 100981 (2022). doi: 10.1016/j.ijcha.2022.100981 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60. Chicco D., Jurman G.: The Matthews correlation coefficient (MCC) should replace the ROC AUC as the standard metric for assessing binary classification. BioData Mining 16(1), 1–23 (2023). doi: 10.1186/s13040-023-00322-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
PLOS Digit Health. doi: 10.1371/journal.pdig.0000459.r001

Decision Letter 0

Martin G Frasch, Nan Liu

25 Apr 2023

PDIG-D-23-00039

Identifying prognostic factors for survival in intensive care unit patients with SIRS or sepsis by machine learning analysis on electronic health records

PLOS Digital Health

Dear Dr. Chicco,

Thank you for submitting your manuscript to PLOS Digital Health. After careful consideration, we feel that it has merit but does not fully meet PLOS Digital Health's publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript within 60 days Jun 24 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at digitalhealth@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pdig/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

* A rebuttal letter that responds to each point raised by the editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

* A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

* An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

We look forward to receiving your revised manuscript.

Kind regards,

Nan Liu

Academic Editor

PLOS Digital Health

Journal Requirements:

1. We ask that a manuscript source file is provided at Revision. Please upload your manuscript file as a .doc, .docx, .rtf or .tex.

2. Please insert an Ethics Statement at the beginning of your Methods section, under a subheading 'Ethics Statement'. It must include:

1) The name(s) of the Institutional Review Board(s) or Ethics Committee(s)

2) The approval number(s), or a statement that approval was granted by the named board(s)

3) (for human participants/donors) - A statement that formal consent was obtained (must state whether verbal/written) OR the reason consent was not obtained (e.g. anonymity). NOTE: If child participants, the statement must declare that formal consent was obtained from the parent/guardian.

3. Please provide separate figure files in .tif or .eps format only and remove any figures embedded in your manuscript file. Please also ensure that all files are under our size limit of 10MB.

For more information about figure files please see our guidelines:

https://journals.plos.org/digitalhealth/s/figures

https://journals.plos.org/digitalhealth/s/figures#loc-file-requirements

4. We have noticed that you have uploaded Supporting Information files, but you have not included a list of legends. Please add a full list of legends for your Supporting Information files after the references list.

Additional Editor Comments (if provided):

Thank you for submitting this study. While the manuscript has merits, some clarifications such as methodology and clinical utility are needed, as suggested by the reviewers.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Does this manuscript meet PLOS Digital Health’s publication criteria? Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe methodologically and ethically rigorous research with conclusions that are appropriately drawn based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

--------------------

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: No

--------------------

3. Have the authors made all data underlying the findings in their manuscript fully available (please refer to the Data Availability Statement at the start of the manuscript PDF file)?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception. The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

--------------------

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS Digital Health does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

--------------------

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The study by Mollura et. al., applied supervised machine-learning models to data collected from patients admitted to intensive care unit to highlight differences in clinical features for survival.

There are a number of major issues that need to be addressed:

Abstract:

1. "However, major differences between these two populations when estimating patients’ risk after ICU admission are not fully elucidated."

Which subsequent risk is being referred to in this sentence?

Author summary:

2. "However, distinguishing between a systemic inflammatory response syndrome (SIRS) and sepsis is not straightforward and the major differences related with patients’ risk are not fully elucidated."

The authors indicate the inability to differentiate between SIRS and sepsis as the major concern.

And this point is reiterated in the introduction.

Why the focus on prognostic factors for survival after either SIRS or sepsis?

There is some disconnect.

Introduction:

3. Why the distinction between patients with SIRS and sepsis?

Will it make a difference if the authors just identified the cohort as patients admitted to ICU, in general?

The aim of the study is to predict survival in ICU patients.

Methods:

4. The dataset did not have information on comorbid conditions at the time of ICU admission. The presence of comorbid conditions impacts on survival of ICU patients. This was not accounted in the models. Does this impact on the models?

5. "It is important to notice that these data were collected originally for administrative health reasons, and not for scientific purpose." The authors considered this a strength and did not highlight potential limitations of using such administrative data for a scientific purpose.

6. Missing data and how missing data was handled was not discussed in the manuscript.

7. Association between features and survival - how was this assessed? The sentence does not provide the details: "Association between the input features and patient survival was also explored with classical statistical approaches."

8. A study flow diagram will be needed to help explain what was done.

9. Was the data split for the 2 analyses: SIRS-related analysis using data from 816 patients and sepsis-related analysis using data from 441 patients?

10. Statistically, was the study adequately powered?

11. Can the authors clarify which model is "Decision Tree (Tree)"? Random forest and XGBoost are decision tree-based algorithms or models.

12. Could the authors explain why this was done and provide reference for this: "Features were standardized removing the median and dividing by the interquartile range, both estimated on the training set, and five machine learning classifiers were used to develop patient survival prediction models."

13. Is Monte Carlo stratified cross-validation a cross-validation approach or a feature importance or ranking approach?

14. Will it be helpful if area-under the curve and calibration plots are provided to assess the performance of the models in addition to MCC?

15. Page 6 has Fig 2 showing ROC curves - there is no mention of ROC curve in the methods section.

Minor concern:

1. The authors use the word "multivariate" a number of times in the manuscript. This should be "multivariable".

Reviewer #2: Mollura et al present in this paper a machine learning approach to predict survival of patients admitted to the ICU having systemic inflammatory response syndrome or sepsis. The authors investigate the performance of five machine learning models and subsequently estimate the importance of the input features. The paper is well written and states clearly the intentions. However, I have some concerns in relation to how the results are generated methodologically and therefore the overall value of the work is somewhat unclear. I also have some doubts on the actual clinical relevance of the paper. I acknowledge that a prototype of a method can be valuable, even if it in its present form would not be finalized for implementation in an ICU electronic patient record system.

Major concerns:

I really appreciate that the authors use a correlation coefficient for evaluating model performance – and in this way take into account all elements of the confusion matrix. Nevertheless, I find it somewhat misleading that the outcome encoded as 1 is survival and not death (table 1 says that the feature ‘mortality’ encodes death as 1, correct me if this means that you encoded death in the outcome with 1 - in that case you can ignore the rest of this point). Using survival as target outcome is counterintuitive, since the rare event usually is the one encoded in this way. To this end I believe that the design makes AUROC, MCC and NPV informative metrics only. Performances in the supplementary tables are exceptionally good, with AUPRC values of > 95% in most cases. An AUPRC of 0.998% implies you are extremely good at predicting the positive class, but in this case the positive class is survival, which accounts for the majority of the population (~97% for the SIRS population). This would have not been the case if death was encoded as 1. Even if this is not an error, I believe that this decision makes the quantitative interpretation of the metrics misleading. Another note is that for the metrics that are not “areas under …” it is not mentioned explicitly what the probability threshold used for the calculation is. Calibration curves are not included at all, which makes the clinical utility of the model less clear.

It is not stated at what timepoint the features were included and at what timepoint the survival was assessed. The authors also do not explain how multiple measurements were aggregated. This is very important in order to understand the study design. Also, no information about imputation and missing values has been included.

The authors did not provide any information about which hyperparameters were used to run the five different ML models. This is also a very important aspect of the study design.

I do not fully understand the way the feature importance was calculated. RFE should routinely remove the least important feature (evaluated using MCC, according to what the authors state), until a desired number of features is reached. In Results, the MCC is reported for each model with only one feature knocked out at a time. This makes the whole feature interpretation difficult to assess as most of the features are not independent or even included into one another (e.g. age into APACHE). This causes features like age, which is usually a very strong predictor of ICU survival, not important because even if explicitly removed from the feature set, it is still implicitly used for prediction by including the APACHE feature. The same argument is also valid for other features.

It is unclear whether multiple test correction has been used to evaluate the associations between features and survival ?

The code used for the analysis has not been provided. This is a very important aspect.

Minor concerns:

I can see that LOS-ICU and diagnosis have not been included in the feature set. While I understand the reason this has been done for LOS-ICU (leakage of information as this feature is only available after the outcome actually happens), I do not fully understand why ‘diagnosis’ has not been used.

The statement on normalization is mentioned twice: “Features were standardized removing the median and dividing by the interquartile range, …” and immediately after “Features were rescaled before feeding them to the classifier by removing the median and dividing by the interquartile range …”.

The feature ‘sex’ is indicated as gender in the data set section, please standardize.

In my opinion, this sentence does not really belong to the dataset section : “In this retrospective study, we developed predictive models of patient survival using machine learning algorithms and we evaluated the importance of features associated with patient survival using machine learning and biostatistics approaches. “

The authors used violin plots. These plots have the disadvantage of showing the full tail of a distribution. In Figure 1 for example many models have MCC scores lower than 0 (it is a bit surprising to have such a wide range of MCC values) . I would suggest instead to use a boxplot, to limit the visualization to the interquartile range and outliers.

Conclusion: while the paper has many valuable aspects and is well written, the unclear methodology and the lack of clear technical descriptions make it very hard to assess the value of the approach.

--------------------

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

Do you want your identity to be public for this peer review? If you choose “no”, your identity will remain anonymous but your review may still be made public.

For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

--------------------

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLOS Digit Health. doi: 10.1371/journal.pdig.0000459.r003

Decision Letter 1

Martin G Frasch, Nan Liu

31 Jul 2023

PDIG-D-23-00039R1

Identifying prognostic factors for survival in intensive care unit patients with SIRS or sepsis by machine learning analysis on electronic health records

PLOS Digital Health

Dear Dr. Chicco,

Thank you for submitting your manuscript to PLOS Digital Health. After careful consideration, we feel that it has merit but does not fully meet PLOS Digital Health's publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript within 60 days Sep 29 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at digitalhealth@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pdig/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

* A rebuttal letter that responds to each point raised by the editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

* A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

* An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

We look forward to receiving your revised manuscript.

Kind regards,

Nan Liu

Academic Editor

PLOS Digital Health

Journal Requirements:

Additional Editor Comments (if provided):

Thank you for the revision. While we see some parts of the manuscript have been improved, there are still major concerns, such as calibration and clinical relevance, among others. Please carefully address reviewers' concerns and comments, provide detailed responses, and make necessary further revisions. It is important to ensure your research results are translatable into clinical practice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #2: (No Response)

--------------------

2. Does this manuscript meet PLOS Digital Health’s publication criteria? Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe methodologically and ethically rigorous research with conclusions that are appropriately drawn based on the data presented.

Reviewer #1: Partly

Reviewer #2: Partly

--------------------

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: No

--------------------

4. Have the authors made all data underlying the findings in their manuscript fully available (please refer to the Data Availability Statement at the start of the manuscript PDF file)?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception. The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

--------------------

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS Digital Health does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

--------------------

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: The authors have in technical terms replied to all the points, however, calibration for example is about making a method work in practice or at least substantiating that it has that potential. The are also other points I believe are not fully clear and overall the paper is not ready for publication in my view.

My main concerns are in particular the two aspects, model calibration and feature interpretation. I understand and agree that calibration curves in cases where the dataset is very imbalanced are often far from the actually observed probabilities. But in these cases one would expect to see mitigation with Platt scaling or isotonic regression on the probabilities. If the decision is to proceed without calibration adjustment, I would still include those calibration curves in the supplementary and specify in the Discussion this as a limitation. The model as it stands can only be used to classify two groups depending on a single decision threshold and cannot provide an individualized survival probability – and this is in my view a limiting factor. Another comment related to the revised version is that the authors did not use any hyperparameter search. As they have shown in the supplementary materials, they used the models with the default parameters as directly provided by scikit-learn. This may not be a problem per-se but I wonder if they could have obtained better performances if they had explored other model configurations. This again makes this translational value unclear. It would even be unethical to use a method in the clinic that was suboptimal, given the data, and we do not know if this is the case.

The feature interpretation is still not clear to me. The response to my previous question “In particular, we use Recursive Feature Elimination, i.e. we remove iteratively only one feature at a time to analyze the observed loss in prediction performances after training and testing again the algorithm. ” seems not to be correct. RFE does not iteratively remove only one feature at a time. Reference 42 used to explain the method says .. “(a) running RF to determine initial importance scores, (b) removing the bottom 3% of variables with the lowest importance scores from the data set (3% was chosen because of the high computational demands of using a lower threshold; this resulted in a total of 324 RF runs), and (c) assigning ranks to removed variables according to the order in which they were removed and their most recent importance scores (i.e., importance scores are only compared within runs, not between runs). This was performed iteratively using the reduced data set created in step two until 3% of the "number of remaining variables" rounds to zero (i.e., no further variables could be removed).” This definition is also consistent with how sklearn (which was not used by the authors to calculate RFE in their notebook) defines it (https://scikit-learn.org/stable/modules/generated/sklearn.feature_selection.RFE.html). The method "removes one feature at a time" to my knowledge would not account for correlated variables as it is done in reference 42 with RFE.

This paper does not really propose a novel framework (as the authors state in the Discussion), and overall I think the implementation of the standard methods along with the limited clinical utility of it makes the paper somewhat incremental.

--------------------

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

Do you want your identity to be public for this peer review? If you choose “no”, your identity will remain anonymous but your review may still be made public.

For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

--------------------

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLOS Digit Health. doi: 10.1371/journal.pdig.0000459.r005

Decision Letter 2

Martin G Frasch, Nan Liu

26 Oct 2023

PDIG-D-23-00039R2

Identifying prognostic factors for survival in intensive care unit patients with SIRS or sepsis by machine learning analysis on electronic health records

PLOS Digital Health

Dear Dr. Chicco,

Thank you for submitting your manuscript to PLOS Digital Health. After careful consideration, we feel that it has merit but does not fully meet PLOS Digital Health's publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript within 60 days Dec 25 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at digitalhealth@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pdig/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

* A rebuttal letter that responds to each point raised by the editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

* A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

* An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

We look forward to receiving your revised manuscript.

Kind regards,

Nan Liu

Academic Editor

PLOS Digital Health

Journal Requirements:

Additional Editor Comments (if provided):

Thank you for addressing Reviewer 2's comments. We noticed that Reviewer 1's comments were accidentally stored in the wrong place, thus you were unable to see them. Below please find Reviewer 1's comments for Revision 1 & Revision 2. Apologies for any inconvenience.

----------------------------------------

Reviewer 1 comments on Revision 1

Thank you to Mollura and colleagues for the effort in addressing the initial questions and suggestions - very much appreciated.

I am, however, unsure about the clinical relevance and how this study could possible possible impact patient care. My main concerns are:

1. From the response, the data being used for this study is from the study: https://doi.org/10.1371/journal.pone.0148699. This data was collected over a decade ago. Patient characteristics and disease profiles change based on advances in medical practice. Over the past decade, there has been significant advances and patient profiles have changed. What is the impact of this on the study findings? Risk predictions models, for example, are revised/updated to deal which such changes.

2. The research idea and concept is great, but is the data being used the most appropriate to address this research question?

3. Calibration curves are highlighting significant imbalanced data. The models are consistently underestimating real probabilities. Could this be discussed/highlighted in the discussion section? How will this possibly impact on clinical implications?

----------------------------------------

Reviewer 1 comments on Revision 2

My review was not made available to the authors in the decision letter, hence these have not been addressed.

Additionally, based on authors' response to concerns from the other reviewer, I am not convinced about the clinical importance and how the findings can be translational at this stage.

Major concerns:

1. Small sample size

2. No external validation of the models that have been developed. The current model after further sensitivity analysis is overfitting.

3. Authors choosing to change labels of methods supposedly used in the analysis without evidence to support the label.

Many thanks for the opportunity to review this.

Kind regards

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: (No Response)

Reviewer #2: All comments have been addressed

--------------------

2. Does this manuscript meet PLOS Digital Health’s publication criteria? Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe methodologically and ethically rigorous research with conclusions that are appropriately drawn based on the data presented.

Reviewer #1: (No Response)

Reviewer #2: Yes

--------------------

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: No

Reviewer #2: Yes

--------------------

4. Have the authors made all data underlying the findings in their manuscript fully available (please refer to the Data Availability Statement at the start of the manuscript PDF file)?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception. The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

--------------------

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS Digital Health does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

--------------------

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: Unfortunately some sections were modified but not marked in blue. Nevertheless I think the authors addressed all the comments I had. I would add in limitations that the SFE method is not fully informative when features are highly interdependent – since the contribution of a feature that is very important may still result as not important during SFE because of its contribution through other covariates that depend on it. With this additional remark in the Discussion I believe the manuscript is ready for publication.

--------------------

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

Do you want your identity to be public for this peer review? If you choose “no”, your identity will remain anonymous but your review may still be made public.

For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

--------------------

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLOS Digit Health. doi: 10.1371/journal.pdig.0000459.r007

Decision Letter 3

Martin G Frasch, Nan Liu

5 Feb 2024

Identifying prognostic factors for survival in intensive care unit patients with SIRS or sepsis by machine learning analysis on electronic health records

PDIG-D-23-00039R3

Dear Dr. Chicco,

We are pleased to inform you that your manuscript 'Identifying prognostic factors for survival in intensive care unit patients with SIRS or sepsis by machine learning analysis on electronic health records' has been provisionally accepted for publication in PLOS Digital Health.

Before your manuscript can be formally accepted you will need to complete some formatting changes, which you will receive in a follow-up email from a member of our team. 

Please note that your manuscript will not be scheduled for publication until you have made the required changes, so a swift response is appreciated.

IMPORTANT: The editorial review process is now complete. PLOS will only permit corrections to spelling, formatting or significant scientific errors from this point onwards. Requests for major changes, or any which affect the scientific understanding of your work, will cause delays to the publication date of your manuscript.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they'll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact digitalhealth@plos.org.

Thank you again for supporting Open Access publishing; we are looking forward to publishing your work in PLOS Digital Health.

Best regards,

Nan Liu

Academic Editor

PLOS Digital Health

***********************************************************

Reviewer Comments (if any, and for reference):

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix

    Text A: Formulas of the confusion matrix rates. Text B: Model Hyperparameters. Text C: Model performances. Text D: Model Calibration. Text E: Single Feature Elimination. Text F: Sensitivity Analysis: Hyperparameter Optimization.

    (PDF)

    pdig.0000459.s001.pdf (227.7KB, pdf)
    S1 Fig. Average calibration curves across the MCCV runs for the best (left column) and worst (right column) for SEPSIS (upper row) and SIRS (lower row).

    (EPS)

    pdig.0000459.s002.eps (1.5MB, eps)
    Attachment

    Submitted filename: reply_to_reviewers.pdf

    pdig.0000459.s003.pdf (348.1KB, pdf)
    Attachment

    Submitted filename: reply_to_reviewers2.pdf

    pdig.0000459.s004.pdf (85.1KB, pdf)
    Attachment

    Submitted filename: 3rd reply to reviewers.pdf

    pdig.0000459.s005.pdf (99.7KB, pdf)

    Data Availability Statement

    The dataset is publically available at the following URL: https://figshare.com/articles/dataset/_C_Reactive_Protein_and_Hemogram_Parameters_for_the_Non_Sepsis_Systemic_Inflammatory_Response_Syndrome_and_Sepsis_What_Do_They_Mean_/1644426.


    Articles from PLOS Digital Health are provided here courtesy of PLOS

    RESOURCES