Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2022 Nov 12;22(22):8757. doi: 10.3390/s22228757

Interpretable Machine Learning for Inpatient COVID-19 Mortality Risk Assessments: Diabetes Mellitus Exclusive Interplay

Heydar Khadem 1,*, Hoda Nemat 1, Jackie Elliott 2,3, Mohammed Benaissa 1
Editors: Christoph M Friedrich, Sebastian Zaunseder
PMCID: PMC9692305  PMID: 36433354

Abstract

People with diabetes mellitus (DM) are at elevated risk of in-hospital mortality from coronavirus disease-2019 (COVID-19). This vulnerability has spurred efforts to pinpoint distinctive characteristics of COVID-19 patients with DM. In this context, the present article develops ML models equipped with interpretation modules for inpatient mortality risk assessments of COVID-19 patients with DM. To this end, a cohort of 156 hospitalised COVID-19 patients with pre-existing DM is studied. For creating risk assessment platforms, this work explores a pool of historical, on-admission, and during-admission data that are DM-related or, according to preliminary investigations, are exclusively attributed to the COVID-19 susceptibility of DM patients. First, a set of careful pre-modelling steps are executed on the clinical data, including cleaning, pre-processing, subdivision, and feature elimination. Subsequently, standard machine learning (ML) modelling analysis is performed on the cured data. Initially, a classifier is tasked with forecasting COVID-19 fatality from selected features. The model undergoes thorough evaluation analysis. The results achieved substantiate the efficacy of the undertaken data curation and modelling steps. Afterwards, SHapley Additive exPlanations (SHAP) technique is assigned to interpret the generated mortality risk prediction model by rating the predictors’ global and local influence on the model’s outputs. These interpretations advance the comprehensibility of the analysis by explaining the formation of outcomes and, in this way, foster the adoption of the proposed methodologies. Next, a clustering algorithm demarcates patients into four separate groups based on their SHAP values, providing a practical risk stratification method. Finally, a re-evaluation analysis is performed to verify the robustness of the proposed framework.

Keywords: COVID-19, diabetes mellitus, machine learning, SHAP

1. Introduction

Shortly after the outbreak of coronavirus disease-2019 (COVID-19), pre-existing diabetes mellitus (DM) was recognised as a risk factor for the new disease [1,2]. Subsequently, extensive research has been underway to study this vulnerability [3]. For instance, adopting logistic regression (LR) analysis, Sourij et al. investigated the prognostic prediction in hospitalised COVID-19 patients with DM. They also offered a simple yet effective score for forecasting the risk of fatal outcomes from age and on-admission values of arterial occlusive disease, c-reactive protein (CRP), estimated glomerular filtration rate, and aspartate aminotransferase [4].

DM comorbidity was later declared a leading cause of in-hospital COVID-19 mortality in some studies [5]. As an example, in a uni-centre retrospective study, Ciardullo et al. deployed LR to perform death prediction analysis for 373 hospitalised COVID-19 patients with DM from diabetes status, comorbid conditions, and laboratory data. Based on the results achieved, the authors affirmed DM as an independent culprit for in-hospital COVID-19 mortality [6].

Although the COVID-19 susceptibility of DM patients has been well documented, explaining this vulnerability remains a challenge [3,7]. The use of explainable machine learning (ML) is one strategy to contribute to addressing this challenge [8,9,10,11].

In general, ML algorithms carry excellent potency in discovering intricate correlated interactions [8,12]. These tools have found practical implementation in COVID-19 research. As a representative, using machine learning techniques, Kar et ll. studied 63 clinical and laboratory factors in relation to 1393 subjects hospitalised for COVID-19 to forecast the probability of mortality at 7 and 28 days. As a result, they generated an effective bespoke death risk score [13].

Integrating underlying ML pipelines with model interpretation frameworks promotes the transparency of the analysis, offsetting the block-box nature of plain ML algorithms [14,15,16]. SHapley Additive exPlanations (SHAP) is an exemplar of elaborate ML explainability techniques [17,18]. SHAP employs the classical notion of Shapley values from cooperative game theory to measure the contribution of input data in forming a given output by the model [19]. The measured SHAP values for a particular input feature indicate the deviation from the average prediction when conditioned on that feature [18].

SHAP analysis has seen promising applications in COVID-19 risk assessment research [19,20,21]. For example, Pan et al. designed ML models dressed with SHAP analysis for COVID-19 prognosis assessment in individuals hospitalised in intensive care units [20].

In a recent publication, we developed machine learning pipelines incorporated with interpretation components for mortality risk prediction and stratification in hospitalised COVID-19 patients with and without DM in parallel [22]. For this purpose, a set of features collected at the point of hospital admission for both groups were investigated. Consequently, the generated risk assessment models possessed potential application to triage systems for both groups and enabled inter-cohort comparative analysis. In the original dataset, though, there existed a pool of historical, on-admission, and inpatient variables collated only for the DM group. These features were either DM-relevant, or the primary investigations did not persuade the clinical data acquisition team to collate for the non-DM cohort. Due to the significance of the topic and the value of further knowledge discovery in this field, the current sequel study is conducted to create risk assessment platforms comparable to those in the earlier study for the same DM cohort by scrutinising only the abovementioned DM-exclusive data opted out of the former investigation.

First, the clinical data are cleansed and prepared for formal ML modelling analysis. After that, an ML model is constructed for inpatient fatality risk assessments. In-depth evaluation and interpretation analyses are performed on the model, and the results obtained are discussed in detail. This follow-on work initially recruits similar main methods contrived in the prior paper, focusing on new findings, discussions, and applications. This homogeneity facilitates the analogical study of the two relevant articles. Next, some compartments in the primary skeleton of the pipelines are replaced with new units, and the investigations are re-conducted. This complementary analysis further inspects the robustness of the infrastructure utilised in the two studies by a side-by-side comparison of the new and old outcomes.

The remainder of the paper is organised as follows. The clinical data utilised in this work are outlined in Section 2. In Section 3, data pre-treatment steps undertaken before the conventional ML modelling analysis are explained. Section 4 describes the primary methodologies implemented for mortality risk assessments. The results achieved and the associated discussion is represented in Section 5. Section 6 reports further stability investigations on the proposed work frames. Finally, Section 7 summarises and concludes the work.

2. Material

The source of the clinical data used in this paper is the dataset primarily described in [23]. The present research explores demographic, clinical, and laboratory data from 156 individuals in the main dataset with confirmed COVID-19 and DM comorbidity. All these participants were admitted to Sheffield Teaching Hospitals (Sheffield, UK) between 29 February 2020 and 1 May 2020. Of the 156 patients, 103 survived, and 51 died due to COVID-19; the other three died due to causes other than COVID-19, according to their death certificates. Table 1 and Table 2 summarise the statistical characteristics of the data used in this work. Table 1 includes categorical variables’ information encompassing the name of categories and the number of recorded data in each category. In Table 2, the mean and standard deviation (SD) of numerical variables are presented alongside the frequency of records for each feature.

Table 1.

A summary of properties of the categorical clinical data used in this article. For each feature, the categories’ names and the number of patients with recorded data in each category are given.

Feature Category Count Feature Category Count
CKD Yes 131 PC-Dizziness Yes 7
No 21 No 145
Radiology-RPLD Yes 12 PC-Headache Yes 5
No 132 No 147
Radiology-Consolidation on report Yes 99 PC-Hyperglycaemia Yes 12
No 46 No 140
Radiology-Worsening consolidation Yes 34 DM type Type 1 12
No 25 Type 2 140
Radiology-CT chest or CTPA Yes 5 DM complications Yes 127
No 143 No 26
Radiology-PE Yes 1 PVD Yes 47
No 5 No 105
Diabetes autoantibodies Yes 1 Peripheral neuropathy Yes 67
No 152 No 84
PC-Fever Yes 64 Background retinopathy Yes 73
No 88 No 64
PC-Cough Yes 81 Preproliferative retinopathy Yes 15
No 72 No 122
PC-SOB Yes 66 Proliferative retinopathy Yes 17
No 86 No 120
PC-Chest pain Yes 11 Previous foot ulcer Yes 28
No 141 No 124
PC-Abdominal pain Yes 8 Active foot ulceration Yes 8
No 144 No 144
PC-Diarrhoea Yes 21 VRII treatment during admission Yes 23
No 131 No 129
PC-Myalgia Yes 16 DNAR Yes 109
No 136 No 41

Note. CKD: chronic kidney disease; CT: computed tomography; CTPA: computed tomography pulmonary angiogram; DNAR: do not attempt resuscitation; PC: presenting complaint; PE: pulmonary embolism; PVD: peripheral vascular disease; RPLD: reported pre-existing lung disease; SOB: shortness of breath; VRII: variable rate intravenous insulin infusion.

Table 2.

A summary of properties of the numerical clinical data used in this article. For each feature, mean and standard deviation, together with the number of patients that a value is recorded for the feature, are given.

Feature Mean ± SD Count Feature Mean ± SD Count
BGV-pH (mmol/L) 7.39 ± 0.09 80 LV-ALPO4 (g/L) 87.27 ± 49.08 100
BGV-HCO3 (mmol/L) 22.43 ± 4.33 79 HV-ALPO4 (g/L) 134.35 ± 87.84 100
BGV-Lactate (mmol/L) 2.09 ± 1.71 80 LAYBA-Albumin (g/L) 39.82 ± 5.17 149
BGV-Na (mmol/L) 134.83 ± 5.45 80 LV-Albumin (g/L) 30.15 ± 4.91 104
BGV-K (mmol/L) 4.13 ± 0.86 80 HV-Albumin (g/L) 37.4 ± 4.2 104
BGV-Cl (mmol/L) 99.65 ± 10.91 80 LV-CRP (mg/dL) 49.37 ± 57.93 133
BGV-Anion Gap (mmol/L) 16.86 ± 10.44 79 HV-CRP (mg/dL) 165.89 ± 124.04 133
LAYBA-HbA1c (mmol/mol) 61.34 ± 17 144 LV-Procalcitonin (µg/L) 0.58 ± 1.04 68
FI-HbA1c (mmol/mol) 68.01 ± 19.52 74 HV-Procalcitonin (µg/L) 3.04 ± 11.98 68
LAYBA-Hb (g/L) 123.9 ± 20.89 146 LV-Ferritin (µg/L) 1019.87 ± 1363.47 45
LV-Hb (g/L) 106.5 ± 20.22 140 HV-Ferritin (µg/L) 1596.16 ± 2861.80 45
HV-Hb (g/L) 125.91 ± 18.29 140 OA-Troponin (ng/L) 49.07 ± 70.64 25
LAYBA-WCC (g/L) 8.09 ± 2.95 147 LV-Troponin (ng/L) 90.75 ± 224.18 32
LV-WCC (g/L) 6.03 ± 2.84 140 HV-Troponin (ng/L) 108.94 ± 237.01 32
HV-WCC (g/L) 11.01 ± 5.69 140 LAYBA-UACR 35.07 ± 76.42 62
LAYBA- NEUT (109/L) 5.35 ± 2.47 147 LAYBA-Vitamin D (ng/mL) 44.25 ± 25 44
LV- NEUT (109/L) 4.24 ± 2.43 140 LAYBA-PT (s) 12.26 ± 4.12 92
HV- NEUT (109/L) 8.72 ± 5.05 140 LV-PT (s) 12.08 ± 2.46 65
LAYBA-LYM (109/L) 1.76 ± 1.01 147 HV-PT (s) 16.54 ± 14.43 65
LV- LYM (109/L) 0.93 ± 1.41 140 LAYBA-APTT (s) 27.18 ± 5.48 93
HV- LYM (109/L) 1.66 ± 1.93 140 LV-APTT (s) 26.41 ± 4.77 66
LaV- LYM (109/L) 1.34 ± 1.45 151 HV-APTT (s) 33.27 ± 11.81 66
LAYBA-MN (109/L) 0.69 ± 0.32 147 LAYBA-Fibrinogen (g/L) 5.14 ± 1.42 94
OA-MN (109/L) 0.68 ± 0.41 148 LV-Fibrinogen (g/L) 5.43 ± 1.24 66
LV-MN (109/L) 0.38 ± 0.24 140 HV-Fibrinogen (g/L) 6.44 ± 1.1 66
HV-MN (109/L) 0.87 ± 0.46 140 LV-D-dimer (µg/L) 5313.47 ± 10,366.62 15
LAYBA-Platelets (1/mL) 253.84 ± 102.75 147 HV-D-dimer (µg/L) 5397.80 ± 10,339.12 15
LV-Platelets (1/mL) 196.06 ± 82.68 140 FI-BGL (mmol/L) 10.63 ± 5.84 152
HV-Platelets (1/mL) 332.41 ± 158.11 140 FI-Ketones (mmol/L) 1.05 ± 1.71 54
LAYBA-Na (mmol/L) 138.21 ± 3.45 150 Preadmission-RR (1/min) 23.89 ± 8.04 150
LV-Na (mmol/L) 133.49 ± 4.56 142 Preadmission-Saturations (%) 90.5 ± 9.96 142
HV-Na (mmol/L) 140.15 ± 5.77 142 Preadmission-Temperature (°C) 37.47 ± 1.21 149
LAYBA-K (mmol/L) 4.56 ± 0.54 150 Preadmission-SBP (mmHg) 139.27 ± 24.98 150
LV-K (mmol/L) 3.86 ± 0.55 140 Preadmission-DBP (mmHg) 75.56 ± 14.14 150
HV-K (mmol/L) 4.94 ± 0.86 140 Preadmission-Pulse (1/min) 90.65 ± 24.15 150
LAYBA-Urea (mmol/L) 8.85 ± 5.45 150 DM duration (years) 14.15 ± 10.8 145
LV-Urea (mmol/L) 7.75 ± 5.5 142 FI-RR (1/min) 26.26 ± 7.88 152
HV-Urea (mmol/L) 14.51 ± 9.28 142 FI-Saturations (%) 91.94 ± 6.24 152
LAYBA-Creatinine (µmol/L) 140.88 ± 125.46 150 FI-FiO2 (%) 47.16 ± 22.41 103
LV-Creatinine (µmol/L) 142.98 ± 156.87 142 FI-Temperature (°C) 37.1 ± 1.4 152
HV-Creatinine (µmol/L) 223.4 ± 244.63 142 FI-SBP (mmHg) 127.74 ± 31.82 152
eGFR 2.62 ± 1.1 152 FI-Pulse (1/min) 94.47 ± 22.33 152
LAYBA-Bilirubin (µmol/L) 7.55 ± 4.08 149 HR-O2 (%) 55.46 ± 22.85 124
LV-Bilirubin (µmol/L) 6.73 ± 3.93 101 LV-BGL (mmol/L) 4.92 ± 2.34 152
HV-Bilirubin (µmol/L) 12.86 ± 11.02 101 HV-BGL (mmol/L) 16.51 ± 6.65 152
LAYBA-ALT (u/L) 20.37 ± 13.93 142 Average BGL (mmol/L) 9.55 ± 3.02 152
LV-ALT (u/L) 29.38 ± 41.64 68 RBGLR below 3 mmol/L 0.01 ± 0.03 151
HV-ALT (u/L) 76.91 ± 160.6 68 RBGLR 4—10 mmol/L 0.61 ± 0.3 151
LAYBA-TP(g/L) 68.75 ± 7.36 149 RBGLR 10.1—14 mmol/L 0.22 ± 0.19 151
LV-TP (g/L) 59.62 ± 7.51 102 RBGLR 14—21 mmol/L 0.11 ± 0.15 151
HV-TP (g/L) 70.02 ± 6.6 102 RBGLR 21.0—27.8 mmol/L 0.03 ± 0.07 151
LAYBA-ALPO4 (g/L) 99.43 ± 45 149 RBGLR above 27.8 mmol/L 0.01 ± 0.03 151

Note. ALT: alanine transaminase; ALPO4: alkaline phosphatase; APTT: activated partial thromboplastin time; BGL: blood glucose level; BGV: blood gas value; Cl: chloride; CRP: c-reactive protein; DBP: diastolic blood pressure; D-dimer: disseminated intravascular coagulation; DM: diabetes mellitus; eGFR: estimated glomerular filtration rate; FI: first inpatient; FiO2: fraction of inspired oxygen; HbA1c: glycated haemoglobin; HCO3: bicarbonate; HR: highest requirement; HV: highest value; K: potassium; LaV: last value; LAYBA: latest available within one year before admission; LV: lowest value; LYM: lymphocytes; MN: monocytes; Na: sodium; NEUT: neutrophils; O2: oxygen; OA: on admission; pH: potential of hydrogen; PT: prothrombin time; RBGLR: ratio of blood glucose level readings; RR: respiratory rate; SBP: systolic blood pressure; SD: standard deviation; TP: total protein; WCC: white cell count. Note. eGFR > 90 = Stage 1, 60–89 = Stage 2, 30–59 = Stage 3, 15–29 = Stage 4, <15 = Stage 5-Stage CKD.

3. Data Curation

The following four pre-treatment stages are undertaken to prepare the data for the ensuing ML modelling analysis.

3.1. Cleaning

In the first data-cleaning step, tainted entities and features are excluded from the rest of the analysis. First, the three individuals with reported non-COVID-19 mortality reasons (as discussed in Section 2) are omitted from the rest of the analysis. Next, features and participants with a high missingness rate are discarded. For this purpose, an inclusion criteria of having a missingness rate of no more than 50% is determined for both features and individuals [24]. Initially, attributes with missing rates larger than the 50% threshold are discarded. Next, the same criterion is applied to data contributors. As a result, the following features are omitted from the rest of the analysis: FI-HbA1c, LV-ALT, HV-ALT, LV-Procalcitonin, HV-Procalcitonin, LV-Ferritin, HV-Ferritin, OA-Troponin, LV-Troponin, HV-Troponin, LAYBA-UACR, LAYBA-Vitamin D, LV-PT, HV-PT, LV-APTT, HV-APTT, LV-Fibrinogen, HV-Fibrinogen, LV-D-dimer, HV-D-dimer, and FI-Ketones. However, no further individual is obviated from the rest of the analysis, as no one holds more than 50% missingness after discarding the abovementioned high-missing rate features.

3.2. Subsetting

After the cleaning phase, data that have met the inclusion criteria and qualified for the subsequent analysis are subdivided into training and testing sets as per the requirements of upcoming supervised ML analysis. For data subsetting, 70% of the cases are allocated as the training set and 30% as the testing set. Stratified random sampling carries out the train-test split process to take into account the distribution of classes. All model training and hyperparameter tuning operations are undertaken on training sets only, with testing sets remaining unseen for evaluation and model interpretation analysis.

3.3. Pre-Processing

Three pre-processing steps are conducted to render the data more suitable for ML analysis: outlier treatment, missing value imputation, and feature transformation.

Initially, leveraging the winsorisation technique, we shift the numerical variables placed outside the 5th to 95th percentile to the corresponding boundary. This confinement pre-empts extreme values of skewing the results.

Next, the missing values for numerical features are treated using the k-nearest neighbour data imputation technique [25], configuring the number of neighbours as five. Exploring all non-missing features, the algorithm selects five data entities from the training set with the most congruency with a given data contributor. Then the average of these akin points is used to interpolate the missing values of the given data instance. For categorical variables, the most repeated value is used to fill in the missing values.

Lastly, features are transformed into a more digestible form for ML algorithms. Categorical attributes are converted to numerical form using the binary encoding technique. The numerical features are standardised. The mean of the training set is subtracted from each feature, and then the results are scaled to unit variance by dividing them by the standard deviation of the training set.

3.4. Feature Elimination

After the data curation steps, a voting feature selection is performed on the pre-processed data to reduce the input size and help preclude the occurrence of a dimensionality curse. First, regular LR, gradient boosting (GB), and AdaBoost (AB) models, which all have already succeeded in applications to COVID-19 research, are fine-tuned. To this end, the random search approach is used to select the hyperparameter values delivering the highest five-fold cross-validation accuracy on the training set. The outcomes of hyperparameter tuning are given in Table A1, Appendix. Next, the recursive feature elimination technique is enfolded around each model, forming a voting system. Each voting system then shortlists 15 features (approximately one-tenth the number of data points, a common practice in ML modelling) by investigating training data only. The features picked by at least two voting systems are then used as predictors to generate the final mortality risk prediction model. The shortlisted features encompass LAYBA-NEUT, HV-NEUT, LaV-LYM, LAYBA-MN, OA-MN, LV-Platelets, HV-Platelets, LV-CRP, LAYBA-PT, FI-BGL, FI-RR, FI-FiO2, and HR-O2.

4. Modelling

This section develops explainable ML models for mortality risk assessment analysis from the selected features. Prior to representing model implementations, providing a brief description of SHAP theory and calculations is of use.

4.1. Preliminary

As a game-theoric model agnostic method, SHAP simulates the formation of outputs by an ML model as a game. In this gamification process, the input features have the role of involved players. Subsequently, the payoff for each player in the game is calculated as Equation (1) [18], based on the principles of Shapley value [19]. On elucidating the formula, the SHAP value of a particular feature for a given individual is calculated by integrating the payoff shares for the feature in all possible coalitions with other variables. The payoff share of the feature in each coalition is determined by calculating the difference between the whole payoff of the coalition with and without the given feature included and then dividing the outcome between the members of the coalition equally.

SHAPx(f)=F:fF(|F|×(N|F|))1×(x^Fx^F\f) (1)

f: a given feature; x: a given data point; SHAPx(f): SHAP value of variable f for x (the payoff of feature f in the designed game); F: all permutations of feature with f included; |F|: the size of F (number of features in F); N: the whole number of features in the models; x^F: the model’s output for x using the feature subset F; x^F\f: the model’s output for x from the feature subset F excluding f.

4.2. Mortality Risk Prediction

The first risk assessment analysis is to forecast in-hospital COVID-19 mortality from selected features. For this purpose, a random forest (RF) classifier, which has proven its capability in COVID-19 risk assessment research [26], is fine-tuned using the same approach explained in Section 3.4. The results of this optimisation analysis are presented in Table A1, Appendix. The fine-tuned RF classifier is then trained on the entire training set to predict inpatient death due to COVID-19. Following that, the generated model undergoes careful evaluation analysis employing four widely used metrics: accuracy, area under the receiver operating characteristic curve (AUC), sensitivity, and specificity. After evaluating the mortality risk prediction model, SHAP is leveraged to interpret the model globally and locally.

4.3. Mortality Risk Stratification

The second risk assessment analysis is to stratify the in-hospital mortality risk of patients. In order to do so, SHAP clustering [27], an extension of SHAP analysis, is deployed. The k-means [28], an algorithm used in previous COVID-19 research [29,30], is employed on SHAP values to search for meaningful clusters of individuals. The algorithm clusters the subjects into an optimised number of groups [22] with identical variance by optimising a criterion known as inertia [28]. For deciding the number of clusters, the heuristic elbow method is employed. Values of 1 to 9 are explored, and the one resulting in an elbow point, based on inertia values achieved, is chosen [31]. According to the outcome of elbow analysis shown in Figure 1, four is determined as the number of clusters, as the diagram has the sharpest break point for this value.

Figure 1.

Figure 1

A schematic of elbow analysis operated to decide the number of clusters.

5. Results and Discussion

This section presents the results of model evaluation and interpretation analysis alongside the corresponding discussion. First, the outcomes of mortality risk prediction analysis are given and then those of mortality risk stratification analysis.

5.1. Mortality Risk Prediction

5.1.1. Evaluation

The generated mortality risk prediction model provides these evaluation results across the testing set: 97% accuracy, 78% AUC, 78% sensitivity, and 80% specificity. Such practical evaluation results support the overall effectiveness of the implemented methodologies, including the feature selection, and hyperparameter tuning processes coupled with the final RF classifier. These dependable outcomes also backend the following SHAP-based analysis.

5.1.2. Global Interpretations

The next results to present for the mortality risk prediction model are the outcome of the interpretation analysis. In this regard, first, the results of global interpretation analysis are reported, followed by those of local interpretation analysis.

Figure 2 illustrates the results of global interpretations for the generated mortality risk prediction model in two plots. Both plots represent the features in descending order as per their overall influence on the model’s outcomes.

Figure 2.

Figure 2

Global interpretation plots for the developed inpatient COVID-19 mortality prediction model. (A) Bee swarm SHAP values plot, (B) SHAP summary importance plot. The bee swarm plot shows all SHAP values in accord with predictors values. The summary plot presents predictors in descending order based on their overall importance on the model’s outcomes derived from mean absolute SHAP values. Note. BGL: blood glucose level; CRP: c-reactive protein; FI: first inpatient; FiO2: fraction of inspired oxygen; HR: highest requirement; HV: highest value; LaV: last value; LAYBA: latest available within one year before admission; LV: lowest value; LYM: lymphocytes; MN: monocytes; NEUT: neutrophils; OA: on admission; O2: oxygen; PT: prothrombin time; RR: respiratory rate; SHAP: SHapley Additive exPlanations.

The bee swarm plot in Figure 2A shows SHAP values and their relative association with predictions of death. Each point on the scheme represents a feature value from the testing set. The values of features are colour-coded from blue to red, encoding low to high values. A positive SHAP value for each point denotes the adverse effect of the feature, viz, its contribution level to a higher risk of death. In contrast, a negative SHAP value indicates the protective effect of the relevant feature, i.e., decreasing the risk of death.

The bar chart in Figure 2B is the variable importance plot for the developed mortality risk prediction model. The plot summarises the features’ overall impacts on the model outputs according to their mean absolute SHAP values, represented by the length of the bars.

According to Figure 2A, the predictors positively associated with mortality risk predictions are HR-O2, FI-FiO2, HV-NEUT, FI-RR, LV-CRP, LAYBA-PT, OA-MN, LAYBA-NEUT, and LAYBA-MN. By comparing the positive and negative SHAP values of these variables, it is noticeable that HR-O2, FI-FiO2, HV-NEUT, and LAYBA-PT show greater adverse than protective effects. In other words, the sinister roles of higher values for these features are relatively more significant than the protective roles of lower values. On the other hand, with similar explanations, it can be inferred that FI-RR carries a stronger protective than adverse power. The rest of the aforementioned variables have comparable protective and adversarial influences. On the other side, based on the plot, it can also be seen that the modalities negatively associated with the prediction of death comprise LaV-LYM, HV-Platelets, and LV-Platelets. Overall, the first two variables possess more substantial sinister impacts (in lower values) than protective impacts (in higher values), whereas the last one holds stronger protective effects (in higher values) than sinister effects (in lower values).

Furthermore, based on Figure 2A, one noteworthy inference is that a more influential feature does not necessarily have stronger adverse and protective power at once. To exemplify, notwithstanding the greater overall importance of HR-O2 over FI-FiO2, the latter possesses a more substantial adverse impact than the former on average. This deduction is formed based on predominantly bigger positive SHAP values for FI-FiO2 compared to HR-O2.

Additionally, according to the plots in Figure 2, HR-O2, FI-FiO2, and HV-NEUT form the top three influential features with considerably higher impacts than others. Therefore, undesired measures for these features may be an indicator of high death risk. These findings underscore the importance of careful inpatient surveillance and the monitoring of peak values of oxygen requirement and NEUT, along with the imperative role of immediate inspection of FI-FiO2 after admission for COVID-19 patients with DM.

Another notable point is that two features from the patients’ historical profiles, LAYBA-PT and LAYBA-NEUT, have shown considerable mortality predictivity power even in the presence of many on-admission and during-admission data. This observation stresses the potential utility of accessing and considering the history profile of COVID-19 patients with DM, and specifies two features as candidates with high priority for consideration in this respect.

5.1.3. Local Interpretations

The next outcome to be presented entails the results of local interpretation analysis for the mortality risk prediction model. In this respect, Figure 3 shows the waterfall plots for two randomly selected examples of data entities, one from individuals with death and the other from those with survival as the outcomes of their admissions. These plots start at the base from E[f(x)], representing the average risk of death according to the training set. Next, each arrow illustrates the influence of a feature, i.e., the feature’s SHAP value, towards forming the specific prediction for the given entry. The positive associations of the given feature with increased mortality risk prediction are exhibited by red rightward arrows and the negative associations with blue leftward arrows. Finally, at the top of the plot, the model’s output for the given sample is represented by f(x). It merits mentioning that each arrow’s length denotes the level of impact from its relevant feature, i.e., absolute SHAP value. Moreover, the arrows are displayed in ascending order from the bottom to the top of the plots according to their size.

Figure 3.

Figure 3

The local interpretation plots for the developed inpatient COVID-19 mortality prediction model. (A) an example of patients with survival as the outcome of admission, (B) an example of patients with death as the outcome of admission. The plots start from the bottom with a predefined prediction for the risk of death equal to the average death rate in the training set. Next, the arrows with an ascending order show how each feature has contributed to the formation of a final prediction specified for the given data instance shown at the top of the plot. Note. BGL: blood glucose level; CRP: c-reactive protein; FI: first inpatient; FiO2: fraction of inspired oxygen; HR: highest requirement; HV: highest value; LaV: last value; LAYBA: latest available within one year before admission; LV: lowest value; LYM: lymphocytes; MN: monocytes; NEUT: neutrophils; OA: on admission; O2: oxygen; PT: prothrombin time; RR: respiratory rate; SHAP: SHapley Additive exPlanations.

One immediate recognition from both plots in Figure 3 is that the grade and order for the features’ impacts on local interpretations are different from global interpretations. This evidence shows how local interpretations can evolve the transparency of the analysis by explaining the formation of each specific outcome through localising and contrasting the effect of the components, as opposed to giving a generic explanation based on all outcomes.

For the death instance represented in Figure 3A, relatively high values for features FI-FiO2, HR-O2, HV-NEUT, and LAYBA-PT have been the most effective predictors of a fatal outcome. In this regard, it is worth remarking that, in line with the aforementioned discussion, FI-FiO2 had a more adverse impact than HR-O2, the most influential feature overall.

For the survival case reported in Figure 3B, features with protective impacts are HR-O2, FI-FiO2, LaV-LYM, and HV-Platelets. It is worth highlighting that this case has received a non-fatal outcome prediction, whilst its influential feature, HV-NEUT, shows an adverse impact. Moreover, for this data instance, HR-O2 and Fi-FO2 both deliver a protective effect, with the former’s being stronger. This perception is also in line with the overall higher protective influence of HR-O2, as discussed before.

5.2. Mortality Risk Stratification

Table 3 outlines the results of the SHAP clustering analysis, including the distribution of patients in the generated clusters, the rate of mortality outcome in each cluster, and a summary of the statistical characteristics of predictors within clusters. Based on the table, it can be apprehended that the clustering approach has made an appropriate risk stratification system by forming four categories with disparate characteristics.

Table 3.

The results of the performed SHAP clustering to generate a mortality risk stratification system.

Characteristics Cluster 1 Cluster 2 Cluster 3 Cluster 4
Count 21 7 10 8
Mortality rate (%) 0.00 ± 0.00 0.43 ± 0.53 0.70 ± 0.48 0.75 ± 0.46
HR-O2 37.31 ± 9.85 37.57 ± 10.83 80.00 ± 0.00 80.00 ± 0.00
FI-FiO2 36.49 ± 10.06 34.03 ± 10.55 78.00 ± 6.32 32.08 ± 6.87
HV-NEUT 6.08 ± 1.97 14.95 ± 3.46 8.99 ± 4.36 7.47 ± 3.32
FI-RR 22.74 ± 4.95 25.57 ± 7.35 31.80 ± 6.36 28.62 ± 6.7
LV-CRP 29.97 ± 21.68 42.67 ± 41.62 75.09 ± 62.95 37.35 ± 33.85
LaV-LYM 1.25 ± 0.5 1.18 ± 0.49 1.10 ± 0.4 0.99 ± 0.62
LAYBA-PT 11.57 ± 0.89 11.36 ± 0.58 12.32 ± 1.87 11.58 ± 0.51
OA-MN 0.72 ± 0.32 0.68 ± 0.3 0.48 ± 0.24 0.63 ± 0.3
LAYBA-NEUT 4.57 ± 1.56 5.19 ± 0.97 5.61 ± 1.73 5.32 ± 1.48
LV-Platelets 189.2 ± 69.29 175 ± 69.63 210.26 ± 70.85 151.43 ± 42.15
LAYBA-MN 0.71 ± 0.25 0.54 ± 0.15 0.64 ± 0.23 0.64 ± 0.13
HV-Platelets 272.1 ± 113.01 420.00 ± 157.79 366.96 ± 152.35 269.68 ± 103.64
FI-BGL 9.67 ± 4.13 13.09 ± 6.42 10.59 ± 4.37 9.85 ± 4.95

Note. BGL: blood glucose level; CRP: c-reactive protein; FI: first inpatient; FiO2: fraction of inspired oxygen; HR: highest requirement; HV: highest value; LaV: last value; LAYBA: latest available within one year before admission; LV: lowest value; LYM: lymphocytes; MN: monocytes; NEUT: neutrophils; OA: on admission; O2: oxygen; PT: prothrombin time; RR: respiratory rate; SD: standard deviation.

In terms of mortality rates, cluster 1 poses a zero mortality rate, cluster 2 has a moderate mortality rate, and clusters 3 and 4 have relatively high mortality rates. Further distinctive patterns can be found based on the feature distributions within clusters, specifically for the more critical variables. For example, a pronounced discriminator between clusters 1 and 2 compared to clusters 3 and 4 is that the first two have an average HR-O2 considerably lower than the other two. Additionally, comparing clusters 1 and 2, a prominent discrepancy is that the former, in general, includes patients with more desired values for HV-NEUT and LV-CRP. One more pattern to mention is that a significant distinguisher between clusters 3 and 4 is the relatively higher average values for FI-FiO2 and LV-CRP in cluster 3.

6. Complementary Analysis

This section presents some extra analysis embarked on for robustness assessments. The following set of amendments is applied to the compartments of the proposed learning environment. The data are reshuffled in their entirety, and a new round of 30–70 stratified random samplings is performed to reallocate training and testing sets. In addition, missing values are interpolated using a different technique by applying the iterative imputation. In order to do so, a Bayesian ridge regressor is set as the estimator for the numerical variables and an RF classifier for the categorical variables. In addition, the mortality risk prediction modelling is performed again using a support vector classifier (SVC). The SVC is fine-tuned using the random search approach described in Section 3.4, and the results are presented in Table A1, Appendix. Following these updates, the evaluation, interpretation, and clustering analyses are re-conducted. The new results are presented concisely below, and the consistency of the proposed core workflow in producing practical outcomes in line with previously discussed findings is inspected.

The features selected in the updated analysis are BGV-Na, BGV-Cl, LV-NEUT, HV-NEUT, LaV-LYM, LV-Platelets, HV-Platelets, LV-Albumin, LV-CRP, Preadmission-SBP, DM duration, FI-RR, FI-FiO2, and HR-O2. In comparison, the top six most important features, according to the previous analysis (HR-O2, FI-FiO2, HV-NEUT, FI-RR, LV-CRP, and LaV-LYM), are all shortlisted in the renewed analysis as well.

Furthermore, the updated fatality risk prediction model yields these new evaluation outcomes over the testing set: 87% accuracy, 92% AUC, 72% sensitivity, and 74% specificity. Similar to the primary analysis, these results are practical, with an outcome of more than 70% for every metric.

Moreover, Figure 4 illustrates the global interpretation plots for the renewed models. According to the plots, HR-O2 and FI-FiO2 are the first and second most important features, similar to the primary analysis. In addition, the top five ranks are occupied by the same features in the original and renewed analysis.

Figure 4.

Figure 4

Global interpretation plots for the updated inpatient COVID-19 mortality prediction model. (A) Bee swarm SHAP values plot, (B) SHAP summary importance plot. The bee swarm plot shows all SHAP values in accord with predictors values. The summary plot presents predictors in descending order based on their overall importance on the model’s outcomes derived from mean absolute SHAP values. Note. BGV: blood gas value; Cl: chloride; CRP: c-reactive protein; DM: diabetes mellitus; FI: first inpatient; FiO2: fraction of inspired oxygen; HR: highest requirement; HV: highest value; LaV: last value; LV: lowest value; LYM: lymphocytes; Na: sodium; NEUT: neutrophils; O2: oxygen; RR: respiratory rate; SBP: systolic blood pressure; SHAP: SHapley Additive exPlanations.

Furthermore, Table 4 shows the results of the new SHAP clustering analysis. For brevity, only the top four important features are shown in the table. As can be seen, similar to the primary analysis, the new SHAP clustering analysis successfully groups patients into four distinguishable categories.

Table 4.

The results of the performed updated SHAP clustering to generate a mortality risk stratification system.

Characteristics Cluster 1 Cluster 2 Cluster 3 Cluster 4
Count 23 10 7 6
Mortality rate (%) 0.04 ± 0.021 0.50 ± 0.53 0.71 ± 0.49 0.83 ± 0.41
HR-O2 40.07 ± 10.69 41.31 ± 10.70 80 ± 0.00 80 ± 0.00
FI-FiO2 34.07 ± 7.69 33.67 ± 7.16 80.00 ± 0.00 51.60 ± 11.36
HV-NEUT 5.68 ± 1.91 14.59 ± 3.60 9.04 ± 4.09 9.07 ± 4.76
FI-RR 22.78 ± 4.17 26.10 ± 7.29 28.71 ± 6.82 30.17 ± 6.46
LV-CRP 28.83 ± 14.59 38.44 ± 33.25 89.35 ± 62.95 106.84 ± 58.60

Note. FI: first inpatient; FiO2: fraction of inspired oxygen; HR: highest requirement; HV: highest value; LV: lowest value; NEUT: neutrophils; O2: oxygen; RR: respiratory rate; SD: standard deviation.

All in all, based on the discussion above, there is a significant agreement between the results of both the updated and the original analysis. This alignment promises the robustness of the core interpretable ML workflow proposed for mortality risk assessment in COVID-19 patients.

7. Conclusions

Inpatient COVID-19 mortality risk assessment specifically designed for patients with pre-existing DM was performed in this work. This goal was achieved by investigating a set of clinical features exclusively pertaining to DM and COVID-19 interplay for 156 individuals. Initially, the clinical data of the studied subjects were carefully pre-treated for the subsequent standard ML modelling analysis. After that, a mortality risk prediction model was created, exercising established ML pipelines. Evaluation analysis was then performed on the generated model. The results underpinned the effectiveness of the data treatment and modelling analysis. Afterwards, the generated model was interpreted globally and locally using SHAP. These interpretations help extend the transparency of the analysis. Next, a mortality risk stratification system was developed upon the outcomes of the SHAP analysis. Finally, an extra analysis was performed to further examine the stability of the core pipelines, where the outcomes corroborated this. The analysis reported in this work can be applied to online surveillance of hospitalised patients. The findings suggest some critical features to be reviewed more carefully in this monitoring process. To further expand upon this area of knowledge, future work could include more rigorous scrutiny of SHAP clustering results by devising a nested model interpretation mechanism.

Acknowledgments

We want to thank Ahmed Iqbal, Marni Greig, Muhammad Fahad Arshad, Thomas H Julian, and Sher Ee Tan for their efforts in collecting the clinical data used in this paper.

Abbreviations

AB: AdaBoost; ALT: alanine transaminase; ALPO4: alkaline phosphatase; APTT: activated partial thromboplastin time; BGL: blood glucose level; BGV: blood gas value; CKD: chronic kidney disease; Cl: chloride; COVID-19: coronavirus disease-2019; CRP: c-reactive protein; CT: computed tomography; CTPA: computed tomography pulmonary angiogram; DBP: diastolic blood pressure; D-dimer: disseminated intravascular coagulation; DM: diabetes mellitus; DNAR: do not attempt resuscitation; eGFR: estimated glomerular filtration rate; FI: first inpatient; FiO2: fraction of inspired oxygen; GB: gradient boosting; HbA1c: glycated haemoglobin; HCO3: bicarbonate; HR: highest requirement; HV: highest value; K: potassium; LaV: last value; LAYBA: latest available within one year before admission; LR: logistic regression; LV: lowest value; LYM: lymphocytes; ML: machine learning; MN: monocytes; Na: sodium; NHS: National health Service; NEUT: neutrophils; O2: oxygen; OA: on admission; PC: presenting complaint; PE: pulmonary embolism; pH: potential of hydrogen; PT: prothrombin time; PVD: peripheral vascular disease; RBGLR: ratio of blood glucose level readings; RPLD: reported pre-existing lung disease; RF: random forest; RR: respiratory rate; SBP: systolic blood pressure; SD: standard deviation; SHAP: SHapley Additive exPlanations; SOB: shortness of breath; SVC: support vector classifier; TP: total protein; VRII: variable rate intravenous insulin infusion; WCC: white cell count.

Appendix A

In this section, the results of hyperparameter tuning are given. Table A1 represents the outcomes of randomised fine-tuning analysis on hyperparameters for all models in the paper. For each model, a search space is studied for the associated hyperparameters then the random search approach is conducted to select hyperparameter values according to the highest performance on the training set.

Table A1.

The results of the conducted randomised hyperparameter tuning processes for the classifiers in the article.

Model Hyperparameter Search Space Selected
LR Regularisation strength {0, 0.10, 0.20, …, 1} 0.40
class weight {0, 1, …, 10} 7
Maximum number of iterations {1000, 2000, …, 10,000} 7000
GB Learning rates {0.01, 0.02, …, 1} 0.10
Number of boosting stages {20, 40, …, 200} 160
Minimum number of samples required to split an internal node {1, 2, …, 10} 2
Minimum number of samples required to be at a leaf node {{1, 2, …, 10} 6
Maximum depth of the individual estimators {1, 2, …, 10} 9
AB Maximum number of estimators at which boosting is terminated {10, 20, …, 100} 90
Learning rates {0.01, 0.02, …, 1} 1.58
RF number of trees {50, 100, …, 500} 20
Maximum depth of the tree {1, 2, …, 10} 3
Minimum number of samples required to split an internal node, {1, 2, …, 10} 4
Minimum number of samples required to be at a leaf node {1, 2, …, 10} 6
Maximum number of leaf nodes {1, 2, …, 10} 3
minimum impurity decrease {0, 0.001, 0.002, …, 0.010} 0.004
Cost complexity pruning factor {0.01, 0.02, …, 0.10} 0.01
Minimum weighted fraction of the sum total of weights {0.01, 0.02, …, 0.10} 0.01
SVC Class weight {0, 1, …, 10} 6
Maximum integration {100, 200, …, 10,000} 2400

Note. AB: AdaBoost; LR: logistic regression; GB: gradient boosting; RF: random forest; SVC: support vector classifier.

Author Contributions

H.K.: conceptualisation, methodology, software, validation, formal analysis, investigation, data curation, writing the original draft, review and editing, visualisation. H.N.: conceptualisation, methodology, software, validation, formal analysis, investigation, data curation, review and editing. J.E.: conceptualisation, project administration, resources, validation, review & editing, supervision. M.B.: conceptualisation, methodology, validation, investigation, review and editing, supervision. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

For analysis, an anonymised dataset is used, and the study is conducted under National Health Service (NHS) ethics as approved by the East-Midlands-Leicester South Research Ethics Committee (20/EM/0145).

Informed Consent Statement

For the analysis, we coded in Python (3.6.7) [32]. The libraries used include; Pandas [33], NumPy [34], and Sklearn [35]. The implementation source code is publicly available on this Gitlab repository: https://gitlab.com/Heydar-Khadem/DM-interplay-COVID19.git.

Conflicts of Interest

The authors declare that they have no known competing financial interest or personal relationships that could have appeared to influence the work reported in this paper.

Funding Statement

This research received no external funding.

Footnotes

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Zhou K., Sun Y., Li L., Zang Z., Wang J., Li J., Liang J., Zhang F., Zhang Q., Ge W., et al. Eleven Routine Clinical Features Predict COVID-19 Severity Uncovered by Machine Learning of Longitudinal Measurements. Comput. Struct. Biotechnol. J. 2021;19:3640–3649. doi: 10.1016/j.csbj.2021.06.022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Onder G., Rezza G., Brusaferro S. Case-Fatality Rate and Characteristics of Patients Dying in Relation to COVID-19 in Italy. JAMA. 2020;323:1775–1776. doi: 10.1001/jama.2020.4683. [DOI] [PubMed] [Google Scholar]
  • 3.Wargny M., Potier L., Gourdy P., Pichelin M., Amadou C., Benhamou P.-Y., Bonnet J.-B., Bordier L., Bourron O., Chaumeil C., et al. Predictors of Hospital Discharge and Mortality in Patients with Diabetes and COVID-19: Updated Results from the Nationwide CORONADO Study. Diabetologia. 2021;64:778–794. doi: 10.1007/s00125-020-05351-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Sourij H., Aziz F., Bräuer A., Ciardi C., Clodi M., Fasching P., Karolyi M., Kautzky-Willer A., Klammer C., Malle O., et al. COVID-19 Fatality Prediction in People with Diabetes and Prediabetes Using a Simple Score upon Hospital Admission. Diabetes Obes. Metab. 2021;23:589–598. doi: 10.1111/dom.14256. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Corona G., Pizzocaro A., Vena W., Rastrelli G., Semeraro F., Isidori A.M., Pivonello R., Salonia A., Sforza A., Maggi M. Diabetes Is Most Important Cause for Mortality in COVID-19 Hospitalized Patients: Systematic Review and Meta-Analysis. Rev. Endocr. Metab. Disord. 2021;22:275–296. doi: 10.1007/s11154-021-09630-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Ciardullo S., Zerbini F., Perra S., Muraca E., Cannistraci R., Lauriola M., Grosso P., Lattuada G., Ippoliti G., Mortara A., et al. Impact of Diabetes on COVID-19-Related in-Hospital Mortality: A Retrospective Study from Northern Italy. J. Endocrinol. Investig. 2021;44:843–850. doi: 10.1007/s40618-020-01382-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Shah H., Khan M.S.H., Dhurandhar N.V., Hegde V. The Triumvirate: Why Hypertension, Obesity, and Diabetes Are Risk Factors for Adverse Effects in Patients with COVID-19. Acta Diabetol. 2021;58:831–843. doi: 10.1007/s00592-020-01636-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Campbell T.W., Wilson M.P., Roder H., MaWhinney S., Georgantas R.W., Maguire L.K., Roder J., Erlandson K.M. Predicting Prognosis in COVID-19 Patients Using Machine Learning and Readily Available Clinical Data. Int. J. Med. Inform. 2021;155:104594. doi: 10.1016/j.ijmedinf.2021.104594. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Dennis J.M., Mateen B.A., Sonabend R., Thomas N.J., Patel K.A., Hattersley A.T., Denaxas S., McGovern A.P., Vollmer S.J. Diabetes and COVID-19 Related Mortality in the Critical Care Setting: A Real-Time National Cohort Study in England. 2020. [(accessed on 5 June 2022)]. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3615999. [DOI] [PMC free article] [PubMed]
  • 10.Haimovich A.D., Ravindra N.G., Stoytchev S., Young H.P., Wilson F.P., van Dijk D., Schulz W.L., Taylor R.A. Development and Validation of the Quick COVID-19 Severity Index: A Prognostic Tool for Early Clinical Decompensation. Ann. Emerg. Med. 2020;76:442–453. doi: 10.1016/j.annemergmed.2020.07.022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Zheng B., Cai Y., Zeng F., Lin M., Zheng J., Chen W., Qin G., Guo Y. An Interpretable Model-Based Prediction of Severity and Crucial Factors in Patients with COVID-19. Biomed Res. Int. 2021;2021:8840835. doi: 10.1155/2021/8840835. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Lalmuanawma S., Hussain J., Chhakchhuak L. Applications of Machine Learning and Artificial Intelligence for COVID-19 (SARS-CoV-2) Pandemic: A Review. Chaos Solitons Fractals. 2020;139:110059. doi: 10.1016/j.chaos.2020.110059. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Kar S., Chawla R., Haranath S.P., Ramasubban S., Ramakrishnan N., Vaishya R., Sibal A., Reddy S. Multivariable Mortality Risk Prediction Using Machine Learning for COVID-19 Patients at Admission (AICOVID) Sci. Rep. 2021;11:12801. doi: 10.1038/s41598-021-92146-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Khadem H., Nemat H., Elliott J., Benaissa M. Signal Fragmentation Based Feature Vector Generation in a Model Agnostic Framework with Application to Glucose Quantification Using Absorption Spectroscopy. Talanta. 2022;243:123379. doi: 10.1016/j.talanta.2022.123379. [DOI] [PubMed] [Google Scholar]
  • 15.Mauer E., Lee J., Choi J., Zhang H., Hoffman K.L., Easthausen I.J., Rajan M., Weiner M.G., Kaushal R., Safford M.M., et al. A Predictive Model of Clinical Deterioration among Hospitalized COVID-19 Patients by Harnessing Hospital Course Trajectories. J. Biomed. Inform. 2021;118:103794. doi: 10.1016/j.jbi.2021.103794. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Bhatt S., Cohon A., Rose J., Majerczyk N., Cozzi B., Crenshaw D., Myers G. Interpretable Machine Learning Models for Clinical Decision-Making in a High-Need, Value-Based Primary Care Setting. NEJM Catal. Innov. Care Deliv. 2021;2 doi: 10.1056/CAT.21.0008. [DOI] [Google Scholar]
  • 17.Lundberg S.M., Erion G.G., Lee S.-I. Consistent Individualized Feature Attribution for Tree Ensembles. arXiv. 20181802.03888 [Google Scholar]
  • 18.Lundberg S., Lee S.-I. A Unified Approach to Interpreting Model Predictions; Proceedings of the 31th Conference on Neural Information Processing Systems; Long Beach, CA, USA. 4–9 December 2017; pp. 4765–4774. [Google Scholar]
  • 19.Shapley L.S. A Value for N-Person Games. Contrib. Theory Games. 1953;2:307–317. [Google Scholar]
  • 20.Pan P., Li Y., Xiao Y., Han B., Su L., Su M., Li Y., Zhang S., Jiang D., Chen X., et al. Prognostic Assessment of COVID-19 in the Intensive Care Unit by Machine Learning Methods: Model Development and Validation. J. Med. Internet Res. 2020;22:e23128. doi: 10.2196/23128. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Hathaway Q.A., Roth S.M., Pinti M.V., Sprando D.C., Kunovac A., Durr A.J., Cook C.C., Fink G.K., Cheuvront T.B., Grossman J.H., et al. Machine-Learning to Stratify Diabetic Patients Using Novel Cardiac Biomarkers and Integrative Genomics. Cardiovasc. Diabetol. 2019;18:78. doi: 10.1186/s12933-019-0879-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Khadem H., Nemat H., Eissa M.R., Elliott J., Benaissa M. COVID-19 Mortality Risk Assessments for Individuals with and without Diabetes Mellitus: Machine Learning Models Integrated with Interpretation Framework. Comput. Biol. Med. 2022;144:105361. doi: 10.1016/j.compbiomed.2022.105361. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Iqbal A., Arshad M., Julian T., Tan S., Greig M., Elliott J. Higher Admission Activated Partial Thromboplastin Time, Neutrophil-Lymphocyte Ratio, Serum Sodium, and Anticoagulant Use Predict in-Hospital Covid-19 Mortality in People with Diabetes: Findings from Two University Hospitals in the UK. Diabet. Med. 2021;178:108955. doi: 10.1016/j.diabres.2021.108955. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Zwart D.L., Langelaan M., van de Vooren R.C., Kuyvenhoven M.M., Kalkman C.J., Verheij T.J., Wagner C. Patient Safety Culture Measurement in General Practice. Clinimetric Properties of “SCOPE.” BMC Fam. Pract. 2011;12:117. doi: 10.1186/1471-2296-12-117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Jonsson P., Wohlin C. An Evaluation of K-Nearest Neighbour Imputation Using Likert Data; Proceedings of the 10th International Symposium on Software Metrics; Chicago, IL, USA. 11–17 September 2004; pp. 108–118. [Google Scholar]
  • 26.Wang J., Yu H., Hua Q., Jing S., Liu Z., Peng X., Luo Y. A Descriptive Study of Random Forest Algorithm for Predicting COVID-19 Patients Outcome. PeerJ. 2020;8:e9945. doi: 10.7717/peerj.9945. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Forte J.C., Yeshmagambetova G., van der Grinten M.L., Hiemstra B., Kaufmann T., Eck R.J., Keus F., Epema A.H., Wiering M.A., van der Horst I.C.C. Identifying and Characterizing High-Risk Clusters in a Heterogeneous ICU Population with Deep Embedded Clustering. Sci. Rep. 2021;11:12109. doi: 10.1038/s41598-021-91297-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.MacQueen J. Some Methods for Classification and Analysis of Multivariate Observations; Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, University of California; Berkeley, CA, USA. 7 January 1966; pp. 281–297. [Google Scholar]
  • 29.Abdullah D., Susilo S., Ahmar A.S., Rusli R., Hidayat R. The Application of K-Means Clustering for Province Clustering in Indonesia of the Risk of the COVID-19 Pandemic Based on COVID-19 Data. Qual. Quant. 2021;56:1283–1291. doi: 10.1007/s11135-021-01176-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Hutagalung J., Ginantra N.L.W.S.R., Bhawika G.W., Parwita W.G.S., Wanto A., Panjaitan P.D. COVID-19 Cases and Deaths in Southeast Asia Clustering Using K-Means Algorithm. J. Phys. Conf. Ser. 2021;1783:012027. doi: 10.1088/1742-6596/1783/1/012027. [DOI] [Google Scholar]
  • 31.Syakur M.A., Khotimah B.K., Rochman E.M.S., Satoto B.D. Integration K-Means Clustering Method and Elbow Method for Identification of the Best Customer Profile Cluster. IOP Conf. Ser. Mater. Sci. Eng. 2018;336:012017. doi: 10.1088/1757-899X/336/1/012017. [DOI] [Google Scholar]
  • 32.Van Rossum G., Drake F.L. Python 3 Reference Manual. CreateSpace; Scotts Valley, CA, USA: 2009. [Google Scholar]
  • 33.McKinney W. Data Structures for Statistical Computing in Python; Proceedings of the the 9th Python in Science Conference; Austin, TX, USA. 28 June–3 July 2010; pp. 51–56. [Google Scholar]
  • 34.Harris C.R., Millman K.J., van der Walt S.J., Gommers R., Virtanen P., Cournapeau D., Wieser E., Taylor J., Berg S., Smith N.J., et al. Array Programming with NumPy. Nature. 2020;585:357–362. doi: 10.1038/s41586-020-2649-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Pedregosa F., Varoquaux G., Gramfort A., Michel V., Thirion B., Grisel O., Blondel M., Prettenhofer P., Weiss R., Dubourg V., et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011;12:2825–2830. [Google Scholar]

Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES