Skip to main content
Frontiers in Cardiovascular Medicine logoLink to Frontiers in Cardiovascular Medicine
. 2024 Oct 22;11:1419551. doi: 10.3389/fcvm.2024.1419551

Prediction of in-hospital mortality risk for patients with acute ST-elevation myocardial infarction after primary PCI based on predictors selected by GRACE score and two feature selection methods

Nan Tang 1, Shuang Liu 1, Kangming Li 1, Qiang Zhou 1, Yanan Dai 1, Huamei Sun 1, Qingdui Zhang 1, Ji Hao 1, Chunmei Qi 1,*
PMCID: PMC11534735  PMID: 39502196

Abstract

Introduction

Accurate in-hospital mortality prediction following percutaneous coronary intervention (PCI) is crucial for clinical decision-making. Machine Learning (ML) and Data Mining methods have shown promise in improving medical prognosis accuracy.

Methods

We analyzed a dataset of 4,677 patients from the Regional Vascular Center of Primorsky Regional Clinical Hospital No. 1 in Vladivostok, collected between 2015 and 2021. We utilized Extreme Gradient Boosting, Histogram Gradient Boosting, Light Gradient Boosting, and Stochastic Gradient Boosting for mortality risk prediction after primary PCI in patients with acute ST-elevation myocardial infarction. Model selection was performed using Monte Carlo Cross-validation. Feature selection was enhanced through Recursive Feature Elimination (RFE) and Shapley Additive Explanations (SHAP). We further developed hybrid models using Augmented Grey Wolf Optimizer (AGWO), Bald Eagle Search Optimization (BES), Golden Jackal Optimizer (GJO), and Puma Optimizer (PO), integrating features selected by these methods with the traditional GRACE score.

Results

The hybrid models demonstrated superior prediction accuracy. In scenario (1), utilizing GRACE scale features, the Light Gradient Boosting Machine (LGBM) and Extreme Gradient Boosting (XGB) models optimized with BES achieved Recall values of 0.944 and 0.954, respectively. In scenarios (2) and (3), employing SHAP and RFE-selected features, the LGB models attained Recall values of 0.963 and 0.977, while the XGB models achieved 0.978 and 0.99.

Discussion

The study indicates that ML models, particularly the XGB optimized with BES, can outperform the conventional GRACE score in predicting in-hospital mortality. The hybrid models' enhanced accuracy presents a significant step forward in risk assessment for patients post-PCI, offering a potential alternative to existing clinical tools. These findings underscore the potential of ML in optimizing patient care and outcomes in cardiovascular medicine.

Keywords: in-hospital mortality, percutaneous coronary intervention, ST-elevation myocardial infarction, global registry of acute coronary events, machine learning prediction, feature selection

1. Introduction

Cardiovascular disease (CVD) constitutes a dominant global health challenge, particularly accentuated within low- and middle-income countries (LMICs). The growing prevalence of CVD risk factors within these regions obviously increases the burden of mortality associated with this disease (13). Myocardial infarction (MI) is a severe medical condition stemming from a sudden reduction in blood flow to the heart, resulting in tissue damage. Clinical manifestations typically include chest pain, shortness of breath, and weakness (4, 5). Preventative measures mostly contain lifestyle changes and pharmacological interventions (6). Treatment modalities include the management of beta-blockers, diuretics, ACE inhibitors, calcium channel blockers, and nitrates.

The effective management of ST-segment elevation myocardial infarction (STEMI) is considered important in inpatient care, a fact emphasized by the guidance provided in the 2012 and 2017 ESC Guidelines. These guidelines prioritize early reperfusion therapy, particularly through main percutaneous coronary intervention (PCI), for optimal STEMI treatment. The diagnosis of STEMI poses challenges due to its potential to represent conditions, requiring careful consideration of various clinical factors during electrocardiogram interpretation (7, 8). Furthermore, STEMI rises as a complication of infective endocarditis, associated with a distinguished 30-day mortality rate (9). Timely diagnosis and immediate restoration of blood flow, preferably through primary PCI, are key steps in reducing myocardial damage and preventing complications following STEMI (10).

Despite the developments in PCI technologies, the in-hospital mortality (IHM) subsequent to PCI in emergency cases persists at a remarkably high rate. A study conducted by Moroni (11) clarified that IHM often correlates with pre-existing serious cardiovascular conditions, with procedural complications attributing to a minority of cases. This suggests an imperative for enhanced treatment modalities for severe cardiovascular situations, particularly in addressing cardiogenic shock. However, the utility of procalcitonin (PCT) as a prognostic indicator in these conditions remains controversial. Covino et al. (12) observed that early assessment of PCT in patients with intra-abdominal infection (IAI) did not yield a significant impact on IHM, while Dutta et al. (13) highlighted the potential of PCT levels in predicting mortality in disapprovingly ill surgical patients. Within the spectrum of STEMI, Dawson et al. (14) reported a lack of substantial reduction in IHM despite changes in technical characteristics. These findings emphasize the demand for further research activities and targeted interventions aimed at justifying IHM following PCI in emergency scenarios.

In contemporary clinical practice, a multitude of risk grading tools are employed to assess the risk of IHM among patients. Notable among these are the History, Electrocardiogram, Age, Risk factors, initial Troponin (HEART) score, the Thrombolysis in Myocardial Infarction (TIMI) score, and the Global Registry of Acute Coronary Events (GRACE) score, as identified by Liu (15). Nevertheless, the efficacy of these tools can fluctuate across diverse patient populations, with certain instruments demonstrating suboptimal performance in present-day practice (16). Within main care backgrounds, there is an observable trend toward utilizing routine healthcare data for risk grading. However, comprehensive documentation regarding the specific tools applied and their performance remains lacking (17). Additionally, within the intensive care situation, there is incredulity regarding the relevance and reliability of scales employed to measure job stressors. This underscores the imperative for further investigation and scholarly inquiry in this domain (18).

Regarding the GRACE scale, despite advancements in treatment approaches, it continues to be a critical tool for evaluating the risk of adverse outcomes in cases of serious coronary syndromes (19). Continuous monitoring of mortality rates in coronary care elements using the GRACE score indicates that while it generally performs adequately, there are still areas where improvements can be made (20). Additionally, research has shown that the GRACE score is effective in predicting major cardiac events in patients presenting with chest pain and suspected acute coronary syndrome (21). Moreover, a modified version of the GRACE score, identified as the angiographic GRACE score, has been developed and validated as a beneficial tool for predicting IHM, specifically in Japanese patients with acute myocardial infarction (22).

Over the past few decades, Data Mining (DM) and Machine Learning (ML) have emerged as influential tools in medicine, particularly in predicting and diagnosing cognitive diseases (23). These methods have been applied to a wide variety of medical conditions, including type 2 diabetes, hypertension, cardiovascular disease, renal diseases, liver diseases, mental illness, and child health (24). The usage of ML in medical informatics has seen a significant increase, with Support Vector Machine (SVM) and Random Forest (RF) being the most popular algorithms for classification problems (25). However, there is no single algorithm that is universally suitable for diagnosing or predicting diseases, and the combination of different processes often yields the greatest results (26).

ML models have shown potential in predicting IHM following PCI in patients with serious STEMI. Studies conducted by Li (27) and Yang (28) employed data from the Chinese Acute Myocardial Infarction (CAMI) registry to develop prediction models, achieving both high performance and interpretability. Moreover, Deng (29) applied a RF algorithm to forecast both no-reflow and IHM in STEMI patients undergoing key PCI, demonstrating superior discrimination. Falcao (30) identified predictors of IHM in patients with STEMI undergoing pharmacoinvasive treatment, including age, comorbidities, and practical success. Additionally, Tanık (31) found that the PRECISE-DAPT score, a predictive tool for bleeding risk, was independently associated with IHM in STEMI patients undergoing primary PCI. Furthermore, Bai (32) compared the performance of various ML models in predicting 1-year mortality in STEMI patients with hyperuricemia, with the CatBoost model showing the highest accuracy. To validate the accuracy of ML models, particularly the XGBoost model, in predicting 1-year mortality in patients with anterior STEMI, Li (33) conducted further research. Collectively, these studies highlight the significant potential of ML in improving risk prediction for STEMI patients post-PCI, offering valuable insights into prognosis and treatment strategies. However, there is a gap in existing literature related to ML prediction model development based on imperative features of patients rather than those four parameters leading to GRACE score development. Also, integrating currently developed optimization algorithms for enhanced prediction accuracy by hybrid and ensemble approaches are the innovative methods which their absence is strongly felt in the literature review.

This study aims to introduce a new approach to investigate the risk factors contributing to IHM in patients with MI following PCI, applying advanced ML techniques. The research methodology involved gathering datasets related to various features of patients to assess their impact on the mortality risk of patients utilizing classifiers like Extreme Gradient Boosting (XGB), Light Gradient Boosting (LGB), Stochastic Gradient Boosting (SGB), and Histogram Gradient Boosting (HGB). Monte Carlo Cross-Validation (MCCV) was used to select the best prediction models based on their Accuracy. Techniques, for instance, Recursive Feature Elimination (RFE) and Shapley Additive Explanations (SHAP), were employed to identify important features for classification. Three different scenarios were designed to predict the risk of IHM within 30 days to provide clinicians with an estimate of patient survivability or mortality likelihood pre-treatment. The first scenario studies the efficacy of the traditional GRACE scale system (including Age, patient age, heart rate (HR), systolic blood pressure (SBP), and acute heart failure (AHF) class), widely entrenched within hospital protocols. The second and third scenarios employ a subclass of features selected via the Shapley Additive explanations (SHAP) and Recursive Feature Elimination (RFE) methods, respectively. All analysis conducted in Python programming software. By comparing the prediction performance of base single models and their hybrid framework (optimized with meta-heuristic algorithms such as Augmented Gray Wolf Optimizer (AGWO), Bald Eagle Search Optimization (BES), Golden Jackal Optimizer (GJO), and Puma Optimizer (PO)) utilizing these scenarios, the study aims to give valuable insights to enhance risk assessment strategies and patient care paradigms for MI patients undergoing PCI intervention.

2. Classification and model selection based on machine learning techniques

The boosting approach involves utilizing a “weak” or “base” learning algorithm repeatedly, each time with a different subset of training examples (or a varied distribution or weighting over the examples). In each iteration, the base learning algorithm generates a new weak prediction rule. After numerous rounds, the boosting algorithm combines these weak rules into a single prediction rule, aiming for a substantially improved level of accuracy compared to any individual weak rule (Figure 1). This iterative process enhances the overall predictive power of the model (34).

Figure 1.

Figure 1

Boosting approach in ML.

2.1. Extreme Gradient Boost (XGB)

The Extreme Gradient Boost Classifier (XGBC) represents a sophisticated implementation of the gradient boosting technique, employing an ensemble approach to combine multiple sets of base learners (trees) to establish a strong model capable of making significant predictions (35). XGBC offers various advantages, including the ability to leverage parallel processing for improved computational efficiency, providing flexibility in setting objectives, incorporating built-in cross-validation, and effectively addressing splits in the presence of negative loss. With these advantages, XGBC emerges as a highly suitable choice for analyzing classification data. Applying a tree-based methodology, XGBC constructs decision trees to classify training data, facilitating the achievement of specific target outcomes (36).

The gradient boosting procedure encompasses the subsequent sequential steps:

  • The initialization of the boosting algorithm involves the description of the function F0(x) (Equation 1):
    F0(x)=argmini1nL(yi,) (1)
  • The iterative calculation includes the derivation of the gradient of the loss function (Equation 2):
    rim=α[(L(yi,F(Xi)))F(xi)]F(X)=Fm1(x) (2)

Where α is the learning rate.

  • Subsequently, each hm(x) is fitted based on the gradient developed at each iterative step:

  • The purpose of the multiplicative factor ym for each terminal node is executed, and subsequently, the boosted model Fm(x) is formulated (Equation 3):
    Fm(x)=Fm1(x)+mhm(x) (3)

2.2. Light Gradient Boosting (LGB)

LGB is a rapid training outline that blends decision tree algorithms with boosting methods. It prioritizes speed, using histogram-based techniques to accelerate training and conserve memory (37). Different from traditional trees, LGB employs leaf-wise tree growth, efficiently identifying high-branching gain leaves to optimize performance (38, 39).

The calculation procedures of LGB, delineated step by step in (40), involve finding a projected function f^(x) that approximates the function f(x) based on the given training dataset X={(xi,yi)}i=1m. The primary objective is to minimize the expected values of specific loss functions, signified as L(y,f(x)) (Equation 4).

f^(x)argminfEy,xL(y,f(x)) (4)

In the process of approximating the final model, LGB will integrate a combination of multiple regression trees, represented as t=1Tft(x) (Equation 5).

fT(X)=t=1Tft(X) (5)

The regression trees signified as wq(x),q{1,2,,N}, denote decision rules, where N is the number of leaves in each tree. q signifies the decision rule, and w is a vector representing the weights of the leaf nodes. The model is incrementally trained at step t in an additive manner.

Γtj=1NL(yi,Ft1(xi)+ft(xi)) (6)

The Newton's method is employed to rapidly estimate the objective function, and (Equation 6) is simplified by eliminating the constant term:

Γtj=1N(gift(xi)+12hift2(xi)) (7)

In the given equation, gi and hi denote the first- and second-order gradient statistics of the loss functions. If the sample set for leaf j is denoted as Ij, then (Equation 7) can be transformed into (Equation 8):

Γt=j=1J((iϵIjgi)ωj+12(iϵIjhi+λ)ωj2) (8)

Equations 9, 10 are employed to calculate the optimal leaf weights ωj and the extreme values of ΓK concerning the tree structure q(x):

ωj=iϵIjgiiϵIjhi+λ (9)
ΓT=12j=1J(iϵIjgi)2iϵIjhi+λ (10)

The term ωj signifies the weight function assessing the effectiveness of the tree structure q(x). Ultimately, the objective function is derived by consolidating the splits.

G=12((iϵIlgi)2iϵIlhi+λ+(iϵIrgi)2iϵIrhi+λ+(iϵIgi)2iϵIhi+λ) (11)

The objective function is defined as the sum of the splits with Il and Ir representing the samples in the left and right branches, respectively (Equation 11).

2.3. Histogram-based Gradient Boosting (HGB)

Histograms are valuable tools for visualizing data distribution and frequency, especially with repetitive data. Grouping input data into bins, as in histograms, enhances model flexibility. Combining histogram-based methods with gradient boosting leads to strong ML ensembles, yielding high-performance models (41). HGBoost employs numeral-based data structures like histograms instead of sorted continuous values during tree-building, enabling it to capture complex nonlinear relationships in datasets effectively. This integration of gradient boosting with histogram-based techniques allows HGBoost to excel in modeling and optimizing feature connections (42).

Histogram-based Gradient Boosting Classification (HGBC) represents a difficult iteration of gradient boosting, employing decision trees as fundamental models and leveraging histograms to achieve outstanding improvements in computational efficiency. Observed remarks show that this methodology yields superior outcomes, diminishes ensemble size, and expedites inference, rendering it an attractive proposition for tackling intricate datasets within academic investigations (43).

2.4. Stochastic Gradient Boosting Machines (SGB)

Friedman (44) proposed Stochastic Gradient Boosting Machines (SGB), a method extensively employed in both classification and regression tasks. Decision stumps or regression trees serve as common choices for weak classifiers within SGB. The main aim of SGB is to train weak learners to minimize loss functions, such as mean square errors, with subsequent weak learners benefiting from the residuals of preceding ones for training.

Consequently, there is a reduction in the value of the loss function for the present weak learners. Employing the bagging technique serves to mitigate correlation among these learners, with each undergoing training on subsets sampled without replacement from the entirety of the dataset. The final prediction is then derived through the amalgamation of predictions generated by this cohort of weak learners (45).

2.5. Monte-Carlo cross-validation (MCCV) for model selection

Numerous methods, such as the Akaike information criterion (46) and Cp statistics (47), tackle the task of model selection. Nevertheless, cross validation (CV) emerges as a standout approach (4851), arranging a predictive perspective in this process. In CV, upon selecting a model (α), the n samples (referred to as S) undergo a division.

The initial component, identified as the calibration set (Sc), consists of nc samples applied for fitting the model, represented by the submatrix XαSc and sub-vector YSc. The subsequent section termed the validation set (Sv), comprises nv=nnc samples dedicated to model validation, depicted by the submatrix XαSv and sub-vector YSv. This arrangement leads to a total of (nnv) possible sample divisions. In each division, the model is fitted using the nc samples from the standardization set Sc, resulting in the estimation β^αSc. Treating the samples in the validation set as if they were future data points, the fitted model predicts the response vector ySv (Equation 12).

y^αSv=XαSvtβ^αSc (12)

The Accuracy across all samples in the validation set is considered by (Equation 13):

Accuracy(Sv,α)=1nv||ySvy^αSv||2 (13)

The formula involves calculating the Euclidean norm of a vector within a framework where a set S is comprised of elements from various validation sets, each corresponding to different sample splits denoted as (nnv). In this framework, the CV standard is defined by excluding a specific number of samples nv for validation, providing a method for systematically evaluating models on subsets of data.

CVnv(α)=SvSAccuracy(Sv,α)(nnv) (14)

For each αR, the computation of CVnv(α) is conducted. (Equation 14) serves as an estimate for Accuracy within the constraints of finite samples. The CV criterion is focused on identifying the optimal α that maximizes values across all CVnv(α) for αR. As a result, the model is characterized by variables indexed by the integers in α is chosen.

The widely used leave-one-out Cross-Validation (LOO-CV), where nv=1, is extensively applied in chemometrics. However, research findings have shown that models selected through LOO-CV can be inaccurately asymptotic. Although LOO-CV can choose a model with a bias bα=0 that approaches infinity encompassing all non-zero elements in βα, it tends to include unnecessary additional variables in the model (52). This suggests that the model's dimension Pα is not optimally concise, potentially leading to overfitting concerns.

It has been established that, in general, CV, under the conditions nc and nv/n1 (53), the likelihood of selecting the model with the best predictive capability tends toward unity when nv samples are reserved for validation. Consequently, the CVnv(α) benchmark (Equation 14) shows asymptotic consistency. Yet, practically computing CVnv with a large nv is infeasible due to its exponential computational complexity. To tackle this issue, MCCV offers a simple and efficient solution. For a given α, the samples are randomly split into two sets: Sc(i) (of size nc) and Sv(i) (of size nv). This process is repeated N times, defining the repeated MCCV criterion as follows (Equation 15):

MCCVnv(α)=1Nnvi=1NySv(i)y^αSv(i)2 (15)

Employing the Monte Carlo method greatly decreases computational complexity. Theoretically, decreasing the number of samples for model calibration requires increasing the number of repetitions. Typically, it is deemed adequate to set N=n2 to ensure that MCCVnv achieves similar performance to traditional CVnv (54).

In this study, 70% of the samples were considered for the fitting (training) of the prediction models, 30% were allocated for the validation process (testing), and finally, two LGBM and XGBC models with an accuracy of 0.97 and 0.98 have been selected, and in the following, only these two models will be examined in their hybrid version.

3. Detailed data assessment

3.1. Data description and preprocessing

The study used data from patients treated at the Regional Vascular Center of Primorsky Regional Clinical Hospital in Vladivostok from 2015 to 2021. Patients were selected for inclusion in the STEMI and PCI study based on criteria confirmed upon their admission to the hospital. Exclusion criteria comprised non-ST elevation myocardial infarction, unconfirmed STEMI, or the absence of an indication for PCI. Finally, 4,677 patients were included in the study, from which 4,359 patients were in the “Alive” group who did not die within 30 days of the study after PCI, and 318 patients were in the “Die” group who died in hospital. The “Die” group comprised patients who passed away at any point during these 30 days, including those who did not survive to undergo post-PCI assessments. Conversely, the “Alive” group consisted of patients who survived the entire 30-day period and were monitored in the hospital throughout. It is important to note that patients with missing data were excluded from the dataset of those patients with no risk of death due to the abundance of information (the number of samples decreased to 2,709). For 318 patients who experienced IHM after PCI, the Multiple Imputation by Chained Equations (MICE) method was used to handle missing data. MICE achieves multiple imputation by creating multiple complete datasets, analyzing each dataset separately, and then combining the results to reduce the bias that a single imputation method might introduce (55). This method fully considers the uncertainty of the data when dealing with missing data, especially suitable for the complex multivariate data structure in this study. Compared with single imputation, MICE can provide more reliable statistical inference when dealing with a large amount of missing data. Ultimately, a cleaned dataset of 3,027 patients with 41 features, as described below in related categories, was chosen for the prediction task:

3.1.1. Cardiovascular parameters

SPAP (Systolic Pulmonary Arterial Pressure), LVRMI (Left Ventricular Regional Motion Index), EF LV (Left Ventricular Ejection Fraction), ESV (End-Systolic Volume), LVRWTI (Left Ventricular Relative Wall Thickness Index), La1 (Left Atrial Diameter), Ra2 (Right Atrial Diameter), Ra1 (Right Atrium Pressure), PI (Pulsatility Index), EDV (End-Diastolic Volume), La2 (Left Atrial Pressure), SBP (Systolic Blood Pressure), DBP (Diastolic Blood Pressure).

3.1.2. Blood parameters

NEUT (Neutrophils), EOS (Eosinophils), WBC (White Blood Cell count), Hb (Hemoglobin), RBC (Red Blood Cell count), PLT (Platelet count), LYM (Lymphocyte count).

3.1.3. Coagulation parameters

TT (Thrombin Time), INR (International Normalized Ratio), APTT (Activated Partial Thromboplastin Time), PCT (Plateletcrit).

3.1.4. Metabolic parameters

Urea [Blood Urea Nitrogen (BUN)], Glu (Glucose), Cr (Creatinine).

3.1.5. Anthropometric parameters

Age (Patient's Age), Weight (Patient's Weight), Height (Patient's Height), BMI (Body Mass Index).

3.1.6. Diagnostic parameters

Killip class [Killip Classification (classification of heart failure severity)], Form STEMI (STEMI Diagnosis), CKD (chronic kidney disease), AFib (Atrial Fibrillation), Diabetes (Diabetes Mellitus), COPD (Chronic Obstructive Pulmonary Disease), aMI (Acute Myocardial Infarction) And Sex (Patient's Gender).

3.2. Feature selection

3.2.1. Shapley Additive Explanations (SHAP)

SHAP, a method for attributing features additively, draws from both game theory and local explanations (56). The Shapley value has gained popularity as a method for providing interpretable feature attribution in ML models (57). SHAP simplifies inputs by transforming the original inputs x into a simplified representation z through a mapping function x=hx(z). This simplification enables the approximation of the original model f(x) using a linear function of binary variables based on z (Equation 16):

f(x)=g(z)=φ0+i=1Mφizi (16)

z is a binary vector with M elements representing input features, φ0 denotes the attribution value of the model when z is all zeros, calculated as f(hx(0)), and φi represents the attribution value of the ith feature (Equations 17, 18).

φi=SF/{i}|S|!(M|S|!1)!M![fx(S{i})fx(s)] (17)
fx(S)=f(hx1(Z))=E[f(x)|xs] (18)

SHAP stands out due to its three core properties: local accuracy, consistency, and proficiency in handling missing data. It uses the SHAP value φi as a unified metric for additive feature attributions. In the SHAP framework, F represents the subset of non-zero inputs in z, while S indicates the subset of F obtained by excluding the ith feature (58). Known for its model-agnostic nature, SHAP shows impressive adaptability across various ML and DL models, effectively determining the relative importance of individual input features within additive feature attribution methodologies (59). Table 1 reports SHAP values obtained for each feature in the dataset based on each base models and selected features.

Table 1.

SHAP values and selected features from the dataset based on each base model (scenario 2).

NO Parameter SHAP values (LGB) Selected features (LGB) SHAP values (XGB) Selected features (XGB)
1 Sex 0.007 0.026
2 Age 0.135 0.217
3 Height 0.070 0.176
4 Weight 0.073 0.205
5 BMI 0.106 0.181
6 SBP 0.041 0.091
7 DBP 0.100 0.146
8 PBP 0.052 0.090
9 HR 0.291 0.445
10 Cr 0.088 0.189
11 Killip class 0.147 0.259
12 EF LV 0.216 0.305
13 EDV 0.068 0.102
14 ESV 0.135 0.150
15 LVRWTI 0.087 0.149
16 LVRMI 0.328 0.570
17 SPAP 0.419 0.493
18 La1 0.082 0.159
19 La2 0.064 0.111
20 Ra1 0.080 0.130
21 Ra2 0.081 0.147
22 WBC 0.151 0.283
23 NEUT 0.669 0.886
24 LYM 0.051 0.117
25 EOS 0.345 0.436
26 RBC 0.117 0.259
27 Hb 0.138 0.209
28 PLT 0.085 0.118
29 Glu 0.452 0.574
30 Urea 0.481 0.651
31 PCT 0.064 0.086
32 PI 0.080 0.152
33 INR 0.532 0.684
34 TT 0.554 0.569
35 APTT 0.067 0.179
36 Form STEMI 0.033 0.052
37 AFib 0.021 0.025
38 Diabetes 0.007 0.025
39 CKD 0.024 0.084
40 aMI 0.004 0.000
41 COPD 0.006 0.005

Figure 2 illustrates the features identified by the SHAP method for the LGB model, while Figure 3 shows the selected features for the XGB model. In the LGB model, ten features were recognized as essential factors in modeling and forecasting, while the XGB model identified 13 features. After a comprehensive examination of the relationships, it becomes apparent that the correlation between systolic pulmonary arterial pressure and heart rate, along with the correlation between neutrophils and glucose, is direct. Conversely, the relationship between neutrophils and eosinophils shows an inverse trend.

Figure 2.

Figure 2

Feature selection and SHAP analysis for LGB.

Figure 3.

Figure 3

Feature selection and SHAP analysis for XGB.

3.2.2. Recursive Feature Elimination (RFE)

The Recursive Feature Elimination (RFE) selection method (60) fundamentally operates through a recursive procedure wherein features are systematically ranked based on a specified measure of their significance.

A feature ranking criterion that performs well for individual features may not be suitable for assessing feature subsets effectively. Metrics such as Dj(i) or (wi)2 measure the impact of removing single features on the objective function but may struggle when removing multiple features simultaneously, which is crucial for obtaining a concise subset. To overcome this limitation, RFE employs an iterative approach to systematically remove the least relevant features in each iteration. RFE considers potential changes in feature importance across various feature subsets, particularly for highly correlated features. The order of feature elimination determines the final ranking, and the top n features are selected from this ranking for the feature selection process (61). Train the classifier, compute the ranking criterion for all features, and then remove the feature with the smallest ranking criterion.

When features are eliminated one by one, they are correspondingly ranked. However, the features ranked highest (eliminated last) may not necessarily be individually the most relevant. The optimal subset is determined by considering features collectively rather than individually. It is important to note that RFE does not affect correlation methods, as the ranking criterion is computed based solely on information from individual features. Table 2 reports the RFE ranking obtained for each feature in the dataset based on each base models and selected features.

Table 2.

RFE ranking and selected features from the dataset based on each base model (scenario 3).

NO Parameter RFE ranking (LGB) Selected features (LGB) RFE ranking (XGB) Selected features (XGB)
1 Sex 39 39
2 Age 25 23
3 Height 26 34
4 Weight 28 30
5 BMI 4 38
6 SBP 24 35
7 DBP 34 16
8 PBP 33 31
9 HR 13 8
10 Cr 15 17
11 Killip class 30 7
12 EF LV 14 14
13 EDV 27 15
14 ESV 20 13
15 LVRWTI 1 25
16 LVRMI 2 11
17 SPAP 8 6
18 La1 22 10
19 La2 31 19
20 Ra1 32 24
21 Ra2 11 12
22 WBC 9 20
23 NEUT 6 1
24 LYM 17 26
25 EOS 23 5
26 RBC 18 22
27 Hb 10 21
28 PLT 16 36
29 Glu 3 9
30 Urea 7 4
31 PCT 29 29
32 PI 19 32
33 INR 12 2
34 TT 5 3
35 APTT 21 37
36 Form STEMI 37 28
37 AFib 35 27
38 Diabetes 38 33
39 CKD 36 40
40 aMI 40 41
41 COPD 41 18

The features selected using RFE for the LGB and XGB models are visually depicted in Figures 4, 5, respectively. The selected features consist of 6 parameters for the LGB model and 8 for the XGB model. Upon scrutiny of the presented matrix, it becomes apparent that the left ventricular regional motion index and the left ventricular relative wall thickness index, both cardiovascular parameters, display a direct relationship with each other. Additionally, it is remarkable that neutrophils demonstrate a strong correlation with heart rate, systolic pulmonary arterial pressure, and Killip classification. Conversely, thrombin time shows no significant relationship with other selected parameters.

Figure 4.

Figure 4

Feature selection and RFE analysis for LGB.

Figure 5.

Figure 5

Feature selection and RFE analysis for XGB.

In this study, NEUT, TT, BUN, Glu, and SPAP were identified as key factors for the risk of IHM after PCI in patients with STEMI through the above-mentioned feature selection methods. NEUT play a central role in infection and inflammation, and their high levels in MI may indicate inflammatory processes associated with myocardial damage (62). Inflammation not only promotes atherosclerosis but may also lead to plaque rupture, increasing the risk of cardiac events (63). TT is an indicator for assessing the coagulation cascade, and its prolongation may suggest abnormal coagulation factor activity, increasing the risk of thrombosis after myocardial infarction (64). Additionally, prolonged TT may be associated with the use of anticoagulant drugs, which is common in the management of heart diseases. BUN reflects renal insufficiency in heart diseases, which may affect fluid and electrolyte balance, activate the renin-angiotensin-aldosterone system, leading to increased blood pressure and cardiac load, affecting cardiac function and clinical outcomes (65). High blood glucose is an independent risk factor for cardiovascular diseases, and chronic hyperglycemia promotes oxidative stress and inflammatory responses, leading to abnormal vascular endothelial function and accelerated atherosclerosis, exacerbating myocardial injury and the risk of cardiovascular events (66). Elevated SPAP is associated with changes in cardiac structure and function, and after myocardial infarction, it may indicate increased right ventricular load, leading to dysfunction, affecting the heart's pumping ability, increasing the risk of heart failure and death (67). These characteristics affect patient outcomes through various biological pathways, and a deeper understanding of these mechanisms can help better understand the disease development process and develop targeted treatment strategies.

4. Optimization methods

In this study, we combined four metaheuristic algorithms: the Augmented Grey Wolf Optimizer (AGWO), Bald Eagle Search Optimization (BES), Golden Jackal Optimizer (GJO), and Puma Optimizer (PO). These algorithms, each mimicking unique behaviors in nature, possess different search strategies that effectively avoid local optima and demonstrate efficient search capabilities and robustness in complex decision spaces. To optimize model performance, we employed grid search and cross-validation methods to fine-tune hyperparameters. Grid search systematically iterates through predefined hyperparameter values and evaluates each combination using cross-validation. Cross-validation divides the dataset into multiple subsets, using one subset as a test set and the rest as training sets, to assess the model's generalization ability. This study specifically utilized the Monte Carlo Cross-Validation (MCCV) method, which evaluates the performance of optimizers under different hyperparameter settings through random sampling to determine the optimal parameter combination, thereby maximizing the model's predictive accuracy.

4.1. Augmented Grey Wolf Optimizer (AGWO)

The AGWO algorithm emphasizes the search parameter (A), fluctuating the global Grey Wolf Optimization (GWO). It matches gray wolves’ hunting behavior, where a leader, α, directs the pack, supported by secondary wolves, β, aiding in decision-making. α represents the estimated outcomes targeted at resolving the research issue (68). The hunting process is categorized into four different sections as follows (69):

4.1.1. Foraging for prey

Exploring prey locations is enabled through the divergence of search agents, a condition satisfied when |A| surpasses 1. Parameter A, essential in balancing exploration and exploitation, is primarily contingent upon parameter a as described in (Equation 19):

a=2cos(rand)×t/Max_iter (19)
A=2a.r1a (20)
C=2.r2 (21)

The parameter a randomly and nonlinearly transitions from 2 to 1 as the iteration (t) increases, while r1 and r2 represent consistently dispersed random vectors ranging between 0 and 1 (Equations 20, 21). This process continues until it reaches the maximum iteration.

4.1.2. Surrounding the prey

The mathematical formulation relating to the encirclement of prey is described as follows (Equations 22, 23):

D=|C.XpiXi| (22)
Xi+1=XpiA.D (23)

X represents the vector indicating the location of the grey wolf, while Xp signifies the vector demonstrating the location of the prey.

4.1.3. Hunting and tracking

In the proposed AGWO algorithm (Algorithm 1), the strategy for hunting is determined exclusively by the parameters α and β, which are defined in (Equations 2426).

Da=|C1.XaiXi|,Dβ=|C2.XβiXi| (24)
X1=XaiA1.Da,X2=XβiA2.Dβ (25)
X1+i=X1+X2/2 (26)
Algorithm 1.

Pseudocode outlining the AGWO.

Initialize population of grey wolves randomly
Estimate the fitness of each wolf in the population
Repeat until stopping standards are met:
 Update α and β positions based on fitness
 Renew other wolves' positions based on α and β positions
 Apply search operator to explore new solutions
 Assess the fitness of new solutions
 Update α and β positions if necessary
Return the best solution found

4.1.4. Attacking the Pre

The coordinated efforts of search agents may aid in the process of preying on a target; this investigation is conducted when the magnitude of set A is less than one.

4.2. Bald Eagle Search Optimization (BES)

Alsattar et al. introduced the Bald Eagle Search (BES) algorithm, drawing inspiration from the discerning hunting strategy observed in bald eagles (70). This algorithm is arranged around three sequential phases reflective of the bald eagle's hunting behavior. Initially, the algorithm identifies spatial domains characterized by a delicate presence of potential targets. Subsequently, within these delineated spaces, extensive exploration is conducted to determine optimal solutions. Finally, similar to the decisive swooping action of the bald eagle, the algorithm strategically converges towards superior solutions (71). Through this emulation of the bald eagle's hunting strategy, the BES algorithm demonstrates a deliberate and efficient approach to optimization problem-solving (72).

4.2.1. Space selection stage

During this phase, bald eagles strive to select a search area abundant with food, aiming for optimal conditions. Here is the mathematical representation of this stage (Equation 27):

Xnew,i=Xbest+βr(XmeanXi) (27)

β control's location changes; r is a random number between 0 and 1. Xnew,i is a new position, Xbest is the best position found, Xmean is the average position of all eagles and Xi is the current eagle's position.

4.2.2. Searching-in-space stage

During this stage, the bald eagle conducts a methodical search across various directions within the designated space to locate potential prey. It strategically assesses optimal hunting positions and plans its swooping maneuvers accordingly. This stage can be succinctly described in mathematical terms as (Equations 2834):

Xnew,i=Xi+f(i)(XiXi+1)+g(i)(XiXmean) (28)
g(i)=gr(i)(max|gr|) (29)
f(i)=fr(i)(max|fr|) (30)
gr(i)=r(i).sin(φ(i)) (31)
fr(i)=r(i).cos(φ(i)) (32)
φ(i)=β.π.rand (33)
r(i)=φ(i)+S.rand (34)

S quantifies the total number of search attempts, while β denotes the angle delineating the direction of the search. The term rand encompasses a numerical value within the inclusive range of 0 to 1.

4.2.3. Swooping stage

In the final phase, each bald eagle begins a swinging motion from a superior location toward its predefined prey. The mathematical definition of this behavior in this phase is presented as follows (Equations 3541):

Xnew,i=rand.Xbest+g1(i).(XiB1.Xmean)+f1(i).(XiB2.Xbest) (35)
g1(i)=gr(i)(max|gr|) (36)
f1(i)=fr(i)(max|fr|) (37)
gr(i)=r(i).sin(φ(i)) (38)
fr(i)=r(i).cos(φ(i)) (39)
φ(i)=β.π.rand (40)
r(i)=φ(i) (41)
B1,B2ϵ[1,2]

.

The comprehensive depiction of the BES algorithm is accessible through the subsequent pseudocode (Algorithm 2), and the flowchart of BES is illustrated in Figure 6.

Algorithm 2.

Pseudocode outlining the Bald Eagle Search Optimization.

Randomly assign initial values: Xi for n points.
Determine the fitness values of the initial points: f(Xi)
While (until the termination criteria are met)
  Selecting Space
  For (every individual i within the population)
  Xnew=Xbest+βr(XmeanXi)
   If f(Xnew)<f(Xi)
   Xi=Xnew
   If f(Xnew)<f(Xbest)
   .Xbest=Xnew
   End If
   End If
  End For
  Searching in Space
  For (each individual denoted as i within the population)
  Xnew=Xi+f(i)(XiXi+1)+g(i)(XiXmean)
   If f(Xnew)<f(Xi)
   .Xi=Xnew
   If f(Xnew)<f(Xbest)
   .Xbest=Xnew
   End If
   End If
  End For
  Swooping and Descending
  For (each memberi among the population)
  .Xnew=rand.Xbest+g1(i).(XiB1.Xmean)+f1(i).(XiB2.Xbest)
   If f(Xnew)<f(Xi)
   .Xi=Xnew
   If f(Xnew)<f(Xbest)
   .Xbest=Xnew
   End If
   End If
  End For
  Set k=k+1
End While
Figure 6.

Figure 6

The flowchart of BES.

4.3. Golden Jackal Optimizer (GJO)

The Golden Jackal Optimizer (GJO) represents a recent advancement in swarm-based optimization methodologies strategically developed to optimize diverse engineering systems and processes (73). Drawing inspiration from the collaborative hunting tactics observed in golden jackals, the GJO includes three important subprocesses: Prey Exploration, Surrounding, and Attacking (74, 75). Within this section, the mathematical formulation of the GJO is clarified.

At the beginning of the optimization process, the generation of a set of prey location matrices is initiated, achieved via the randomization method described in (Equation 42):

[Y1,1Y1,jY1,nY2,1Y2,jY2,nYN1,1YN,1YN1,jYN1,nYN,jYN,n] (42)

The method that the golden jackal hunts, where the E value is greater than 1, is illustrated numerically. N is the number of prey populations at this stage, and n is the total number of variables.

Y1(t)=YM(t)E.|YM(t)rl.prey(t)| (43)
Y2(t)=YFM(t)E.|YFM(t)rl.prey(t)| (44)

In the given equation, t represents the iteration number, YM(t) and YFM(t) denote the positions of male and female golden jackals, respectively, while prey(t) represents the prey's position vector. The updated positions of the golden jackals are Y1(t) and Y2(t), respectively. The variable E signifies the prey's evading energy, calculated using a specific formula (Equations 45, 46):

E=E1.E0 (45)
E1=c1.(1(tT)) (46)

The equation assesses the ability of prey to avoid predators, considering several aspects. Firstly, a random value within the range of −1 to 1, denoted as E0, represents the prey's starting energy level. The parameter T signifies the maximum number of iterations, while c1 is a constant value typically set to 1.5. E1 indicates how quickly the prey's energy decreases over time.(Equations 47, 48) apply the distance between the golden jackal and the prey, expressed as |YM(t)rl.prey(t)|, where rl denotes a vector of random numbers resulting from the Levy flight function.

rl=0.05.LF(y) (47)
LF(y)=0.01×(μ×σ)/(|v(1/β)|),σ={Γ(1+β)×sin(πβ/2)Γ(1+β2)×β×(2β1)} (48)

The calculation uses random values for u and v that fall between 0 and 1, and it also includes a constant b that is often set to 1.5 by default.

Y(t+1)=Y1(t)+Y2(t)2 (49)

The formula calculates the prey's updated location, Y(t+1), based on the positions of the male and female golden jackals.

The reduced capability of the prey to evade emerges when it faces violence from the golden jackals. This mathematical expression illustrates a decline in evading energy when |E| is less than or equal to 1.

Y1(t)=YM(t)E.|rl.YM(t)prey(t)| (50)
Y2(t)=YFM(t)E.|rl.YFM(t)rl.prey(t)| (51)

The comprehensive depiction of the GJO algorithm is outlined in the pseudocode provided below (Algorithm 3) and Figure 7 illustrates the flowchart of GJO.

Algorithm 3.

Pseudocode delineation of the Golden Jackal Optimizer.

Inputs: The population size N and maximum number of iterations T
Outputs: The position of the prey and its corresponding fitness value
Initiate the random population of prey denoted as Yi for i=1,2,...,N
 While (t<T)
Compute the fitness values of the prey
Y1 = determine the best position of the male jackal among the prey individuals
Y2 = identify the second-best position of the female jackal among the prey individuals
  For (each member of the prey population)
  Modify the evading energy E in accordance with (Equations 47, 49)
  Adjust the variable rl by applying (Equations 49, 50)
   If (|E|1) (Exploration Stage)
   Refine the prey's location in space via the use of (Equations 43, 44, 49).
   If (|E|>1) (Exploitation Stage)
   Adjust the prey's position based on (Equations 4951).
  End For
.  t=t+1
 End While
Return Y1

Figure 7.

Figure 7

The flowchart of GJO.

4.4. Puma optimizer (PO)

The Puma algorithm was subjected to review by Abdollah Zadeh et al. (76), and its description is as follows:

4.4.1. Inspiration

The Puma, also called cougar or mountain lion, is a large American feline found across a vast range from the Andes to Canada. It is known for its adaptability, nocturnal nature, and ambush hunting style, preying on deer, rodents, and occasionally domestic animals (7779). Pumas prefer dense scrub and rocky habitats, establish large territories, and display typical territorial behavior (80). They typically capture large prey every two weeks, relocating it for feeding over several days. Pumas are solitary, except for mothers and cubs, and rarely encounter each other except to share prey or in small communities centered around a dominant male's territory (81).

4.4.2. Mathematical representation

This section presents the PO algorithm, which draws inspiration from the hunting behaviors of pumas. Different from conventional meta-heuristic optimizers, PO introduces a unique mechanism for transitioning between the exploration and exploitation phases. It conceptualizes the best solution as a male puma and views the entire optimization space as a puma's territory, with other solutions representing female pumas. Purposeful and intelligent phase selection guides solutions through exploration or exploitation in each iteration. Drawing from puma behavior, diverse optimization approaches are employed in each phase, enhancing the algorithm's efficiency.

4.4.2.1. Puma-inspired intelligence (phase transition mechanism)

The algorithm, inspired by puma behavior, features an exploitation phase for revisiting known hunting grounds and an exploration phase for discovering new territories. It incorporates a sophisticated mechanism resembling an advanced hyper-heuristic algorithm, integrating diversity and intensification components for scoring. The phase transition section adopts two approaches inspired by puma intelligence: inexperienced pumas explore new territories while targeting promising areas for ambush.

4.4.2.1.1. Inexperienced phase

In its early stages, the puma lacks experience and often engages in exploration activities simultaneously due to its unfamiliarity with its environment and lack of awareness of hunting locations within its territory. Conversely, it seeks hunting opportunities in favorable areas. In the Puma algorithm, during the initial three iterations, both exploration and exploitation operations are carried out concurrently until initialization is completed in the phase transition phase. In this section, as the exploitation and exploration phases are selected in each iteration, only two functions (f1 and f2) are applied and calculated using (Equations 5255):

f1Explor=PF1.(SeqcostExplore1SeqTime) (52)
f1Exploit=PF1.(SeqcostExploit1SeqTime) (53)
f2Explor=PF2.(SeqcostExplore1+SeqcostExplore2+SeqcostExplore3SeqTime1+SeqTime2+SeqTime3) (54)
f2Exploit=PF2.(SeqcostExploit1+SeqcostExploit2+SeqcostExploit3SeqTime1+SeqTime2+SeqTime3) (55)

The values of Seqcost, associated with both exploitation and exploration phases, are determined using (Equations 5255), while SeqTime remains constant at 1. PF1 and PF2, parameters with predetermined values, are used to prioritize the functions f1 and f2 during the optimization process.

SeqCostExplore1=|CostBestInitialCostExlore1| (56)
SeqCostExplore2=|CostExplore2CostExlore1| (57)
SeqCostExplore3=|CostExplore3CostExlore2| (58)
SeqCostExploit1=|CostBestInitialCostExloit1| (59)
SeqCostExploit2=|CostExploit2CostExloit1| (60)
SeqCostExploit3=|CostExploit3CostExloit2| (61)

In Equations 56, 61, the term CostBestInitial represents the cost of the initial optimal solution generated during the initialization phase. Additionally, six variables, namely CostExlore1, CostExlore2, CostExlore3, CostExloit1, CostExloit2, and CostExloit3, denote the costs associated with the best solutions obtained from the exploitation and exploration phases across three repetitions (Equations 5760).

After evaluating the functions f1 and f2 following the third iteration, a decision is made to exclusively pursue either exploration or exploitation phases. The positive experiences of other Pumas influence this choice. To determine which phase to prioritize, the coordinates of both the exploitation and exploration points are computed by applying (Equations 62, 63):

ScoreExplore=(PF1.f1Explor)+(PF2.f2Explor) (62)
ScoreExploit=(PF1.f1Exploit)+(PF2.f2Exploit) (63)

After computing ScoreExplore and ScoreExploit using (Equations 62, 63), the system determines whether to proceed with the exploration or exploitation phase based on their values. If ScoreExploitScoreExplore, the exploitation stage is entered; otherwise, the exploration step is chosen. However, a serious consideration arises at the end of the third iteration: each step independently generates solutions exceeding the total population size. To address this, the total cost of solutions from both phases is calculated at the end of the third iteration. Only the best solutions from the entire pool are retained, ensuring that the population size remains constant by replacing the current solutions.

4.4.2.1.2. Experienced and Skilled phase

After three generational iterations, the Pumas acquire a satisfactory level of experience to opt for a singular optimization phase for subsequent iterations. Within this phase, three distinct scoring functions, namely f1, f2, and f3, are applied. The main function, f1, prioritizes either the exploration or exploitation phase based on their comparative performance, with a particular emphasis on the exploration phase. This function is determined using (Equation 52).

f1texploit=PF1.|CostoldexploitCostnewexploitTtexploit| (64)
f1texploit=PF1.|CostoldexploreCostnewexploreTtexplore| (65)

Equations 64, 65 define f1texploit and f1texploit for the exploitation and exploration phases at iteration t. Costoldexploit and Costnewexplore are costs before and after improving the current selection, while Ttexplore and Ttexploit indicate unselected iterations. PF1, set between 0 and 1, determines the importance of the first function: advanced values prioritize it.

The second function gives preference to the phase that outperforms the other, focusing on resonance. It assesses good performances sequentially, aiding in the selection of the exploitation phase. (Equations 66, 67) are employed to calculate this function.

f2texploit=PF2.|(Costold.1exploitCostnew,1exploit)+(Costold.2exploitCostnew,2exploit)+(Costold.3exploitCostnew,3exploit)Tt.1exploit+Tt.2exploit+Tt.3exploit| (66)
f2texplore=PF2.|(Costold.1exploreCostnew,1explore)+(Costold.2exploreCostnew,2explore)+(Costold.3exploreCostnew,3explore)Tt.1explore+Tt.2explore+Tt.3explore| (67)

Equations 66, 67 introduce functions for exploration and exploitation in an optimization process, with costs representing solution performance. Updates to solution costs are tracked across current and past selections. Iteration counts capture unselected iterations between selections. The parameter PF2 influences the effectiveness of the exploration-exploitation balance. Overall, these elements form a framework for optimizing strategies.

The third function in the selection mechanism emphasizes diversity by increasing in value when its priority rises and decreasing when it declines. It ensures that less frequently selected phases still have a chance to be chosen, preventing the algorithm from getting trapped in local optima. This function is depicted in (Equations 68, 69).

f3texploit={ifselected,f3texploit=0otherwise,f3texploit+PF3 (68)
f3texplore={ifselected,f3texplore=0otherwise,f3texplore+PF3 (69)

Equations 68, 69 define functions f3texploit and f3texplore separately, representing the third function in exploitation and exploration stages over iterations signified by t. (Equation 54) specifies that if a stage is not chosen, the value of its corresponding third function increases by a parameter PF3 in each iteration; otherwise, it is set to zero. PF3 is a user-adjustable parameter ranging between 0 and 1, determining the likelihood of selecting a stage. A higher PF3 increases the chances of selecting the stage with a lower score and vice versa. (Equations 70, 71) compute the cost associated with changing stages.

Ftexploit=(αtexploit.(f1texploit))+(αtexploit.(f2texploit))+(δtexploit.(lc.f3texploit)) (70)
Ftexplore=(αtexplore.(f1texplore))+(αtexplore.(f2texplore))+(δtexplore.(lc.f3texplore)) (71)
c={{|CostoldCostnew|}exploitation,{|CostoldCostnew|}exploration},0lc (72)
αtexplore,exploit={ifFexploit>Fexplore,αexploit=0.99,αexplore=αexplore0.01,0.01otherwise,αexplore=0.99,αexploit=αexploit0.01,0.01 (73)
δtexploit=1αtexploit (74)
δtexplore=1αtexplore (75)

Equations 70, 71 determine final costs for exploitation and exploration phases, with parameters a and d varying based on phase results, prioritizing diversity. (Equation 73) penalizes parameter a of the phase with higher cost, adjusting it linearly by 0.01. This approach, as discussed in (82), relies on lc, representing non-zero cost differences between exploitation and exploration phases (Equation 72).

4.4.2.2. Exploration

In the exploration phase, inspired by the behavior of pumas searching for food, a random search is conducted within the territory. Pumas either explore new areas or approach other pumas to potentially share prey. Initially, the entire population is sorted in ascending order, and then each puma refines its solutions using (Equations 74, 75).

Ifrand1>0.5,Zi,G=RDim(UbLB)+LBOtherwise,Zi,G=Xa,G+G.(Xa,GXb,G)+G.(((Xa,GXb,G)(Xc,GXd,G))+((Xc,GXd,G)(Xe,GXf,G))) (76)
G=2.rand21 (77)

Equations 76, 77 involves randomly generating numbers within specified bounds and dimensions for problem-solving. Depending on certain conditions, one of two equations is selected to produce a new solution. This solution is then used to improve the current solution (Equations 7881).

Xnew={Zi.G,ifj=jrandorrand3UXa,G,otherwise (78)
NC=1U (79)
p=NCNpop (80)
ifCostXnew<CostXi,U=U+P (81)
4.4.2.3. Exploitation

In the exploitation stage, the PO algorithm employs two operators inspired by puma behaviors: ambush hunting and sprinting. Pumas, in nature, typically ambush prey from concealed positions or chase them down in open spaces. (Equation 82) simulates the behavior of chasing prey, reflecting one of these hunting strategies.

Xnew={ifrand40.5,Xnew=(mean(Soltotal)Npop).X1r(1)β×Xi1+(α.rand5)otherwise,ifrand6L,Xnew=Pumamale+(2.rand7).exp(rand1).X2rXiotherwise,Xnew=(2×rand8)×F1.R.X(i)+F2.(1R).Pumamale)(2.rand91+randn2)Pumamale (82)

Equation 82 in the PO algorithm embodies two strategies inspired by puma behaviors: fast running and ambush hunting. During exploitation, if a randomly generated number exceeds 0.5, the fast-running strategy is applied; otherwise, the ambush strategy is chosen. These strategies involve different movements towards prey, with various parameters and random factors influencing the process.

The Puma optimizer stands out for its higher implementation complexity compared to other optimizers due to its multiple phases and operations involved in creating intelligent systems. In each iteration, the cost function is applied only once for each search agent, ensuring acceptable computational complexity, as detailed in the relevant section.

4.5. Hybrid models’ development

AGWO, BES, GJO, and PO optimizers integrated with base models to supplement the efficacy of the selected models. As presented in Tables 3, 4, the fine tunned hyperparameters in the hybridization process for LGBC and XGBC are reported. For instance, the hyperparameters n_estimators, max_depth, and learning_rate are crucial for optimizing ensemble methods like Gradient Boosting Machines. n_estimators define the number of trees in the ensemble, with more trees generally improving performance but increasing computational cost and overfitting risk. max_depth limits the depth of each tree, balancing the ability to capture complex patterns with the risk of overfitting; deeper trees can capture more details but may overfit, while shallower trees might underfit. learning_rate, specific to boosting algorithms, scales the contribution of each tree, with lower rates enhancing robustness and preventing overfitting but requiring more iterations.

Table 3.

The results of hyperparameters tunning in LGBC-based hybrid models development.

Prediction scenarios Model Hyperparameter
num_leaves max_depth learning_rate n_estimators max_bin
Scenario 1 LGAG 321 285 0.168 317 681,000
LGBE 131 148 0.361 119 289,000
LGGJ 792 45 0.747 796 844,000
LGPO 613 999 0.504 999 1,000
Scenario 2 LGAG 654 213 0.566 813 685,100
LGBE 999 648 0.736 889 753,000
LGGJ 981 539 0.746 822 788,114
LGPO 274 921 0.497 867 66,000
Scenario 3 LGAG 213 286 0.126 681 385,140
LGBE 140 411 0.651 214 293,000
LGGJ 442 845 0.740 226 922,419
LGPO 299 633 0.771 77 253,000

Table 4.

The results of hyperparameters tunning in XGBC-based hybrid models development.

Prediction scenarios Model Hyperparameter
n_estimators max_depth learning_rate colsample_bytree subsample reg_alpha reg_lambda
Scenario 1 XGAG 581 48 0.237 0.371 0.382 0.14 0.641
XGBE 628 134 0.196 0.657 0.658 0.007 0.189
XGGJ 414 282 0.474 0.902 0.894 0.334 0.005
XGPO 341 999 0.927 0.999 0.873 0.001 0.999
Scenario 2 XGAG 168 34 0.2781 0.391 0.617 0.236 0.468
XGBE 345 315 0.16 0.152 0.294 0.067 0.065
XGGJ 478 167 0.933 0.465 0.590 0.236 0.005
XGPO 999 234 0.99 0.269 0.999 0.999 0.001
Scenario 3 XGAG 691 31 0.017 0.681 0.914 0.642 0.260
XGBE 333 371 0.324 0.578 0.484 0.143 0.143
XGGJ 201 393 0.208 0.232 0.393 0.635 0.005
XGPO 394 595 0.614 0.488 0.667 0.001 0.072

Furthermore, Figure 8 illustrates the convergence of hybrid models based on LGB across all three scenarios over 200 iterations. In the second scenario, the initial iterations for the hybrid models commence with a modest Accuracy of approximately 0.5, whereas in the third scenario, they begin with a higher Accuracy of around 0.6. Remarkably, the LGBE (S3) model achieves a remarkable accuracy of 0.97 within approximately 140 iterations. The convergence patterns of XGB-based hybrid models are depicted in Figure 9. Initially, the models display an accuracy of approximately 0.6. The XGBE (S3) model attains an Accuracy of nearly one after 125 iterations. Furthermore, the XGAG (S1) model achieves an Accuracy of 0.91 by the 110th iteration, indicating the weakest performance of features in scenario (1) in training hybrid models.

Figure 8.

Figure 8

The convergence plot of LGB-based hybrid models in all three scenarios.

Figure 9.

Figure 9

The convergence plot of XGB-based hybrid models in all three scenarios.

5. Analysis of results

5.1. Metrics for evaluating predictions

The importance of performance evaluation criteria in assessing ML algorithms is highlighted in the article, emphasizing the need to select metrics tailored to the specific problem. For comprehensive comparative analysis in classification tasks, widely recognized measures such as Accuracy, Precision, Recall, F1-Score, Correlation Coefficient (MCC), and Heidke Skill Score (HSS) are employed.

Accuracy serves as the primary metric for evaluating the accuracy of predictions. Precision, Recall, and F1-Score complement Accuracy, especially in scenarios with imbalanced data distributions. Precision measures the Accuracy of positive predictions, while Recall identifies all relevant instances within a class. The F1-Score combines both Precision and Recall to provide a balanced assessment. The MCC evaluates the reliability of binary classifications by considering true positives, true negatives, false positives, and false negatives. Higher MCC scores indicate more accurate predictions. MCC is particularly useful for assessing classifiers, especially in cases of unbalanced datasets, as it treats both positive and negative samples equally. These metrics, defined by (Equations 8387):

Accuracy=TP+TNTP+TN+FP+FN (83)
Precision=TPTP+FP (84)
Recall=TPR=TPP=TPTP+FN (85)
F1Score=2×Recall×PrecisionRecall+Precision (86)
MCC=TP×TNFP×FN(TP+FP)(TP+FN)(TN+FP)(TN+FN) (87)

TP represents the number of true positives, TN stands for the total of true negatives, FP indicates the number of false positives, and FN denotes the count of false negatives.

The HSS is a statistical metric devised by meteorologist Paul Heidke to evaluate the accuracy of categorical forecasts, primarily in meteorology (83). It involves comparing observed and forecasted categorical outcomes, taking into account hits, correct rejections, false alarms, and misses. The HSS formula provides a comprehensive assessment of predictive skills (Equation 88).

HSS=2×(TPFNFPTN)(TP+TN)×(TN+FN)+(TP+FP)×(FP+FN) (88)

HSS is a metric used in meteorology to assess the accuracy of categorical weather forecasts. It compares observed and forecasted events. A score of 1 indicates perfect agreement, and 0 suggests performance equivalent to random chance.

5.2. Findings and discussion

The results are presented across three scenarios. In the first scenario, the GRACE Scale was applied, incorporating four parameters: HR, Age, SBP, and Killip Class, which are traditionally employed in hospitals (84). Table 5 provides a comprehensive comparison of performance metrics, encompassing Accuracy, Precision, Recall, F1-Score, MCC, and HSS, for the LGBM model alongside its hybrid models (LGAG, LGBE, LGGJ, and LGPO) and the XGBC model with its hybrid versions (XGAG, XGBE, XGGJ, and XGPO) across scenario (1) during both training and testing phases and for all data. Especially, the XGBE model displayed remarkable performance, achieving an Accuracy of 0.954, outperforming other models. Close behind, the LGBE and XGPO models each attained an Accuracy of 0.944. Particular significance was the superior performance demonstrated by the BES optimizer.

Table 5.

Estimation metrics results for models’ prediction performance based on scenario (1).

Model Phase Index values
Accuracy Precision Recall F1-score MCC HSS
LGBM Train 0.902 0.877 0.902 0.876 0.446 0.117
Test 0.890 0.853 0.890 0.868 0.126
All 0.893 0.821 0.906 0.862 0.133
LGAG Train 0.913 0.908 0.913 0.882 0.298 0.210
Test 0.924 0.914 0.924 0.902 0.349
All 0.916 0.909 0.916 0.888 0.312
LGBE Train 0.949 0.946 0.949 0.945 0.677 0.623
Test 0.932 0.925 0.932 0.927 0.521
All 0.944 0.939 0.944 0.940 0.633
LGGJ Train 0.926 0.917 0.926 0.916 0.495 0.452
Test 0.925 0.913 0.925 0.916 0.434
All 0.926 0.916 0.926 0.916 0.477
LGPO Train 0.940 0.935 0.940 0.936 0.618 0.570
Test 0.927 0.919 0.927 0.921 0.479
All 0.936 0.930 0.936 0.931 0.580
XGBC Train 0.914 0.904 0.914 0.908 0.445 0.430
Test 0.913 0.907 0.913 0.909 0.410
All 0.913 0.905 0.913 0.908 0.435
XGAG Train 0.923 0.912 0.923 0.908 0.446 0.406
Test 0.928 0.917 0.928 0.918 0.448
All 0.924 0.913 0.924 0.911 0.446
XGBE Train 0.959 0.958 0.959 0.957 0.747 0.698
Test 0.943 0.938 0.943 0.939 0.604
All 0.954 0.952 0.954 0.951 0.706
XGGJ Train 0.937 0.932 0.937 0.928 0.574 0.521
Test 0.931 0.921 0.931 0.923 0.486
All 0.935 0.928 0.935 0.926 0.548
XGPO Train 0.947 0.943 0.947 0.942 0.658 0.620
Test 0.939 0.933 0.939 0.935 0.571
All 0.944 0.940 0.945 0.940 0.633

In the second scenario, the features selected by SHAP were used, which included ten parameters for the LGBM model and 13 parameters for the XGBC model. Table 6 presents the results of evaluation metrics for the two mentioned single models and their hybrid versions based on scenario (2). The LGBM model was characterized by its relatively lower performance, evidenced by an Accuracy score of 0.921. Conversely, the LGBE model emerged as a standout performer within the domain of LGBM hybrid models, showing notable efficacy with an Accuracy score of 0.963. Especially, the XGBC model displayed the highest level of performance, boasting an impressive Accuracy value of 0.978, thereby establishing itself as the benchmark against which all other models are measured.

Table 6.

Estimation metrics results for models’ prediction performance based on scenario (2).

Model Phase Index values
Accuracy Precision Recall F1-score MCC HSS
LGBM Train 0.916 0.922 0.917 0.919 0.550 0.561
Test 0.933 0.936 0.933 0.934 0.594
All 0.921 0.926 0.921 0.923 0.562
LGAG Train 0.934 0.936 0.934 0.935 0.632 0.638
Test 0.942 0.946 0.942 0.943 0.656
All 0.936 0.939 0.936 0.937 0.639
LGBE Train 0.966 0.968 0.966 0.967 0.817 0.793
Test 0.957 0.958 0.957 0.957 0.734
All 0.963 0.965 0.963 0.964 0.794
LGGJ Train 0.947 0.947 0.947 0.947 0.694 0.664
Test 0.932 0.935 0.932 0.933 0.590
All 0.942 0.943 0.942 0.943 0.664
LGPO Train 0.961 0.962 0.961 0.961 0.780 0.747
Test 0.946 0.946 0.946 0.946 0.662
All 0.957 0.957 0.957 0.957 0.747
XGBC Train 0.944 0.943 0.944 0.943 0.674 0.642
Test 0.924 0.933 0.924 0.928 0.573
All 0.938 0.939 0.938 0.939 0.642
XGAG Train 0.950 0.949 0.950 0.949 0.709 0.691
Test 0.938 0.945 0.938 0.941 0.652
All 0.946 0.948 0.946 0.947 0.691
XGBE Train 0.981 0.980 0.981 0.980 0.885 0.865
Test 0.972 0.972 0.973 0.972 0.821
All 0.978 0.978 0.978 0.978 0.867
XGGJ Train 0.958 0.957 0.958 0.957 0.756 0.732
Test 0.944 0.950 0.944 0.946 0.682
All 0.953 0.955 0.953 0.954 0.733
XGPO Train 0.966 0.968 0.966 0.967 0.817 0.793
Test 0.957 0.958 0.957 0.957 0.734
All 0.963 0.965 0.963 0.964 0.794

The features selected by RFE were applied in the third scenario, comprising six features in the LGBM-based models and eight features in the XGBC-based model. According to Table 7, the XGBE model was the peak performer, boasting an exceptional Accuracy score of 0.990. Following closely, the LGBE model secured the second position with a commendable Accuracy of 0.977, while the XGPO model secured the third rank with an Accuracy score of 0.975. In contrast, the LGBM simple model presented the least impressive performance among the models under analysis.

Table 7.

Estimation metrics results for models’ prediction performance based on scenario (3).

Model Phase Index values
Accuracy Precision Recall F1-score MCC HSS
LGBM Train 0.932 0.931 0.932 0.931 0.606 0.610
Test 0.939 0.940 0.939 0.940 0.621
All 0.934 0.937 0.934 0.934 0.610
LGAG Train 0.939 0.933 0.939 0.933 0.599 0.594
Test 0.948 0.964 0.948 0.944 0.636
All 0.942 0.936 0.942 0.936 0.609
LGBE Train 0.980 0.980 0.980 0.980 0.882 0.864
Test 0.971 0.971 0.971 0.971 0.818
All 0.977 0.977 0.977 0.977 0.864
LGGJ Train 0.956 0.953 0.956 0.953 0.723 0.706
Test 0.956 0.953 0.956 0.953 0.695
All 0.956 0.953 0.956 0.953 0.715
LGPO Train 0.967 0.966 0.967 0.966 0.801 0.793
Test 0.965 0.965 0.965 0.965 0.778
All 0.966 0.965 0.966 0.966 0.794
XGBC Train 0.942 0.940 0.942 0.941 0.653 0.643
Test 0.938 0.941 0.938 0.939 0.625
All 0.941 0.940 0.941 0.940 0.643
XGAG Train 0.955 0.955 0.955 0.955 0.741 0.728
Test 0.949 0.952 0.949 0.950 0.696
All 0.953 0.954 0.953 0.953 0.728
XGBE Train 0.992 0.991 0.992 0.991 0.951 0.938
Test 0.986 0.986 0.986 0.985 0.907
All 0.990 0.990 0.990 0.990 0.939
XGGJ Train 0.965 0.963 0.965 0.963 0.787 0.777
Test 0.963 0.962 0.963 0.963 0.762
All 0.964 0.963 0.964 0.963 0.779
XGPO Train 0.980 0.980 0.980 0.980 0.886 0.854
Test 0.964 0.964 0.964 0.964 0.773
All 0.975 0.975 0.975 0.975 0.854

In general, based on the comparative representations presented in Figures 1014, it is evident that the models from the third scenario outperform those from the first and second scenarios according to the metrics of Accuracy, Precision, Recall, F1-Score, and MCC.

Figure 10.

Figure 10

Graphical comparison of accuracy metric for the three scenarios in prediction models.

Figure 14.

Figure 14

Graphical comparison of MCC metric for the three scenarios in prediction models.

Figure 11.

Figure 11

Graphical comparison of precision metric for the three scenarios in prediction models.

Figure 12.

Figure 12

Graphical comparison of recall metric for the three scenarios in prediction models.

Figure 13.

Figure 13

Graphical comparison of F1-score metric for the three scenarios in prediction models.

Table 8 displays the evaluation criteria values used to assess the effectiveness of the models in distinguishing between the Alive and Die classes for the first scenario, while Tables 9, 10 present these metric values for the second and third scenarios, respectively.

Table 8.

The results of the evaluation criteria for assessing the effectiveness of the constructed models in classifying patients in scenario (1).

Model Phase Index values
Precision Recall F1-score
LGBM Alive 0.301 0.109 0.160
Die 0.913 0.974 0.943
LGAG Alive 0.826 0.134 0.230
Die 0.918 0.997 0.956
LGBE Alive 0.777 0.563 0.653
Die 0.956 0.983 0.969
LGGJ Alive 0.695 0.377 0.489
Die 0.938 0.983 0.960
LGPO Alive 0.724 0.518 0.604
Die 0.952 0.980 0.965
XGBC Alive 0.551 0.419 0.476
Die 0.941 0.965 0.953
XGAG Alive 0.720 0.317 0.440
Die 0.933 0.987 0.959
XGBE Alive 0.844 0.630 0.722
Die 0.963 0.988 0.975
XGGJ Alive 0.777 0.430 0.553
Die 0.944 0.987 0.965
XGPO Alive 0.799 0.546 0.649
Die 0.955 0.986 0.970

Table 9.

The results of the evaluation criteria for assessing the effectiveness of the constructed models in classifying patients in scenario (2).

Model Phase Index values
Precision Recall F1-score
LGBM Alive 0.572 0.641 0.605
Die 0.962 0.950 0.956
LGAG Alive 0.648 0.701 0.673
Die 0.969 0.961 0.965
LGBE Alive 0.778 0.852 0.813
Die 0.985 0.975 0.980
LGGJ Alive 0.687 0.704 0.696
Die 0.969 0.967 0.968
LGPO Alive 0.767 0.775 0.771
Die 0.977 0.977 0.976
XGBC Alive 0.662 0.690 0.676
Die 0.968 0.964 0.966
XGAG Alive 0.702 0.739 0.720
Die 0.973 0.968 0.970
XGBE Alive 0.936 0.824 0.876
Die 0.982 0.994 0.988
XGGJ Alive 0.739 0.778 0.758
Die 0.977 0.972 0.974
XGPO Alive 0.778 0.852 0.813
Die 0.985 0.975 0.980

Table 10.

The results of the evaluation criteria for assessing the effectiveness of the constructed models in classifying patients in scenario (3).

Model Phase Index values
Precision Recall F1-score
LGBM Alive 0.649 0.644 0.647
Die 0.963 0.964 0.964
LGAG Alive 0.786 0.518 0.624
Die 0.952 0.985 0.968
LGBE Alive 0.891 0.863 0.877
Die 0.986 0.989 0.987
LGGJ Alive 0.854 0.637 0.730
Die 0.963 0.989 0.976
LGPO Alive 0.856 0.771 0.811
Die 0.977 0.987 0.982
XGBC Alive 0.701 0.651 0.675
Die 0.964 0.971 0.968
XGAG Alive 0.943 0.764 0.754
Die 0.976 0.973 0.971
XGBE Alive 0.970 0.919 0.944
Die 0.992 0.997 0.994
XGGJ Alive 0.846 0.754 0.797
Die 0.975 0.986 0.980
XGPO Alive 0.869 0.866 0.868
Die 0.986 0.987 0.986

In all three scenarios, the models demonstrated higher accuracy in predicting and classifying patients in the Die class compared to the Alive class. Comparing the performance of the models in the Alive class in the first scenario, the XGBE model displayed superior performance with a Precision of 0.844, representing a 12.36% decrease compared to its Precision in the Die class. Conversely, the LGBE model outperformed the LGPO model with a Precision of 0.777. Moving to the second and third scenarios, the XGBE model achieved Precision values of 0.936 and 0.970, respectively, showing improved performance by 9.83% and 12.99% compared to the first scenario. Furthermore, the LGBE model maintains consistent performance in the second scenario, with a marginal difference of 0.13%, while in the third scenario, it demonstrated superior performance with a 12.79% increase.

In the first scenario, the XGBE model achieved the maximum performance in the Die class with a Precision of 0.963, while the LGBE, XGPO, and LGPO models displayed nearly identical performance in this class, with Precision values of 0.956, 0.955, and 0.952, respectively. Moving to the second scenario, the XGPO model demonstrated superior performance in classifying patients in the Die class with a Precision of 0.985, while the XGBE model ranked third with a slight difference of 0.31%. Lastly, in the third scenario, the XGBE model surpassed all others with an impressive Precision of 0.992 in the Die class, securing the top position. The LGBE model followed closely behind with a Precision of 0.986, earning the second rank.

Figure 15 presents a visual comparison of the models introduced in this research across scenarios (1), (2), and (3), using Precision, Recall, and F1-score metrics. In the LGBM and XGBC basic models, the Recall values are lower than those of other hybrid models in the Alive class, with values of 0.109 and 0.419 for the first scenario, 0.641 and 0.690 for the second scenario, and 0.644 and 0.651 for the third scenario, respectively. The lowest Recall value is attributed to the LGBM model in scenario (1) for the classification of Alive patients, while the highest value is recorded for the XGBE model in the third scenario and LGAG in the first scenario in the Die class, both with a value of 0.997.

Figure 15.

Figure 15

Comparative visual display of evaluation metrics for models across three scenarios in the Die and alive classes.

Figure 16 displays the confusion matrix depicting the classification performance in scenario (1), using the four features introduced by the GRACE Scale. This visual representation offers insights into the model's classification outcomes across various diagnostic categories. The LGBM model showed the highest error rate in misclassifying individuals from the Alive class into the Die group, with 253 patients misclassified. Following closely, the LGAG model ranked next, committing a similar error with 246 misclassified patients. Conversely, the LGAG model demonstrated the lowest error rate, misclassifying only eight deceased patients into the Alive class.

Figure 16.

Figure 16

Confusion matrices depicting the accuracy of individual models within scenario (1).

Additionally, the XGBC model incorrectly classified 97 dead patients into the Alive group. In contrast, the LGBE model showcased superior performance compared to other hybrid models based on LGBM, with 124 and 46 misclassifications in the Alive and Die classes, respectively. Similarly, the XGBE model exhibited the lowest misclassification rate compared to other XGBC-based hybrid models.

Figure 17 depicts the correct and incorrect classification results of the models based on scenario (2), while Figure 18 represents those based on scenario (3). In the second scenario, SHAP was employed to identify effective features in modeling, whereas the third scenario employed RFE, resulting in an obvious increase in model accuracy. In scenario (2), as illustrated in Figure 17, the LGBM model continued to display the highest misclassification rate in the Alive class, speciously placing 102 patients in the Die class; however, it had enhanced its performance by 59.68% in correctly classifying the group of living patients. Conversely, the LGBE and XGPO models demonstrated the lowest errors in classifying living patients, misclassifying only 42 patients while correctly classifying 242 patients. The XGBE model excelled in classifying dead individuals, accurately classifying 2,727 patients while misclassifying only 16 patients.

Figure 17.

Figure 17

Confusion matrices depicting the accuracy of individual models within scenario (2).

Figure 18.

Figure 18

Confusion matrices depicting the accuracy of individual models within scenario (3).

In scenario (3), as delineated in Figure 18, notable discrepancies appear in the classification of alive patients. Specifically, the LGAG model shows a significant degree of error, misclassifying 137 patients. Similarly, the LGBM model demonstrates a considerable level of misclassification, with 99 patients incorrectly assigned to the Alive class. Contrarily, the XGBE model displays admirable performance, achieving 261 correct classifications and 23 misclassifications within the Alive group. Impressively, the XGBE model makes minimum errors, with only eight deceased patients erroneously categorized as Alive.

In general, the models in scenario (1) show the weakest performance, while the highest performance is observed in the third scenario. The application of scenario (1) in hospitals entails a high risk, as it relies only on four features: HR, Age, SBP, and Killip Class. Conversely, in scenario (2), the models employ ten features for LGBM and 13 features for XGB, leading to significantly higher accuracy compared to predictions based on the GRACE score. In scenario (3), the efficiency of the models surpasses that of scenarios (1) and 2 despite using fewer features 6 for the LGBM model and 8 for the XGB model. It is noteworthy that despite the reduced number of parameters, higher accuracy has been achieved. Upon comparing the two models, it can be concluded that the XGBE model offered the highest accuracy with eight features. This level of accuracy allows hospitals and healthcare professionals to predict the probability of survival more accurately, thereby reducing in-hospital mortality rates and tailoring treatments accordingly.

On the other hand, scenario (3) demanded a diminished set of parameters in comparison to scenario (2), thereby reducing the time required for testing. Such efficiency is particularly admirable in the context of patients’ serious conditions, where timely intervention is paramount. Moreover, the efficient testing regimen of scenario three not only hastens decision-making but also mitigates financial burdens. The decreased number of requisite tests translates to lower costs incurred by both patients and healthcare facilities, emphasizing the compelling value proposition of the model's heightened accuracy.

Figures 19, 20 depict HSS values for models based on LGBM and XGBC, respectively, to assess the accuracy of the predictions. In Figure 19, the mean HSS value for the third scenario approximates 0.7, while for the second scenario, it is around 0.65. Notably, the overall mean HSS value for the first scenario is approximately 0.4. This delineates that in scenario (1), the predictive accuracy stands at roughly 40%, which deviates from acceptable performance standards. Conversely, as depicted in Figure 20, the mean HSS value is about 0.5, highlighting the models’ lack of precision in scenario one concerning patient prediction and classification accuracy. Moreover, the mean HSS value for XGBC-based models in scenarios (2) and (3) averages approximately 0.67 and 0.71, respectively. Collectively, these findings prove the superior performance of models in scenario (3), revealing their exceptional forecasting capabilities and optimal operational efficiency.

Figure 19.

Figure 19

The chart illustrates the HSS values of LGBM models across three scenarios.

Figure 20.

Figure 20

The chart illustrates the HSS values of XGBC models across three scenarios.

5.3. Comparative analysis

In this section, for comparing the Accuracy of predictions conducted by the best developed model (XGBE in the third scenario) in the study by those models in existing literature, the metric results are reported in Table 11. The results reveal that the Accuracy, Precision, and F1-score of the XGBE were 3% to 5% higher than the developed Catboost in the previous study.

Table 11.

Comparison results between the accuracy of the best developed model with models in existing literature.

Developed model Reference Accuracy Precision F1-score
Categorical boosting (Catboost) (32) 0.96 0.95 0.97
XGBE (XGB optimized with BEO) This study 0.99 0.99 0.99

6. Conclusion

Cardiovascular disease presents a significant global health challenge, especially in low-income countries, contributing to increased mortality rates. Myocardial infarction (MI) arises from reduced blood flow to the heart, leading to tissue damage and symptoms like chest pain and shortness of breath. Effective management of ST-segment elevation myocardial infarction (STEMI) was critical, with early reperfusion therapy, particularly through percutaneous coronary intervention (PCI), prioritized for optimal care. This study employed advanced machine learning (ML) techniques to investigate risk factors influencing in-hospital mortality (IHM) in MI patients following PCI. Many ML classifiers, such as Extreme Gradient Boosting (XGB), Light Gradient Boosting (LGB), Stochastic Gradient Boosting (SGB), and Histogram Gradient Boosting (HGB), were used, and Monte Carlo cross-validation (MCCV) assisted in selecting top-performing models. Three scenarios were designed to evaluate forecast accuracy, one of which (scenario 1) was based on the traditional GRACE scaling system which can be calculated using online calculators available on medical websites or through electronic health record systems. The objective of this study was to provide insights to improve risk assessment and patient care strategies for MI patients undergoing PCI by using more imperative features of the patients rather than those utilized in traditional methods (GRACE score), which are extracted by feature selection methods. Additionally, meta-heuristic algorithms, including Gray Wolf Optimizer (AGWO), Bald Eagle Search Optimization (BES), Golden Jackal Optimizer (GJO), and Puma Optimizer (PO), were employed to enhance prediction accuracy.

In the evaluation of scenario (1) using the F1-Score standard, the LGBE and XGBE models demonstrated superior performance with values of 0.940 and 0.951, respectively. In the second scenario, these values increased to 0.964 and 0.978, indicating an improvement of 2.4% and 2.76% in model performance. Moreover, in scenario (3), these models showed further performance enhancements, with F1-score values increasing by 3.79% and 3.9%. The MCC value for the LGBE and XGBE models in the third scenario reached the highest level, with scores of 0.864 and 0.939, respectively. Despite scenario (1)'s reliance on only four features and its consequent weak performance, scenarios (2) and (3) demonstrate improved accuracy by applying more parameters. Especially, scenario (3) surpasses the others in efficiency despite employing fewer features, with the XGB model achieving the highest accuracy using eight features. This improved accuracy enables hospitals to predict survival probabilities more precisely, thereby reducing in-hospital mortality rates and permitting tailored treatments. Scenario (3)'s streamlined parameter testing process makes it the preferred choice, offering swift decision-making and cost reductions while ensuring accurate forecasts, particularly critical in serious patient conditions. Furthermore, the model constructed in this study can be integrated into clinical decision support systems, such as electronic health record (EHR) systems, to automatically provide risk scores when assessing STEMI patients, assisting doctors in considering the patient's IHM risk when choosing treatment strategies. Thus, a personalized treatment plan can be developed based on the patient's IHM risk level. For example, in high-risk patients, more proactive preventive treatment measures, such as early cardiac rehabilitation programs or intensified medication therapy, can be considered. At the same time, the predictive results of the model can serve as a basis for discussion among multidisciplinary teams, promoting communication and collaboration among medical personnel with different professional backgrounds, and jointly developing the best treatment plan for the patient.

7. Limitations

The main limitation of this study is the single-center nature of the data source, which may limit the assessment of the model's generalizability. Additionally, although we have established an effective predictive model, we have not conducted detailed analyses on different patient subgroups, which may affect the model's applicability within specific subgroups. Future studies will address these limitations by collecting multicenter data and performing subgroup analyses to improve the model's generalizability and accuracy.

Funding Statement

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported by the Xuzhou Medical University Affiliated Hospital of Science and Technology Development Excellent Talent Fund Project (Grant No. XYFY202249).

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.

Author contributions

NT: Conceptualization, Writing – original draft, Data curation, Formal Analysis, Investigation, Methodology. SL: Data curation, Formal Analysis, Investigation, Writing – review & editing. KL: Data curation, Formal Analysis, Investigation, Writing – review & editing. QiaZ: Data curation, Formal Analysis, Investigation, Writing – review & editing. YD: Data curation, Formal Analysis, Investigation, Writing – review & editing. HS: Data curation, Formal Analysis, Funding acquisition, Investigation, Writing – review & editing. QinZ: Data curation, Formal Analysis, Investigation, Writing – review & editing. JH: Funding acquisition, Writing – review & editing. CQ: Funding acquisition, Writing – review & editing.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fcvm.2024.1419551/full#supplementary-material

Datasheet1.pdf (53.9KB, pdf)

References

  • 1.Kelly BB, Narula J, Fuster V. Recognizing global burden of cardiovascular disease and related chronic diseases. Mt Sinai J Med. (2012) 79(6):632–40. 10.1002/msj.21345 [DOI] [PubMed] [Google Scholar]
  • 2.Wurie HR, Cappuccio FP. Cardiovascular disease in low- and middle-income countries: an urgent priority. Ethn Health. (2012) 17(6):543–50. 10.1080/13557858.2012.778642 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Roth GA, Huffman MD, Moran AE, Feigin V, Mensah GA, Naghavi M, et al. Global and regional patterns in cardiovascular mortality from 1990 to 2013. Circulation. (2015) 132(17):1667–78. 10.1161/CIRCULATIONAHA.114.008720 [DOI] [PubMed] [Google Scholar]
  • 4.Murphy A, Goldberg S. Mechanical complications of myocardial infarction. Am J Med. (2022) 135(12):1401–9. 10.1016/j.amjmed.2022.08.017 [DOI] [PubMed] [Google Scholar]
  • 5.Salari N, Morddarvanjoghi F, Abdolmaleki A, Rasoulpoor S, Khaleghi AA, Hezarkhani LA, et al. The global prevalence of myocardial infarction: a systematic review and meta-analysis. BMC Cardiovasc Disord. (2023) 23(1):206. 10.1186/s12872-023-03231-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Wei Q, Xiao Y, Du L, Li Y. Advances in nanoparticles in the prevention and treatment of myocardial infarction. Molecules. (2024) 29(11):2415. 10.3390/molecules29112415 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Ibanez B, James S, Agewall S, Antunes MJ, Bucciarelli-Ducci C, Bueno H, et al. 2017 ESC guidelines for the management of acute myocardial infarction in patients presenting with ST-segment elevation: the task force for the management of acute myocardial infarction in patients presenting with ST-segment elevation of the European Society of Cardiology (ESC). Eur Heart J. (2018) 39(2):119–77. 10.1093/eurheartj/ehx393 [DOI] [PubMed] [Google Scholar]
  • 8.Contessotto P, Spelat R, Ferro F, Vysockas V, Krivickienė A, Jin C, et al. Reproducing extracellular matrix adverse remodelling of non-ST myocardial infarction in a large animal model. Nat Commun. (2023) 14(1):995. 10.1038/s41467-023-36350-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Nazir S, Elgin E, Loynd R, Zaman M, Donato A. ST-elevation myocardial infarction associated with infective endocarditis. Am J Cardiol. (2019) 123(8):1239–43. 10.1016/j.amjcard.2019.01.033 [DOI] [PubMed] [Google Scholar]
  • 10.Mitsis A, Gragnano F. Myocardial infarction with and without ST-segment elevation: a contemporary reappraisal of similarities and differences. Curr Cardiol Rev. (2021) 17(4):e230421189013. 10.2174/1573403X16999201210195702 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Moroni F, Gurm HS, Gertz Z, Abbate A, Azzalini L. In-hospital death among patients undergoing percutaneous coronary intervention: a root-cause analysis. Cardiovasc Revasc Med. (2022) 40:8–13. 10.1016/j.carrev.2022.01.021 [DOI] [PubMed] [Google Scholar]
  • 12.Covino M, Fransvea P, Rosa F, Cozza V, Quero G, Simeoni B, et al. Early procalcitonin assessment in the emergency department in patients with intra-abdominal infection: an excess or a need? Surg Infect (Larchmt). (2021) 22(8):787–96. 10.1089/sur.2020.373 [DOI] [PubMed] [Google Scholar]
  • 13.Nazer R, Albarrati A, Ullah A, Alamro S, Kashour T. Intra-abdominal hypertension in obese patients undergoing coronary surgery: a prospective observational study. Surgery. (2019) 166(6):1128–34. 10.1016/j.surg.2019.05.038 [DOI] [PubMed] [Google Scholar]
  • 14.Dawson LP, Warren J, Mundisugih J, Nainani V, Chan W, Stub D, et al. Trends and clinical outcomes in patients undergoing primary percutaneous revascularisation for ST-elevation myocardial infarction: a single centre experience. Heart Lung Circ. (2018) 27(6):683–92. 10.1016/j.hlc.2017.06.722 [DOI] [PubMed] [Google Scholar]
  • 15.O'Rielly CM, Harrison TG, Andruchow JE, Ronksley PE, Sajobi T, Robertson HL, et al. Risk scores for clinical risk stratification of emergency department patients with chest pain but No acute myocardial infarction: a systematic review. Can J Cardiol. (2023) 39(3):304–10. 10.1016/j.cjca.2022.12.028 [DOI] [PubMed] [Google Scholar]
  • 16.Gerber Y, Weston SA, Enriquez-Sarano M, Jaffe AS, Manemann SM, Jiang R, et al. Contemporary risk stratification after myocardial infarction in the community: performance of scores and incremental value of soluble suppression of tumorigenicity-2. J Am Heart Assoc. (2017) 6(10):e005958. 10.1161/JAHA.117.005958 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Girwar SM, Jabroer R, Fiocco M, Sutch SP, Numans ME, Bruijnzeels MA. A systematic review of risk stratification tools internationally used in primary care settings. Health Sci Rep. (2021) 4(3):e329. 10.1002/hsr2.329 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Laurent A, Lheureux F, Genet M, Martin Delgado MC, Bocci MG, Prestifilippo A, et al. Scales used to measure job stressors in intensive care units: are they relevant and reliable? A systematic review. Front Psychol. (2020) 11:245. 10.3389/fpsyg.2020.00245 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Shuvy M, Beeri G, Klein E, Cohen T, Shlomo N, Minha S, et al. Accuracy of the global registry of acute coronary events (GRACE) risk score in contemporary treatment of patients with acute coronary syndrome. Can J Cardiol. (2018) 34(12):1613–7. 10.1016/j.cjca.2018.09.015 [DOI] [PubMed] [Google Scholar]
  • 20.Ciambrone G, Higa CC, Gambarte J, Novo F, Nogues I, Borracci RA. Continuous monitoring of coronary care mortality using the global registry for acute coronary events (GRACE) score. Crit Pathw Cardiol. (2020) 19(3):126–30. 10.1097/HPC.0000000000000208 [DOI] [PubMed] [Google Scholar]
  • 21.Chen J, Cazenave A, Dahle C, Llovel W, Panet I, Pfeffer J, et al. Applications and challenges of GRACE and GRACE follow-on satellite gravimetry. Surv Geophys. (2022) 43(1):305–45. 10.1007/s10712-021-09685-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Mitarai T, Tanabe Y, Akashi YJ, Maeda A, Ako J, Ikari Y, et al. A novel risk stratification system “angiographic GRACE score” for predicting in-hospital mortality of patients with acute myocardial infarction: data from the K-ACTIVE registry. J Cardiol. (2021) 77(2):179–85. 10.1016/j.jjcc.2020.08.010 [DOI] [PubMed] [Google Scholar]
  • 23.Bratić B, Kurbalija V, Ivanović M, Oder I, Bosnić Z. Machine learning for predicting cognitive diseases: methods, data sources and risk factors. J Med Syst. (2018) 42(12):243. 10.1007/s10916-018-1071-x [DOI] [PubMed] [Google Scholar]
  • 24.Saberi-Karimian M, Khorasanchi Z, Ghazizadeh H, Tayefi M, Saffar S, Ferns GA, et al. Potential value and impact of data mining and machine learning in clinical diagnostics. Crit Rev Clin Lab Sci. (2021) 58(4):275–96. 10.1080/10408363.2020.1857681 [DOI] [PubMed] [Google Scholar]
  • 25.Aromolaran O, Aromolaran D, Isewon I, Oyelade J. Machine learning approach to gene essentiality prediction: a review. Brief Bioinform. (2021) 22(5):bbab128. 10.1093/bib/bbab128 [DOI] [PubMed] [Google Scholar]
  • 26.Subrahmanya SVG, Shetty DK, Patil V, Hameed BMZ, Paul R, Smriti K, et al. The role of data science in healthcare advancements: applications, benefits, and future prospects. Ir J Med Sci. (2022) 191(4):1473–83. 10.1007/s11845-021-02730-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Zhao J, Zhao P, Li C, Hou Y. Optimized machine learning models to predict in-hospital mortality for patients with ST-segment elevation myocardial infarction. Ther Clin Risk Manag. (2021) 17:951–61. 10.2147/TCRM.S321799 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Yang J, Li Y, Li X, Tao S, Zhang Y, Chen T, et al. A machine learning model for predicting in-hospital mortality in Chinese patients with ST-segment elevation myocardial infarction: findings from the China myocardial infarction registry. J Med Internet Res. (2024) 26:e50067. 10.2196/50067 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Deng L, Zhao X, Su X, Zhou M, Huang D, Zeng X. Machine learning to predict no reflow and in-hospital mortality in patients with ST-segment elevation myocardial infarction that underwent primary percutaneous coronary intervention. BMC Med Inform Decis Mak. (2022) 22(1):109. 10.1186/s12911-022-01853-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Falcão FJ, Alves CM, Barbosa AH, Caixeta A, Sousa JM, Souza JA, et al. Predictors of in-hospital mortality in patients with ST-segment elevation myocardial infarction undergoing pharmacoinvasive treatment. Clinics (Sao Paulo). (2013) 68(12):1516–20. 10.6061/clinics/2013(12)07 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Tanik VO, Cinar T, Arugaslan E, Karabag Y, Hayiroglu MI, Cagdas M, et al. The predictive value of PRECISE-DAPT score for in-hospital mortality in patients with ST-elevation myocardial infarction undergoing primary percutaneous coronary intervention. Angiology. (2019) 70(5):440–7. 10.1177/0003319718807057 [DOI] [PubMed] [Google Scholar]
  • 32.Bai Z, Lu J, Li T, Ma Y, Liu Z, Zhao R, et al. Clinical feature-based machine learning model for 1-year mortality risk prediction of ST-segment elevation myocardial infarction in patients with hyperuricemia: a retrospective study. Comput Math Methods Med. (2021) 2021:7252280. 10.1155/2021/7252280 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Li YM, Jiang LC, He JJ, Jia KY, Peng Y, Chen M. Machine learning to predict the 1-year mortality rate after acute anterior myocardial infarction in Chinese patients. Ther Clin Risk Manag. (2020) 16:1–6. 10.2147/TCRM.S236498 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Wojciechowski S, Majchrzak-Górecka M, Biernat P, Odrzywołek K, Pruss Ł, Zych K, et al. Machine learning on the road to unlocking microbiota’s potential for boosting immune checkpoint therapy. Int J Med Microbiol. (2022) 312(7):151560. 10.1016/j.ijmm.2022.151560 [DOI] [PubMed] [Google Scholar]
  • 35.Li S, Song S, Huang G. Prediction reweighting for domain adaptation. IEEE Trans Neural Netw Learn Syst. (2017) 28(7):1682–95. 10.1109/TNNLS.2016.2538282 [DOI] [PubMed] [Google Scholar]
  • 36.Li Y, Li M, Li C, Liu Z. Forest aboveground biomass estimation using landsat 8 and sentinel-1A data with machine learning algorithms. Sci Rep. (2020) 10(1):9952. 10.1038/s41598-020-67024-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Kobayashi Y, Yoshida K. Quantitative structure-property relationships for the calculation of the soil adsorption coefficient using machine learning algorithms with calculated chemical properties from open-source software. Environ Res. (2021) 196:110363. 10.1016/j.envres.2020.110363 [DOI] [PubMed] [Google Scholar]
  • 38.Anjum M, Khan K, Ahmad W, Ahmad A, Amin MN, Nafees A. New SHapley additive ExPlanations (SHAP) approach to evaluate the raw materials interactions of steel-fiber-reinforced concrete. Materials (Basel). (2022) 15(18):6261. 10.3390/ma15186261 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Hu X, Yin S, Zhang X, Menon C, Fang C, Chen Z, et al. Blood pressure stratification using photoplethysmography and light gradient boosting machine. Front Physiol. (2023) 14:1072273. 10.3389/fphys.2023.1072273 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Liao H, Zhang X, Zhao C, Chen Y, Zeng X, Li H. LightGBM: an efficient and accurate method for predicting pregnancy diseases. J Obstet Gynaecol. (2022) 42(4):620–9. 10.1080/01443615.2021.1945006 [DOI] [PubMed] [Google Scholar]
  • 41.Nematollahi MA, Jahangiri S, Asadollahi A, Salimi M, Dehghan A, Mashayekh M, et al. Body composition predicts hypertension using machine learning methods: a cohort study. Sci Rep. (2023) 13(1):6885. 10.1038/s41598-023-34127-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Xue T, Zhu T, Peng W, Guan T, Zhang S, Zheng Y, et al. Clean air actions in China, PM2.5 exposure, and household medical expenditures: a quasi-experimental study. PLoS Med. (2021) 18(1):e1003480. 10.1371/journal.pmed.1003480 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Huo S, Nelde A, Meisel C, Scheibe F, Meisel A, Endres M, et al. A supervised, externally validated machine learning model for artifact and drainage detection in high-resolution intracranial pressure monitoring data. J Neurosurg. (2024) 141(2):509–17. 10.3171/2023.12.JNS231670 [DOI] [PubMed] [Google Scholar]
  • 44.Su X, Bai M. Stochastic gradient boosting frequency-severity model of insurance claims. PLoS One. (2020) 15(8):e0238000. 10.1371/journal.pone.0238000 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Nakapraves S, Warzecha M, Mustoe CL, Srirambhatla V, Florence AJ. Prediction of mefenamic acid crystal shape by random forest classification. Pharm Res. (2022) 39(12):3099–111. 10.1007/s11095-022-03450-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Willis L, Lee E, Reynolds KJ, Klik KA. The theory of planned behavior and the social identity approach: a new look at group processes and social norms in the context of student binge drinking. Eur J Psychol. (2020) 16(3):357–83. 10.5964/ejop.v16i3.1900 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.GBD 2015 Disease and Injury Incidence and Prevalence Collaborators. Global, regional, and national incidence, prevalence, and years lived with disability for 310 diseases and injuries, 1990–2015: a systematic analysis for the global burden of disease study 2015. Lancet. (2016) 388(10053):1545–602. 10.1016/S0140-6736(16)31678-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Ranalli MG, Salvati N, Petrella L, Pantalone F. M-quantile regression shrinkage and selection via the lasso and elastic net to assess the effect of meteorology and traffic on air quality. Biom J. (2023) 65(8):e2100355. 10.1002/bimj.202100355 [DOI] [PubMed] [Google Scholar]
  • 49.Li L, Feng CX, Qiu S. Estimating cross-validatory predictive p-values with integrated importance sampling for disease mapping models. Stat Med. (2017) 36(14):2220–36. 10.1002/sim.7278 [DOI] [PubMed] [Google Scholar]
  • 50.Bell G. Fluctuating selection: the perpetual renewal of adaptation in variable environments. Philos Trans R Soc Lond B Biol Sci. (2010) 365(1537):87–97. 10.1098/rstb.2009.0150 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Heard NA, Holmes CC, Stephens DA, Hand DJ, Dimopoulos G. Bayesian coclustering of anopheles gene expression time series: study of immune defense response to multiple experimental challenges. Proc Natl Acad Sci U S A. (2005) 102(47):16939–44. 10.1073/pnas.0408393102 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Shan G. Monte carlo cross-validation for a study with binary outcome and limited sample size. BMC Med Inform Decis Mak. (2022) 22(1):270. 10.1186/s12911-022-02016-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Li L, Chow SC, Smith W. Cross-validation for linear model with unequal variances in genomic analysis. J Biopharm Stat. (2004) 14(3):723–39. 10.1081/BIP-200025679 [DOI] [PubMed] [Google Scholar]
  • 54.Cao Y, Lin H, Wu TZ, Yu Y. Penalized spline estimation for functional coefficient regression models. Comput Stat Data Anal. (2010) 54(4):891–905. 10.1016/j.csda.2009.09.036 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Slade E, Naylor MG. A fair comparison of tree-based and parametric methods in multiple imputation by chained equations. Stat Med. (2020) 39(8):1156–66. 10.1002/sim.8468 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Meyers RL, Maibach R, Hiyama E, Häberle B, Krailo M, Rangaswami A, et al. Risk-stratified staging in paediatric hepatoblastoma: a unified analysis from the children’s hepatic tumors international collaboration. Lancet Oncol. (2017) 18(1):122–31. 10.1016/S1470-2045(16)30598-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Bloch L, Friedrich CM, Alzheimer’s Disease Neuroimaging Initiative. Data analysis with shapley values for automatic subject selection in Alzheimer’s disease data sets using interpretable machine learning. Alzheimers Res Ther. (2021) 13(1):155. 10.1186/s13195-021-00879-4 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Kim D, Handayani MP, Lee S, Lee J. Feature attribution analysis to quantify the impact of oceanographic and maneuverability factors on vessel shaft power using explainable tree-based model. Sensors (Basel). (2023) 23(3):1072. 10.3390/s23031072 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Yi F, Yang H, Chen D, Qin Y, Han H, Cui J, et al. XGBoost-SHAP-based interpretable diagnostic framework for Alzheimer’s disease. BMC Med Inform Decis Mak. (2023) 23(1):137. 10.1186/s12911-023-02238-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Darst BF, Malecki KC, Engelman CD. Using recursive feature elimination in random forest to account for correlated variables in high dimensional data. BMC Genet. (2018) 19(Suppl 1):65. 10.1186/s12863-018-0633-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Alanni R, Hou J, Azzawi H, Xiang Y. A novel gene selection algorithm for cancer classification using microarray datasets. BMC Med Genomics. (2019) 12(1):10. 10.1186/s12920-018-0447-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 62.Arruda-Olson AM, Reeder GS, Bell MR, Weston SA, Roger VL. Neutrophilia predicts death and heart failure after myocardial infarction: a community-based study. Circ Cardiovasc Qual Outcomes. (2009) 2(6):656–62. 10.1161/CIRCOUTCOMES.108.831024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Han YC, Yang TH, Kim DI, Jin HY, Chung SR, Seo JS, et al. Neutrophil to lymphocyte ratio predicts long-term clinical outcomes in patients with ST-segment elevation myocardial infarction undergoing primary percutaneous coronary intervention. Korean Circ J. (2013) 43(2):93–9. 10.4070/kcj.2013.43.2.93 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Shin HC, Jang JS, Jin HY, Seo JS, Yang TH, Kim DK, et al. Combined use of neutrophil to lymphocyte ratio and C-reactive protein level to predict clinical outcomes in acute myocardial infarction patients undergoing percutaneous coronary intervention. Korean Circ J. (2017) 47(3):383–91. 10.4070/kcj.2016.0327 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Mhurchu CN, Rodgers A, Pan WH, Gu DF, Woodward M, Asia Pacific Cohort Studies Collaboration. Body mass index and cardiovascular disease in the Asia-pacific region: an overview of 33 cohorts involving 310000 participants. Int J Epidemiol. (2004) 33(4):751–8. 10.1093/ije/dyh163 [DOI] [PubMed] [Google Scholar]
  • 66.Mehta L, Devlin W, McCullough PA, O'Neill WW, Skelding KA, Stone GW, et al. Impact of body mass index on outcomes after percutaneous coronary intervention in patients with acute myocardial infarction. Am J Cardiol. (2007) 99(7):906–10. 10.1016/j.amjcard.2006.11.038 [DOI] [PubMed] [Google Scholar]
  • 67.Bucholz EM, Rathore SS, Reid KJ, Jones PG, Chan PS, Rich MW, et al. Body mass index and mortality in acute myocardial infarction patients. Am J Med. (2012) 125(8):796–803. 10.1016/j.amjmed.2012.01.018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.Saberi S, Rezaie B. Robust adaptive direct speed control of PMSG-based airborne wind energy system using FCS-MPC method. ISA Trans. (2022) 131:43–60. 10.1016/j.isatra.2022.04.035 [DOI] [PubMed] [Google Scholar]
  • 69.Premkumar M, Sinha G, Ramasamy MD, Sahu S, Subramanyam CB, Sowmya R, et al. Augmented weighted K-means grey wolf optimizer: an enhanced metaheuristic algorithm for data clustering problems. Sci Rep. (2024) 14(1):5434. 10.1038/s41598-024-55619-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Abou El Ela AA, El-Sehiemy RA, Shaheen AM, Shalaby AS, Mouafi MT. Reliability constrained dynamic generation expansion planning using honey badger algorithm. Sci Rep. (2023) 13(1):16765. 10.1038/s41598-023-43622-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Lakshmi GVN, Jayalaxmi A, Veeramsetty V. Optimal placement of distributed generation based on DISCO’s financial benefit with loss and emission reduction using hybrid Jaya-Red Deer optimizer. Electr Eng (Berl). (2023) 105(2):965–77. 10.1007/s00202-022-01709-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 72.Sayed GI, Soliman MM, Hassanien AE. A novel melanoma prediction model for imbalanced data using optimized SqueezeNet by bald eagle search optimization. Comput Biol Med. (2021) 136:104712. 10.1016/j.compbiomed.2021.104712 [DOI] [PubMed] [Google Scholar]
  • 73.Wang Z, Mo Y, Cui M, Hu J, Lyu Y. An improved golden jackal optimization for multilevel thresholding image segmentation. PLoS One. (2023) 18(5):e0285211. 10.1371/journal.pone.0285211 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Liu W, Liu T, Liu Z, Luo H, Pei H. A novel deep learning ensemble model based on two-stage feature selection and intelligent optimization for water quality prediction. Environ Res. (2023) 224:115560. 10.1016/j.envres.2023.115560 [DOI] [PubMed] [Google Scholar]
  • 75.Jiang S, Yue Y, Chen C, Chen Y, Cao L. A multi-objective optimization problem solving method based on improved golden jackal optimization algorithm and its application. Biomimetics (Basel). (2024) 9(5):270. 10.3390/biomimetics9050270 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 76.Karim FK, Khafaga DS, Eid MM, Towfek SK, Alkahtani HK. A novel bio-inspired optimization algorithm design for wind power engineering applications time-series forecasting. Biomimetics (Basel). (2023) 8(3):321. 10.3390/biomimetics8030321 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.Ruprecht J, Eriksson CE, Forrester TD, Spitz DB, Clark DA, Wisdom MJ, et al. Variable strategies to solve risk-reward tradeoffs in carnivore communities. Proc Natl Acad Sci U S A. (2021) 118(35):e2101614118. 10.1073/pnas.2101614118 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Engebretsen KN, Beckmann JP, Lackey CW, Andreasen A, Schroeder C, Jackson P, et al. Recolonizing carnivores: is cougar predation behaviorally mediated by bears? Ecol Evol. (2021) 11(10):5331–43. 10.1002/ece3.7424 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Iriarte JA, Franklin WL, Johnson WE, Redford KH. Biogeographic variation of food habits and body size of the America puma. Oecologia. (1990) 85(2):185–90. 10.1007/BF00319400 [DOI] [PubMed] [Google Scholar]
  • 80.Wu ZS, Fu WP, Xue R. Nonlinear inertia weighted teaching-learning-based optimization for solving global optimization problem. Comput Intell Neurosci. (2015) 2015:292576. 10.1155/2015/292576 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 81.Walling A. From cutie to cougar. Fam Med. (2022) 54(4):306–7. 10.22454/FamMed.2022.443767 [DOI] [PubMed] [Google Scholar]
  • 82.Spicher N, Kukuk M. Delineation of electrocardiograms using multiscale parameter estimation. IEEE J Biomed Health Inform. (2020) 24(8):2216–29. 10.1109/JBHI.2019.2963786 [DOI] [PubMed] [Google Scholar]
  • 83.Lang Z, Wen QH, Yu B, Sang L, Wang Y. Forecast of winter precipitation type based on machine learning method. Entropy (Basel). (2023) 25(1):138. 10.3390/e25010138 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84.Shakhgeldyan KI, Kuksin NS, Domzhalov IG, Rublev VY, Geltser BI. Interpretable machine learning for in-hospital mortality risk prediction in patients with ST-elevation myocardial infarction after percutaneous coronary interventions. Comput Biol Med. (2024) 170:107953. 10.1016/j.compbiomed.2024.107953 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Datasheet1.pdf (53.9KB, pdf)

Data Availability Statement

The original contributions presented in the study are included in the article/Supplementary Material, further inquiries can be directed to the corresponding author.


Articles from Frontiers in Cardiovascular Medicine are provided here courtesy of Frontiers Media SA

RESOURCES