Skip to main content
JMIR Medical Informatics logoLink to JMIR Medical Informatics
. 2019 Oct 23;7(4):e14993. doi: 10.2196/14993

Exploiting Machine Learning Algorithms and Methods for the Prediction of Agitated Delirium After Cardiac Surgery: Models Development and Validation Study

Hani Nabeel Mufti 1,2,3,, Gregory Marshal Hirsch 4, Samina Raza Abidi 5, Syed Sibte Raza Abidi 6
Editor: Gunther Eysenbach
Reviewed by: Anis Davoudi, Darlinton Carvalho, Didi Surian, Balaji Polepalli Ramesh
PMCID: PMC6913743  PMID: 31558433

Abstract

Background

Delirium is a temporary mental disorder that occasionally affects patients undergoing surgery, especially cardiac surgery. It is strongly associated with major adverse events, which in turn leads to increased cost and poor outcomes (eg, need for nursing home due to cognitive impairment, stroke, and death). The ability to foresee patients at risk of delirium will guide the timely initiation of multimodal preventive interventions, which will aid in reducing the burden and negative consequences associated with delirium. Several studies have focused on the prediction of delirium. However, the number of studies in cardiac surgical patients that have used machine learning methods is very limited.

Objective

This study aimed to explore the application of several machine learning predictive models that can pre-emptively predict delirium in patients undergoing cardiac surgery and compare their performance.

Methods

We investigated a number of machine learning methods to develop models that can predict delirium after cardiac surgery. A clinical dataset comprising over 5000 actual patients who underwent cardiac surgery in a single center was used to develop the models using logistic regression, artificial neural networks (ANN), support vector machines (SVM), Bayesian belief networks (BBN), naïve Bayesian, random forest, and decision trees.

Results

Only 507 out of 5584 patients (11.4%) developed delirium. We addressed the underlying class imbalance, using random undersampling, in the training dataset. The final prediction performance was validated on a separate test dataset. Owing to the target class imbalance, several measures were used to evaluate algorithm’s performance for the delirium class on the test dataset. Out of the selected algorithms, the SVM algorithm had the best F1 score for positive cases, kappa, and positive predictive value (40.2%, 29.3%, and 29.7%, respectively) with a P=.01, .03, .02, respectively. The ANN had the best receiver-operator area-under the curve (78.2%; P=.03). The BBN had the best precision-recall area-under the curve for detecting positive cases (30.4%; P=.03).

Conclusions

Although delirium is inherently complex, preventive measures to mitigate its negative effect can be applied proactively if patients at risk are prospectively identified. Our results highlight 2 important points: (1) addressing class imbalance on the training dataset will augment machine learning model’s performance in identifying patients likely to develop postoperative delirium, and (2) as the prediction of postoperative delirium is difficult because it is multifactorial and has complex pathophysiology, applying machine learning methods (complex or simple) may improve the prediction by revealing hidden patterns, which will lead to cost reduction by prevention of complications and will optimize patients’ outcomes.

Keywords: delirium, cardiac surgery, machine learning, predictive modeling

Introduction

Background

Delirium or acute confusion is a temporary mental disorder that occurs among hospitalized patients [1]. The Society of Thoracic Surgeons defines delirium as a mental disturbance marked by illness, confusion, and cerebral excitement, with a comparatively short course [2], developing over a short period (usually from hours to days) and which tends to fluctuate during the day [3]. Delirium symptoms range from a disturbance in consciousness (eg, coma) to cognitive disorders involving disorientation and hallucinations. Delirium has a wide range of presentations, from extremely dangerous agitation to depression-like isolation and, on the basis of its presentation, it has 3 distinct subclasses—that is, hyperactive, hypoactive, and mixed [4]. This diversity of possible presentations, along with its sudden onset and unpredictable course, makes early detection challenging. Royston and Cox state that “from the patient’s point of view, delirium and subsequent cognitive decline are among the most feared adverse events following surgery” [5]. The diversity of delirium’s presentation, along with its sudden onset and unpredictable course, makes its early detection difficult; however, the ability to predict delirium in patients can play a fundamental role in initiating preventive measures that can significantly improve outcomes.

Patients undergoing cardiac surgery are at higher risk of developing delirium [6-9]. Several studies demonstrated a negative association between postoperative delirium and an increased morbidity and mortality [7-10]. Of particular concern is the strong relationship between delirium and postoperative infections in cardiac surgery patients [7,9,11].

Given the undesirable consequences of delirium on surgical outcomes, it is deemed useful to predict the potential incidence of delirium in patients to pre-emptively administer and plan for therapeutic interventions to deal with delirium and in turn improve the surgical outcomes. Typically, predictive models for delirium use a range of clinical variables, applied to conventional statistical methods, mainly logistic regression (LR) [12-14]. The current predictive models for delirium generally present a simplified linear weighted representation of the statistical significance of the clinical variables toward the prediction of delirium [15].

However, we argue that the prediction of delirium is quite complex given the multiplicity of reasons and confounding factors contributing to the manifestation of delirium in patients. Data mining methods can be used to uncover underlying relationships between variables to develop predictive models that can categorize the patient population into ones that have the propensity to develop delirium versus those that are less likely to develop delirium. Sometimes, these relationships or patterns cannot be easily explained yet appear to be essential and have a significant contribution to the improvement of the predictive model’s performance, even if it is minimal (eg, a 0.01% improvement in a model’s performance means that for every 1000 patients, 1 extra life is saved or a complication is prevented or an accident is avoided).

Artificial intelligence in health care, particularly the use of machine learning methods, provides a purposeful opportunity to discover such underlying patterns and correlations by mining the data leading to the learning of data-driven prediction models. Machine learning models have been successfully applied in medical data [16-22] to solve a wide range of clinical issues, such as myocardial infarction [23], atrial fibrillation [24], trauma [25], breast cancer [26-28], Alzheimer [29-31], cardiac surgery [22,32], and others [20,21,33-35].

The main objective of this study was to develop predictive models to pre-emptively predict the manifestation of agitated delirium in patients after cardiac surgery. Although discovering underlying hidden patterns is interesting and can be done using the data mining methods used in this work, this was not our main objective as the pathophysiology of delirium is considered multifactorial and complex to start with. The rationale is that if we can identify based on preoperative clinical parameters which patients are likely to develop postoperative delirium, then clinicians can initiate preventive and therapeutic measures in a timely fashion, to mitigate the undesirable effects of delirium. Our approach for predictive modeling is to investigate machine learning methods to learn the prediction models using retrospective clinical data for around 5500 patients over a 7-year period who received cardiac surgery at Queen Elizabeth II Health Sciences Center (QEII HSC) in Halifax, Canada. In this paper, several machine learning models were explored, including artificial neural networks (ANN), Bayesian belief networks (BBN), decision trees (DT), naïve Bayesian (NB), LR, random forest (RF), and support vector machines (SVM).

Related Work

Although the prevalence of postoperative delirium is low (10%-25%), it is associated with cognitive deterioration coupled with a set of complications in surgical patients. The complexity of delirium comes from its relation to multiple risk factors and the accompanying uncertainty of its pathophysiology [10,11,36]; this leads to challenges in pre-emptively identifying patients that are likely to develop postoperative delirium. Several authors have indicated that delirium is associated with adverse outcomes and advocate early recognition to ensure preventive measures can be applied in a timely and effective manner [3,7,9,10,13,14,37]. Some of the proposed preventive interventions that have been shown to reduce the incidence of delirium in high-risk patients include early mobilization and use of patient’s personal aids (reading glasses, hearing aid, etc) [38]. However, the pre-emptive identification of postoperative delirium is clinically challenging.

A structured PubMed search using the PubMed Advanced Search Builder with the structure (“delirium”) AND “predictive model”, will result in only 38 items. If we direct our attention to all the research published focusing on delirium and cardiac surgery, query structure (“delirium”) AND “cardiac surgery”, we will get 485 items. If we combine all the 3 terms, query structure ((“delirium”) AND “cardiac surgery”) AND “predictive model”, we will narrow the results down to 4 items.

In recognition of the importance of delirium within the cardiac surgical population, some have attempted to develop a predictive model. In this work, we decided to focus on articles that were published in English and focused on developing a predictive model for the prediction of delirium after cardiac surgery in adult patients. The initial search resulted in 38 articles. After reviewing the articles’ abstracts, we excluded articles that were not written in English, not about cardiac surgery patients, and in which no statistical model was developed. We ended up with 16 articles that were available for review. Multimedia Appendix 1 represents a summary of most relevant studies that attempted to develop a model for the prediction of delirium after cardiac surgery on adult patients.

For patients who underwent cardiac surgery, Afonso et al [12] conducted a prospective observational study on 112 consecutive adult cardiac surgical patients. Patients were evaluated twice daily for delirium using Richmond Agitation-Sedation Scale (RASS) and confusion assessment method for the intensive care unit (CAM-ICU), and the overall incidence of delirium was 34%. Increased age and the surgical procedure duration were found to be independently associated with postoperative delirium. Similarly, Bakker et al [13] prospectively enrolled 201 cardiac surgery patients aged 70 years and above. They found that a low Mini-Mental State Exam score and a higher preoperative creatinine were independent predictors of postoperative delirium [13]. Unfortunately, both of these models were based on a small sample size (<250 patients) and did not have a validation cohort.

Research in the use of machine learning–based prediction models to detect delirium is rather limited, especially for cardiac surgery. Kramer et al [39] developed predictive models using a large dataset comprising medical and geriatrics patients that had the diagnosis of delirium in their discharge code and a control group of randomly selected patients from the same period who did not develop delirium. The prediction models performed well with the highest performance achieved by the RF model (receiver operating characteristic-area under the curve [ROC-AUC]≈91%). Although they argue that their data were imbalanced, they used the ROC-AUC as their evaluation metric, which does not consider the class imbalance. Davoudi et al [40] applied 7 different machine learning methods on data extracted from the electronic health (eHealth) record of patients undergoing major surgery in a large tertiary medical center to predict delirium; they found an incidence of 3.1%. They were able to achieve a ROC-AUC ranging from 71% to 86%. Owing to the class imbalance secondary to the low incidence of delirium and to improve the model’s performance, they applied data-level manipulation using over- and undersampling, which did not result in a significant improvement (ROC-AUC ranging from 79% to 86%). Lee et al [41] published a nice systematic review and identified 3 high-quality ICU delirium risk prediction models: the Katznelson model, the original PRE-DELIRIC (PREdiction of DELIRium in ICu patients), and the international recalibrated PRE-DELIRIC model. All of these models used LR modeling as the primary technique for creating the predictive model. In the same paper by Lee et al [41], they externally validated these models on a prospective cohort of 600 adult patients that underwent cardiac surgery in a single institution. After updating, recalibrating, and applying decision curve analysis (DCA) to the models, they concluded that the recalibrated PRE-DELIRIC risk model is slightly more helpful. They argue that available models of predicting delirium after cardiac surgery have only modest accuracy. The current models are suboptimal for routine clinical use. Corradi et al [42] developed a predictive model using a large dataset (~78,000 patients) over 3 years in a single center using a good number of feature set (~128 variables). Their model had very good accuracy and the ROC-AUC ~90% on their test dataset. They used the CAM to detect delirium in the intensive care (CAM-ICU) and regular patient wards. Lee et al [41] conducted a systematic review in search for prediction models for delirium specifically designed for cardiac surgery patients. They found only 3 high-quality models and externally validated them on a local population of 600 patients. They used several metrics to evaluate the recalibrated models on the validation cohort (ROC-AUC, Hosmer–Lemeshow test, Nagelkerke’s R2, Brier score, and DCA). In their analysis, the recalibrated PRE-DELIRIC prediction model performed better when compared with the Katznelson model. However, based on the DCA and the expected net benefit of both models, there appears to be limited clinical utility of any of the models.

Methods

Data Sources and Study Population

This single-center retrospective cohort study included patients who underwent cardiac surgery at the QEII HSC in Halifax, Canada, between January 2006 and December 2012. Over those 7 years, 7209 patients underwent cardiac surgery. The Maritime Heart Center (MHC) registry was used to create the dataset. The MHC registry is a prospectively collected, detailed clinical database on all cardiac surgical cases performed at the MHC since March 1995 with more than 20,000 patients and 500 different variables. The final dataset included 5584 patients who met our inclusion criteria and were successfully discharged (home, other institution closer to home, nursing home, or rehabilitation facility).

Delirium in the acquired database is coded as a binary outcome (Yes/No) and is defined as short-lived mental disturbance marked by illusions, confusion, or cerebral excitement, requiring temporary medical and/or physical intervention or a consultation, or extending the patient’s hospital stay. Intraoperative management varied depending on the anesthetist preferences and the patient clinical status. Although most patients were managed in a systematic approach based on standard of care, in the ICU, CAM-ICU was used to trigger further investigations if delirium was suspected. If delirium had been suspected after transfer from the ICU, the diagnosis was confirmed using different diagnostic criteria and screening tools (eg, Mini-Mental State Exam and CAM).

Full ethics approval was obtained from the Capital Health Research Ethics Board, in keeping with the Tri-Council Policy Statement: Ethical Conduct for Research Involving Humans. Informed consent was waved by the ethics board as the study did not involve therapeutic interventions or potential risks to the involved subjects.

Predictive Modeling: Methodology and Methods

Our aim was to develop a prediction model that can identify patients who are at risk of developing delirium after cardiac surgery. We investigated relevant machine learning methods, each with a specific learning algorithm to correlate the patient presurgery variables with a probabilistic determination of delirium as per the observations noted in the cardiac surgery dataset. The rationale for working with multiple machine learning methods was to determine the effectiveness of the different methods and then to select the best performing model that can be used in a clinical setting to predict postoperative delirium in new patients.

We pursued the standard data mining methodology comprising 6 steps as shown in Multimedia Appendix 2. These steps are as follows: (1) data acquisition: This step involved the procurement of the required dataset from the source (in this case from the MHC), while complying with data access and secondary data usage protocols; (2) data preprocessing: This step involved the cleaning of the dataset by removing incomplete records and next identifying the significant features/variables to develop the prediction models; (3) modeling strategy set-up: This step involved the formulation of the modeling strategy in terms of data partitioning into training dataset (N=4476; 80% of original) and test dataset (N=1117; 20% of original), data presentation during training, model evaluation criteria; (4) class imbalance and training dataset class optimization: This step was introduced to address the target class imbalance in the original dataset, so as to minimize the effect of the dominant class on the performance of the predictive models. We explored data level techniques, such as over- and undersampling, to address the class imbalance in the final training dataset, resulting in the final balanced training dataset (n=1014). (5) model learning: This step involved setting up different model configurations—that is, setting up the model parameters for the candidate machine learning methods—and learning the models by presenting the preprocessed training data (step 2) as per the modeling strategy (step 3). As model learning is an exploratory exercise where different model configurations and multiple instantiations of the model are pursued to account for the probabilistic nature of machine learning methods and to avoid overfitting, 10-fold stratified cross-validation was used; and (6) model evaluation: In this step, the learnt models are evaluated (against the predefined criteria) for their effectiveness to predict delirium using the test data.

Data Preprocessing and Variables Selection

Characteristics of patients who developed delirium postoperatively were compared with patients who did not. The mean and standard deviation were used for continuous variables that had a normal distribution and were compared using the 2-sided t test. Continuous variables that were not normally distributed were reported using the median and interquartile range and were compared using the Wilcoxon rank sum test. Categorical variables were reported as frequencies and percentages and were analyzed by 2 (Chi-square) or Fisher exact test as appropriate. The Kruskal-Wallis test was used for ordinal variables. Next, exploratory data analysis followed by univariate LR analysis was applied to isolate key perioperative variables with significant influence on postoperative agitated delirium.

All measures of significance are 2-tailed, and a P value <.05 was considered statistically significant. Statistical analysis and the assessment of model’s performance was conducted using the R-Software, version 3.1.0 (R Project for Statistical Computing) [43]. On the basis of univariate LR analysis, 22 variables were used to generate the machine learning–based predictive models.

The basic premise of any DT model is that it recursively split features based on the target variable’s purity. The ultimate goal of the algorithm is to optimize each split on maximizing the homogeneity of the grouping at each split (also known as purity) [44-46]. A node having multiple classes is impure, whereas a node having only 1 class is pure. One of the useful features of RF is its ability to identify relevant variables by assigning variable importance measure to the input variables [44-46]. Variable importance in RF can be measured using either misclassification error, Gini index, or cross-entropy. Most machine learning experts discourage the use of misclassification error in tree-based models because it is not differentiable and, hence, less amenable to numerical optimization [44,46]. In addition, cross-entropy and the Gini index are more sensitive to changes in the node probabilities than the misclassification rate. Both Gini index and cross-entropy apply probability to gauge the disorder of grouping by the target variable. However, they are a bit different, and the results can vary. The Gini index measures how often a randomly chosen element from the set would be incorrectly labeled, starting with the assumption that the node is impure (Gini index=1) and subtracting the probabilities of the target variable. If the node is composed of a single class (also known as pure), then the Gini index will be 0. On the other hand, cross-entropy is more computationally heavy because of the log in the equation. Instead of utilizing simple probabilities, this method takes the log of the probabilities (usually the log base 2; any log base can be used, but it has to be consistent for the sake of comparison between different tree-based models). The entropy equation uses logarithms because of many advantageous properties (mainly the additive property) that can be very beneficial in imbalanced class distributions and multiclass target variables [44,46]. A cross-entropy of 1 indicates a highly disorganized node (impure node), whereas a cross-entropy of 0 indicates a highly organized node (pure node).

In RF, each tree in the forest is grown fully (unpruned) using bootstrap samples of the original dataset, the out-of-bag (OOB) samples are used as test samples. A random subset of variables k from the original input variables space K (where k<K) is used at each node. On the basis of a specific measure (eg, mean decrease in impurity, Gini index, and mean decrease in accuracy), variables are selected, and the process is repeated to the end of the tree. The performance of each tree is computed over the corresponding OOB sample. For each variable, its importance is calculated as the mean relative decrease across the forest of trees performance when the observations of this variable in the OOB sample are randomly permuted. As the Waikato Environment for Knowledge Acquisition software (WEKA) was used in this work to develop the RF model, it applies the cross-entropy method as its default method for variables importance ranking.

The Issue of Outcome Class Imbalance

In our dataset, the outcome class distribution is notably imbalanced (only 11.4% of patients developed delirium). Typically, classification algorithms tend to predict the majority class very well but perform poorly on the minority class due to 3 main reasons [47-49]: (1) the goal of minimizing the overall error (maximize accuracy), to which the minority class contributes very little; (2) algorithm’s assumption that classes are balanced; and (3) the assumption that impact of making an error is equal.

Several data manipulation techniques can be applied to reduce the impact of this class imbalance: at the data level (oversampling minority class or undersampling the majority class) or at the algorithm level (applying different costs to each class) [47-49]. Although data manipulation methods can improve a model’s performance, these methods do have some drawbacks [49]. At the data-level manipulation, oversampling tends to artificially increase the number of the minority class by creating modified copies; it tends to overfit the results to the training set and consequently is likely to poorly generalize. On the other hand, because undersampling discards some of the majority class observations, it essentially bears the risk of losing some potentially important hidden information. Algorithm level manipulation involves some trial and error and can be sensitive to training data changes.

In real life, class imbalance cannot be avoided as it is a result of the nature of the problem and domain (eg, natural disasters and patient death). In our dataset, oversampling led to overfitting on the training dataset with suboptimal generalization when applied to the imbalanced dataset. As postoperative delirium is linked with a wide range of complications (from a minor temporary confusion that totally resolves with no sequalae to the other extreme of sepsis and death), it is very hard to associate it to a specific cost. As such, given the intent of this study, we decided to apply random subsampling to balance the training dataset and have equal representation of outcome classes, thus optimizing the training dataset for the models. We used the SpreadSubSample filter in WEKA [46] to produce a random subsample by undersampling the majority class (which can be done by either specifying a ratio or the number of observations). In our case, we specified a ratio of 1:1. By doing so, the filter generates a new balanced dataset by decreasing the number of the majority class instances, which reduces the difference between the minority and the majority classes. Undersampling is considered an effective method for dealing with class imbalance [50]. In this approach, a subset of the majority class is used to learn the model. Many of the majority class examples are ignored; the training set becomes more balanced, which makes the training more efficient. The most common type of undersampling is random majority undersampling (RUS). In RUS, observations from the majority class are randomly removed. The final balanced training dataset (N=1014, 1:1 delirium) was used to develop the models.

Training With 10-Fold Cross-Validation and Test Datasets

In predictive modeling, it is a common practice to separate the data into training and test dataset. In an effort to avoid overfitting and overestimating the model’s performance, the test dataset is only used to evaluate the performance of the prediction model [44,46,51,52]. The problem of evaluating the model on the training dataset is that it may exhibit high prediction ability (overfitting), yet it fails when asked to predict new observations. To address this issue, cross-validation is commonly used to (1) estimate the generalizability of an algorithm and (2) optimize the algorithm performance by adjusting the parameters [44,46,51-53]. We applied stratified 10-fold cross-validation on the balanced training dataset (50% delirium). The test dataset was preserved imbalanced to simulate the real clinical scenario and evaluate the behavior of different methods. Several metrics were used, that are immune to class imbalance, to appraise the final model’s performance on the test dataset [44,46,47,49,51,52].

Results

Development of Prediction Models: Experiments and Results

We investigated a range of relevant predictive modeling methods—that is, function-based models (LR, ANN, and SVM), Bayesian models (NB and BBN), and tree-based models (C4.5 DT and RF)—to generate 7 prediction models (all developed using the same balanced dataset). All models were generated and tested using the WEKA software, version 3.7.10 [54]. The setting of the prediction models and the optimization steps that were applied in this research are available in Multimedia Appendix 3. These predictive modeling algorithms were chosen based on 2 main reasons: (1) their noted effectiveness in solving medical-related classification problems and (2) a strong theoretical background that supports predictive modeling via data classification [11,16,19,20,22,23,25,30,31,39,46,52,55-63]. Experiments were conducted on a MacBook Pro (Apple Inc; 15-inch, 2017) with a 3.1-GHz Intel Core i7 processor and a 16 GB RAM 2133 MHz, running a MacOS High Sierra Version 10.13.

General Patients’ Characteristics and Important Variables in the Dataset

Given the above definitions and procedures, agitated delirium was documented in 11.4% patients (n=661). The majority of patients were men (74%). Coronary artery bypass graft (CABG) was the most commonly performed procedure (67%). Almost 56% stayed in the ICU for 24 hours or less. Only 2% suffered a permanent stroke. Patients who developed postoperative agitated delirium were older and had a significantly higher incidence of comorbid diseases. A higher proportion of patients who developed agitated delirium underwent a combined procedure (CABG plus valve). The median stay in the cardiovascular intensive care unit in hours was 4 times higher for patients who developed agitated delirium postoperatively, compared with patients who did not (P<.001). Univariate analysis of in-hospital mortality did not show any statistical significance (in-hospital mortality: 4.1% vs 3.6%; P=.57; Table 1).

Table 1.

Patient characteristics (N=4467).

Patient characteristics Delirium P value

No (n=3960) Yes (n=507)
Preoperative characteristics
  Age (years)     <.001
    Mean (SD) 66 (11) 72 (10)
    Range 19-95 25-91
  Male gender, n (%) 2942 (74.3) 386 (76.1) .36
  Hypertension, n (%) 2970 (75) 401 (79) .04
  Diabetes mellitus, n (%) 1426 (36) 223 (44) <.001
  Cerebrovascular disease, n (%) 436 (11) 112 (22) <.001
  Chronic obstructive pulmonary disease, n (%) 531 (13.4) 104 (20.5) <.001
  Frail, n (%) 238 (6) 49 (9.7) .002
  Ejection fraction <30%, n (%) 436 (11) 106 (21) <.001
  Preoperative atrial fibrillation, n (%) 424 (10.7) 102 (20.1) <.001
  EURO IIa score >5%, n (%) 717 (18.1) 231 (45.6) <.001
  Urgency, n (%)

<.001
    Elective (admitted from home) 1901 (48) 198 (39)
    Need surgery during hospitalization 1742 (44) 223 (44)
    Urgent/emergent (life threatening) 317 (8) 91 (18)
Intraoperative characteristics, n (%) <.001
  Procedure


    Coronary artery bypass graft 2744 (69.3) 291 (57.4)
    Aortic valve replacement 622 (15.7) 93 (18.3)
    Mitral valve surgeryb 170 (4.3) 20 (4)
    CABG+AVRc 325 (8.2) 79 (15.6)
    CABG+MVd surgery 51 (1.3) 3.4 (17)
  Repeat sternotomy 230 (5.8) 59 (11.6) <.001
In-hospital morbidity, n (%) <.001
  Reintubation 79 (2) 48 (9.5)
  New postoperative atrial fibrillation 1247 (31.5) 217 (42.8)
  Pneumonia 174 (4.4) 101 (20)
  Sepsis 40 (1) 35 (6.9)
  Deep sternal wound infection 24 (0.6) 15 (3)
  Blood products transfusion within 48 hours from surgery 990 (25) 269 (53)
  Length of stay after surgery <1 week 2257 (57) 66 (13)
  Discharged home 3513 (88.7) 301 (59.4)

aEURO II: European System for Cardiac Operative Risk Evaluation II.

bMitral valve replacement or repair.

cCABG+AVR: coronary artery bypass graft + aortic valve replacement.

dCABG+MV: coronary artery bypass graft + mitral valve.

Univariate LR analysis of all pre-, intra-, and postoperative variables that can contribute to the development of delirium was performed using appropriate statistical tests in the R-Software. Univariate LR was applied on all candidate variables with a P value of less than .05 in univariate LR analysis to extract odds ratio (OR) with 95% CI generated for each candidate variable. The candidate variables were ranked based on the how low is the actual P value, the Akaike information criterion (lower is better), and impact of variable on postoperative delirium (signified by the OR). Then WEKA was used to generate variable importance using the RF model. WEKA applies the cross-entropy method to assess purity of the candidate variables with the RF algorithm as its default method, as it is more sensitive to class imbalance. Variables that appear higher at the trees are considered more relevant [44,51,52,63]. This is represented by the percentage of decrease of impurity (or increase of purity) of the final model based on adding this specific attribute. The number of times the candidate variable appeared in any location in all of the created tree models through the RF ensemble model process is also a criterion used in WEKA. The more times a variable is being selected in the RF creation process, the higher likelihood of it being important for the classification of the final target variable. This is also reflected in the decrease of impurity measure as the more decrease in impurity, the higher number of times that variable appears, which can imply its importance. Table 2 displays the importance of each input variable used in our RF model and its rank compared with the univariate LR analysis.

Table 2.

List of candidate variables based on univariate logistic regression analysis compared with random forest.

Variable Type Unit Univariate logistic regression analysisa Random forest



OR (95% CI) P value Rank Decrease of impurity, % Nodes using that attribute, n Rank
Age (years) Continuous Years 1.1 (1.03-1.07) <.001 1b 43 3238 1
Mechanical ventilation >24 hours Categorical Yes/no 5.8 (3.9-8.6) <.001 3 21 297 21
Preoperative creatinine clearance Continuous μmol/L 0.97 (0.96-0.98) <.001 1b 39 2544 4
Length of stay in the ICUc Ordinal d 2 26 590 20

>72 hours 7.6 (4.9-11.9) <.001

24-72 hours 1.7 (0.9-2.8) <.001
Procedure other than isolated CABGe Categorical Yes/no 2.9 (1.8-2.5) <.001 6 28 370 15
Blood product within 48 hours Categorical Yes/no 2.9 (2.0-4.2) <.001 5 28 452 14
Intraoperative TEEf Categorical Yes/No 2.0 (1.3-3.1) .002 10 27 568 18
EURO IIg score Continuous Percent 1.07 (1.05-1.09) <.001 17 41 2716 2
Preoperative hemoglobin Continuous gm/dL 0.98 (0.97-0.99) <.001 1b 40 2766 3
Preoperative A-Fibh Categorical Yes/no 2.3 (1.4-3.6) <.001 7 35 486 6
Timing of IABPi Ordinal 4 29 329 12

Preoperative 1.4 (0.6-2.9) .42

Intraoperative 6.8 (1.9-23.1) .002
Intraoperative inotropes Categorical Yes/no 2.1 (1.4-3.0) <.001 8 27 514 17
COPDj Categorical Yes/no 1.7 (1.1-2.7) .02 14 33 689 9
CVDk Categorical Yes/no 1.8 (1.1-2.9) .01 13 29 516 13
DMl Categorical Yes/no 0.9 (0.6-1.4) .79 16 39 995 5
Frail Categorical Yes/no 2.0 (1.1-3.5) .03 12 30 381 11
History of turn down Categorical Yes/no 8.2 (2.8-24.3) <.001 1b 21 93 22
EFm categories Ordinal 9 33 89 10

30%-50% 1.4 (0.9-2.1) .18

<30% 2.1 (1-4.2) .04
Gender Categorical Yes/no 1.2 (0.8-1.9) .47 16 35 752 7
Aortic stenosis Ordinal 14 26 899 16

Moderate 1.4 (0.6-2.8) .43

Severe 1.6 (1.1-2.5) .01  
Mitral insufficiency Ordinal 15 26 899 19

Moderate 1.4 (0.9-2.1) .07

Severe 2.3 (1.02-4.6) .03
Postoperative arrhythmias Categorical Yes/no 1.8 (1.3-2.8) .002 11 34 746 8

aAnalysis was done using univariate logistic regression with a P value of <.05 considered to be statistically significant.

bThese variables were all equally ranked as 1st because they had almost equal odds ratios and a P value of <.001.

cICU: intensive care unit.

dNot applicable.

eCABG: coronary artery bypass graft.

fTEE: transesophageal echo.

gEURO II: European System for Cardiac Operative Risk Evaluation II.

hA-Fib: atrial fibrillation.

iIABP: intra-aortic balloon pump.

jCOPD: chronic obstructive pulmonary disease.

kCVD: cerebrovascular disease.

lDM: diabetes mellitus.

mEF: ejection fraction.

Prediction Model’s Performance Evaluation

There exist several metrics to evaluate the performance of a predictive model, whereby predictive accuracy is the most commonly used metric as it relates a model’s ability to correctly identify observation assignments, irrespective of the class distribution. However, in the presence of a noted class imbalance in the dataset, this measure can be misleading because the minority class (positive cases in our dataset) has a smaller influence of the model’s output, and as such the model will tend to favor the majority class [47]. In our dataset, there is a significant imbalance of the outcome of interest distribution (delirium: 11.4% positive cases).

To provide a more robust evaluation of the prediction model’s performance, in the presence of the class imbalance in our dataset, we used the evaluation measures of F1 measure, ROC-AUC, and precision-recall curve area under the curve (PRC-AUC) [44,46,47,51,52]. The ROC-AUC was primarily used to assess the classifier’s general performance (model discrimination=how well the predicted risks distinguish between patients with and without disease) [64]. The F1 score was primarily used as the harmonic mean of precision and recall [46,52]. The F1 score provides the most reliable assessment of a model’s prediction performance, while considering the worst-case prediction scenario for a classifier (model calibration=evaluates the reliability of the estimated risks: if we predict 10%, on average 10/100 patients should have the disease) [64].

Sensitivity (recall) is considered a measure of completeness (the percentage of positive cases that have been correctly identified as positive). Positive predictive value (precision, PPV) is considered a measure of exactness (the percentage of cases labeled by the classifier as positive that are indeed positive) [46,52]. The PRC-AUC is a useful measure in the presence of class imbalance, and the outcome of interest is to identify the minority class [65,66]. The PRC identifies the PPV for each corresponding value on the sensitivity scale (model calibration). As the PRC is dependent on the class representation in the dataset, it provides a simple visual representation of the model’s performance across the whole spectrum of sensitivities. By doing so, it can aid in identifying the best model (based on the trade of being either exact vs complete, ideally optimizing both) [66]. In addition, the PRC enables comparing models at predetermined recall thresholds (eg, the best precision at 50% recall). This adds more fixability in choosing the best model based on the domain and problem in hand.

As our primary interest was to identify patients who were more likely to develop delirium (minority class) while accounting for the class imbalance in the test dataset, we decided to evaluate the models using the ROC-AUC as a measure of the model discrimination in conjunction with F1 score and PRC-AUC as measures of the model calibration. Tables 3 and 4 present the prediction performance of all prediction models based on the test data. Figure 1 illustrates the ROC-AUCs and PRC-AUCs for the developed models.

Table 3.

Comparison of model’s performance metrics applied on the balanced training dataset using 10-fold cross-validation and the imbalanced test dataset to predict delirium after cardiac surgery. Performance metrics: accuracy, sensitivity, specificity, positive predictive value, negative predictive value, and Cohen kappa. All measures are reported out of 100% with standard deviation in brackets as a measure of variability.

Model Accuracy Δa Sensitivity Δ Specificity Δ PPVb Δ NPVc Δ Kappa Δ
Dataset: 10-fold cross-validation applied on the balanced training dataset (N=1014, delirium=50%)

ANNd 71.7 (4.3) nse 71.8 (7) +f 71.6 (7) g 71.7 (5) 71.7 (7) 43.3 (9) ns

BBNh 71.3 (4.4) ns 72.2 (7) + 71.2 (7) 69.9 (5) 71.3 (7) 43.1 (9) ns

DTi 70.1 (4.3) ns 68.1 (7) ns 72.9 (9) ns 72.9 (5) ns 72.6 (9) ns 43.3 (8) ns

LRj 73.3 (4.4) Bk 69.8 (7) B 76.7 (7) B 75 (5) B 75.6 (6) B 44.5 (9) B

NBl 73.0 (4.2) ns 64.8 (7) ns 79.5 (5) + 74.4 (5) ns 79.5 (5) + 42.9 (8) ns

RFm 72.5 (4.4) ns 74.3 (7) + 71.7 (7) 72.1 (4) ns 72.8 (7) 45.7 (9) ns

SVMn 71.3 (4.5) ns 60.2 (8) 83.8 (5) + 77.8 (5) + 83.1 (5) + 43.2 (9) ns
Dataset: Imbalanced test dataset (N=1117, delirium=11.4%)

ANN 74.3 (3.2) ns 67.7 (5) + 72.9 (5) ns 24.3 (14) ns 94.6 (5) ns 22.85 (9) ns

BBN 74.1 (3.8) ns 68.7 (9) + 70.8 (9) 22.9 (15) ns 94.5 (6) ns 21.81 (11) ns

DT 74.4 (5.4) ns 66.9 (10) + 75.4 (10) ns 25.8 (17) ns 94.7 (10) ns 24.97 (13) ns

LR 75.6 (4.7) B 64.6 (9) B 77.1 (7) B 26.5 (16) B 94.4 (8) B 22.6 (13) B

NB 71.7 (3.1) 66.1 (12) ns 72.4 (8) 23.5 (18) ns 94.3 (9) ns 21.55 (10) ns

RF 75.4 (3.4) ns 72.4 (4) + 72.4 (4) 25.2 (8) + 95.3 (4) + 24.69 (7) ns

SVM 78.9 (2.1) + 62.2 (4) ns 81.1(3.2) + 29.7 (12) + 94.4 (6) ns 29.33 (9) +

aChange compared to base model (B).

bPPV: positive predictive value.

cNPV: negative predictive value.

dANN: artificial neural networks.

ens: not a statistically significant change in performance (P≥.05).

fStatistically significant improvement of performance metric (P<.05).

gStatistically significant deterioration of performance metric (P<.05).

hBBN: Bayesian belief networks.

iDT: J48 decision tree.

jLR: logistic regression.

kB: base comparator (reference) algorithm.

lNB: naïve Bayesian.

mRF: random forest.

nSVM: support vector machines.

Table 4.

Comparison of model’s performance metrics applied on the balanced training dataset using 10-fold cross-validation and the imbalanced test dataset to predict delirium after cardiac surgery. Performance metrics: receiver operator curve-area under the curve, harmonic mean of precision and recall, and precision-recall curve-area under the curve. All measures are reported out of 100% with standard deviation in brackets as a measure of variability.

Model ROC-AUCa F1 scoreb PRC-AUCc


Yesd Δe Nof Δ Avgg Δ Yes Δ No Δ Avg Δ
Dataset: 10-fold cross-validation applied on the balanced training dataset (N=1014, delirium=50%)

ANNh 80.4 (4) nsi 71.7 (5) ns 71.7 (5) ns 71.7 (5) ns 78.5 (5) ns 80.1 (5) ns 79.3 (5) ns

BBNj 77.4 (4) k 70.1 (5) ns 69.1 (5) ns 69.6 (5) ns 75.3 (5) ns 77.3 (5) ns 76.3 (5)

DTl 77.2 (4) ns 70.9 (4) ns 72.4 (4) ns 71.7 (4) ns 74.4 (5) ns 73.8 (5) ns 73.8 (5)

LRm 81.4 (4) Bn 72.3 (5) B 74.2 (5) B 73.2 (5) B 79.8 (5) B 81 (5) B 80.4 (5) B

NBo 79.9 (4) ns 72.7 (5) ns 73.2 (5) ns 73 (5) ns 78.1 (5) ns 79.8 (5) ns 78.9 (5) ns

RFp 81.3 (4) ns 74.1 (5) ns 72.6 (5) ns 73.3 (5) ns 78.8 (5) ns 81 (5) ns 79.9 (5) ns

SVMq 81.1 (5) ns 67.2 (6) 74.4 (6) ns 71.1 (6) 80.4 (5) ns 80.5 (5) ns 80.4 (5) ns
Dataset: Imbalanced test dataset (N=1117, delirium=11.4%)

ANN 78.2 (6) ns 35.8 (9) ns 82.4 (9) ns 77.1 (9) ns 30.4 (9) +r 96.2 (9) ns 88.7 (9) ns

BBN 77.3 (6) ns 34.3 (8) ns 82.9 (8) ns 76.6 (8) ns 30.7 (8) + 95.8 (8) ns 88.4 (8) ns

DT 74.6 (7) 37.3 (8) ns 83.9 (8) ns 78.6 (8) ns 25.3 (8) ns 94.3 (8) ns 86.5 (8) ns

LR 77.5 (5) B 37.6 (11) B 84.9 (11) B 79.5 (11) B 27.1 (10) B 97.1 (10) B 88.4 (10) B

NB 75.6 (8) ns 34.7 (10) ns 81.9 (10) ns 76.6 (10) ns 28.7 (9) ns 95.6 (9) ns 88.0 (9) ns

RF 78.0 (4) ns 37.4 (8) ns 82.3 (8) ns 77.2 (8) ns 28.3 (8) ns 96.3 (8) ns 88.6 (8) ns

SVM 77.2 (6) ns 40.2 (7) + 87.2 (7) + 81.9 (7) + 29.6 (9) + 96.0 (9) ns 88.4 (9) ns

aROC-AUC: receiver operator curve-area under the curve.

bF1 score: harmonic mean of precision and recall.

cPRC-AUC: precision-recall curve-area under the curve.

dYes: positive instances or patients who developed delirium.

eChange compared to base model (B)

fNo: negative instances or patients who did not develop delirium.

gAvg: weighted average measured as the sum of all values in that metric, each weighted according to the number of instances with that particular class label by multiplying that value by the number of instances in that class, then divided by the total number of instances in the dataset.

hANN: artificial neural networks.

ins: not a statistically significant change in performance (P≥.05).

jBBN: Bayesian belief networks.

kStatistically significant deterioration of performance metric (P<.05).

lDT: J48 decision tree.

mLR: logistic regression.

nB: base comparator (reference) algorithm.

oNB: naïve Bayesian.

pRF: random forest.

qSVM: support vector machines.

rStatistically significant improvement of performance metric (P<.05).

Figure 1.

Figure 1

Receiver-operator curves (ROC) and precision-recall curves (PRC) for the training dataset using 10-fold cross-validation and test datasets. (A) ROC for training using 10-fold cross-validation. (B) ROC for test dataset. (C) PRC for training using 10-fold cross-validation. (D) PRC for test dataset. ANN: artificial neural networks; BBN: Bayesian belief networks; DT: J48 decision tree; LR: logistic regression; NB: naïve Bayesian; RF: random forest, SVM: support vector machines; P:N: positive to negative ratio.

When comparing the prediction performance using the ROC-AUC (Figure 1) for the test dataset, it may be noted that the prediction performance of all the prediction models on the test dataset is quite similar, except for DT, which was lower. This indicates that there is no obvious difference in the discriminative power of the classification models—that is, the ability of a model to distinguish between positive cases from negative ones. However, given the class imbalance in our dataset, this result might not be representative of a model’s true predictive power; hence, a further examination of the results was needed to identify the best performing model given the class imbalance.

As LR was the most commonly used algorithm to predict the manifestation of postoperative delirium in the medical literature [8,12,13,40,41,67-74], we developed a multivariate step-wise LR model that identified 8 variables as significant predictors of postoperative agitated delirium (Multimedia Appendix 2). The main purpose of developing the LR model was to give medical experts, who are not familiar with machine learning algorithms, an algorithm that they are acquainted with and use as a comparator.

In our study, for every 100 patients who developed delirium, the RF model had the best sensitivity and was able to correctly identify 72 patients (see Tables 3 and 4). The SVM model had the best PPV (out of 100 patients who were labeled positive by SVM, 30 were actually positive) and the best accuracy, specificity, and kappa. The PRC-AUC and F1 scores for SVM were the best out of all models (29.2% and 40.2%, respectively), with moderate discrimination (ROC-AUC=77.2 %). We also examined the relationship between precision (PPV) and recall (sensitivity) at different thresholds (see Table 5). At 50% sensitivity (recall), the RF model had the best precision, 37%). At 75% sensitivity (recall), RF was the best model with a precision of 25% followed by ANN with a PPV of 24%.

Table 5.

Precision of each model for all datasets at different recall thresholds.

Recall threshold (%) Model precision (%)


ANNa BBNb DTc LRd NBe RFf SVMg
Dataset: Training with 10-fold cross-validation

~25 87 83 81 88 85 87 91

~50 80 78 81 82 78 83 83

~75 71 70 69 73 70 72 72
Dataset: Test







~25 35 40 31 36 32 32 36

~50 34 33 30 34 31 37 34

~75 24 22 21 23 20 25 23

aANN: artificial neural networks.

bBBN: Bayesian belief networks.

cDT: J48 decision tree.

dLR: logistic regression.

eNB: naïve Bayesian.

fRF: random forest.

gSVM: support vector machine.

On the basis of our experiments using the PRC-AUC and PRC analysis, the RF and ANN models demonstrated the ability to distinguish patients at risk of developing delirium (minority class) when compared with the other models. ANN is considered to be a black box as it is difficult to explain, especially to people who are nonexperts, not familiar with the principles and motivation behind the ANN algorithm, and do not know how the algorithm reaches its decision and activation thresholds. However, major work has been conducted over the last decade and is still ongoing to enhance the expandability of ANN by unlocking the black box to allow accountability [75-77]. Numerous techniques have been developed and were successfully applied [78-80], giving some transparency to the model and making it more human interpretable.

Discussion

Principal Findings

Patients undergo high-risk interventions with the expectation of improving their quality of life. It is highly undesirable that any medical intervention, inadvertently, negatively impacts their cognitive functions and in turn quality of life, especially if an adverse outcome is preventable.

With the paradigm shift in health care emphasizing the patient’s quality of life after an intervention [81], innovative approaches are needed to both pre-emptively identify and effectively treat delirium. Given the availability of long-term surgical outcome data and advance machine learning methods, it is now possible to investigate the formulation of data-driven prediction models to pre-emptively identify patients susceptible to postsurgery delirium. LR-based prediction models to detect delirium have been developed using patient data from electronic medical records—in one study advanced text mining has been applied to abstract relevant data from clinical notes [82], and in another study attribute-based triggers were used [57]. We contend that with the availability of large volumes of patient data (before, during, and after the medical intervention), there are practical opportunities to develop data-driven prediction models to detect postoperative delirium in patients. Such artificial intelligence–based machine learning–based models are quite capable of identifying hidden yet important relations among variables and representing them in terms of a mathematical model that can be applied to classify/predict the output for new scenario. The artificial intelligence-based machine learning approach is rather different from the traditional statistical data analysis approaches; however, recently such methods have been applied to improve early and precise detection of diseases [16,21,25,27-29], including the prediction of outcomes after cardiac surgery [22,32,83].

In our study, we investigated the development of delirium prediction models using long-term (over 5 years) surgical outcomes data for over 5000 patients. We developed several prediction models, while addressing the underlying class imbalance issue, and compared their performance on an independent test set. Except for SVM (ROC-AUC=71.7%), the ROC-AUC of the predictive models was at least 75%, indicating a good general performance by predicting the correct classification most of the time [84,85]. Using the F1 score and the PRC-AUC, which are more sensitive to class imbalance, we were able to demonstrate that the SVM followed by the BBN models offered the best prediction performance in correctly identifying adult patients at risk of developing agitated delirium after cardiac surgery (F1 score: 40.2 and 34.4 and PRC-AUC: 30.7 and 29.6; respectively).

Our predictive models had a worse performance when compared with the findings of Kumar et al [39] (ROC-AUC of the RF model ~91%). Although they argue that their data were imbalanced, they used the ROC-AUC as their evaluation metric, which does not consider the class imbalance. On the other hand, PRC-AUC inherently accounts for class distribution (the probability is conditioned on the model estimate of the class label, which will vary if the model is applied on a population with different baseline distributions). It is more useful if the goal is improving the prediction of positive class in an imbalanced population with known baseline probability (eg, document retrieval, fraud detection, and medical complications) [44,46,48,51,66].

Compared with the findings of Corradi et al [42] (ROC-AUC of the RF model ~91% and PRC-AUC ~61 %), our model was worse. Although they included a lot of physiological parameters, they did not include any laboratory parameters. In addition, they applied the algorithm on all patients within the study period (medical and surgical). Most of the variables used were correlated—that is, they were a function of each other (eg, RASS and mechanical ventilation, RASS score and vasopressors, and dementia and the Charleston Comorbidity Index)—which likely impacted the generalizability of the model.

The paper published by Davoudi et al [40] is the only paper that is closely related to our work as they were specifically addressing the question of predicting delirium after major surgery and had a large cohort of patients who underwent cardiothoracic surgery (6890 patients, 13%). They were able to achieve an ROC-AUC ranging from 79% to 86%, which was close to the ROC-AUC we were able to achieve (71.7%-78%). Unfortunately, it is not clear what type of delirium they were capturing and the urgency of surgery these patients were undergoing. Also, only 13% of these patients underwent cardiothoracic surgery. They mainly relied on the ROC-AUC to compare the model’s performance, which is insensitive to the target class imbalance.

Lee et al [41] conducted a unique systematic review in 2017, addressing the issue of predictive models for discovering delirium after cardiac surgery. They were only able to identify 3 high-quality models (Katznelson, Original PRE-DELIRIC, and the recalibrated PRE-DELIRIC). As the original PRE-DELIRIC was recently externally validated, they externally validated the Katznelson and recalibrated PRE-DELIRIC model on a local population dataset of 600 patients. Several metrics were used to evaluate the model’s discrimination and calibration. All metrics for recalibrated PRE-DELIRIC model outperformed the Katznelson model (see Multimedia Appendix 1). However, these metrics cannot distinguish clinical utility. To identify clinical utility of these models, they performed DCA to ascertain the clinical utility of each model. The main advantage of DCA is that it incorporates preferences (patient or physician) represented as threshold probability of choosing or denying a treatment, across a range of probabilities [41]. The net benefit (the expected benefit of offering or denying a treatment at that threshold) of each algorithm was evaluated. Based on the DCA analysis, both models had limited clinical utility, with the recalibrated PRE-DELIRIC having marginally better performance at low thresholds between 20% and 40%. Regrettably, they used already validated models that are based on LR. They mentioned very limited information about the validation cohort (such as mean age, gender distribution, and type of cardiac surgery). In addition, they did not address the significant class imbalance (delirium=13.8%). Finally, the use of DCA to evaluate clinical utility of the models is very innovative but it can be only applied to evaluate models that were developed by the same algorithm but have different parameters. Its applicability across different modeling algorithms is still not clear. One of the essential assumptions of DCA is that the predicted probability and threshold probability are independent. In the case of delirium, it would be very difficult to assert that independence, as delirium is multifactorial, and there is no clear mechanism to its development. Violating this assumption might significantly affect the results and interpretation of the DCA.

To our knowledge, this is the first paper that explicitly attempts to develop several predictive models using machine learning methodology and compare their performance for the sole purpose of proactively predicting agitated delirium in adult patients undergoing cardiac surgery. A notable aspect of our work is the use of multiple performance evaluation measures to evaluate the different facets of a prediction model with respect to its prediction performance. We demonstrated the importance of using different metrics when analyzing model’s performance (eg, F1 score and PRC-AUC) and the importance of visual analysis of the curves across different probabilities (eg, PRC). Using a static or single measure, like ROC-AUC or accuracy, might lead to false assumptions and incorrect decisions, especially in the presence of class imbalance in the dataset [66].

An important factor in the selection of a prediction model is its interpretability (clarity) to the users (especially health care providers) who are particularly keen to know the basis for a recommendation/decision when it is derived from a computational model. One of the drawbacks of ANN and SVM is that they are not easy to explain, that is, how the output was produced (ie, they are regarded as black box models). This inability to explain the model and its predictions tends to raise a degree of skepticism among health care practitioners regarding the prediction produced [46,52,56]. However, the application of additional methods to decipher the ANN and SVM models’ decision logic in terms of understandable production rules that illustrate a correlation between clinical attribute values and the output class can increase their acceptance and subsequent use by medical practitioners [75-80]. Other machine learning methods, such as the BBN model provides a simple but elegant graphical representation of the problem space that can be interpreted by health care professionals.

Predicting delirium is a challenging problem, but with a significant health outcome and system use impact. Given the complexity of how and why delirium manifests in certain patients, the ability to correctly identify if not all but even a fair number of the potential patients who are at risk of developing delirium will be a significant improvement from the current state where patients are diagnosed with delirium only after it starts, and hence, the administration of appropriate interventions is delayed. To address this challenging problem, we investigated the application of machine learning methods to predict postoperative delirium after cardiac surgery. Our methodology involved addressing the target class imbalance and employing appropriate evaluation metrics to measure the prediction performance from a clinical utility perspective. We argue that with the increased use of eHealth records and auxiliary data collection tools, the volume of health data being collected is reaching the level of big data. This brings relief to the need to apply advance machine learning techniques to analyze the data for improved and effective data-driven decision support [86,87] that would enable timely intervention for negative outcomes [5,56,87] to improve health outcomes and in turn enhance patient safety and satisfaction.

Limitations

We recognize that our study has certain limitations. First, as postoperative complications (including delirium) in our database are captured as binary outcomes (yes/no) but without a time stamp, it was hard to determine if agitated delirium was a secondary phenomenon (eg, because of infection, uncontrolled pain, and prolonged mechanical ventilation) or because of a pre-existing medical comorbidity. Second, the prevalence of agitated delirium was only 11.4%. This low representation is most likely because of the definition of delirium in the source database (only agitated subtype). This can potentially limit the ability to generalize the developed models to other types of delirium [10,11]. Third, there exist more advance machine learning software than what were available in WEKA, but we chose WEKA because of its open source, flexibility, and ease of use [54]; and finally, the study is based on a retrospective design and hence may suffer from the pitfalls associated with such a design.

Clinical Equipoise and Key Messages

The key messages of this paper are as follows:

  • From a clinical standpoint:

    • Patients undergoing cardiovascular surgical procedures are at higher risk of developing agitated delirium due to several factors, including surgical complexity, comorbidities, and age [7,8].

    • Preventing delirium should be the goal, especially if patients at risk were identified. This will mitigate its negative sequalae and improve the patient’s quality of life. Some of the proposed preventive interventions that have been shown to reduce the incidence of delirium in high-risk patients include early mobilization, use of patient’s personal aids (reading glasses, hearing aid, etc), pharmacological interventions (the use of less sedatives and addressing pain), and improving sleep environment especially in the intensive care [38,88-91].

  • From a predictive modeling perspective:

    • Addressing class imbalance on the training dataset (a common feature of medical datasets) could enhance the machine learning model’s performance in identifying patients likely to develop postoperative delirium.

    • Keeping an open mind and exploring different modeling methodologies will enable the selection of the most appropriate model that can generate the best results.

    • The PRC offers a more intuitive and direct measure of the model performance that is representative of its true performance, especially in the presence of class imbalance.

Conclusions and Future Research

Postoperative agitated delirium is associated with major morbidity that impacts the patient postoperative recovery. Cardiac surgery patients are at high risk of developing postoperative delirium. To improve health outcomes of cardiac surgery, the current approach to address the effects of delirium is a preventive program of care [88-91], such as ABCDE, which involves awakening and breathing coordination for liberation from sedation and mechanical ventilation, choosing sedatives that are less likely to increase risk of delirium, delirium management, and finally, early mobility and exercise [36]. As much as the ABCDE approach provides a road map of how to manage delirium, it does not provide mechanisms to identify patients at risk of developing delirium. Hence, the ABCDE approach serves as an after-the-event management strategy, while leaving a gap in terms of a proactive prevention strategy for delirium. Our ability to predict delirium in patients, and in turn proactively administer therapeutic and behavioral therapies to mitigate the negative effects of delirium, will lead to significant improvements in health outcomes, patient satisfaction and quality of life, and health system cost saving.

In this study, we pursued the development of prediction models using preoperative clinical data to establish a mapping between the patient’s preoperative clinical variables and the onset of postoperative delirium. We investigated machine learning methods to develop a viable postoperative delirium prediction model which can be operationalized in a clinical setting as a delirium screening tool to proactively identify patients at risk of developing postcardiac surgery agitated delirium. We posit that the use and operationalization of delirium predictive model can significantly reduce the incidence of delirium by enabling the administration of preventive measures in a timely manner. In this paper, we presented work detailing the development of data-driven delirium prediction models with a reasonable accuracy. Furthermore, the work contributes 3 findings that are useful for future efforts to develop advanced delirium prediction models—that is, (1) addressing class imbalance on the training dataset will enhance the machine learning model’s performance in identifying patients likely to develop postoperative delirium, (2) when evaluating the model’s performance, selecting unsuitable measures can influence model interpretation and its utility, and (3) the PRC offers a more intuitive and direct measure of the model’s performance that is representative of its true performance, especially in the presence of class imbalance.

In our future research, we will attempt to apply feature extraction to identify key features to enhance the model’s performance. At the same time, we will attempt to isolate modifiable features that are clinically relevant so that personalized interventions can be started in a timely fashion. We will also attempt to apply evolutionary computations to optimize classifiers parameters. Another interesting application is the use of deep learning methods to create new features or feature sets to boost the model’s performance and accuracy.

In conclusion, we argue that any improvement in our ability to predict delirium using prediction models, even if numerically small, is of consequential clinical significance—this situation is like 2 drugs that have the same treatment profile, but one drug has fewer side effects, and hence, the ability to precisely select the right drug has an impact on patient safety. When dealing with complex medical problems, such as delirium, we posit that the application of advanced machine learning methods might actually improve disease prediction capabilities which in turn will enhance opportunities for preventive, personalized, and precise medical interventions that would improve the patient’s quality of life after surgery.

Acknowledgments

This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

Abbreviations

ANN

artificial neural networks

BBN

Bayesian belief networks

CABG

coronary artery bypass graft

CAM

confusion assessment method

CAM-ICU

confusion assessment method for the intensive care unit

DCA

decision curve analysis

DT

decision trees

eHealth

electronic health

ICU

intensive care unit

LR

logistic regression

MHC

Maritime Heart Center

NB

naïve Bayesian

OOB

out-of-bag

OR

odds ratio

PPV

positive predictive value (precision)

PRC-AUC

precision-recall curve-area under the curve

QEII HSC

Queen Elizabeth II Health Sciences Center

RASS

Richmond Agitation-Sedation Scale

RF

random forest

ROC-AUC

receiver operator curve-area under the curve

RUS

random majority undersampling

SVM

support vector machines

WEKA

Waikato Environment for Knowledge Acquisition

Appendix

Multimedia Appendix 1

Summary of relevant related work.

Multimedia Appendix 2

Machine learning methodology and forest plot for logistic regression.

Multimedia Appendix 3

Setting of the prediction models and the optimization steps that were applied on the used algorithms.

Footnotes

Authors' Contributions: HNM conceived the presented research question, devised the project and the main conceptual ideas, conducted the literature review, applied for research ethics, designed and performed the experiments, analyzed the data, derived the predictive models, assessed their performance on the test dataset, and took the lead in writing the manuscript. SSRA and SRA verified the analytical methods. GMH verified clinical relevance and literature review. SSRA, SRA, and GMH equally cosupervised this work. All authors discussed the results and reviewed to the final manuscript.

Conflicts of Interest: None declared.

References

  • 1.Koster S, Oosterveld FG, Hensens AG, Wijma A, van der Palen J. Delirium after cardiac surgery and predictive validity of a risk checklist. Ann Thorac Surg. 2008 Dec;86(6):1883–7. doi: 10.1016/j.athoracsur.2008.08.020. [DOI] [PubMed] [Google Scholar]
  • 2.The Society of Thoracic Surgeons. [2019-10-08]. STS National Database https://www.sts.org/registries-research-center/sts-national-database.
  • 3.American Psychiatric Association . Practice Guideline for the Treatment of Patients with Delirium. Washington, DC: American Psyciatric Assosiation; 2010. [Google Scholar]
  • 4.American Psychiatric Association . Diagnostic And Statistical Manual Of Mental Disorders. Fifth Edition. Washington, DC: American Psychiatric Publishing; 2013. Diagnostic and statistical manual of mental disorders : DSM-52013. [Google Scholar]
  • 5.Royston D, Cox F. Anaesthesia: the patient's point of view. Lancet. 2003 Nov 15;362(9396):1648–58. doi: 10.1016/S0140-6736(03)14800-3. [DOI] [PubMed] [Google Scholar]
  • 6.Martin B, Buth KJ, Arora RC, Baskett RJ. Delirium as a predictor of sepsis in post-coronary artery bypass grafting patients: a retrospective cohort study. Crit Care. 2010;14(5):R171. doi: 10.1186/cc9273. https://ccforum.biomedcentral.com/articles/10.1186/cc9273. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Martin B, Buth KJ, Arora RC, Baskett RJ. Delirium: a cause for concern beyond the immediate postoperative period. Ann Thorac Surg. 2012 Apr;93(4):1114–20. doi: 10.1016/j.athoracsur.2011.09.011. [DOI] [PubMed] [Google Scholar]
  • 8.Gottesman R, Grega M, Bailey M, Pham L, Zeger S, Baumgartner W, Selnes O, McKhann G. Delirium after coronary artery bypass graft surgery and late mortality. Ann Neurol. 2010 Mar;67(3):338–44. doi: 10.1002/ana.21899. http://europepmc.org/abstract/MED/20373345. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Smulter N, Lingehall HC, Gustafson Y, Olofsson B, Engström KG. Delirium after cardiac surgery: incidence and risk factors. Interact Cardiovasc Thorac Surg. 2013 Nov;17(5):790–6. doi: 10.1093/icvts/ivt323. http://europepmc.org/abstract/MED/23887126. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Cavallazzi R, Saad M, Marik PE. Delirium in the ICU: an overview. Ann Intensive Care. 2012 Dec 27;2(1):49. doi: 10.1186/2110-5820-2-49. http://europepmc.org/abstract/MED/23270646. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Andrejaitiene J, Sirvinskas E. Early post-cardiac surgery delirium risk factors. Perfusion. 2012 Mar;27(2):105–12. doi: 10.1177/0267659111425621. [DOI] [PubMed] [Google Scholar]
  • 12.Afonso A, Scurlock C, Reich D, Raikhelkar J, Hossain S, Bodian C, Krol M, Flynn B. Predictive model for postoperative delirium in cardiac surgical patients. Semin Cardiothorac Vasc Anesth. 2010 Sep;14(3):212–7. doi: 10.1177/1089253210374650. [DOI] [PubMed] [Google Scholar]
  • 13.Bakker RC, Osse R, Tulen J, Kappetein A, Bogers A. Preoperative and operative predictors of delirium after cardiac surgery in elderly patients. Eur J Cardiothorac Surg. 2012 Mar;41(3):544–9. doi: 10.1093/ejcts/ezr031. [DOI] [PubMed] [Google Scholar]
  • 14.Stransky M, Schmidt C, Ganslmeier P, Grossmann E, Haneya A, Moritz S, Raffer M, Schmid C, Graf BM, Trabold B. Hypoactive delirium after cardiac surgery as an independent risk factor for prolonged mechanical ventilation. J Cardiothorac Vasc Anesth. 2011 Dec;25(6):968–74. doi: 10.1053/j.jvca.2011.05.004. [DOI] [PubMed] [Google Scholar]
  • 15.Stoltzfus JC. Logistic regression: a brief primer. Acad Emerg Med. 2011 Oct;18(10):1099–104. doi: 10.1111/j.1553-2712.2011.01185.x. doi: 10.1111/j.1553-2712.2011.01185.x. [DOI] [PubMed] [Google Scholar]
  • 16.Koh HC, Tan G. Data mining applications in healthcare. J Healthc Inf Manag. 2005;19(2):64–72. [PubMed] [Google Scholar]
  • 17.Lisboa PJ. A review of evidence of health benefit from artificial neural networks in medical intervention. Neural Netw. 2002 Jan;15(1):11–39. doi: 10.1016/s0893-6080(01)00111-3. [DOI] [PubMed] [Google Scholar]
  • 18.Bertsimas D, Bjarnadóttir MV, Kane MA, Kryder JC, Pandey R, Vempala S, Wang G. Algorithmic prediction of health-care costs. Oper Res. 2008;56(6):1382–92. doi: 10.1287/opre.1080.0619. [DOI] [Google Scholar]
  • 19.Bell LM, Grundmeier R, Localio R, Zorc J, Fiks AG, Zhang X, Stephens TB, Swietlik M, Guevara JP. Electronic health record-based decision support to improve asthma care: a cluster-randomized trial. Pediatrics. 2010 Apr;125(4):e770–7. doi: 10.1542/peds.2009-1385. [DOI] [PubMed] [Google Scholar]
  • 20.Tomar D, Agarwal S. A survey on Data Mining approaches for Healthcare. Int J Bio Sci Bio Tech. 2013 Oct 31;5(5):241–66. doi: 10.14257/ijbsbt.2013.5.5.25. [DOI] [Google Scholar]
  • 21.Fei Y, Hu J, Gao K, Tu J, Li W, Wang W. Predicting risk for portal vein thrombosis in acute pancreatitis patients: a comparison of radical basis function artificial neural network and logistic regression models. J Crit Care. 2017 Jun;39:115–23. doi: 10.1016/j.jcrc.2017.02.032. [DOI] [PubMed] [Google Scholar]
  • 22.Santelices LC, Wang Y, Severyn D, Druzdzel MJ, Kormos RL, Antaki JF. Development of a hybrid decision support model for optimal ventricular assist device weaning. Ann Thorac Surg. 2010 Sep;90(3):713–20. doi: 10.1016/j.athoracsur.2010.03.073. http://europepmc.org/abstract/MED/20732482. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Baxt WG. Use of an artificial neural network for the diagnosis of myocardial infarction. Ann Intern Med. 1991 Dec 1;115(11):843–8. doi: 10.7326/0003-4819-115-11-843. [DOI] [PubMed] [Google Scholar]
  • 24.Artis SG, Mark R, Moody G. Detection of Atrial Fibrillation Using Artificial Neural Networks. Proceedings Computers in Cardiology; CinC'91; September 23-26, 1991; Venice, Italy, Italy. 1991. [DOI] [Google Scholar]
  • 25.Eftekhar B, Mohammad K, Ardebili HE, Ghodsi M, Ketabchi E. Comparison of artificial neural network and logistic regression models for prediction of mortality in head trauma based on initial clinical data. BMC Med Inform Decis Mak. 2005 Feb 15;5:3. doi: 10.1186/1472-6947-5-3. https://bmcmedinformdecismak.biomedcentral.com/articles/10.1186/1472-6947-5-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Belciug S, Gorunescu F, Salem AB, Gorunescu M. Clustering-Based Approach for Detecting Breast Cancer Recurrence. Proceedings of the 2010 10th International Conference on Intelligent Systems Design and Applications; ISDA'10; November 29-December 1 2010; Cairo, Egypt. Intelligent Systems Design and Applications (ISDA), 2010 10th International Conference on: 2010. [DOI] [Google Scholar]
  • 27.Salama GI, Abdelhalim M, Zeid MA. Breast cancer diagnosis on three different datasets using multi-classifiers. Int J Comp Inf Technol. 2012;1(1):36–43. https://www.ijcit.com/archives/volume1/issue1/Paper010105.pdf. [Google Scholar]
  • 28.Ayer T, Chhatwal J, Alagoz O, Kahn CE, Woods RW, Burnside ES. Informatics in radiology: comparison of logistic regression and artificial neural network models in breast cancer risk estimation. Radiographics. 2010 Jan;30(1):13–22. doi: 10.1148/rg.301095057. http://europepmc.org/abstract/MED/19901087. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Joshi S, Shenoy D, Vibhudendra SG, Rrashmi P, Venugopal K, Patnaik L. Classification of Alzheimer's Disease and Parkinson's Disease by Using Machine Learning and Neural Network Methods. Proceedings of the 2010 Second International Conference on Machine Learning and Computing; ICMLC'10; February 9-11, 2010; Singapore. 2010. pp. 218–22. [DOI] [Google Scholar]
  • 30.Escudero J, Zajicek J, Ifeachor E. Early detection and characterization of Alzheimer's disease in clinical scenarios using Bioprofile concepts and K-means. Conf Proc IEEE Eng Med Biol Soc. 2011;2011:6470–3. doi: 10.1109/IEMBS.2011.6091597. [DOI] [PubMed] [Google Scholar]
  • 31.Ramani RG, Sivagami G. Parkinson disease classification using data mining algorithms. Int J Comp App. 2011;32(9):17–22. doi: 10.5120/3932-5571. [DOI] [Google Scholar]
  • 32.Nilsson J, Ohlsson M, Thulin L, Höglund P, Nashef SA, Brandt J. Risk factor identification and mortality prediction in cardiac surgery using artificial neural networks. J Thorac Cardiovasc Surg. 2006 Jul;132(1):12–9. doi: 10.1016/j.jtcvs.2005.12.055. https://linkinghub.elsevier.com/retrieve/pii/S0022-5223(06)00124-3. [DOI] [PubMed] [Google Scholar]
  • 33.Kazemi Y, Mirroshandel SA. A novel method for predicting kidney stone type using ensemble learning. Artif Intell Med. 2018 Jan;84:117–26. doi: 10.1016/j.artmed.2017.12.001. [DOI] [PubMed] [Google Scholar]
  • 34.Edelstein P. Emerging directions in analytics. Predictive analytics will play an indispensable role in healthcare transformation and reform. Health Manag Technol. 2013 Jan;34(1):16–7. [PubMed] [Google Scholar]
  • 35.Moradi M, Ghadiri N. Different approaches for identifying important concepts in probabilistic biomedical text summarization. Artif Intell Med. 2018 Jan;84:101–16. doi: 10.1016/j.artmed.2017.11.004. [DOI] [PubMed] [Google Scholar]
  • 36.Brummel NE, Girard TD. Preventing delirium in the intensive care unit. Crit Care Clin. 2013 Jan;29(1):51–65. doi: 10.1016/j.ccc.2012.10.007. http://europepmc.org/abstract/MED/23182527. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Reade MC, Finfer S. Sedation and delirium in the intensive care unit. N Engl J Med. 2014 Jan 30;370(5):444–54. doi: 10.1056/NEJMra1208705. [DOI] [PubMed] [Google Scholar]
  • 38.Ettema RG, van Koeven H, Peelen LM, Kalkman CJ, Schuurmans MJ. Preadmission interventions to prevent postoperative complications in older cardiac surgery patients: a systematic review. Int J Nurs Stud. 2014 Feb;51(2):251–60. doi: 10.1016/j.ijnurstu.2013.05.011. [DOI] [PubMed] [Google Scholar]
  • 39.Kramer D, Veeranki S, Hayn D, Quehenberger F, Leodolter W, Jagsch C, Schreier G. Development and validation of a multivariable prediction model for the occurrence of delirium in hospitalized gerontopsychiatry and internal medicine patients. Stud Health Technol Inform. 2017;236:32–39. [PubMed] [Google Scholar]
  • 40.Davoudi A, Ozrazgat-Baslanti T, Ebadi A, Bursian A, Bihorac A, Rashidi P. Delirium Prediction using Machine Learning Models on Predictive Electronic Health Records Data. Proceedings of the 2017 IEEE 17th International Conference on Bioinformatics and Bioengineering; BIBE'17; October 23-25, 2017; Washington, DC, USA. 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Lee A, Mu J, Joynt G, Chiu C, Lai V, Gin T, Underwood M. Risk prediction models for delirium in the intensive care unit after cardiac surgery: a systematic review and independent external validation. Br J Anaesth. 2017 Mar 1;118(3):391–9. doi: 10.1093/bja/aew476. https://linkinghub.elsevier.com/retrieve/pii/S0007-0912(17)30208-8. [DOI] [PubMed] [Google Scholar]
  • 42.Corradi JP, Thompson S, Mather JF, Waszynski CM, Dicks RS. Prediction of incident delirium using a random forest classifier. J Med Syst. 2018 Nov 14;42(12):261. doi: 10.1007/s10916-018-1109-0. [DOI] [PubMed] [Google Scholar]
  • 43.Global Biodiversity Information Facility. 2014. [2019-10-08]. R: a language and environment for statistical computing https://www.gbif.org/tool/81287/r-a-language-and-environment-for-statistical-computing.
  • 44.Hastie T, Tibshirani R, Friedman J. The Elements Of Statistical Learning: Data Mining, Inference, And Prediction. New York: Springer; 2008. [Google Scholar]
  • 45.Breiman L. Random forests. Mach Learn. 2001;45(1):5–32. https://link.springer.com/article/10.1023/A:1010933404324. [Google Scholar]
  • 46.Witten IH, Frank E, Hall MA. Data Mining: Practical Machine Learning Tools And Techniques. Third Edition. Burlington, Massachusetts: Morgan Kaufmann; 2011. [Google Scholar]
  • 47.Ganganwar V. An overview of classification algorithms for imbalanced datasets. Int J Emerg Technol Adv Eng. 2012;2:42–7. https://www.researchgate.net/publication/292018027_An_overview_of_classification_algorithms_for_imbalanced_datasets. [Google Scholar]
  • 48.Chawla NW. Data Mining and Knowledge Discovery Handbook. New York: Springer; 2005. Data mining for imbalanced datasets: An overview; pp. 853–67. [Google Scholar]
  • 49.He H, Garcia EA. Learning from imbalanced data. IEEE Trans Knowl Data Eng. 2009;21(9):1263–84. doi: 10.1109/tkde.2008.239. [DOI] [Google Scholar]
  • 50.Pyle D. Data Preparation for Data Mining. Burlington, Massachusetts: Morgan Kaufmann; 1999. [Google Scholar]
  • 51.James G, Witten D, Hastie T, Tibshirani R. An Introduction to Statistical Learning: with Applications in R. New York: Springer; 2013. [Google Scholar]
  • 52.Han J, Kamber M, Jian PP. Data Mining Concepts And Techniques. Burlington, Massachusetts: Morgan Kaufmann; 2011. [Google Scholar]
  • 53.Refaeilzadeh P, Tang L, Liu H. Cross-validation. Encyclopedia Database Syst. 2016:1–7. doi: 10.1007/978-1-4899-7993-3_565-2. [DOI] [Google Scholar]
  • 54.Hall M, Frank E, Holmes G, Pfahringer B, Reutemann P, Witten IH. The WEKA data mining software. SIGKDD Explor Newsl. 2009;11(1):10. doi: 10.1145/1656274.1656278. [DOI] [Google Scholar]
  • 55.Cooper CG, Dash D, Levander J, Wong W, Hogan W, Wagner M. Bayesian biosurveillance of disease outbreaks. Proceedings of the 20th conference on Uncertainty in artificial intelligence; UAI'04; July 7-11, 2004; Banff, Canada. 2004. pp. 94–103. [Google Scholar]
  • 56.Tufféry S. Data Mining and Statistics for Decision Making. Hoboken, New Jersey: Wiley; 2011. [Google Scholar]
  • 57.Moon K, Jin Y, Jin T, Lee S. Development and validation of an automated delirium risk assessment system (Auto-DelRAS) implemented in the electronic health record system. Int J Nurs Stud. 2018 Jan;77:46–53. doi: 10.1016/j.ijnurstu.2017.09.014. [DOI] [PubMed] [Google Scholar]
  • 58.Mao Y, Chen Y, Hackmann G, Chen M, Lu C, Kollef M. Early Deterioration Warning for Hospitalized Patients by Mining Clinical Data. Int J Knowl Discov Bioinformatics. 2011;2:1–20. doi: 10.4018/jkdb.2011070101. [DOI] [Google Scholar]
  • 59.Watt EW, Bui AA. Evaluation of a dynamic bayesian belief network to predict osteoarthritic knee pain using data from the osteoarthritis initiative. AMIA Annu Symp Proc. 2008 Nov 6;:788–92. http://europepmc.org/abstract/MED/18999030. [PMC free article] [PubMed] [Google Scholar]
  • 60.Krishna G, Kumar B, Orsu N, B S. Performance analysis and evaluation of different data mining algorithms used for cancer classification. Int J Adv Res Artif Intell. 2013;2(5) doi: 10.14569/ijarai.2013.020508. [DOI] [Google Scholar]
  • 61.Zhou F, Jin L, Dong J. Premature ventricular contraction detection combining deep neural networks and rules inference. Artif Intell Med. 2017 Jun;79:42–51. doi: 10.1016/j.artmed.2017.06.004. [DOI] [PubMed] [Google Scholar]
  • 62.Haddawy P, Hasan AI, Kasantikul R, Lawpoolsri S, Sa-Angchai P, Kaewkungwal J, Singhasivanon P. Spatiotemporal Bayesian networks for malaria prediction. Artif Intell Med. 2018 Jan;84:127–38. doi: 10.1016/j.artmed.2017.12.002. [DOI] [PubMed] [Google Scholar]
  • 63.Boucekine M, Loundou A, Baumstarck K, Minaya-Flores P, Pelletier J, Ghattas B, Auquier P. Using the random forest method to detect a response shift in the quality of life of multiple sclerosis patients: a cohort study. BMC Med Res Methodol. 2013 Feb 15;13:20. doi: 10.1186/1471-2288-13-20. https://bmcmedresmethodol.biomedcentral.com/articles/10.1186/1471-2288-13-20. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 64.Steyerberg EW. Clinical Prediction Models: A Practical Approach to Development, Validation, and Updating. New York: Springer; 2008. [Google Scholar]
  • 65.Murphy KP. Machine Learning: A Probabilistic Perspective. Cambridge: MIT Press; 2012. [Google Scholar]
  • 66.Saito T, Rehmsmeier M. The precision-recall plot is more informative than the ROC plot when evaluating binary classifiers on imbalanced datasets. PLoS One. 2015;10(3):e0118432. doi: 10.1371/journal.pone.0118432. http://dx.plos.org/10.1371/journal.pone.0118432. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67.Inouye SK, Charpentier PA. Precipitating factors for delirium in hospitalized elderly persons. Predictive model and interrelationship with baseline vulnerability. J Am Med Assoc. 1996 Mar 20;275(11):852–7. doi: 10.1001/jama.1996.03530350034031. [DOI] [PubMed] [Google Scholar]
  • 68.Inouye SK, Viscoli CM, Horwitz RI, Hurst LD, Tinetti ME. A predictive model for delirium in hospitalized elderly medical patients based on admission characteristics. Ann Intern Med. 1993 Sep 15;119(6):474–81. doi: 10.7326/0003-4819-119-6-199309150-00005. [DOI] [PubMed] [Google Scholar]
  • 69.O'Keeffe ST, Lavan JN. Predicting delirium in elderly patients: development and validation of a risk-stratification model. Age Ageing. 1996 Jul;25(4):317–21. doi: 10.1093/ageing/25.4.317. [DOI] [PubMed] [Google Scholar]
  • 70.van den Boogaard M, Pickkers P, Slooter AJ, Kuiper MA, Spronk PE, van der Voort PH, van der Hoeven JG, Donders R, van Achterberg T, Schoonhoven L. Development and validation of PRE-DELIRIC (PREdiction of DELIRium in ICu patients) delirium prediction model for intensive care patients: observational multicentre study. Br Med J. 2012 Feb 9;344:e420. doi: 10.1136/bmj.e420. http://europepmc.org/abstract/MED/22323509. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Katznelson R, Djaiani GN, Borger MA, Friedman Z, Abbey SE, Fedorko L, Karski J, Mitsakakis N, Carroll J, Beattie WS. Preoperative use of statins is associated with reduced early delirium rates after cardiac surgery. Anesthesiology. 2009 Jan;110(1):67–73. doi: 10.1097/ALN.0b013e318190b4d9. [DOI] [PubMed] [Google Scholar]
  • 72.Isfandiaty R, Harimurti KF, Setiati SF, Roosheroe A. Incidence and predictors for delirium in hospitalized elderly patients: a retrospective cohort study. Acta Med Indones. 2012 Oct;44(4):290–7. http://www.inaactamedica.org/archives/2012/23314969.pdf. [PubMed] [Google Scholar]
  • 73.Carrasco MP, Villarroel L, Andrade M, Calderón J, González M. Development and validation of a delirium predictive score in older people. Age Ageing. 2014 May;43(3):346–51. doi: 10.1093/ageing/aft141. [DOI] [PubMed] [Google Scholar]
  • 74.Chaiwat O, Chanidnuan M, Pancharoen W, Vijitmala K, Danpornprasert P, Toadithep P, Thanakiattiwibun C. Postoperative delirium in critically ill surgical patients: incidence, risk factors, and predictive scores. BMC Anesthesiol. 2019 Mar 20;19(1):39. doi: 10.1186/s12871-019-0694-x. https://bmcanesthesiol.biomedcentral.com/articles/10.1186/s12871-019-0694-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 75.Lisboa PJ. Interpretability in Machine Learning – Principles and Practice. Proceedings of the International Workshop on Fuzzy Logic and Applications; WILF'13; November 19-22, 2013; Genoa, Italy. 2013. pp. 15–21. [Google Scholar]
  • 76.Goodman B, Flaxman S. European union regulations on algorithmic decision-making and a 'Right to Explanation'. AI Mag. 2017;38(3):50–7. doi: 10.1609/aimag.v38i3.2741. [DOI] [Google Scholar]
  • 77.Vellido A, Martín-Guerrero J, Lisboa PJ. CiteSeer. 2012. [2019-10-08]. Making Machine Learning Models Interpretable http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.431.5382.
  • 78.Intrator O, Intrator N. Interpreting neural-network results: a simulation study. Comput Stat Data Anal. 2001;37(3):373–93. doi: 10.1016/s0167-9473(01)00016-0. [DOI] [Google Scholar]
  • 79.Guidotti R, Monreale A, Ruggieri S, Turini F, Giannotti F, Pedreschi D. A survey of methods for explaining black box models. ACM Comput Surv. 2019;51(5):1–42. doi: 10.1145/3236009. [DOI] [Google Scholar]
  • 80.Montavon G, Samek W, Müller K. Methods for interpreting and understanding deep neural networks. Digit Signal Process. 2018;73:1–15. doi: 10.1016/j.dsp.2017.10.011. [DOI] [Google Scholar]
  • 81.Buth KJ, Gainer RA, Legare J, Hirsch GM. The changing face of cardiac surgery: practice patterns and outcomes 2001-2010. Can J Cardiol. 2014 Feb;30(2):224–30. doi: 10.1016/j.cjca.2013.10.020. [DOI] [PubMed] [Google Scholar]
  • 82.Mikalsen K, Soguero-Ruiz C, Jensen K, Hindberg K, Gran M, Revhaug A, Lindsetmo R, Skrøvseth SO, Godtliebsen F, Jenssen R. Using anchors from free text in electronic health records to diagnose postoperative delirium. Comput Methods Programs Biomed. 2017 Dec;152:105–14. doi: 10.1016/j.cmpb.2017.09.014. [DOI] [PubMed] [Google Scholar]
  • 83.Wise ES, Hocking KM, Brophy CM. Prediction of in-hospital mortality after ruptured abdominal aortic aneurysm repair using an artificial neural network. J Vasc Surg. 2015 Jul;62(1):8–15. doi: 10.1016/j.jvs.2015.02.038. https://linkinghub.elsevier.com/retrieve/pii/S0741-5214(15)00289-X. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84.Fawcett T. An introduction to ROC analysis. Pattern Recognit Lett. 2006;27(8):861–74. doi: 10.1016/j.patrec.2005.10.010. [DOI] [Google Scholar]
  • 85.Fawcett T. ROC graphs: Notes and practical considerations for researchers. Mach Learn. 2004;31:1–38. https://www.hpl.hp.com/techreports/2003/HPL-2003-4.pdf. [Google Scholar]
  • 86.O'Connor PJ, Sperl-Hillen JM, Rush WA, Johnson PE, Amundson GH, Asche SE, Ekstrom HL, Gilmer TP. Impact of electronic health record clinical decision support on diabetes care: a randomized trial. Ann Fam Med. 2011;9(1):12–21. doi: 10.1370/afm.1196. http://www.annfammed.org/cgi/pmidlookup?view=long&pmid=21242556. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Berwick DM, Hackbarth AD. Eliminating waste in US health care. J Am Med Assoc. 2012 Apr 11;307(14):1513–6. doi: 10.1001/jama.2012.362. [DOI] [PubMed] [Google Scholar]
  • 88.Hsieh SJ, Ely EW, Gong MN. Can intensive care unit delirium be prevented and reduced? Lessons learned and future directions. Ann Am Thorac Soc. 2013 Dec;10(6):648–56. doi: 10.1513/AnnalsATS.201307-232FR. http://europepmc.org/abstract/MED/24364769. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89.O'Hanlon S, O'Regan N, Maclullich AM, Cullen W, Dunne C, Exton C, Meagher D. Improving delirium care through early intervention: from bench to bedside to boardroom. J Neurol Neurosurg Psychiatry. 2014 Feb;85(2):207–13. doi: 10.1136/jnnp-2012-304334. [DOI] [PubMed] [Google Scholar]
  • 90.Cerejeira J, Mukaetova-Ladinska E. A clinical update on delirium: from early recognition to effective management. Nurs Res Pract. 2011;2011:875196. doi: 10.1155/2011/875196. doi: 10.1155/2011/875196. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91.Trogrlić Z, van der Jagt M, Bakker J, Balas MC, Ely EW, van der Voort PH, Ista E. A systematic review of implementation strategies for assessment, prevention, and management of ICU delirium and their effect on clinical outcomes. Crit Care. 2015 Apr 9;19:157. doi: 10.1186/s13054-015-0886-9. https://ccforum.biomedcentral.com/articles/10.1186/s13054-015-0886-9. [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Multimedia Appendix 1

Summary of relevant related work.

Multimedia Appendix 2

Machine learning methodology and forest plot for logistic regression.

Multimedia Appendix 3

Setting of the prediction models and the optimization steps that were applied on the used algorithms.


Articles from JMIR Medical Informatics are provided here courtesy of JMIR Publications Inc.

RESOURCES