Skip to main content
Health Data Science logoLink to Health Data Science
. 2021 Dec 23;2021:9808426. doi: 10.34133/2021/9808426

Self-Correcting Recurrent Neural Network for Acute Kidney Injury Prediction in Critical Care

Hao Du 1, Ziyuan Pan 2, Kee Yuan Ngiam 3, Fei Wang 4, Ping Shum 2, Mengling Feng 1,
PMCID: PMC10904062  PMID: 38487505

Abstract

Background. In critical care, intensivists are required to continuously monitor high-dimensional vital signs and lab measurements to detect and diagnose acute patient conditions, which has always been a challenging task. Recently, deep learning models such as recurrent neural networks (RNNs) have demonstrated their strong potential on predicting such events. However, in real deployment, the patient data are continuously coming and there is no effective adaptation mechanism for RNN to incorporate those new data and become more accurate.Methods. In this study, we propose a novel self-correcting mechanism for RNN to fill in this gap. Our mechanism feeds prediction errors from the predictions of previous timestamps into the prediction of the current timestamp, so that the model can “learn” from previous predictions. We also proposed a regularization method that takes into account not only the model’s prediction errors on the labels but also its estimation errors on the input data.Results. We compared the performance of our proposed method with the conventional deep learning models on two real-world clinical datasets for the task of acute kidney injury (AKI) prediction and demonstrated that the proposed model achieved an area under ROC curve at 0.893 on the MIMIC-III dataset and 0.871 on the Philips eICU dataset.Conclusions. The proposed self-correcting RNNs demonstrated effectiveness in AKI prediction and have the potential to be applied to clinical applications.

1. Introduction

Electronic health record (EHR) data are accumulative, routinely collected patient observations from hospitals or clinical institutes. In the case of intensive care unit (ICU), EHR includes not only the static information such as patient demographics and discrete time series data such as medication and diagnosis but also continuous multivariate time series data such as vital signs and laboratory measurements. In order to detect and diagnose acute (and usually deadly) patients’ conditions, ICU intensivists need to continuously monitor high-dimensional vital signs and lab measurements [1]. Example acute conditions include acute kidney injury, acute hypertension, acute organ failure, and acute septic shocks. It has always been challenging to track all indicative changes in various patients’ data to diagnose these acute conditions accurately in time. Predictive models developed with ICU EHR data provide an opportunity for early detection of ICU acute conditions, which can lead to in-time and better care. In this study, we propose a self-correcting deep learning framework with accumulative ICU data to predict these acute conditions. Without the loss of generality, we will focus on the prediction of acute kidney injury (AKI), while the predictive modeling of other conditions can be handled similarly.

AKI is a sudden onset of renal failure or kidney damage, occurring in at least 5% of hospitalized patients. It is associated with a significant increase on mortality, length of stay (LOS), and hospital cost under a wide range of conditions [2]. AKI is a good study case for disease risk predictive modeling because (1) the precise definition for AKI allows temporal anchoring of events. AKI can be defined with the urine output criteria in a short diagnostic time frame. These criteria require dynamic modelling over a rolling 6 to 24 hours’ window [3]. The hourly recorded EHR data in ICU makes it possible to identify the onset of AKI events accurately and efficiently. (2) If detected and managed in time, AKI is potentially avoidable and reversible in the process of a few hours to several days. A previous trial showed that AKI alerts in the ICU led to an increase in the frequency of therapeutic interventions for AKI patients and that AKI patients in the “risk” phase were more likely to return to baseline renal function within 8 hours [4]. An accurate, automated, and early AKI detection system could potentially prevent AKI events, thus reducing mortality, shortening LOS, avoiding the development of chronic kidney disease, and creating quality of care indicators.

AKI prediction with utilization of features from EHR data is attracting widespread research interest. In particular, much research in recent years has focused on predictive modeling on a broad population to identify high-risk subjects as early as possible [5]. Initially, AKI prediction was modeled by standard statistical modeling methods, including logistic regression, discriminant analysis, or decision tree algorithm [69]. Data were accumulated using sliding window method, and the prediction was generated at a specified interval (per hour, two hours, day, shift, etc.). Recently, a number of studies have been carried out utilizing recurrent neural networks (RNNs) in clinical diagnosis and prediction. RNNs are one branch of neural networks, which are powerful to process sequential data [10]. In RNNs, hidden units connect to each other by forming a directed cycle. Each output value is dependent on the previous computations. In traditional RNNs, the network can only look back to a few steps due to the gradient vanishing and exploding problems [10]. To address these limitations, variants of RNNs, such as LSTM [11] and GRU [12], are proposed and well utilized in clinical prediction problems. These variants have the hidden state with the forget gate that decide what to keep in and erase from the memory of the network.

In addition to LSTMs, pooling and word embedding were used in DeepCare [13] to model illness states of patients and to predict patients’ outcomes. DeepCare is an end-to-end deep dynamic memory neural network. DeepCare introduced time parameterizations to handle irregularly timed events and utilized accumulative temporal data by moderating the forgetting and consolidation of memory cells. DeepCare demonstrated improved disease progression modeling accuracy and risk prediction compared to Markov models and plain RNNs. The limitation that is shared by both Doctor AI and DeepCare is that, as they continue to predict patients’ disease progression, they lack a feedback mechanism to allow the models to learn and improve from their previous prediction mistakes.

Specifically, on AKI prediction, a number of studies have been carried out. Hurry et al. employed Bayesian networks on AKI prediction by predicting the likelihood of AKI onset based on longitudinal patient data on the MIMIC II database [14]. In addition, Nogueira et al. applied Markov chain model on the PhysioNet dataset to predict the future state of the patients based on the current medical state and ICU type. The common limitation of these studies is that the proposed methods are all dependent on hand-engineered features and expert knowledge, where the hidden states of the patient condition were not effectively modeled.

In the setting of ICU, or more in general the setting of in-patient care, after patients’ admission, their situation often evolves rapidly over time. Therefore, patients’ EHR data are dynamic time series in nature. The challenges of longitudinal EHR data, including event temporality, high dimensionality, and irregular sampling, have been investigated in many studies using RNNs [15]. In addition to these challenges, patients’ EHR data also have an accumulative characteristic: as patients stay longer in the hospital, more data are collected about their disease progression, and thus a more accurate modeling of patients’ physiological states is possible. Conventional RNNs were not optimized to accumulate information in time series data. The accumulated error between prediction and the patient’s status are not specifically modeled across the patients’ ICU stay. In addition, an effective ICU acute condition predictor is expected to enhance itself through self-correcting and learning from its accumulated prediction errors. This self-correcting mechanism is lacking in conventional RNN models.

To address the above limitations, we propose a variant of RNN to predict the onset of patients’ AKI in ICU. In this pilot study, we validate the effectiveness of our proposed self-correcting model with two actual ICU patient EHR datasets from the US. In the next phase, we plan to validate and deploy our algorithm in our local hospital.

The main contributions of this study are summarized as follows:

  • (1)

    Our method utilized the accumulative data of patients in ICU instead of a snapshot of the patient’s condition to improve the performance of AKI prediction

  • (2)

    We developed a novel accumulative self-correcting mechanism by modeling the accumulated errors in the model when the prediction is incorrect

  • (3)

    We proposed a regularization method for our model, which takes into account not only the model’s prediction error on the label but also its estimation errors on the future input data. Such regularization reduces the variance of the model and improves the efficiency of the self-correcting mechanism

  • (4)

    Our proposed method has been validated in two real-world large-scale ICU datasets. It was shown to outperform traditional RNNs. In addition, the method is in-progress of being validated locally with data from our own hospital

2. Materials and Methods

2.1. Problem Definition

For any ICU patient and any time point t during his/her ICU stay, our goal is to predict whether the ICU patient may develop AKI in next 6 hours, i.e., t+6, based on all his/her data of this ICU stay accumulated up till time t . In this study, AKI was defined according to the most commonly used RIFLE criteria [16]. A patient was detected with AKI if his/her urine output is less than 0.5 mL/kg/h for ≥6 h.

Based on this definition, for a patient, the AKI actual onset label at any time t, denoted as yt, can only be observed at time step t+6. In the traditional RNN, the correctness of the predicted y^t would not affect the prediction in the future time step; although at time step t+6, we will know whether our predicted y^t=yt. In this study, we want to fully utilize all the observed data including the label yt in order to continuously improve the accuracy of the model. Therefore, we designed the novel self-correcting mechanism to further enhance the conventional RNN model.

2.2. Data and Data Preprocessing

We applied our proposed method on the Medical Information Mart for Intensive Care III (MIMIC-III) and Phillips eICU Collaborative Research Dataset. MIMIC-III dataset [17] consists of medical records of over 40,000 ICU patients between 2001 and 2012. Data in MIMIC-III include demographic information, vital signs, medication records, laboratory measurements, observations, fluid balance, procedure codes, diagnostic codes, imaging reports, hospital length of stay, and survival data. We further validated the performance of our proposed method with the Phillips eICU Collaborative Research Dataset. The eICU dataset is populated with data from a combination of multiple ICUs across the United States [18]. The dataset covers 200859 ICU patients admitted in 2014 and 2015. For this project, we extracted the following variables from both the MIMIC-III and eICU datasets:

  • (1)

    Demographic information (static variables): age and gender

  • (2)

    Comorbidities (static variables): ICD-9 [19] defined as comorbidity conditions of congestive heart failure, cardiac arrhythmias, valvular disease, pulmonary circulation, peripheral vascular, hypertension, paralysis, neurological disorder, chronic pulmonary diseases, diabetes, hypothyroidism, renal failure, liver diseases, peptic ulcer, aids, lymphoma, metastatic cancer, rheumatoid arthritis, coagulopathy, obesity, fluid electrolyte, anemias, alcohol abuse, drug abuse, psychoses, and depression

  • (3)

    Vital signs (time series variables): mean arterial blood pressure, heart rate, respiration rate, and temperature

  • (4)

    Lab measurements (time series variables): bilirubin, BUN (blood urea nitrogen), creatinine, glucose, HCO 3 (serum bicarbonate), HCT (hematocrit), K (serum potassium), lactate, Mg (serum magnesium), Na (serum sodium), PaCO 2 (partial pressure of arterial CO 2), PaO 2 (partial pressure of arterial O 2), pH, platelets, troponin, and WBC (white blood cell count)

  • (5)

    Fluids (time series variables): urine output and fluid balance

  • (6)

    Interventions (time series variables): usage of mechanical ventilation, vasopressor, and sedative medications

For the extracted time series variables, the vital signs were regularly collected on hourly basis. But the lab measurements, fluid information, and interventions were collected with random time windows. For these variables, we transformed the data into regularly sample time series, where the time gap between two data point is always one hour. For the time steps where there was no recorded data, data were imputed with the weighted average value of the nearest data points. One single data point was obtained on each hour for each patient admitted to ICU. The extracted data are illustrated in Figure 1. We then transformed the data into feature vectors of each patient, normalized the feature vectors with minimal and maximum values in the cohort, and fed them into the models.

Figure 1.

Figure 1

Visualization of data extracted for AKI prediction.

2.3. Self-Correcting RNN Models

2.3.1. General Idea

Figure 2 graphically illustrated the proposed self-correcting RNN framework. Compared to the traditional multilayer RNN, we created a feedback loop between each time step t and t6,t7T. At each time step t, we have y^t6, representing the predicted label from our model 6 hours ago, and yt6 the true label. Discrepancies between the predicted y^t6 and the label yt6 are fed into the the feed forward layers. Then, the output of the feed forward layers will be fed into each RNN layer as a part of the input. We believe this can provide additional information about the correctness of previous hidden states.

Figure 2.

Figure 2

Self-correcting RNN. Circles are used for RNN cells (either LSTM or GRU), while diamond-shaped units are used for input and output. Italic letters (e.g., xt and yt) denote the predicted values, while bold capital letters (e.g., Xt and Yt) denote the actual values.

Note that, for the initial time steps t16, there is no feedback sent to the RNN from the previous time step, as we will need at least 6 hours of data to obtain the true label of AKI. In these cases, a default state is sent to each RNN layer instead. This default state is trained by backpropagation as neural network parameters. Like the normal RNN models, this model can be applied in both classification tasks and regression tasks. Cross-entropy loss and mean square error can be used as the loss functions for classification tasks and regression tasks respectively.

2.3.2. GRU and LSTM Fundamentals

Gated recurrent unit (GRU) and long short-term memory (LSTM) are the two most commonly used variants of RNN. They have been shown to work well on modeling sequential data with long-term dependencies.

The common property shared between GRU and LSTM is the additive update process. The values of the gates depend on the input and the previous state. And update process is controlled by the gates together with the input and previous state. The function ht=RNNht1,xt performed by LSTM or GRU can therefore be divided into two steps: gates=RNNgateht1,xt and ht=RNNstategates,ht1,xt. The joint distribution of a GRU or LSTM network factorizes as [20,21]

(1)py1:T,h1:T,gate1:Tx1:T,h0=t=1Tpythtphtgatet,ht1,xtpgatetht1,xt,

where h0 denotes the initial states of the GRU/LSTM layers.

Note that these probability distributions modeled by the RNN are all deterministic and that this is the joint distribution of a single-layer RNN. The joint distribution of multilayer RNN factorizes into more components.

2.3.3. Self-Correcting Mechanism

As mentioned in General Idea, at time step t, we will get the true label yt6 (i.e., whether the patient develops AKI). And we want this information to be used to improve the accuracy of the prediction at the current time step. Therefore, the joint probability modeled by the neural network should be

(2)py^1:T,h1:T,gate1:Tx1:T,h0,y1:T6=t=7Tpy^thtphtgatet,ht1,xt,y^t6,yt6pgatetht1,xt,y^t6,yt6t=16py^thtphtgatet,ht1,xtpgatetht1,xt,

where y^t denotes the output of the neural network and yt denotes the label. yt6 (both the prediction and the actual result) is part of the feature vector we fed into the RNN model at t as the input, together with the original input data. For t<6,yt6 is not available. We replace yt6 with trained constants in that case.

2.3.4. Self-Correcting RNN with Regularization

Note that for t16, the factorized joint distributions of Equations (1) and (2) are the same. And the difference between Equations (1) and (2) is that, in Equation (2), the probability of gatet and ht is conditioned on y^t6 and yt6. Based on this probability model, we designed the neural network shown in Figure 2. Both y^t6 and yt6 are fed into each RNN layer. We call it “self-correcting RNN” because the update of hidden state from ht1 to ht is based on the label yt6 and the predicted value y^t6 from the past time step. If the neural network makes a wrong prediction at the past time step, the hidden state is expected to be updated accordingly. One challenge with the self-correcting RNN is that the label and predicted value used to update ht are from 6 time steps ago. The model may achieve better performance if we can minimize the time gap. To further improve the self-correcting RNN, we designed the regularization method for it. The self-correcting RNN with regularization is shown in Figure 3. Instead of only predicting y^t1 at time step t1, the model predicts x^t as well. Then, at time step t, the predicted x^t and the input xt will be fed into the feed forward layers together with y^t6 and yt6. The main difference is that the probability distribution of gatet and ht now conditions on x^t as well. When we train the model, we add the mean squared error between x^t and xt to the total loss after multiplying it with a certain coefficient. So, the model will learn to predict xt by backpropagation. This regularization method boost the performance of the self-correcting RNN in the following two ways:

  • (1)

    It minimizes the time gap of the self-correcting mechanism.

  • (2)

    It enforces the model to predict xt+1 instead of only yt. More information needs to be captured by the hidden state, and hence, the variance of the model decreases.

Figure 3.

Figure 3

Self-correcting RNN with regularization. Circles are used for RNN cells (either LSTM or GRU), while diamond shaped units are used for input and output. Italic letters (e.g., xt and yt) denote the predicted values, while bold capital (e.g., Xt and Yt) letters denote the actual values.

2.3.5. Self-Correcting Regression RNN with Regularization

All the models described above are classification models because the final result should be a binary value represents whether the patient will develop AKI in the next 6 hours. And the actual AKI label depends on whether the value of urine output/weight is larger than 0.5 mL/kg/h. So, we also designed a self-correcting RNN model with regularization for the urine output regression problem. The structure of this model is the similar to the one shown in Figure 3, except that it predicts the next-6-hour urine output instead of the AKI label and then predicts the label based on the patient’s weight and the predicted urine output. So, it becomes a regression model. In the regression models,Yt is the urine output. And the predicted and actual values of Yt6 are part of the feature vector for t. And the other data (vital signs, lab measurements, etc.) in the feature vector is the same as the classification models. And the mean square error is used for backpropagation.

2.3.6. Stop-Gradient Technique for Feedback Loop

Another challenge with the proposed self-correcting models is that the gradient of yt is affected by the errors at the future time steps. Let Jθ denote the cost function of the parameters θ of the neural network. When we train the RNN models using gradient descent algorithm, the partial derivative J/y^t is first calculated and then backpropagated through the time. In the traditional RNN, the partial derivative of y^t is J/y^t=Jt/y^t [22], where Jt is the cross-entropy loss between y^t and the label yt. The partial derivative of yt is not affected by the loss at the other time steps. This is the desired property of the RNN. In our self-correcting RNN model, the partial derivative of yt is

(3)Jy^t=Jty^t+i=6Tt6Jt+iy^t.

This is because the loss at the future time step can be backpropagated through the RNN layers and feed forward layers, and finally to yt, as shown in Figure 4. And this is not what we desired. Intuitively, the problem is that output layer does not only need to predict y^t accurately but also need to generate the y^t such that the value will later lead to more accurate y^t+6 given the current neural network parameters, since Jt+6/y^t is also a component of J/y^t. This is an undesired property and may potentially affect the performance of the model. The y^t predicted by the model should only be used for the self-correcting mechanism to boost the prediction accuracy at future time step.

Figure 4.

Figure 4

Diagram of backpropagation of the error. Dotted lines indicate the direction of backpropagation, and the dotted lines in red show how the errors in the future time steps are backpropagated to y^ through the feedback loop connection. The green dotted line indicates where we apply the stop-gradient technique.

To tackle this issue, we truncate the gradient right before feeding the feedback into the feedback network, as shown in Figure 4. This is referred to as the stop-gradient technique for the self-correcting models.

3. Results

3.1. Experiment Setup

Systematic experiments were conducted to compare the performance of our proposed models and the previously proposed RNN approaches (the baseline models). Two datasets—the MIMIC-III and the eICU datasets—were chosen to validate our proposed methods as they are representative datasets with the richest critical care EHR data. Our models were trained and tested on these two datasets separately.

In this study, we only included patients who stayed in ICU for at least 12 hours. This criterion was set based on two considerations: (1) patients, who were discharged or died within the first 12 hours of ICU stay, were very unique patients that do not fit our clinical application and (2) our proposed self-correcting mechanism only starts from t=7 (the 7th hour in ICU) onward, and thus we will need a number of time steps for the model to be stabilized. In addition, we have also removed patients whose data for the selected variables was not recorded for at least once during the ICU stay. With these inclusion and exclusion criteria, we ended up with about 25,000 patients out of the 40,000 MIMIC patients and about 11,000 patients out of 200,859 eICU patients. We also eliminated the outliers in the extracted time series data (e.g., negative heart rate and unreasonably high body temperature) by removing the extreme data points at the upper or lower one percentiles. We used 5-fold cross-validation to evaluate the performance of all models.

For each of the two datasets, we trained three variations of our proposed model and also four existing commonly used RNN models as the baseline models. The baseline models included multilayer GRU model, multilayer GRU model with attention mechanism, multilayer LSTM model, and multilayer LSTM model with attention mechanism. We used the self-attention [23] mechanism with same dimension as outputs of each RNN layer. And the three proposed models are self-correcting RNN, self-correcting RNN with regularization, and self-correcting regression RNN with regularization. The proposed self-correcting mechanism added complexity to the RNN. Therefore, to rule out the impacts of added complexity, we built the baseline models with the number of layers from 1 to 3 and the number of neurons in each layer from 64 to 160, shown in Table 1. For each baseline model, we compared the best performance across different architectures with the proposed models built with 2 layers of GRU and 128 neurons in each layer. For all these RNN models, we built the models with the same architecture: two layers of GRU and 128 neurons in each layer. We applied the same settings for dropout, gradient clipping, and reweighted loss function to address the imbalanced dataset. We chose Adam optimization as the optimization method for all the models we trained, as it combines the advantages of Adagrad and RMSProp optimization. We trained all the RNN models with the same initial learning rate and the same decay rate. All the models were trained until they converged. We also adopted dropout method in all the RNN models with the dropout rate of 10% to prevent overfitting. All the seven models converged.

Table 1.

Baseline performance with different numbers of layers and neurons.

Model No. of layers No. of neurons MIMIC eICU
AUC F1 AUC F1
Multilayer GRU (baseline) 1 64 0.761 0.396 0.739 0.735
128 0.769 0.415 0.750 0.742
160 0.773 0.436 0.762 0.746
2 64 0.769 0.415 0.728 0.730
128 0.775 0.428 0.753 0.743
160 0.776 0.440 0.757 0.746
3 64 0.769 0.422 0.723 0.726
128 0.776 0.434 0.744 0.737
160 0.779 0.438 0.753 0.741

Multilayer GRU with attention (baseline) 1 64 0.757 0.394 0.761 0.745
128 0.762 0.422 0.785 0.759
160 0.766 0.429 0.785 0.758
2 64 0.760 0.414 0.758 0.745
128 0.773 0.425 0.793 0.765
160 0.760 0.409 0.803 0.773
3 64 0.763 0.413 0.757 0.745
128 0.768 0.428 0.796 0.768
160 0.765 0.420 0.814 0.781

Multilayer LSTM (baseline) 1 64 0.733 0.268 0.779 0.754
128 0.752 0.393 0.788 0.759
160 0.748 0.384 0.790 0.762
2 64 0.754 0.403 0.781 0.756
128 0.762 0.404 0.795 0.766
160 0.755 0.396 0.803 0.769
3 64 0.754 0.404 0.780 0.756
128 0.755 0.358 0.803 0.770
160 0.758 0.400 0.811 0.777

Multilayer LSTM with attention (baseline) 1 64 0.755 0.383 0.790 0.761
128 0.760 0.403 0.806 0.773
160 0.754 0.394 0.811 0.778
2 64 0.753 0.403 0.808 0.777
128 0.756 0.398 0.823 0.788
160 0.761 0.403 0.824 0.788
3 64 0.747 0.371 0.811 0.779
128 0.747 0.357 0.824 0.789
160 0.752 0.398 0.825 0.788

To ensure the reproducibility of our research, our models were tested on two open datasets, MIMIC and eICU, and all the source code for the proposed models will be released online to public as well.

3.2. Performance Results

For each dataset, we measured the performance of the models based on area under ROC curve (AUC) and F1 score. We only calculated the AUC for time steps t>6, as our proposed self-correcting mechanism starts at t=7. And we got the result using 5-fold cross-validation. The AUCs were reported in Table 2, and the F1 scores of the models were summarized in Table 3.

Table 2.

AUC of the four models on MIMIC-III and eICU.

Model MIMIC eICU
Multilayer GRU (baseline) 0.779 0.762
Multilayer GRU with attention (baseline) 0.773 0.814
Multilayer LSTM (baseline) 0.762 0.811
Multilayer LSTM with attention (baseline) 0.760 0.825
Self-correcting RNN 0.889 0.837
Self-correcting regression RNN with regularization 0.886 0.861
Self-correcting RNN with regularization 0.893 0.871

Proposed methods.

Table 3.

F1 scores of the four models on MIMIC-III and eICU.

Model MIMIC eICU
Multilayer GRU (baseline) 0.440 0.746
Multilayer GRU with attention (baseline) 0.429 0.781
Multilayer LSTM (baseline) 0.404 0.777
Multilayer LSTM with attention (baseline) 0.403 0.789
Self-correcting RNN 0.720 0.780
Self-correcting regression RNN with regularization 0.735 0.803
Self-correcting RNN with regularization 0.738 0.808

Proposed methods.

As shown in the tables, all our proposed models outperformed the baseline RNN models over both the MIMIC and eICU datasets. The self-correcting RNN with regularization achieved the best performance. For the MIMIC dataset, our self-correcting RNN with regularization model achieved 0.893 as AUC and 0.738 as F1 score, corresponding to an over 15% improvement comparing with the best baseline models. For the eICU dataset, our self-correcting RNN with regularization model achieved 0.871 as AUC and 0.808 as F1 score, corresponding to close to 10% improvement.

4. Discussions

4.1. The Benefits of Self-Correcting Mechanism

Similar RNN methods, such as directed acyclic graph LSTM (DAG-LSTM) and graph state LSTM, have been applied to natural language processing (NLP) tasks [24, 25]. Compared with DAG-LSTM, the proposed self-correcting RNN does not include bidirectional RNN layers that will introduce patient information in the future. In addition, compared with the graph state LSTM that models state transitions for each word, the self-correcting mechanism uses the information from t6 to t1 to predict the onset of AKI at time t.

All the three self-correcting models outperformed the traditional baseline RNN models. From the results on the MIMIC-III dataset, there is a huge difference between the AUC of self-correcting RNN and the traditional multilayer GRU. It is because the additional information provided by the feedback network is helpful for the RNN update process. These results verified our hypothesis that the self-correcting mechanism could boost our model’s performance.

4.2. The Benefits of the Proposed Regularization Method

Self-correcting RNN with regularization achieved the highest AUC. And on the eICU dataset, the self-correcting regression RNN with regularization also achieved much higher AUC than the self-correcting RNN model. It indicates that the proposed regularization method helps to further improve the performance of the models by enforcing the model to predict the future input. Our experiment on the eICU dataset verifies that the self-correcting RNN model with regularization has a smaller performance gap between training and testing data, as compared to the one without regularization. It proves that the regularization method reduces the variance of the model.

4.3. Why Attention Mechanism Does Not Work

In terms of AUC, the models with attention perform slightly better on eICU than the ones without it but achieved worse results on MIMIC-III. It shows that attention does not help in our case. Attention mechanism is generally helpful in the cases where the output at a time step largely depends on the input at one or a few previous time steps. However, in our case, the output (AKI status) needs to be deduced from the trend of input data and is not just related to the input value at one time step. RNN is able to capture the trend of the data in the hidden states. And our proposed self-correcting mechanism is able to correct the error in the hidden states. So, it works better in our case.

4.4. The Benefit of the Stopped Gradient for the Feedback

To verify that the benefits from the proposed stop-gradient technique, we trained a self-correcting RNN with regularization on MIMIC-III without applying the stop-gradient technique. The AUC of this model was 0.852, while the AUC was 0.893 with the stop-gradient technique. The difference in the AUC verified that the stop-gradient technique is critical to the self-correcting models, because it prevent the gradient of yt from being affected by the future errors.

4.5. The Implications of the Proposed Method

The proposed method has the potential to be adopted in clinical applications to prevent AKI and allow intervention to occur in a timely manner. In the clinical settings, the intensivists need to monitor several clinical measurements and prognosis to detect AKI early. The proposed method demonstrated effectiveness in modeling high-dimensional, time series data to alert the clinicians the onset of AKI. The alert is personalized for each patient by modelling the prognosis of multiple clinical variables and the alert time window could be as early as 6 hours, which may enable clinicians to prevent AKI with early treatments. Further studies will be required to validate the effectiveness of the proposed method in prospective trials.

5. Conclusions

We proposed a novel self-correcting enhancement to RNNs to better predict the onset of acute conditions in ICU. The proposed self-correcting mechanism made the update process of the hidden state of RNN to be dependent on the previous predicted output and the corresponding label. The additional information provided by the previous predicted output and label helped to boost the performance of model. We also proposed a regularization method for our model, which takes into account not only the model’s prediction errors on the labels but also its estimation errors on the input data. The regularization method reduces the variance of the model and also reduces the time gap for the self-correcting mechanism. The proposed model can be applied on both classification and regression tasks. Our proposed models were tested on real-world large-scale ICU datasets MIMIC-III and eICU and were shown to constantly outperform the baseline multilayer GRU model. Moreover, although we focus on the prediction of acute kidney injury as an example, the proposed model can be easily generalized to predict other acute conditions in ICU. This is the first phase of our project. Inspired by the achieved promising results, we plan to move on further validate the proposed algorithm in our local hospital with the ultimate goal to deploy it as an decision support tool.

Acknowledgment

This work is supported by the National University of Singapore start-up grant with award number R-608-000-172-133.

Data Availability

The MIMIC and eICU datasets used in this study are freely available databases for critical care research. The databases could be accessed via https://physionet.org/content/mimiciii/1.4/ and https://eicu-crd.mit.edu/. The source codes for the experiments are publicly available at https://github.com/nus-mornin-lab/AKI_HDS.

Authors’ Contributions

H.D. contributed to the study design, data extraction, model design and experiments, data analysis, and manuscript preparation. Z.P contributed to the study design, data extraction, model design and experiments, data analysis, and manuscript preparation. K.Y.N contributed to the study design and clinical data interpretation. F.W contributed to the model design and experiments, data analysis, and manuscript preparation. P.S contributed to the model design and experiments, data analysis, and manuscript preparation. M.F contributed as a Project Director and Principal investigator and contributed to the study design, model design and experiments, and manuscript preparation. All authors have approved the final version of the manuscript submitted. Hao Du and Ziyuan Pan contributed equally to this work.

References

  • 1.Harty J., “Prevention and management of acute kidney injury,” Ulster Medical Journal, vol. 83, no. 3, p. 149, 2014 [PMC free article] [PubMed] [Google Scholar]
  • 2.Chertow G. M., Burdick E., Honour M., Bonventre J. V., and Bates D. W., “Acute kidney injury, mortality, length of stay, and costs in hospitalized patients,” Journal of the American Society of Nephrology, vol. 16, no. 11, pp. 3365–3370, 2005 [DOI] [PubMed] [Google Scholar]
  • 3.Sutherland S. M., Goldstein S. L., and Bagshaw S. M., “Leveraging big data and electronic health records to enhance novel approaches to acute kidney injury research and care,” Blood Purification, vol. 44, no. 1, pp. 68–76, 2017 [DOI] [PubMed] [Google Scholar]
  • 4.Colpaert K., Hoste E. A., Steurbaut K., Benoit D., Van Hoecke S., De Turck F., and Decruyenaere J., “Impact of real-time electronic alerting of acute kidney injury on therapeutic intervention and progression of RIFLE class,” Critical Care Medicine, vol. 40, no. 4, pp. 1164–1170, 2012 [DOI] [PubMed] [Google Scholar]
  • 5.Weisenthal S., Liao H., Ng P., and Zand M., “Sum of previous inpatient serum creatinine measurements predicts acute kidney injury in rehospitalized patients,”, 2017,https://arxiv.org/abs/1712.01880.
  • 6.Thakar C. V., Arrigain S., Worley S., Yared J.-P., and Paganini E. P., “A clinical score to predict acute renal failure after cardiac surgery,” Journal of the American Society of Nephrology, vol. 16, no. 1, pp. 162–168, 2005 [DOI] [PubMed] [Google Scholar]
  • 7.Palomba H., de Castro I., Neto A., Lage S., and Yu L., “Acute kidney injury prediction following elective cardiac surgery: Akics score,” Kidney International, vol. 72, no. 5, pp. 624–631, 2007 [DOI] [PubMed] [Google Scholar]
  • 8.Brown J. R., Cochran R. P., Leavitt B. J., Dacey L. J., Ross C. S., MacKenzie T. A., Kunzelman K. S., Kramer R. S., Hernandez Jr F., Helm R. E., Westbrook B. M., Dunton R. F., Malenka D. J., O'Connor G. T., and Northern New England Cardiovascular Disease Study Group, “Multivariable prediction of renal insufficiency developing after cardiac surgery,” Circulation, vol. 116, no. 11_supplement, pp. I139–I143, 2007 [DOI] [PubMed] [Google Scholar]
  • 9.Mohamadlou H., Lynn-Palevsky A., Barton C., Chettipally U., Shieh L., Calvert J., Saber N. R., and Das R., “Prediction of acute kidney injury with a machine learning algorithm using electronic health record data,” Canadian Journal of Kidney Health and Disease, vol. 5, 2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Miotto R., Wang F., Wang S., Jiang X., and Dudley J. T., “Deep learning for healthcare: review, opportunities and challenges,” Briefings in Bioinformatics, vol. 19, no. 6, pp. 1236–1246, 2018 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Hochreiter S., and Schmidhuber J., “Long short-term memory,” Neural Computation, vol. 9, no. 8, pp. 1735–1780, 1997 [DOI] [PubMed] [Google Scholar]
  • 12.Cho K., van Merrienboer B., Gulcehre C., Bahdanau D., Bougares F., Schwenk H., and Bengio Y., “Learning phrase representations using RNN encoder-decoder for statistical machine translation,”, 2014,https://arxiv.org/abs/1406.1078.
  • 13.Pham T., Tran T., Phung D., and Venkatesh S., “Deepcare: a deep dynamic memory model for predictive medicine,” Advances in Knowledge Discovery and Data Mining. PAKDD 2016, Bailey J., Khan L., Washio T., Dobbie G., Huang J., and Wang R., Eds., Springer, Cham, vol. 9652, Lecture Notes in Computer Science, pp. 30–41, 2016 [Google Scholar]
  • 14.Cruz H., Grasnick B., Dinger H., Bier F., and Meinel C., “Early detection of acute kidney injury with bayesian networks,” in Proceedings of the 7th International Symposium on Semantic Mining in Biomedicine, SMBM 2016, Potsdam, Germany, 2016, pp. 29–36 [Google Scholar]
  • 15.Qiao Z., Zhao S., Xiao C., Li X., Qin Y., and Wang F., “Pairwise-ranking based collaborative recurrent neural networks for clinical event prediction,” in Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI), Stockholm, Sweden, 2018, pp. 3520–3526 [Google Scholar]
  • 16.Bellomo R., Ronco C., Kellum J. A., Mehta R. L., Palevsky P., and Acute Dialysis Quality Initiative workgroup, “Acute renal failure - definition, outcome measures, animal models, fluid therapy and information technology needs: the second international consensus conference of the acute dialysis quality initiative (adqi) group,” Critical care, vol. 8, no. 4, pp. R204–R212, 2004 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Johnson A. E., Pollard T. J., Shen L., Lehman L. W. H., Feng M., Ghassemi M., Moody B., Szolovits P., Anthony Celi L., and Mark R. G., “MIMIC-III, a freely accessible critical care database,” Scientific Data, vol. 3, no. 1, p. 160035, 2016 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Moody G. B., Mark R. G., and Goldberger A. L., “PhysioNet: a web-based resource for the study of physiologic signals,” IEEE Engineering in Medicine and Biology Magazine, vol. 20, no. 3, pp. 70–75, 2001 [DOI] [PubMed] [Google Scholar]
  • 19.AM AssociationInternational Classification of Diseases, 9th revision, clinical modification: physician ICD-9-CM, 2005: volumes 1 and 2, color-coded, illustrated, Amer Medical Assn, vol. 1, 2004 [Google Scholar]
  • 20.Fraccaro M., Sønderby S. K., Paquet U., and Winther O., “Sequential neural models with stochastic layers,”, https://arxiv.org/abs/1605.07571.
  • 21.Chung J., Kastner K., Dinh L., Goel K., Courville A. C., and Bengio Y., “A recurrent latent variable model for sequential data,”, http://arxiv.org/abs/1506.02216.
  • 22.Chen G., “A gentle tutorial of recurrent neural network with error backpropagation,”, http://arxiv.org/abs/1610.02583.
  • 23.Vaswani A., Shazeer N., Parmar N., Uszkoreit J., Jones L., Gomez A. N., Kaiser L., and Polosukhin I., “Attention is all you need,” Advances in Neural Information Processing Systems, pp. 5998–6008, 2017
  • 24.Song L., Zhang Y., Wang Z., and Gildea D., “N-ary relation extraction using graph state LSTM,”, 2018,https://arxiv.org/abs/1808.09101.
  • 25.Peng N., Poon H., Quirk C., Toutanova K., and Yih W.-T., “Cross-SentenceN-ary relation extraction with graph LSTMs,” Transactions of the Association for Computational Linguistics, vol. 5, pp. 101–115, 2017 [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The MIMIC and eICU datasets used in this study are freely available databases for critical care research. The databases could be accessed via https://physionet.org/content/mimiciii/1.4/ and https://eicu-crd.mit.edu/. The source codes for the experiments are publicly available at https://github.com/nus-mornin-lab/AKI_HDS.


Articles from Health Data Science are provided here courtesy of AAAS Science Partner Journal Program

RESOURCES