Skip to main content
IEEE Journal of Translational Engineering in Health and Medicine logoLink to IEEE Journal of Translational Engineering in Health and Medicine
. 2019 Nov 12;8:1900111. doi: 10.1109/JTEHM.2019.2952610

A Multi-Task Group Bi-LSTM Networks Application on Electrocardiogram Classification

Qiu-Jie Lv 1, Hsin-Yi Chen 1,, Wei-Bin Zhong 1, Ying-Ying Wang 2, Jing-Yan Song 2, Sai-Di Guo 2, Lian-Xin Qi 2, Calvin Yu-Chian Chen 1,3,4
PMCID: PMC7028438  PMID: 32082952

Abstract

Background: Cardiovascular diseases (CVD) are the leading cause of death globally. Electrocardiogram (ECG) analysis can provide thoroughly assessment for different CVDs efficiently. We propose a multi-task group bidirectional long short-term memory (MTGBi-LSTM) framework to intelligent recognize multiple CVDs based on multi-lead ECG signals. Methods: This model employs a Group Bi-LSTM (GBi-LSTM) and Residual Group Convolutional Neural Network (Res-GCNN) to learn the dual feature representation of ECG space and time series. GBi-LSTM is divided into Global Bi-LSTM and Intra-Group Bi-LSTM, which can learn the features of each ECG lead and the relationship between leads. Then, through attention mechanism, the different lead information of ECG is integrated to make the model to possess the powerful feature discriminability. Through multi-task learning, the model can fully mine the association information between diseases and obtain more accurate diagnostic results. In addition, we propose a dynamic weighted loss function to better quantify the loss to overcome the imbalance between classes. Results: Based on more than 170,000 clinical 12-lead ECG analysis, the MTGBi-LSTM method achieved accuracy, precision, recall and F1 of 88.86%, 90.67%, 94.19% and 92.39%, respectively. The experimental results show that the proposed MTGBi-LSTM method can reliably realize ECG analysis and provide an effective tool for computer-aided diagnosis of CVD.

Keywords: ECG, bidirectional long short-term memory network, attention mechanism, multi-task learning


The Group Bi-LSTM and Residual Group CNN (Res-GCNN) were utilized to build classification model to realize intelligent assistant diagnosis of CVD. The result showed the proposed MTGBi-LSTM method can reliably realize any multi-lead ECG (2-lead or more) analysis and provide an effective tool for computer-aided diagnosis of CVD.

Download video file (27.6MB, mp4)

I. Introduction

Cardiovascular disease is the leading cause of death around the world [1]. The total number of deaths caused by CVD increased from 12.3 million (25.8%) in 1990 to 17.9 million (32.1%) in 2015. Deaths caused by CVD are more prevalent at certain ages and are increasing in most developing countries [2], [3]. It is estimated that up to 90% of CVD may be preventable [4], [5]. Most CVDs can be diagnosed by ECG data. ECG is time-series physiological data. The electrical signal is generated by the sinoatrial node, which passes through the atrioventricular node and Purkinje fibers, and finally reaches the ventricle [6].

The automatic analysis of ECG is of great significance to the early detection and treatment of CVD. With the development of AI, more and more deep learning [7] methods are applied to address medical data [8][10], such as feedforward neural network [11][14], and the Recurrent Neural Networks (RNN) [15] (Long Short-Term Memory (LSTM) [16], [17], Gate Recurrent Unit (GRU) [18]). However, most of previous studies used the MIT-BIH-AR database [19][21], which only has 48 patients. And they only used one-lead ECG data for model training. Since the same person’s ECG data fragments are highly correlated [14], [22], this leads to high accuracy. Once other databases are used or are actually used in clinical practice, the performance of their models will decline dramatically and are not robust.

The multi-lead ECG can provide a full evaluation of the cardiac electrical activity (includes arrhythmias, myocardial infarction, conduction disturbances) [23]. Each lead reflects the electrical signal changes in different parts of the heart. The diagnosis of CVD often needs to be combined with the change of electrical signals in different parts of the heart. This means that each lead signal and correlation among the leads have the same importance for intelligent analysis of ECG.

The accuracy of these methods is not high, and these studies are simply predicting a few types of CVD categories [23], [24]. These methods use the depth of the model in exchange for the performance improvement of the model, without exploring the essential characteristics of the ECG. Previous works based on multi-lead ECG analysis did not pay attention to the correlation among different lead information.

The main contributions of this work can be concluded as follows

  • 1.

    The number of leads for existing ECG data is not uniform, so we design a Multi-Task Group Bi-LSTM (MTGBi-LSTM), which uses any multi-lead ECG (2-lead or more) learning to simulate the assistant diagnostic process of cardiologists. In the training phase, the model can change the number of its groups according to the number of ECG leads input. This is very promising and influential for the clinical environment.

  • 2.

    We propose a Group Bi-LSTM module, a combination of Global Bi-LSTM and Intra-Group Bi-LSTM, which can learn each lead individually and fuse the global information of different leads, especially it can sparse autoencoding ECG data. Its effectively enhances the expressiveness of signal representation with respect to ECG waveform features.

  • 3.

    Last but not least, this is the first time that a multi-task framework has been applied to the CVD-assisted diagnosis. In addition, an attention mechanism for the temporal characteristics of ECG is proposed, which provides discriminative feature representation for multi-task module. And then a dynamic weighted loss function is proposed to solve the unbalanced proportion of positive and abnormal diseases.

II. Methods

For multi-lead ECGs, we propose a Multi-Task Group Bi-LSTM model framework for deep mining of ECG timing features implicit in each lead and between individual leads. On this basis, attention mechanism is added to enable the model to retain key information with emphasis on the premise of guaranteeing comprehensiveness. In addition, through multi-task learning, the model can fully mine the association information between diseases and obtain more accurate diagnostic results. The model architecture is shown in Fig. 1 and Table 1.

FIGURE 1.

FIGURE 1.

Multi-Task Aided Diagnosis Model Based on Group Bi-LSTM. The Global Bi-LSTM (colored in blue) transmits the hidden state of each time step to Intra-group Bi-LSTM (colored in yellow) through the global access structure (colored in red). Inline graphic and Inline graphic obtained after the output feature of Bi-LSTM passes attention structure, the final fully connected layer (FC_1, FC_2) respectively generate the category distribution of two tasks.

TABLE 1. Architectures for MTGBi-LSTM.

layer name Output size MTGBi-LSTM
input Inline graphic
Res-GCNN conv1.x Inline graphic Inline graphic, 64, stride Inline graphic
Inline graphic Inline graphic, 64, stride Inline graphic
Inline graphic Inline graphic, 256, stride Inline graphic
conv2.x Inline graphic Inline graphic, 128, stride Inline graphic
Inline graphic Inline graphic, 128, stride Inline graphic
Inline graphic Inline graphic, 512, stride Inline graphic
conv3.x Inline graphic Inline graphic, 256, stride Inline graphic
Inline graphic Inline graphic, 256, stride Inline graphic DC, dilation 2,256, stride Inline graphicx3 DC, dilation 3,256, stride Inline graphic avg_pool,1x2, stride Inline graphic
Inline graphic Inline graphic, 512, stride Inline graphic
avg_pool Inline graphic Inline graphic, stride Inline graphic
reshape Inline graphic
Group Bi-LSTM The Global Bi-LSTM The Intra-Group Bi-LSTM
Global Bi-LSTM Inline graphic 256-d Intra-Group Bi-LSTM Inline graphic 256-d
Attention Inline graphic Attention Inline graphic
Concat Inline graphic
Multi-Task Learning Task 1 Task 2
FC_1 Inline graphic 512-d FC_2 Inline graphic 512-d

A. Residual Group Convolution Neural Network (Res-GCNN)

The different lead data of the ECG is divided into different groups for Res-GCNN, and the number of leads of the ECG is the number of groups that need to be grouped. Such as, the ECG data is input to the top convolution layer in the size of Inline graphic is the number of leads of ECG, this resulted in the number of Res-GCNN being 12.

The Res-GCNN module of each group is composed of contraction path and expansion path, as show in Fig. 2. And more structural details are in Table 1. In the contraction path, a deep residual network with three residual blocks [25] is designed to encode the single lead data of ECG. The first two residual module has three continuous convolutions layers with kernel size of Inline graphic, Inline graphic, and Inline graphic, and each convolution layer is followed by a rectified linear unit ReLU [26] and Batch Normalization [27] to improve sparsely. In the extended path, the last layer of the last residual block uses three filters of different sizes, which are Inline graphic, Inline graphic dilated convolution (1x3DC, dilation rate is set to 2), Inline graphic DC (dilation rate: 3), and an avg pool to convolute and pool the input data. The design of Res-GCNN module reduces the network parameters and the thickness of the feature map, and makes the feature map keep more high-order features at different levels, enriching the expressive ability of the network [28]. Finally, all sub-layer outputs within each group are cascaded and transmitted to the Group Bi-LSTM section.

FIGURE 2.

FIGURE 2.

The architecture of the Res-GCNN module. The “X3” in the figure represents three consecutive convolutional layers with kernel sizes of Inline graphic, Inline graphic, and Inline graphic. “X2” represents two consecutive similar residual modules.

B. Group Bi-LSTM Module

Group Bi-LSTM is divided into two parts, the global Bi-LSTM and the intra-group Bi-LSTM. The Group Bi-LSTM Module can dynamically change the number of its groups according to the number of ECG leads input, and combine the global Bi-LSTM and intra-group Bi-LSTM autoencoder signal representation to provide high-resolution feature information for the final CVD-assisted diagnosis and recognition.

1). The Global Bi-LSTM

The global Bi-LSTM updates the cell states by combining the long-term state information of the up and down time, after receiving the feature information of all the leads Inline graphic at the current time, and synchronously sends the hidden state Inline graphic of the output to the each intra-group Bi-LSTM. There is only one global Bi-LSTM in the Group Bi-LSTM Module.

In Fig. 3, the global Bi-LSTM removes or adds information to the cell states through three “gates” structure (forget gate, input gate and output gate) [29]. Gate is a way to allow information to pass selectively, including a sigmoid neural network layer and a pointwise multiplication operation. The feature sequence Inline graphic after Res-GCNN is used as the input of global Bi-LSTM.

FIGURE 3.

FIGURE 3.

Cell Structure of Global Bi-LSTM. We employed a peephole connections, i.e., the extra blue connections.

The function of forget gate is to selectively discard the feature information of cell state. After inputting the spatial features of the ECG, we hope to pay more attention to the characteristic information of P wave, QRS wave and T wave. In the whole ECG, some relatively flat features may not affect the classification results, so we need to selectively discard them to extract the essential feature information of ECG.

1).

where Inline graphic is the input value in LSTM, that is, the latent discriminative structural feature maps is extracted by the i-th Res-GCNN module. Inline graphic is the number of leads of ECG. Inline graphic, Inline graphic represent the cell state at the previous moment and the output of the previous cell, respectively. Inline graphic, Inline graphic are the parameters to be learned by LSTM. Inline graphic represents the Sigmoid activation function.

The function of the input gate is to control whether the feature information input at the current time is integrated into the memory unit. When it receives a certain feature of the ECG, it may be important for the prediction of the entire ECG or not.

1).

where Inline graphic are the parameters to be learned by LSTM. Inline graphic represents the state of the cell at the current moment. Inline graphic represents the Tanh activation function. The rest of the symbols have the same meaning as Eq. (1).

The purpose of the output gate Inline graphic is to generate a hidden layer unit Inline graphic from the memory unit Inline graphic. However, not all information in Inline graphic is related to the hidden layer unit Inline graphic. Inline graphic may contain a lot of information that is useless to Inline graphic. Therefore, the role of Inline graphic is to determine which parts of Inline graphic are useful for Inline graphic and which parts are useless.

1).

where Wo, bo are the parameters to be learned by LSTM. Inline graphic represents the output of the cell at the current moment.

2). The Intra-Group Bi-LSTM

The intra-group Bi-LSTM only accepts the feature map information of a single lead Inline graphic. After introducing the inter-lead feature information Inline graphic from the global Bi-LSTM through the global access structure, updates the state of the cell by combining the ECG information at the time before and after. The input of ECG data has n- lead, and the Group Bi-LSTM Module has Inline graphic Intra-Group Bi-LSTM.

In Fig. 4, the intra-group Bi-LSTM removes or adds information to the cell state through three “gate” structures. The forward LSTM in the intra-group Bi-LSTM first accepts the hidden state of the forward LSTM in the global Bi-LSTM at the current time, and then combines with the current input sequence feature data and the hidden state of the previous moment to remove and update the cell state at the current time. Another reverse LSTM is similar.

FIGURE 4.

FIGURE 4.

Cell Structure of Intra-group Bi-LSTM. Inline graphic is the inter-lead feature information introduced from the global Bi-LSTM through the global access structure.

The intra-group Bi-LSTM network computes a mapping from an input sequence Inline graphic to an output sequence Inline graphic by calculating the network unit activations using the following equations iteratively from Inline graphic 1 to Inline graphic:

2).

where Inline graphic represents the input value in the intra-group Bi-LSTM, in other words, ts the latent discriminative structural feature maps is extracted by the i-th Res-GCNN module. It must be noted that it is distinguished from Inline graphic, the input of Group Bi-LSTM module, in Eq. (1). Inline graphic represents the hidden state of the forward LSTM in the Global Bi-LSTM at the current time, that is, the output state of the cell in the global Bi-LSTM in Eq. (1). Inline graphic, Inline graphic represents the Sigmoid and the Tanh activation function, respectively.Inline graphic represents the element level product operation. Inline graphic are the parameters to be learned by LSTM.

C. The Attention Mechanism

It must be noted that we employs the attention mechanism [30] on every Bi-LSTM (including the Global Bi-LSTM, Intra-Group Bi-LSTM). The Attention mechanism of this model includes two parts of calculation process. One part is about the process of calculating the probability distribution of attention, the other part is the process of calculating the final feature based on attention distribution (Fig. 5).

FIGURE 5.

FIGURE 5.

Bi-LSTM Model with Attention Mechanism. Inline graphic represents the sum of the final hidden layer state values for each independent direction in Bi-LSTM, which is called the final state of Bi-LSTM. Inline graphic denotes the attention probability distribution of the hidden layer unit state to the final state at all times. Inline graphic denotes the final ECG feature vector weighted by attention in Bi-LSTM.

Inline graphic represents the sum of the final hidden layer state values for each independent direction in Bi-LSTM, which is called the final state of Bi-LSTM.Inline graphic denotes the attention probability distribution of the hidden layer unit state to the final state at all times, and component Inline graphic denotes the attention probability of Bi-LSTM state Inline graphic to the final state at Inline graphic times. Inline graphic is obtained by adding the states of the each independent directions at that time. Inline graphic denotes the ECG feature vector weighted by attention in Bi-LSTM.

  • 1)
    The attention probability of the output data for the final state at the moment Inline graphic. Enter the hidden state of the Bi-LSTM network layer at each moment into the attention module to generate a weight vector. The weight matrix is got by the following formula:
    graphic file with name M109.gif

    The formula uses the softmax function as the calculation method of attention probability distribution. Inline graphic denotes the number of input sequence elements. Inline graphic is the weight matrix. Inline graphic represents the sum of the final hidden layer state values for each independent direction in Bi-LSTM.Inline graphic represents the sum of two-way hidden layer state values at the moment Inline graphic.

  • 2)
    Inline graphic, the final feature based on attention distribution, is expressed as:
    graphic file with name M116.gif

Inline graphic denotes the number of input sequence elements. Inline graphic represents the attention probability of the output data for the final state at the moment Inline graphic. Inline graphic represents the sum of hidden layer states in two independent directions at the moment Inline graphic.

D. Dynamic Weighted Loss

Considering that in the field of ECG research, since most subjects have normal heartbeats and a small proportion of lesions, the ECG dataset is generally unbalanced. As shown in the distribution of ECGs in Table 3, N class accounted for 61.10% of the complete data set, SA, AA accounted for 18.14% and 7.72%, respectively, while the remaining 12 categories accounted for only 13.04% of the total. Therefore, there is an urgent need for a robust and stable classification method that can cope with different data sets.

TABLE 3. Data Distribution of Disease Types in Task 1.

Task1 Train Set Val Set Test Set Total
Normal 87064 10883 10883 108830
AR 47570 5946 5950 59466
MI 451 56 57 564
VH 6989 875 875 8739
AH 407 51 52 510
Total 142481 17811 17817 178109

In order to solve the problem of this unbalanced data set, most researchers use data augmentation techniques to expand the size of data sets [31]. However, this may take more training time, and for the examples that are difficult to classify (hard examples), the loss corresponding to the training will be very small, resulting in a small gradient change in the inverse calculation, and the convergence of the parameters is limited. Compared with other examples, we need to make this sample produce relatively large loss values.

Our method uses raw input data without adding any extra data. To increase the loss of the hard example, we adjusted the weights of the hard example and easy example. We recommend using the new dynamic weighted loss function defined below. First, we define the set of Inline graphic heartbeat labels in the Inline graphic-th batch as Inline graphic, and the Inline graphic-th label in Inline graphic as Inline graphic.

D.

where Inline graphic represents different categories in different tasks, Inline graphic {N, AR, MI, VH, AH} in task one. Inline graphic {N, SA, AA, JA, VA, HB, AMI, IMI, LMI, LVH, RVH, BVH, LAH, RAH, BAH} in task two. Then, the loss weight of the Inline graphic-th class in the Inline graphic-th batch, Inline graphic is calculated as follows:

D.

Among them, Inline graphic is the batch size and Inline graphic is set to 0.01 to prevent the loss weight equal to 0 when Inline graphic contains only one class. Finally, the weighted cross-entropy loss function of the first batch is calculated, as shown in formula (18).

D.

where Inline graphic are the predictive probability of the Inline graphic-th training instance in the first batch, Inline graphic is the weight matrix of all layers, and Inline graphic is the L2 regularization parameter, which is set to 0.01 in our experiment.

E. Multi-Task Learning

The single-task model focuses on a single task, ignoring other information that may help optimize metrics among diseases. The module trains two tasks related to ECG prediction in parallel, and the feature representation of one task can also be used by other tasks. It promotes the two ECG tasks to learn together. By sharing the network structure before the multi-task module [32], two tasks are trained in parallel using the shared representation. Each task keeps the relevant output on its own path. This resulted the parameters of ECG classification model are reduced, the phenomenon of over-fitting is avoided to some extent, and the generalization ability of the model is improved.

Combining the feature vectors learned from the group Bi-LSTM module. The final fully connected layer respectively generate the category distribution in each time step of two tasks, and at the same time get the loss function of the two tasks. Combining these two loss values according to different weights, the Adam optimizer algorithm is responsible for minimizing the joint loss value. The final classification loss is as follows:

E.

where Inline graphic is a parameter that interpolates between the task1 and task2 losses, Inline graphic, Inline graphic represent the loss functions of task 1 and task 2, respectively.

F. Dataset

The Chinese Cardiovascular Disease Database (CCDD), the largest ECG database all over the world, is developed by the Chinese Academy of Sciences [33]. More than 18,000 recordings in this database are collected from real world with high quality, each ECG record corresponds to one patient, which has been labeled by professional doctors. The records in the database are all 12-lead ECGs, each with a duration of about 10s, digitized at a rate of 500 sampling points per second. More details of screening patient records were contributed in supplementary information 3. Table 2 shows the data distribution of the ECG data sets after screening.

TABLE 2. Data Distribution of Different Disease Types.

Task1 Task2 Number of Records
Normal Rhythm(Normal) Normal Rhythm 108830
Arrhythmia(AR) Sinus Arrhythmia(SA) 32306
Atrial Arrhythmia(AA) 13742
Junctional Arrhythmia(JA) 404
Ventricular Arrhythmia(VA) 3883
Heart Block(HB) 9131
Myocardial Infarction(MI) Anterior Myocardial Infarction(AMI) 132
Inferior Myocardial Infarction(IMI) 320
Lateral Myocardial Infarction(LMI) 112
Ventricular Hypertrophy(VH) Left Ventricular Hypertrophy(LVH) 8531
Right Ventricular Hypertrophy(RVH) 115
Bilateral Ventricular Hypertrophy(BVH) 93
Atrial Hypertrophy(AH) Left Atrial Hypertrophy(LAH) 210
Right Atrial Hypertrophy(RAH) 216
Bilateral Atrial Hypertrophy(BAH) 84
Total 178109

Each ECG has two different label values, which are used as the label values of two tasks. Task1 was labeled as arrhythmia, myocardial infarction, ventricular hypertrophy, atrial hypertrophy and normal rhythm. Each ECG tag was labeled as normal rhythm and 14 subtypes of diseases corresponding to the four diseases in Task 1, including sinus arrhythmia, heart block, anterior myocardial infarction, left ventricular hypertrophy, etc. These 15 subtypes were labeled as Task2 label. The specific ECG distribution is shown in Table 3 and Table 4.

TABLE 4. Data Distribution of Disease Types in Task 2.

Task2 Train Set Val Set Test Set Total
Normal 87064 10883 10883 108830
SA 25844 3231 3231 32306
AA 10993 1374 1375 13742
JA 323 40 41 404
VA 3106 388 389 3883
HB 7304 913 914 9131
AMI 105 13 14 132
IMI 256 32 32 320
LMI 90 11 11 112
LVH 6824 853 854 8531
RVH 91 12 12 115
BVH 74 10 9 93
LAH 168 21 21 210
RAH 172 22 22 216
BAH 67 8 9 84
Total 142481 17811 17817 178109

III. Results and Discussion

In this section, we will present the results of the above experiments to verify the effectiveness of MTGBi-LSTM in multi-lead ECG classification tasks and its advantages over existing CNN-based or RNN-based methods.

A. Performance of MTGBi-LSTM

The model training process was evaluated by using the validation set. The classification accuracy of ECG validation set has been basically smooth and stable after 60000 iterations (Fig. 6). The model has stronger information extraction and fitting ability and its performance approximates the optimal.

FIGURE 6.

FIGURE 6.

Validation performance graph of comparison networks.

The results in Table 5 show that the accuracy, precision, recall, and F1 of MTGBi-LSTM reached the best scores of 0.8886, 0.9067, 0.9419, and 0.9239, respectively. Fig. 7 shows a confusion matrix for the classification results of the MTGBi-LSTM model on the clinical ECG test set. The normal and 14 abnormal rhythms in ECG data set are classified and the best recognition results are obtained. This comparative experiment proves the validity of the MTGBi-LSTM framework proposed by us, which can deeply mine the essential features of ECG and accurately predict it.

TABLE 5. Experimental Results of Various Models.

Class Baseline Bi-LSTM Group Bi-LSTM AGBi-LSTM A&LGBi-LSTM MTGBi-LSTM
Normal 0. 8709 0.9106 0.9297 0.9314 0.9354
SA 0. 7356 0.8329 0.8261 0.8289 0.8458
AA 0. 6053 0.7502 0.8209 0.8272 0.8263
JA 0. 5333 0.4706 0.5052 0.5361 0.5869
VA 0. 5244 0.6024 0.6355 0.6793 0.6918
HB 0. 6513 0.7865 0.7825 0.7965 0.8462
AMI 0. 3846 0.5517 0.4545 0.5218 0.5333
IMI 0 3517 0.4118 0.6572 0.75 0.7188
LMI 0. 3333 0.5 0.5263 0.375 0.4210
LVH 0. 6017 0.6524 0.7137 0.7522 0.7796
RVH 0. 5555 0.6316 0.5883 0.6667 0.6316
BVH 0. 3636 0.56 0.5556 0.5 0.5
LAH 0. 2727 0.4889 0.4 0.5238 0.4390
RAH 0. 4000 0.5263 0.6829 0.6154 0.6487
BAH 0.5000 0.5882 0.5883 0.6667 0.6252
Aggregate Results
Accuracy 0.7841 0.8497 0.8721 0.878 0.8886
Precision 0.8252 0.8720 0.8965 0.9005 0.9067
Recall 0.8724 0.9224 0.9245 0.9307 0.9419
F1 0.8481 0.8965 0.9103 0.9154 0.9239

FIGURE 7.

FIGURE 7.

Normalized confusion matrix of model results as percentages for the 15-classproblems. The percentage of all possible records in each category is displayed on a color gradient scale.

The results obtained from the MTGBi-LSTM model are analyzed in detail. As shown in Table 5, the recognition rate of 14 abnormal diseases in the test set reached 85.92%. Among them, the recognition rate of arrhythmia was 87.95%, and the rate of being misclassified as normal category was 4.44%. The model can screen arrhythmias effectively.

The classification and recognition rate of ventricular hypertrophy was 75.77%. The diagnostic rate of being misclassified as normal category was 9.6%. The experimental results show that the model can detect the status of organic lesions in time and warn the subjects to do further examination to find out the potential causes.

ECG data of myocardial infarction and atrial hypertrophy are too little and the data distribution is extremely unbalanced. For both types of identification, the total recall reached 85.92%, only 7 abnormal records are considered normal by the MTGBi-LSTM model. It urges the subjects to do more detailed examinations in order to determine whether they have cardiovascular disease and has early warning effect.

In many cases, the location of the disease is ambiguous and the abnormal rhythm is very similar. Because of the lack of information about subjects, limited signal duration, cardiologists cannot draw reasonable conclusions from ECG. Similar factors also exist in the analysis of ECG which is partially misclassified.

B. Ablation Study of Framework Component

In order to intuitively show the contribution of the different module to the classification performance, we compared various Bi-LSTM variant models. Baseline Bi-LSTM is a combination of Res-GCNN and common Bi-LSTM. The Group Bi-LSTM is based on the Base Bi-LSTM structure and replaces the common Bi-LSTM with the Group Bi-LSTM module. AGBi-LSTM adds attention mechanism based on Group Bi-LSTM. A&LG-LSTM adds dynamic weighted loss based on AGBi-LSTM, and MTGBi-LSTM adds multi-task learning based on A&LGBi-LSTM. In Table 5, the F1 score that are evaluated under multiple comparative case studies with same testing ECG sample data.

1). Group Bi-LSTM Improves Prediction

The results show that the performance of Group Bi-LSTM is better than that of Baseline Bi-LSTM, in specific, precision, recall and F1 score of Group Bi-LSTM group increased to 0.8720, 0.9224 and 0.8965, respectively. The performance of Group Bi-LSTM is more excels and is a powerful model of feature learning.

The design of Group Bi-LSTM enables the model to deeply exploit the features of single lead and the relationship between each lead of ECG. This means that there is also information flow in the ECG between different leads at the same time. its effectively enhances the quality of feature expression between different ECG leads and the expressiveness of signal representation with respect to ECG waveform features, thus leading to accurate and reliable estimations for CVD. This really is something worth celebrating, a milestone.

2). Attention Mechanism is Useful

The performance of AGBi-LSTM is better than that of Group Bi-LSTM. This shows that the attention mechanism and dynamic weighted loss increases the classification performance. By multiplying the weight vectors in attention mechanism, the influence weights for local temporal features are generated, which can highlight the main waveform features of ECG, and merge the ECG temporal features of each iteration as the main waveform features. Attention probability distribution is used to control the influence of elements in input sequence on elements in output sequence. While retaining more valuable information, reducing the impact of irrelevant or weak correlation information on output data.

The MTGBi-LSTM model combines attention mechanism to selectively learn the final feature map, breaking the limitation of the traditional neural network structure relying on the internal fixed length vector in encoding and decoding, and enhancing the robustness of the model.

3). Dynamic Weighted Loss Has Significant Effect

The results in Table 5 show that the classification performance of the MTGBi-LSTM model has been significantly improved. The F1 values of JA, VA, and LVH have increased by about 4 points. AMI, IMI, RVH, LAH, and, BAH have increased by about 10 points. In particular, the F1 value of LAH increased from 0.4 to 0.5238.

At the same time, we also noticed that the F1 values of LMI, BVH, and RAH have dropped, and there are only about 10 data in the test data set. Manually checking for inconsistencies found that the MTGBi-LSTM classification error overall seemed very reasonable. In some cases, because ECG data records are turbulent, it is not possible to draw reasonable conclusions from the ECG data.

We find that the results are improved when using dynamic weighted loss. It is proved that the dynamic weighted loss function solves the problem of the imbalance of normal and abnormal diseases in the field of intelligent assisted diagnosis.

Multi-Task learning mechanism is important. For the symbol Inline graphic of the multitasking learning mechanism, we evaluate the four values of Inline graphic to verify the classification accuracy on the test set. As Inline graphic decreases, the prediction accuracy increases gradually, and all models trained using multitasking learning mechanisms are superior to those that are not trained (Table 6). Because it provides the best prediction accuracy on the test set, in our complete framework, choose Inline graphic as the training model.

TABLE 6. Evaluation of the Number of ECG Leads and Multi-Task Loss Weight, on the Test Set.
Accuracy Precision Recall F1
Inline graphic,Inline graphic 0.7893 0.7922 0.8046 0.7984
Inline graphic, Inline graphic 0.8324 0.8451 0.8542 0.8496
Inline graphic, Inline graphic 0. 7806 0.8058 0.8456 0.8252
Inline graphic, Inline graphic 0.8029 0.8303 0.8692 0.8493
Inline graphic, Inline graphic 0.8886 0.9067 0.9419 0.9239
Inline graphic, Inline graphic 0.878 0.9005 0.9307 0.9154

Inline graphic is a parameter that interpolates between the task1 and task2 losses, Inline graphic is the number of leads of ECG data.

The local minimum of different tasks may be located at the low point of different local concave planes. The task1 increases the inter-CVD variations by drawing feature maps extracted from different diseases apart, while the task2 reduces the intra-CVD variations by pulling feature maps extracted from the same disease together. Through different forces between two tasks, the hidden layer can jump out of the local minimum, reduce the risk of model over-fitting, and make the classification accuracy of the model higher.

C. Effectiveness for Multi-Lead ECG

We evaluate the ECG data of 3-lead, 8-lead, and 12-lead to prove that the MTGBi-LSTM model can intelligently analyze the ECG data of any number of leads. It also demonstrates that the 12-lead ECG can provide a higher quality assessment analysis.

We performed a comparative experiment on the same ECG data set, as shown in Table 3 and Table 4. The labels and the number of records are the same, except that the number of leads used is different. The 3-lead ECG data is selected from I, II, III and the 8-lead ECG data is selected from II, III, V1, V2, V3, V4, V5, V6.

As seen in Table 6, when Inline graphic, Inline graphic and Inline graphic (Inline graphic) are compared, the precision, Recall, and F1 of the model increase with the number of ECG leads.

The 3-lead based MTGBi-LSTM model predicts both normal and arrhythmia categories, and the predictions for MI, VH, and AH are significantly worse. The ECG information collected by the 3-lead is relatively simple, and the information on the changes of the ST segment and the P wave is not obvious.

The 8-lead based MTGBi-LSTM model clearly achieved better performance than the 3-lead. When Inline graphic, the selected 8-lead are the basic leads of the ECG data, and the remaining four leads of the twelve leads are linearly derived from the eight basic leads. The ECG information collected by the 8 lead is rich, and the MTGBi-LSTM model can extract more highly discriminative ECG essential features.

This shows that our MTGBi-LSTM model can evaluate any multi-lead ECG (2-lead or more) and the 12-lead ECG data based MTGBi-LSTM model achieves the best performance.

D. Performance Comparison

CNN can stimulate low-dimensional local features implied in ECG waveforms into high-dimensional space, and the subsampling of a merge operation commonly used in CNN can inevitably extract peak points in a given merge window. In previous work [14], [22], [35] (Table VII), the researchers used deep feedforward neural network to classify ECG and achieved high performance.

TABLE 7. Comparison of Experimental Results of Different Methods.

Author, Year Database Number of Lead ECG Rhythms Classifier Acc Se Sp F1
A. H. Ribeiro (2019) [23] TNMG 12-lead 1dAVb, RBBB,LBBB,SB,AF,ST CNN / 0.954 0.996 0.914
Andrew Y. Ng(2019) [14] 91,232 ECG records from 53,549 patients single-lead N, AVB, Bigeminy, EAR, IVR, Junctional rhythm, Noise, Sinus rhythm, SVT, Trigeminy, VT, Wenckebach CNN / / / 0.837
A. Sellami(2019) [35] MIT-BIH-AR single-lead N, S, V, F, Q CNN 0.9948 0.9697 0.9987 0.9839
X. Zhai(2018) [36] MIT-BIH-AR single-lead N, S, V, F, Q CNN 0.986 0.938 0.992 0.9642
U. R. Acharya(2017) [37] afdb,cudb,MIT-BIH-AR single-lead Afib, Afl, Vfib, N CNN 0.949 0.9913 0.8144 0.8942
G. Swapna(2018) [38] MIT-BIH-AR single-lead normal, cardiac arrhythmia CNN-LSTM 0.834 / / /
S. Saadatnejad(2019) [21] MIT-BIH-AR single-lead N, LBBB, RBBB, S, V, F, Q WT, LSTM 0.986 0.938 0.992 0.9642
A. Mostayed(2019) [24] CPSCD 12-lead N, AF, I-AVB, LBBB, RBBB, PAC, PVC, STD, STE LSTM / / / 0.7415
O. Yildirim(2019) [39] MIT-BIH-AR single-lead APB, LBBB, N, RBBB, PVC LSTM 0.9911 0.99 0.99 0.99
B. Hou(2019) [40] MIT-BIH-AR single-lead N, LBBB, RBBB, APC, PVC LSTM,SVM 0.9974 0.9935 0.9984 0.9959
F. Liu(2019) [41] MIT-BIH-AR single-lead N, S, V, F, Q DWT,LSTM,CNN 0.995 0.999 0.982 0.9904
A. Elola(2019) [42] OHCA single-lead Pulsed Rhythm, Pulseless Rhythm Bi-GRU / 0.917 0.925 /
Wang LP [43] CCDD 12-lead normal and abnormal MTHC 0.7770 0.9513 0.5462 /
Zhu HH [44] CCDD 12-lead normal and abnormal multi-classifier 0.7251 0.6198 0.9359 /
Jin LP [25] CCDD 8-lead normal and abnormal LCNN 0.8366 0.8384 0.8343 /
Li HH [45] CCDD 8-lead normal and abnormal CNN 0.8477 0.8519 0.8445 /

Abbreviations: N, normal; AVB, atrioventricular block; EAR, ectopic atrial rhythm; IVR, idioventricular rhythm; SVT, supraventricular tachycardia; VT, ventricular tachycardia; 1dAVb, 1st degree AV block; RBBB, right bundle branch block; LBBB, left bundle branch block; SB, sinus bradycardia; AF, atrial fibrillation; ST, sinus tachycardia; Afib, Atrial fibrillation; Afl, Atrial flutter; Vfib, Ventricular flutter; Vfl, Ventricular flutter; S, supraventricular ectopic beats; V, ventricular ectopic beats; F, fusion beats; Q, unclassifiable beats; PAC, Premature atrial contraction; PVC, Premature ventricular contraction; STD, ST-segment depression; STE, ST-segment elevated; I-AVB, First-degree atrioventricular block; APC, atrial premature complexes Dataset Abbreviations: afdb, MIT-BIH atrial fibrillation, cudb, Creighton university ventricular tachyarrhythmia; MIT-BIH-AR. MIT-BIH arrhythmia; TNMG, Telehealth Network of Minas Gerais; SMCA, Seton Medical Center Austin; CPSCD, the Chinese physiological signal challenge dataset; OHCA, Outof-Hospital Cardiac Arrest database Classifier Abbreviations: DWT, discrete wavelet transform. [45]; LSTM, long short-term memory; CNN, convolutional neural network; WT, wavelet transform; SVM, Support Vector Machine. [46]

As a comparison, results from RNN including those in [21], [23], [41] are listed in Table 7. The advantage of RNN modeling is that the circular network can store a certain length of context information, RNN can handle ECGs of any length of time. When the neural activation unit completes some non-linear mapping, the RNN network can continuously update the network state, so RNN can learn long-term dependencies.

Furthermore, although both the CNN-based method and the RNN-based method obtain good results, those study has several important limitations. The CNN can only learn completely fixed time-varying weights, and it can only accept fixed length ECG input. CNN cannot understand the complex time characteristics in ECG data, ignoring the correlation between ECG data before and after.

The RNN-based algorithm simply superimposes the number of layers of LSTM or GRU, or combines with tricks in deep learning to improve model performance. These algorithms use the depth of the model in exchange for the performance improvement of the model, without exploring the essential characteristics of the ECG. Previous works based on multi-lead ECG analysis did not pay attention to the correlation among different lead information. The accuracy of the previous works is far from the requirement of technical reliability in the actual clinical environment.

Different lead electrocardiogram is a projection of the vector cardiogram in different directions. The 12-lead ECG can provide a full evaluation of the cardiac electrical activity, and the detection rate of standard 12-lead clinical ECG is significantly higher than single-lead ECG. Single-lead ECG of different diseases has a similar fluctuation trend. For example, in the single-lead ECG signal with normal sinus beats as the main component, the similarity of adjacent beats is high, when arrhythmia or noise disturbance occurs, the similarity of adjacent beats generally decreases. At this time, it is difficult to determine whether it is arrhythmia or low signal-to-noise ratio. In [14], [21], [34][41], the input dataset is limited to single-lead ECG records, which provides limited signal compared to a standard 12-lead ECG, and it remains to be determined if those algorithm performance would be similar in 12-lead ECG.

The most commonly used data set currently used to design and evaluate ECG algorithms is the MIT-BIH arrhythmia database [19], which consists of 48 and a half hour 2-lead ECG data records. The beat cycle of the database may belong to the same patient, and different beat data may have the same trend. So the model evaluated using the MIT-BIH-AR database is not robust, and the data sample reported in the MIT-BIH-AR database are insufficient and leads to the overfitting problem.

As the amount of digitized ECG data increases, the performance of deep neural network-based machine learning models can approach and possibly exceed human diagnostic capabilities. After evaluating the largest known CCDD database in the world, MTGBi-LSTM model obtains the most advanced performance on more than 17,000 12-lead ECG records, which proves the effectiveness of our algorithm.

Here, we focused on a central challenge in ECG diagnostics, and proposed the intelligent assistant diagnosis model of CVD based on Res-GCNN and GBI-LSTM hybrid structure. It can accept any multi-lead ECG. We propose a group of Bi-LSTM structures that can Computer-Assisted diagnose ECGs with any number of leads. At the same time, attention mechanism is added to each Bi-LSTM structure to break through the limitation that traditional Bi-LSTM structure relies on internal fixed length vectors when encoding and decoding. The most important thing is that MTGBi-LSTM model explores the characteristic expression of different ECG leads for the first time, which is of great significance for the auxiliary diagnosis of multi-lead ECG.

IV. Conclusion

In this paper, we propose a Multi-Task Group Bi-LSTM framework to classify various diseases of ECG in order to realize intelligent assistant diagnosis of CVD. This model employs a group Bi-LSTM and Res-GCNN to improve the learning ability of time and space. Furthermore, the model can deeply exploit the features of single-lead and the relationship between each lead of ECG. Attention mechanism and multi-task learning technology are used to enhance the recognition ability of the model, and ultimately to achieve a variety of disease classification. According to the results shown, compared with other classifiers, the F1 of the experimental model in this paper is up to 0.9239 and the results confirm the high accuracy and generalization ability of the proposed model. Experimental results confirm the MTGBi-LSTM is efficient and powerful, and could match state-of-the-art algorithms.

The intelligent assisted diagnosis of electrocardiogram improves the diagnostic efficiency of cardiologists, and we hope that this technology can be combined with online equipment to serve as an early screening tool for cardiovascular disease in areas with limited medical resources.

Funding Statement

This work was supported in part by the Guangzhou Science and Technology under Grant 201803010072, in part by the Science, Technology, and Innovation Commission of Shenzhen Municipality under Grant JCYJ20170818165305521, and in part by the Start-up Hundred Talent Program from Sun Yat-sen University (SYSU).

References

  • [1].World Health Organization, Mendis S., Puska P., and Norrving B., Global Atlas on Cardiovascular Disease Prevention and Control. Geneva, Switzerland: World Health Organization, 2011, pp. 3–18. [Google Scholar]
  • [2].Wang H.et al. , “Global, regional, and national life expectancy, all-cause mortality, and cause-specific mortality for 249 causes of death, 1980–2015: A systematic analysis for the Global Burden of Disease Study 2015,” Lancet, vol. 388, no. 10053, pp. 1459–1544, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Abubakar I., Tillmann T., and Banerjee A., “Global, regional, and national age-sex specific all-cause and cause-specific mortality for 240 causes of death, 1990-2013: A systematic analysis for the global burden of disease study 2013,” Lancet, vol. 385, no. 9963, pp. 117–171, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [4].McGill H. C., McMahan C. A., and Gidding S. S., “Preventing heart disease in the 21st century: Implications of the pathobiological determinants of atherosclerosis in youth (PDAY) study,” Circulation, vol. 117, no. 9, pp. 1216–1227, Mar. 2008. [DOI] [PubMed] [Google Scholar]
  • [5].O’Donnell M. J.et al. , “Global and regional effects of potentially modifiable risk factors associated with acute stroke in 32 countries (INTERSTROKE): A case-control study,” Lancet, vol. 388, no. 10046, pp. 761–775, 2016. [DOI] [PubMed] [Google Scholar]
  • [6].Rautaharju P. M., “A hundred years of progress in electrocardiography 2: The rise and decline of vectorcardiography,” Ann. Noninvasive Electrocardiol., vol. 1, no. 3, pp. 335–346, 2010. [PubMed] [Google Scholar]
  • [7].LeCun Y., Bengio Y., and Hinton G., “Deep learning,” Nature, vol. 521, no. 7553, p. 436, 2015. [DOI] [PubMed] [Google Scholar]
  • [8].Esteva A.et al. , “Dermatologist-level classification of skin cancer with deep neural networks,” Nature, vol. 542, no. 7639, p. 115, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [9].Gulshan V.et al. , “Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs,” J. Amer. Med. Assoc., vol. 316, no. 22, pp. 2402–2410, 2016. [DOI] [PubMed] [Google Scholar]
  • [10].Kaur G., Singh G., and Kumar V., “A review on biometric recognition,” Int. J. Bio-Sci. Bio-Technol., vol. 6, no. 4, pp. 69–76, 2014. [Google Scholar]
  • [11].Gee A. H., Garcia-Olano D., Ghosh J., and Paydarfar D., “Explaining deep classification of time-series data with learned prototypes,” 2019, arXiv:1904.08935. [Online]. Available: https://arxiv.org/abs/1904.08935 [PMC free article] [PubMed]
  • [12].Lu P.et al. , “Research on improved depth belief network-based prediction of cardiovascular diseases,” J. Healthcare Eng., vol. 2018, May 2018, Art. no. 8954878. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [13].Rajpurkar P., Hannun A. Y., Haghpanahi M., Bourn C., and Ng A. Y., “Cardiologist-level arrhythmia detection with convolutional neural networks,” 2017, arXiv:1707.01836. [Online]. Available: https://arxiv.org/abs/1707.01836
  • [14].Hannun A. Y.et al. , “Cardiologist-level arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network,” Nature Med., vol. 25, no. 1, pp. 65–69, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [15].Cho K.et al. , “Learning phrase representations using RNN encoder-decoder for statistical machine translation,” 2014, arXiv:1406.1078. [Online]. Available: https://arxiv.org/abs/1406.1078
  • [16].Graves A., “Long short-term memory,” in Supervised Sequence Labelling With Recurrent Neural Networks. Berlin, Germany: Springer, 2012, pp. 37–45. [Google Scholar]
  • [17].Zhao Y., Yang R., Chevalier G., Xu X., and Zhang Z., “Deep residual bidir-LSTM for human activity recognition using wearable sensors,” Math. Problems Eng., vol. 2018, Dec. 2018, Art. no. 7316954. [Google Scholar]
  • [18].Chung J., Gulcehre C., Cho K. H., and Bengio Y., “Empirical evaluation of gated recurrent neural networks on sequence modeling,” 2014, arXiv:1412.3555. [Online]. Available: https://arxiv.org/abs/1412.3555
  • [19].Goldberger A. L.et al. , “PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals,” Circulation, vol. 101, no. 23, pp. e215–e220, 2000. [DOI] [PubMed] [Google Scholar]
  • [20].Tan J. H.et al. , “Application of stacked convolutional and long short-term memory network for accurate identification of CAD ECG signals,” Comput. Biol. Med., vol. 94, pp. 19–26, Mar. 2018. [DOI] [PubMed] [Google Scholar]
  • [21].Saadatnejad S., Oveisi M., and Hashemi M., “LSTM-based ECG classification for continuous monitoring on personal wearable devices,” IEEE J. Biomed. Health Inform., to be published. [DOI] [PubMed]
  • [22].Ribeiro A. H.et al. , “Automatic diagnosis of the short-duration 12-lead ECG using a deep neural network: The code study,” 2019, arXiv:1904.01949. [Online]. Available: https://arxiv.org/abs/1904.01949
  • [23].Mostayed A., Luo J., Shu X., and Wee W., “Classification of 12-lead ECG signals with Bi-directional LSTM network,” 2018, arXiv:1811.02090. [Online]. Available: https://arxiv.org/abs/1811.02090
  • [24].Jin L. P. and Dong J., “Deep learning research on clinical electrocardiogram analysis,” Sci. China, Inf. Sci., vol. 45, no. 3, pp. 398–416, 2015. [Google Scholar]
  • [25].He K., Zhang X., Ren S., and Sun J., “Deep residual learning for image recognition,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2016, pp. 770–778. [Google Scholar]
  • [26].Glorot X., Bordes A., and Bengio Y., “Deep sparse rectifier neural networks,” in Proc. 14th Int. Conf. Artif. Intell. Statist., 2011, pp. 315–323. [Google Scholar]
  • [27].Ioffe S. and Szegedy C., “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” in Proc. Int. Conf. Int. Conf. Mach. Learn., 2015, pp. 448–456. [Google Scholar]
  • [28].Szegedy C., Vanhoucke V., Ioffe S., Shlens J., and Wojna Z., “Rethinking the inception architecture for computer vision,” in Proc. Comput. Vis. Pattern Recognit., Jun. 2016, pp. 2818–2826. [Google Scholar]
  • [29].Zhou P.et al. , “Attention-based bidirectional long short-term memory networks for relation classification,” in Proc. 54th Annu. Meeting Assoc. Comput. Linguistics, vol. 2, 2016, pp. 207–212. [Google Scholar]
  • [30].Vaswani A.et al. , “Attention is all you need,” 2017, arXiv:1706.03762. [Online]. Available: https://arxiv.org/abs/1706.03762
  • [31].Zhong Z., Zheng L., Zheng Z., Li S., and Yang Y., “Camera style adaptation for person re-identification,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2018, vol. 28, no. 3, pp. 5157–5166. [Google Scholar]
  • [32].Collobert R. and Weston J., “A unified architecture for natural language processing: Deep neural networks with multitask learning,” in Proc. Int. Conf. Mach. Learn., 2008, pp. 160–167. [Google Scholar]
  • [33].Zhang J.-W., Liu X., and Dong J., “CCDD: An enhanced standard ecg database with its management and annotation tools,” Int. J. Artif. Intell. Tools, vol. 21, no. 5, 2012, Art. no. 1240020. [Google Scholar]
  • [34].Sellami A. and Hwang H., “A robust deep convolutional neural network with batch-weighted loss for heartbeat classification,” Expert Syst. Appl., vol. 122, pp. 75–84, May 2019. [Google Scholar]
  • [35].Zhai X. and Tin C., “Automated ECG classification using dual heartbeat coupling based on convolutional neural network,” IEEE Access, vol. 6, pp. 27465–27472, 2018. [Google Scholar]
  • [36].Acharya U. R., Fujita H., Oh S. L., Hagiwara Y., Tan J. H., and Adam M., “Automated detection of arrhythmias using different intervals of tachycardia ECG segments with convolutional neural network,” Inf. Sci., vol. 405, no. 1, pp. 81–90, Sep. 2017. [Google Scholar]
  • [37].Swapna G., Soman K. P., and Vinayakumar R., “Automated detection of cardiac arrhythmia using deep learning techniques,” Procedia Comput. Sci., vol. 132, pp. 1192–1201, Jan. 2018. [Google Scholar]
  • [38].Yildirim O., Baloglu U. B., Tan R.-S., Ciaccio E. J., and Acharya U. R., “A new approach for arrhythmia classification using deep coded features and LSTM networks,” Comput. Methods Programs Biomed., vol. 176, pp. 121–133, Jul. 2019. [DOI] [PubMed] [Google Scholar]
  • [39].Hou B., Yang J., Wang P., and Yan R., “LSTM based auto-encoder model for ECG arrhythmias classification,” IEEE Trans. Instrum. Meas., to be published.
  • [40].Liu F., Zhou X., Cao J., Wang Z., Wang H., and Zhang Y., “Arrhythmias classification by integrating stacked bidirectional LSTM and two-dimensional CNN,” in Proc. Pacific–Asia Conf. Knowl. Discovery Data Mining. Cham, Switzerland: Springer, 2019, pp. 136–149. [Google Scholar]
  • [41].Elola A.et al. , “Deep neural networks for ECG-based pulse detection during out-of-hospital cardiac arrest,” Entropy, vol. 21, no. 3, p. 305, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [42].Wang L. P., “Study on approach of ECG classification with domain knowledge,” Ph.D. dissertation, East China Normal Univ, Shanghai, China, 2013. [Google Scholar]
  • [43].Zhu H. H., “Research on ECG recognition critical methods and development on remote multi body characteristic signal monitoring system,” Ph.D. dissertationm, Univ. Chin. Acad. Sci, Beijing, China, 2013. [Google Scholar]
  • [44].Huihui L. and Linpeng J., “An ECG classification algorithm based on heart rate and deep learning,” Space Med. Med. Eng., vol. 29, no. 3, pp. 189–194, 2016. [Google Scholar]
  • [45].Ye C., Kumar B. V. K. V., and Coimbra M. T., “Heartbeat classification using morphological and dynamic features of ECG signals,” IEEE Trans. Biomed. Eng., vol. 59, no. 10, pp. 2930–2941, Oct. 2012. [DOI] [PubMed] [Google Scholar]
  • [46].Shahbudin S., Shamsudin S. N., and Mohamad H., “Discriminating ECG signals using support vector machines,” in Proc. IEEE Symp. Comput. Appl. Ind. Electron. (ISCAIE), Apr. 2015, pp. 175–180. [Google Scholar]

Articles from IEEE Journal of Translational Engineering in Health and Medicine are provided here courtesy of Institute of Electrical and Electronics Engineers

RESOURCES