Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2020 Feb 9.
Published in final edited form as: IEEE J Biomed Health Inform. 2019 Oct 7;24(2):457–464. doi: 10.1109/JBHI.2019.2946066

Knowledge-based Analysis for Mortality Prediction from CT Images

Hengtao Guo 1, Uwe Kruger 2, Ge Wang 3, Mannudeep K Kalra 4, Pingkun Yan 5,*
PMCID: PMC7007835  NIHMSID: NIHMS1063197  PMID: 31603807

Abstract

Low-Dose CT (LDCT) can significantly improve the accuracy of lung cancer diagnosis and thus reduce cancer deaths compared to chest X-ray. The lung cancer risk population is also at high risk of other deadly diseases, for instance, cardiovascular diseases. Therefore, predicting the all-cause mortality risks of this population is of great importance. This paper introduces a knowledge-based analytical method using deep convolutional neural network (CNN) for all-cause mortality prediction. The underlying approach combines structural image features extracted from CNNs, based on LDCT volume at different scales, and clinical knowledge obtained from quantitative measurements, to predict the mortality risk of lung cancer screening subjects. The proposed method is referred as Knowledge-based Analysis of Mortality Prediction Network (KAMP-Net). It constitutes a collaborative framework that utilizes both imaging features and anatomical information, instead of completely relying on automatic feature extraction. Our work demonstrates the feasibility of incorporating quantitative clinical measurements to assist CNNs in all-cause mortality prediction from chest LDCT images. The results of this study confirm that radiologist defined features can complement CNNs in performance improvement. The experiments demonstrate that KAMP-Net can achieve a superior performance when compared to other methods.

Index Terms—: Lung cancer, low-dose CT, mortality risk, machine learning and deep learning, convolutional neural network, clinical knowledge

I. Introduction

LOW-Dose CT has proven to be effective for lung cancer screening. For example, the National Lung Screening Trial (NLST) observed a 20% decrease in lung cancer related mortality in at-risk subjects (55 to 74 years, 30 pack-year cigarette-smoking history) [1]. The prevalence of lung cancer is highly correlated with CVDs and both are associated with significant morbidity and mortality [2], [3]. More precisely, both share several risk factors that are predominantly attributed to unhealthy dietary habits, obesity and tobacco use etc. By analyzing the NLST data, Chiles et al. [4] showed that coronary artery calcification (CAC) is strongly associated with mortality. In a different study, the Dutch-Belgian Randomized Lung Cancer Screening Trial (NELSON), it was found that CAC can predict all-cause mortality and cardiovascular events on lung cancer screening LDCT [5]. The work in [6] has also shown significant difference in CAC scores between the survivor and non-survivor groups, indicating that CAC influences the mortality risk of lung cancer patients. Moreover, other factors may also increase the mortality risk. For example, non-surviving NLST subjects tend to have higher fat attenuation and decreased muscle mass, comparing to the surviving ones and there is a strong difference in emphysema severity between survivors and non-survivors [6].

Over the past few years, the application of deep learning, a subdomain of machine learning, has led to a series of breakthroughs producing a paradigm shift that resulted in numerous innovations in medicine, ranging from medical image processing, to computer-assisted diagnosis, to health record analysis. Deep learning has also been applied for automatic calcium scoring from chest LDCT images. For example, Cano-Espinosa et al. [7] proposed to use a convolutional neural network for Agatston score regression from non-contrast chest CT scans without segmenting CAC regions. Recently, de Vos. et al. [8] proposed to (i) use one convolutional network for template image and input CT registration and (ii) use another network for direct coronary calcium regression. Lessmann et al. [9] report that (i) deep neural networks can measure the size of CAC from LDCT and (ii) the use of different filters, during the reconstruction process, can influence the quantification results. Training such networks, however, requires manually labeling the area of calcification from images. This results in significant efforts and only a small number of images can be annotated. This may adversely affect the network performance. Moreover, CAC segmentation does reveal other imaging markers that may predict the mortality risk.

Recently, van Velzen et al. [10] introduced a convolutional autoencoder to extract image features for cardiovascular mortality prediction in a latent space. The features then serve as the input to a separate classifier, for example a neural network, a random forest classifier or a support vector machine, to compute a risk value. However, such a two-phase method may not be able to extract the most distinctive features associated with CVD. Moreover, traditional convolutional neural networks (CNNs) rely on directly extracted image features to perform image classification. This, however, omits clinical knowledge summarized by physicians through their diagnosis. Since various predefined imaging markers have been well recognized as indication of mortality risk, it is advisable to utilize this information for estimating this risk.

This paper hypothesizes that incorporating clinical knowledge into a deep learning based mortality risk prediction produces valuable complementary information which increases the prediction accuracy. To test the hypothesis, we introduce a novel method that combines extracted features from a CNN with clinical knowledge for predicting all-cause mortality risk of lung cancer patients from their LDCT images. More precisely, the method introduced here relies on a dual-stream network (DSN), which takes whole slices as well as cropped cardiac patches as the input for feature extraction. The multiscale input has been demonstrated to have a positive impact on the CNN’s performance [11], as it contains both global image slice information and details of important local areas. The second component of the introduced method is incorporating clinical knowledge that is based on four clinical measurements, including CAC, muscle mass, fat attenuation, and emphysema. Inspired by the work of Fu et al. [12], we employ a support vector machine (SVM) classifier to combine the clinical measurements to generate a combined mortality risk probability. The resultant method is referred to here as the knowledge-based analysis for mortality prediction (KAMP-Net). The experimental results confirm that KAMP-Net predicts mortality more accurately when compared with other competitive networks. The contributions of this paper are summarized as follows.

  1. We utilize deep neural networks for predicting all-cause mortality risk of lung cancer patients by automatically discovering imaging features instead of measuring the extent of CAC as a surrogate index [13], [14].

  2. We introduce a new gray-level image color-coding method to efficiently reuse the seminal deep CNN network structures.

  3. The DSN takes multi-scale image inputs, composed by LDCT slices and cardiac image patches, for both local and global feature extraction.

  4. Our results demonstrate that the DSN-extracted features, when combined with clinical knowledge from predefined imaging marker, can significantly improve the prediction performance.

II. Methods

In this section, we present our proposed method for mortality risk prediction using LDCT images and related clinical measurements.

A. Multi-Channel Image Coding

LDCT images are 3D volume data containing information of internal structures such as organs, bones, blood vessels and soft tissue. The value of each voxel varies from −1000 Hounsfield units (HU) to around 2000 HU. Directly suppressing such a large value range into the typical range processed by deep CNN may result in information loss. To make full use of the anatomical information in CT images, we divide the range of CT image intensity values into three segments, according to the clinical expert knowledge on the intensity distribution of the tissues of interest. Namely, values below −900 HU are extracted and normalized to [0,255] as emphysema-concentrated interval to form the first channel. Similarly, voxels with values in the range of (−900,0] are assigned to the second channel representing fat-concentration intensity interval. CT numbers larger than 300 are typically from very strong calcification so we top off there and normalize all the values in (0,300] to form the third channel. For visualization purpose, the three channels are mapped to red, blue, and green channels of color image as shown in Fig. 2. After separating different anatomical structures to separate channels, the intensity range of different tissue types throughout the CT slice become more balanced. For instance, the coronary artery calcification in the heart region appearing as bright green no longer suppressing other imaging components like fat or emphysema.

Fig. 2:

Fig. 2:

Two examples of anatomical-information based multichannel image coding. The second row is the magnification of the heart region in the first row. With the proposed coding scheme, the large intensity range of CT images can be divided into three smaller segments to highlight the important imaging features. Red areas mostly correspond to emphysema severity; blue areas represent fat attenuation concentrated regions; green areas contain mostly calcification or bone.

B. Network Design and Implementation

As shown in Fig. 1, the deep neural network consists of two streams, which is referred as dual stream networks (DSN). The upper stream extracts global image features from the input axial view image slice, which is manually chosen from LDCT scan. The lower stream takes one automatically selected region of interest (ROI) as input, which contains the most severe calcification in either left anterior descending (LAD), left circumflex (LCX) or left coronary artery (LCA). When there is no obvious calcification, the slices where the LAD is most visible were chosen. The automatic ROI detection is performed by a pre-trained cascaded detector [15]. We implement both networks with 2D convolutions to guarantee a manageable computational burden. The lower stream supplements the upper stream with local detailed visual cues to emphasize the importance of those local regions. The lower stream supplements the upper stream with local detailed visual cues to emphasize the importance of those local regions. The deep residual network (ResNet) [16], which is one of the top performing deep CNNs in various computer vision tasks, has been adopted as the backbone of DSN. By using only the convolutional layers of ResNet, image features can be extracted by ResNet-x, where x denotes the depth of the network. At the end of the convolutional layers, 512 features are extracted by ResNet-18 and 34, and 2048 features are ResNet-50, 101 and 152, respectively. According to our previous work [17], ResNet-34 achieves the best accuracy in the patch-input network, so we chose to use it as the lower stream’s backbone architecture.

Fig. 1:

Fig. 1:

Overview of the proposed KAMP-Net, which combines features extracted using clinical knowledge and features discovered by the deep learning DSN for improved mortality risk prediction.

The proposed KAMP-Net was implemented in Python using the open source PyTorch library [18]. The training loss is defined as the cross-entropy between the prediction probability and ground-truth label as

Loss =1Ni=1N(yi log (pi)+(1yi) log (1pi)) (1)

where N indicates the batch size, yi ∈ {0, 1} is the label of groundtruth of the ith sample and pi is the network-derived probability for class yi after softmax. Training of the network is completed in two stages. The two streams of DSN are first trained separately in stage one and then combined for fine-tuning in stage two.

In the first training stage, we implemented ResNet using the pre-defined structure provided by PyTorch [18]. Instead of generating probabilities for 1000 classes, the only difference between our network and the original ResNet is that the last fully connected (FC) layer outputs the classification probabilities of two categories: deceased or survived. Both patch-wise and slice-wise networks are trained from scratch using Adam optimizer [19] with initial learning rate of 1 × 10−5, which then decays by 0.9 after every five epochs. While many DL-based medical image analysis papers report that networks pre-trained on ImageNet data can achieve better performance [20], we chose to train the network from scratch instead of using networks pre-trained on ImageNet data, because there exists large image appearance difference between natural images from ImageNet and the LDCT lung images. Each sample in our dataset has been labeled either 0 (deceased) or 1 (survived) for training and validation.

In the second training stage, we remove the FC layers of the two sub-network streams pre-trained in stage one and combine the convolutional segments to form DSN. The output feature maps of the two sub-networks are concatenated and fed to a new FC layer, which generates two probabilities for survival and death prediction, respectively. The entire DSN with newly added FC layer is trained for another 200 epochs for fine-tuning with again the learning rate of of 1 × 10−5. As the pre-trained slice-wise and patch-wise networks have already gained the ability to extract informative medical image features, the training of DSN would converge quickly.

C. Integration of Deep Learning and Clinical Knowledge

To further increase the accuracy of mortality prediction from LDCT images, we propose to combine clinical measurements with deep learning. Although CNNs are very powerful in extracting imaging markers, they lack of logical reasoning and high level intelligence of human experts, which makes it difficult for them to figure out connections between seemingly distant concepts. On the other hand, expert defined measurements from the images, including emphysema severity, muscle mass, fat attenuation and coronary artery calcification score, can be useful for this task as shown in the previous work [6]. CAC scores can be quantified in different ways [21], [22] and automatic methods have been presented [23]. In our work, we utilize the CAC risk score, which was graded on a 4-point scale, to denote different severity. The CAC risk score is given by two radiologists, following a blinded and randomized manner. More detailed information about clinical measurements utilized in this work is available in the reference [6]. Those measurements contain high-level information and may not be readily grabbed by the CNNs. These knowledge based features can be complementary to what CNNs extract. We thus propose to combine the two groups of features to achieve more accurate prediction.

However, directly concatenating those measurements with the feature vectors from CNN could have only trivial effects on the prediction results. Since the CNN-extracted feature vector has much higher dimensionality (e.g. 512 for ResNet-34) than the clinical measurement (4 in this case), the latter will be overwhelmed after simple concatenation and contribute little to the risk prediction. To balance the contributions of the two groups of features to the final output, we merge the two groups at a later stage after obtaining the initial probabilities. As shown in Fig. 1, a linear SVM classifier with the four clinical measurements as input is trained for mortality prediction. This SVM classifier will produce the probabilities of being deceased pSVMd or survived pSVMs, which add up to 1. On the DSN side, a softmax activation function is used to generate the probability output. The two sets of probabilities are then combined to obtain the overall chance of survival as

ps=αpDSNs+(1α)pSVMs, (2)

where ps, pDSNs and pSVMs are the combined probability, DSN estimated probability and SVM estimated probability of survival, respectively. The contribution ratio α is a weighting parameter in the range of [0, 1]. The probability of death pd can be computed as 1 − ps.

III. Experimental Results

This section presents experimental results of applying the KAMP-Net model for mortality risk prediction and provide detailed analysis and comparison of its performance.

A. Materials

All the study data used in this work are from the National Lung Screening Trial (NLST) [24], which are managed by the National Cancer Institute Cancer Data Access System. In this large scale clinical trial, NLST compared LDCT with the chest radiography for lung cancer screening in more than 50,000 current or former smokers who met the various inclusion criteria. Our hypothesis of the study is that the analysis of LDCT images acquired for lung cancer screening can effectively predict the all-cause mortality of the subjects by combining the clinical knowledge and advanced deep learning techniques. To efficiently investigate the effects of imaging features and clinical measurements, a balanced study is designed in our work. Following the same protocol used in [6], 180 subjects were selected for the study, where the 90 survived and 90 deceased subjects are equally distributed in a variety of different cancer stages including no cancer. More precisely, each group consists of 49 subjects with stage I, 19 subjects with stage II, and 22 subjects with stage III lung cancers. The motivation is to rule out the influence of cancer stage and determine the effects of other factors, which may cause essential difference between the two groups.

The prediction is formulated as a binary classification problem by using the subject survival or decease status at the end of the follow-up period as the ground truth. The NLST trial uses lung cancer mortality as the primary endpoint of the study but also recorded all-cause mortality during the follow-up. The average follow-up period of the NLST trial is 6.5 years. More specifically, the average number of days of follow-up is 1660 ± 488 for the survivors, and the days to death for the deceased subjects are 894±542. Each patient went through three LDCT lung cancer screening exams, of which the first LDCT scan of each patient is used in this study. The survival label is used as the ground truth for training and evaluating the prediction algorithms.

The size of axial view slices in LDCT volume is 512 × 512 pixels. The number of slices per subject varies between 46 and 245. Three consecutive slices are extracted for each subject, which were manually chosen to be the slices in the CT volume for which the coronary artery is most visible. The use of three consecutive slices from a volume increases the number of slices from 180 to 540 images, i.e. we have a significantly larger set for network training and validation.

Data augmentation has been shown to be an effective approach to improve the performance of deep CNNs [25]. In this paper, data augmentation operations including random cropping and scaling are used for training the networks, which, theoretically, yields an infinite number of samples. The image patches in the size 161×161 pixels are cropped from LDCT images using a pre-trained cascaded detector, which automatically locates a bounding-box over the heart region. Both the input slices and heart region patches are randomly cropped with the size ratio between 0.6 and 0.8. Please note this random ratio is the ratio of the original images. The cropping was conducted such that the aspect ratio is one, i.e. length and height contain the same number of pixels. The cropped cardiac patches are resized to 224 × 224 pixels for network input. The aim of resizing the input images is to fit the design of original ResNet architecture.

Before applying to the ResNets, we conduct image normalization for gray-scale input and color-coded input separately, using their own means and standard deviations. As for the gray-scale images, we applied a single mean and a single standard deviation, which was computed from all the image samples. For the color-coded images, the mean and standard deviation are computed for each channel. For each channel, the normalization is performed by first subtracting its mean and then dividing the difference by its corresponding standard deviation. As a result, the pixel intensity distribution of the images has a mean of 0 and a standard deviation of 1 for each channel. In summary, the normalization for the gray scale slices and color-coded slices are performed separately, but in a consistent manner.

B. Performance Evaluation

Since the available dataset is relatively small, we applied a ten-fold cross validation scheme to our dataset for evaluating the performance of the proposed method and other comparative methods. We shuffle the list of subjects and divide them into 10 parts, where each part contains 18 subjects with 9 deceased and 9 survived. In each fold, one part is left out for testing. Among the remaining nine parts, one part is randomly chosen for validation and the other eight are for training. For each fold, the training is performed using the training set until the network performance is optimized on the validation set. Upon completion of this training process, the performance of the trained network is evaluated on the left-out testing set. The cross-validation continues until each part has been left out. In the testing phase, all three slices of each subject are used. We then compute the average probability and assign this average risk score to the subject. Since we aim to predict the ending points of subjects to be either “survivor” or “nonsurvivor” at the end of the follow-up period, receiver operating characteristic (ROC) curves are drawn to demonstrate the performance. Area under the curve (AUC) scores are used to compare the performances of different methods. When training the networks for each fold, the maximum number of epochs is set to be 200. Fig. 3 shows the training and validation loss curves over a 200-epoch training of the ten-fold cross validation.

Fig. 3:

Fig. 3:

Training and validation loss curves of the ten-fold cross validation. The solid lines represent the mean loss curves over the ten folds and the shadow areas indicate the ranges.

To further evaluate the mutual influence between image-extracted features and the clinical information on the performance of the proposed KAMP-Net, we explore in terms of the DSN weight ratio α through 0 to 1, with an increment of 0.05. As shown in Fig. 4, when the ratio α equals to 0.75, the curve arrives at its peak with the highest AUC score of 0.82 and the lowest standard deviation of 0.07. With ratio α increasing from 0 (pure SVM prediction votes from clinical measurements) to 1 (pure DSN prediction votes from LDCT images), the overall KAMP AUC score experienced a steady increase and then decrease. Such a tendency in this α-AUC curve explicitly shows that there exists a delicate balance point where the votes from DSN and SVM can reach the best performance. At this balance point, the DL-based image features and the medical information from clinical measurements are collaborating with each other as well as complementing each other’s missing clues on predicting one patient’s status.

Fig. 4:

Fig. 4:

Effects of varying α, the weight of probability votes from DSN in Eqn. (2), with increment of 0.05.

The mean ROC curves over ten-fold cross validation of different methods are shown in Fig. 5. The corresponding mean AUC scores and standard deviations are also provided. The SVM model is trained using the four clinical measurements on the same cross validation folds as the DL methods. It can be seen that our proposed KAMP-Net achieves both the highest AUC score and the lowest standard deviation compared to other methods. Our previous work HyRiskNet [17] is included for comparison, which directly concatenates one additional CAC risk score with the high-dimensional deep CNN extracted feature vector.

Fig. 5:

Fig. 5:

Ten-fold cross validation ROC curves and AUC values of HyRiskNet-34, DSN, SVM and the proposed KAMP-Net.

We now compare the performance of KAMP-Net with that of its individual components, i.e. the DSN and the SVM models. Fig. 5 allows comparing the performance of the three models graphically based on the estimated ROC curves. For KAMP-Net, we select α = 0.75. To qualitatively test whether the increase of the AUC value is statistically significant, we test the null hypotheses that AUCKAMP = AUCDSN and AUCKAMP = AUCSVM against the one sided alternative hypotheses AUCKAMP > AUCDSN and AUCKAMP > AUCSVM, respectively. The two tests rely on three samples that store AUC values obtained from the previous 10-fold cross-validatory assessment; one sample stores the 10 AUC values for the KAMP-Net model, whilst the other two samples store the AUC values for the DSN and SVM models. This allows a pairwise comparison involving 10 pairs of values for testing AUCKAMP = AUCDSN as well as AUCKAMP = AUCSVM. First, we confirmed that the two sample differences were drawn from normal distributions by applying the Anderson-Darling test and then applied a standard paired t-test [26]. In both cases, the null hypothesis was rejected and, therefore, concluded that the increase in the risk prediction accuracy by the KAMP-Net model is statistically significant.

It should be noted that even without using any clinical measurements, the current DSN has already outperformed the previous CNN based methods presented in [17], which use only patch image information as input. On the other hand, the performance of SVM shows that these four clinical measurements carry quantification information strongly associated with survival in our experiments. However, it is only a limited set of measurements. When being complemented with deep CNN discovered features, the performance has become even better.

C. Effectiveness of Color-Coding

In this paper, we introduce the color-coding scheme to highlight the anatomical difference for more effective feature extraction. To evaluate the performance, we conducted experiments on all the ResNet network structures available in PyTorch using both the original LDCT image and the color-coded version as inputs, respectively. The experimental results are shown in Table. I. While the networks in Color-Coding group take the 3-channel pre-processed images as input, the networks in the other group just take the single-channel grey scale images as input. Such original images are obtained through directly suppressing the raw slices from LDCT 3D volume to the range [0,255] from a wide range of Hounsfield Units. The two groups of networks were trained on the same ten folds, with the same training strategy and parameters.

Table I:

Comparison of the performances of all ResNets and DSNs with color-coded and grey-scale LDCT image inputs, respectively. Slice-x and Patch-x indicate the ResNet-x architectures taking whole slice image input and cropped cardiac patch input, respectively. Two listed DSNs are composed of Slice-50 and Patch-34 networks trained on color-coding images and gray-scale images, respectively.

Network Color-Coding Grey-Scale
AUC STD AUC STD
Slice-18 0.64 0.04 0.63 0.07
Slice-34 0.68 0.04 0.67 0.06
Slice-50 0.71 0.06 0.68 0.05
Slice-101 0.68 0.06 0.65 0.08
Slice-152 0.66 0.07 0.64 0.07
Patch-18 0.68 0.06 0.65 0.09
Patch-34 0.73 0.07 0.68 0.07
Patch-50 0.69 0.04 0.66 0.04
Patch-101 0.66 0.03 0.65 0.05
Patch-152 0.64 0.04 0.64 0.07
DSN 0.76 0.10 0.70 0.08

To statistically analyze the significance of color-coding, we applied a paired hypothesis test for the two groups of observations. Prior to that, we verified that the sample differences were drawn from a normal distribution by applying the well-known Anderson-Darling test. This allowed the use of the standard t-test [26] for the null hypothesis, which stated that the use of color-coded images does not affect the overall performance compared to gray-scale images, against the one-sided alternative hypothesis that color-coding increases the prediction accuracy when compared to gray-scale images. The computed t-value of the slice-wise section maps to the rejection region and we, therefore, rejected the null hypothesis, which confirms that the use of color-coding led to a statistically significant improvement in the mortality prediction accuracy. This indicates that directly suppressing a whole slice from a large dynamic range to generate input for the networks may result in significant loss of information. Conversely, the introduced color-coding scheme alleviates this problem. In contrast, however, there is no significant difference between the color-coding group and the grey scale group when applying pre-processing to the patch-wise networks. In summary of the results in Table I, we select ResNet-50 and ResNet-34 as the backbone networks for the color-coded input slices and patches for DSN in KAMP-Net, respectively. Two DSNs, composed of Slice-50 and Patch-34 networks trained on color-coding input and gray-scale input respectively, achieve different performance. Such experimental results further indicate the superiority of applying color-coding scheme during the multiscale analysis.

D. Evaluation of Dual Stream Network

We then evaluate the performance of DSN by comparing the network structures as well as training strategies. Table II shows the ROC curves and also AUC values of DSN trained from scratch (SDN-scratch), Slice-50, Patch-34 and DSN. It can be seen that DSN outperforms both Slice-50 and Patch-34 by combining them together and fine-tuning. This indicates that the slice- and patch-networks actually contain complementary information for each other, which leads to improved performance in the final mortality risk prediction. It is also interesting to see that DSN outperformed DSN-scratch by 8% in terms of AUC score. That may be due to the difficulties in training the large concatenated network. The superior performance of our proposed DSN demonstrates the importance of having both well designed networks and good training strategy.

Table II:

Comparison showing the effectiveness of DSN using pre-trained segments. All the following experiments are conducted on color-coded images.

Method AUC STD
DSN-scratch 0.70 0.09
Slice-50 0.71 0.06
Patch-34 0.73 0.07
DSN 0.76 0.10

E. Feature Visualization

To help understand the features extracted by DSN, we compute the class activation map (CAM) by averaging the 512 × 7 × 7 feature maps from the patch-wise network with the corresponding weights of the last FC layer as in [27]. We also used t-Distributed Stochastic Neighbor Embedding (t-SNE) [28] to reduce the dimensionality of the feature maps to 2D for visualization. Fig. 6 shows the projection of validation samples from a randomly selected fold of the ten-fold cross validation scheme into 2D using t-SNE. From the point scattering shown in this figure, we can see that the positive and negative samples are roughly separated from each other, which indicates that DSN has the capability of extracting image features from LDCT images, which are strongly associated with the subject mortality.

Fig. 6:

Fig. 6:

The scatter plot in the center is the t-SNE representation of combined feature vectors Each feature vector is the concatenation of one 2048D feature vector (from slice-input network stream) and one 512D feature vector (from patch-input network stream). Each green dot represents a survivor and a red dot denotes a non-survivor. The green dots and red dots are roughly separated from each other. Eight cases (four with cardiac patches and the other four with whole slices) are randomly selected to show the generated heatmaps.

Fig. 6 also includes several examples with CAMs superimposed on the gray scale images as heatmaps. The closer to red in the heatmaps, the stronger activation there is in the original image, which indicates that information from that area contributes more to the final decision. As it can be seen from Fig. 6, the heatmaps for the deceased subjects predicted correctly by KAMP-Net tend to have strong activation over the coronary artery area in LDCT cropped cardiac areas, especially over the bright calcification region. This finding matches with the clinical literature that CAC is one of the major risk factors for mortality [4]. For survived subjects, the heatmaps suggest that KAMP-Net looks more at surrounding lung tissue and muscles as suggested by our previous work in [6]. For the heatmaps generated from image slices, survivors tend to have strong activation around the vertebral bone. It reflects the fact that subjects with higher bone density tends to be better health condition. In fact, two selected deceased subjects are both experiencing severe emphysema, and their generated heatmaps happen to highlight the emphysema region around the lungs.

IV. Discussions

The developed KAMP-Net is then compared against several other clinically used scoring methods for further validation. The results are shown in Fig. 7. It can be seen that the traditional semi-automatic methods, such as Agatston score [21], Agatston risk, muscle mass and fat attenuation perform similarly and the mean AUC values are in the range of [0.62, 0.65], which are slightly better than random guess. Emphysema severity itself alone cannot serve as a strong predictor (AUC = 0.55), which is consistent with the conclusions of a previous study [6]. It is interesting to see that the visual inspection of CAC by radiologists, with AUC of 0.64, outperforms the semi-automatic CAC scoring methods (Agatston score). This suggests that some information about the condition of cardiovascular vessels is not captured by those scoring methods, but has been taken into account by the radiologists.

Fig. 7:

Fig. 7:

Performance comparison of various methods on all-cause mortality risk prediction.

The significant performance improvement comes from the proposed KAMP-Net as shown in Fig. 7. The deep CNNs in DSN successfully extract and quantify features in cardiac patches and slices from chest LDCT images for all-cause mortality prediction, which couldn’t be directly measured by radiologists. The proposed KAMP-Net (with α = 0.75) achieves the best performance with AUC of 0.82, which improves the prediction performance by 28.1% over the visual inspection of radiologists.

V. Conclusions

In this paper, to accurately predict the all-cause mortality risk of a subject, we propose to combine multi-scale heterogeneous features. Those features are either automatically obtained from the images through training or manually defined by physicians based on their clinical knowledge. It has been shown that the patch-based and slice-based deep CNNs can complement each other in feature extraction for all-cause mortality prediction. Furthermore, incorporating the clinical measurements made by radiologists and summarized by a SVM model has yielded a significant performance improvement. This has led to the introduction of a novel method that combines the use of CNNs and a SVM models, which we have shown to produce a synergistic effect.

Our current study comes with the following limitations.

  1. We manually choose the slices to cover the most significant CAC. In fact, we can improve the consistency of evaluation by automatically extracting the key slices from a 3D volume.

  2. The dataset used in our current work is of limited size. We will enlarge the dataset by including more subjects to evaluate the performance of the proposed method in our future work.

  3. The clinical measurements used in this study are manually acquired. It is, however, recommended to incorporate automatic scoring methods for future work.

  4. Although the color-coding pre-processing of LDCT images has shown to be beneficial, the current thresholds and channel arrangement were manually set, which could be performed automatically in our future work.

VI. Acknowledgments

The authors thank the National Cancer Institute for access to NCI’s data collected by the National Lung Screening Trial. The statements contained herein are solely those of the authors and do not represent or imply concurrence or endorsement by NCI. The authors would also like to thank NVIDIA Corporation for the donation of the Titan Xp GPU used for this research.

This work was supported by National Heart, Lung, and Blood Institute (NHLBI) of the National Institutes of Health (NIH) under award R56HL145172.

Footnotes

The source code of this work is available at https://github.com/DIALRPI/KAMP-Net.

Contributor Information

Hengtao Guo, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY 12180, USA.

Uwe Kruger, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY 12180, USA.

Ge Wang, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY 12180, USA.

Mannudeep K. Kalra, Department of Radiology, Massachusetts General Hospital, Boston, MA 02114, USA

Pingkun Yan, Department of Biomedical Engineering and the Center for Biotechnology and Interdisciplinary Studies, Rensselaer Polytechnic Institute, Troy, NY 12180, USA.

References

  • [1].National Lung Screening Trial Research Team, Aberle DR, Adams AM, Berg CD, Black WC, Clapp JD, Fagerstrom RM, Gareen IF, Gatsonis C, Marcus PM, and Sicks JD, “Reduced lung-cancer mortality with low-dose computed tomographic screening,” The New England Journal of Medicine, vol. 365, no. 5, pp. 395–409, August 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Pope CA III, Burnett RT, Turner MC, Cohen A, Krewski D,Jerrett M, Gapstur SM, and Thun MJ, “Lung cancer and cardiovascular disease mortality associated with ambient air pollution and cigarette smoke: shape of the exposure–response relationships,” Environmental health perspectives, vol. 119, no. 11, pp. 1616–1621, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Omenn GS, Goodman GE, Thornquist MD, Balmes J, Cullen MR, Glass A, Keogh JP, Meyskens FL Jr, Valanis B, Williams JH Jr et al. , “Effects of a combination of beta carotene and vitamin a on lung cancer and cardiovascular disease,” New England journal of medicine, vol. 334, no. 18, pp. 1150–1155, 1996. [DOI] [PubMed] [Google Scholar]
  • [4].Chiles C, Duan F, Gladish GW, Ravenel JG, Baginski SG, Snyder BS, DeMello S, Desjardins SS, Munden RF, and NLST Study Team, “Association of coronary artery calcification and mortality in the national lung screening trial: A comparison of three scoring methods,” Radiology, vol. 276, no. 1, pp. 82–90, July 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [5].Jacobs PC, Gondrie MJA, van der Graaf Y, de Koning HJ,Isgum I, van Ginneken B, and Mali WPTM, “Coronary artery calcium can predict all-cause mortality and cardiovascular events on low-dose ct screening for lung cancer,” American Journal of Roentgenology, vol. 198, no. 3, pp. 505–511, March 2012. [DOI] [PubMed] [Google Scholar]
  • [6].Digumarthy SR, De Man R, Canellas R, Otrakji A, Wang G, and Kalra MK, “Multifactorial analysis of mortality in screening detected lung cancer,” Journal of Oncology, vol. 2018, p. 7, 2018. [Online]. Available: 10.1155/2018/1296246 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [7].Cano-Espinosa C, González G, Washko GR, Cazorla M, and Estépar RSJ, “Automated agatston score computation in non-ecg gated ct scans using deep learning,” in Proceedings of SPIE–the International Society for Optical Engineering, vol. 10574 NIH Public Access, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [8].de Vos BD, Wolterink JM, Leiner T, de Jong PA, Lessmann N, and Išgum I, “Direct automatic coronary calcium scoring in cardiac and chest ct,” IEEE transactions on medical imaging, 2019. [DOI] [PubMed] [Google Scholar]
  • [9].Lessmann N, van Ginneken B, Zreik M, de Jong PA, de Vos BD,Viergever MA, and Isgum I, “Automatic calcium scoring in low-dose chest CT using deep neural networks with dilated convolutions,” IEEE Transactions on Medical Imaging, vol. 37, no. 2, pp. 615–625, February 2018. [DOI] [PubMed] [Google Scholar]
  • [10].van Velzen SGM, Zreik M, Lessmann N, Viergever MA,de Jong PA, Verkooijen HM, and Igum I, “Direct Prediction of Cardiovascular Mortality from Low-dose Chest CT using Deep Learning,” arXiv:1810.02277 [cs], October 2018, arXiv: 1810.02277. [Online]. Available: http://arxiv.org/abs/1810.02277 [Google Scholar]
  • [11].Li G and Yu Y, “Visual saliency based on multiscale deep features,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 5455–5463. [Google Scholar]
  • [12].Fu H, Xu S, abd Lin Yanwu, Wong DWK, Mani B, Mahesh M, Tin A, and Liu J, “Multi-context deep network for angle-closure glaucoma screening in anterior segment OCT,” in Medical Image Computing and Computer Assisted Intervention (MICCAI), October 2018, pp. 356–363. [Google Scholar]
  • [13].Shemesh J, “Coronary artery calcification in clinical practice: what we have learned and why should it routinely be reported on chest CT?” Annals of Translational Medicine, vol. 4, no. 8, April 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [14].Wolterink JM, Leiner T, Viergever MA, and Igum I, “Automatic Coronary Calcium Scoring in Cardiac CT Angiography Using Convolutional Neural Networks,” in Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015, ser. Lecture Notes in Computer Science; Springer, Cham, October 2015, pp. 589–596. [Google Scholar]
  • [15].Viola P, Jones M et al. , “Rapid object detection using a boosted cascade of simple features,” CVPR (1), vol. 1, pp. 511–518, 2001. [Google Scholar]
  • [16].He K, Zhang X, Ren S, and Sun J, “Deep Residual Learning for Image Recognition,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2016, pp. 770–778. [Google Scholar]
  • [17].Yan P, Guo H, Wang G, De Man R, and Kalra MK, “Hybrid deep neural networks for all-cause mortality prediction from LDCT images,” arXiv:1810.08503 [cs.CV], October 2018. [DOI] [PubMed] [Google Scholar]
  • [18].Paszke A, Gross S, Chintala S, Chanan G, Yang E, DeVito Z, Lin Z,Desmaison A, Antiga L, and Lerer A, “Automatic differentiation in pytorch,” in NIPS 2017 Workshop Autodiff, 2017. [Google Scholar]
  • [19].Kingma DP and Ba J, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014. [Google Scholar]
  • [20].Shin H-C, Roth HR, Gao M, Lu L, Xu Z, Nogues I, Yao J,Mollura D, and Summers RM, “Deep convolutional neural networks for computer-aided detection: Cnn architectures, dataset characteristics and transfer learning,” IEEE transactions on medical imaging, vol. 35, no. 5, pp. 1285–1298, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [21].Agatston AS, Janowitz WR, Hildner FJ, Zusmer NR, Viamonte M, and Detrano R, “Quantification of coronary artery calcium using ultrafast computed tomography,” Journal of the American College of Cardiology, vol. 15, no. 4, pp. 827–832, March 1990. [DOI] [PubMed] [Google Scholar]
  • [22].Callister TQ, Cooil B, Raya SP, Lippolis NJ, Russo DJ, and Raggi P, “Coronary artery disease: improved reproducibility of calcium scoring with an electron-beam CT volumetric method.” Radiology, vol. 208, no. 3, pp. 807–814, September 1998. [DOI] [PubMed] [Google Scholar]
  • [23].González G, Washko GR, and Estépar RSJ, “Automated agatston score computation in a large dataset of non ecg-gated chest computed tomography,” in 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI) IEEE, 2016, pp. 53–57. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [24].Chin J, Syrek Jensen T, Ashby L, Hermansen J, Hutter JD, and Conway PH, “Screening for Lung Cancer with Low-Dose CT Translating Science into Medicare Coverage Policy,” New England Journal of Medicine, vol. 372, no. 22, pp. 2083–2085, May 2015. [DOI] [PubMed] [Google Scholar]
  • [25].Krizhevsky A, Sutskever I, and Hinton GE, “ImageNet classification with deep convolutional neural networks,” in Advances in Neural Information Processing Systems (NIPS), 2012. [Google Scholar]
  • [26].Montgomery DC and Runger GC, Applied Statistics and Probability for Engineers, 5th Edition Hoboken, NJ: John Wiley & Sons, 2010. [Google Scholar]
  • [27].Zhou B, Khosla A, Lapedriza A, Oliva A, and Torralba A, “Learning deep features for discriminative localization,” in Computer Vision and Pattern Recognition (CVPR), 2016, pp. 2921–2929. [Google Scholar]
  • [28].van der Maaten L and Hinton G, “Visualizing data using t-SNE,” Journal of Machine Learning Research, vol. 9, no. 11, pp. 2579–2605, 2008. [Google Scholar]

RESOURCES