Skip to main content
PLOS ONE logoLink to PLOS ONE
. 2019 Jul 12;14(7):e0219369. doi: 10.1371/journal.pone.0219369

A CAD system for pulmonary nodule prediction based on deep three-dimensional convolutional neural networks and ensemble learning

Wenkai Huang 1, Yihao Xue 2, Yu Wu 3,*
Editor: Yuchen Qiu4
PMCID: PMC6625700  PMID: 31299053

Abstract

Background

Detection of pulmonary nodules is an important aspect of an automatic detection system. Incomputer-aided diagnosis (CAD) systems, the ability to detect pulmonary nodules is highly important, which plays an important role in the diagnosis and early treatment of lung cancer. Currently, the detection of pulmonary nodules depends mainly on doctor experience, which varies. This paper aims to address the challenge of pulmonary nodule detection more effectively.

Methods

A method for detecting pulmonary nodules based on an improved neural network is presented in this paper. Nodules are clusters of tissue with a diameter of 3 mm to 30 mm in the pulmonary parenchyma. Because pulmonary nodules are similar to other lung structures and have a low density, false positive nodules often occur. Thus, our team proposed an improved convolutional neural network (CNN) framework to detect nodules. First, a nonsharpening mask is used to enhance the nodules in computed tomography (CT) images; then, CT images of 512×512 pixels are segmented into smaller images of 96×96 pixels. Second, in the 96×96 pixel images which contain or exclude pulmonary nodules, the plaques corresponding to positive and negative samples are segmented. Third, CT images segmented into 96×96 pixels are down-sampled to 64×64 and 32×32 size respectively. Fourth, an improved fusion neural network structure is constructed that consists of three three-dimensional convolutional neural networks, designated as CNN-1, CNN-2, and CNN-3, to detect false positive pulmonary nodules. The networks’ input sizes are 32×32×32, 64×64×64, and 96×96×96 and include 5, 7, and 9 layers, respectively. Finally, we use the AdaBoost classifier to fuse the results of CNN-1, CNN-2, and CNN-3. We call this new neural network framework the Amalgamated-Convolutional Neural Network (A-CNN) and use it to detect pulmonary nodules.

Findings

Our team trained A-CNN using the LUNA16 and Ali Tianchi datasets and evaluated its performance using the LUNA16 dataset. We discarded nodules less than 5mm in diameter. When the average number of false positives per scan was 0.125 and 0.25, the sensitivity of A-CNN reached as high as 81.7% and 85.1%, respectively.

Introduction

Lung cancer causes more deaths worldwide than any other type of cancer. In the past 50 years, the incidence of lung cancer has significantly increased; it now ranks first in malignant tumors in males and has become a major threat to life and health. However, when an early diagnosis can be made, the 5-year survival rate increases from 10%–16% to 70%. In lung CT, automatic detection of pulmonary nodules plays an important role in a computer-aided diagnosis (CAD) system. Hawkins et al. [1] determined the predictive value of a CT scan for malignant nodules. Using radio physics, lung cancer screenings can be used to assess the risk of lung cancer development through baseline CT scanning.

Pulmonary nodules develop into tumors through a gradual process. People should pay attention to malignant pulmonary nodules and intervene as soon as possible. When malignant nodules are found, patients can undergo operations and improve their survival rate. Pulmonary nodule detection is a key step in improving the sensitivity of the detection of malignant nodules. This study will improve the accuracy of pulmonary nodule detection by helping doctors to detect pulmonary nodules and judge whether they are benign or malignant.

CAD systems are divided into two subcategories: computer-aided detection (CADe) and computer-aided diagnosis (CADx) systems. A CAD system for detecting pulmonary nodules usually involves four stages: lung segmentation, nodule detection, feature analysis, and false positive elimination [2]. To meet these challenges, many researchers have explored extracting information from pulmonary nodules to improve the sensitivity of nodule detection. In the existing approaches, researchers manually extract the features and properties of nodes in CT images and input them into a classifier or neural network such as a Support Vector Machine (SVM) [3], CNN [4], or fully connected neural network [5] to classify nodules and perform semantic feature analysis [6].

In recent years, deep learning methods have been introduced into the field of medical image analysis, achieving promising results [7], [8]. CNNs are one of the most popular deep learning methods. In this study, a new neural network is proposed to detect pulmonary nodules that can optimize some of the parameters required to detect pulmonary nodules and improve the sensitivity of the detection steps involved in pulmonary nodule detection ina CAD system. Fig 1 shows a basic CNN flowchart for detecting pulmonary nodules.

Fig 1. CNN detection of pulmonary nodules.

Fig 1

Since a CNN first won the ImageNet-Large-Scale-Visual-Recognition-Challenge (ILSVRC)in 2012, they have sparked great achievements in computer vision software applications. In addition, CNNs play an effective role in the field of medical image analysis, especially in detecting nodules and judging their benign vs. malignant qualities.

For nodule detection, Zhao et al. [9] proposed a detection method based onpositron emission tomography (PET) and CT scan data to differentiate malignant from benign pulmonary nodules. Their method combined the characteristics of PET and CT signals and they proposed an improved FP-dimension reduction method to detect pulmonary nodules in PET/CT images using convolutional neural networks. The accuracy of this method reached as high as 95.6%. Shen et al. [10] proposed the Multi-Convolutional Neural Network (MC-CNN) model by adopting a new multicrop aggregation strategy. As different regions are extracted from the convolution feature map, the nodule information is automatically extracted by using maximum pooling at different times. Tajbakhsh et al. [11] compared the performance of two machine learning models, Massive Training Artificial Neural Networks (MTANNs) and CNNs, to detect and classify pulmonary nodules. After using two optimized MTANNs and four different CNN architectures, their experiments showed that MTANNs achieved a better performance. Xia et al. [12] proposed an effective thyroid cancer detection system based on an Extreme Learning Machine (ELM). Due to the characteristics of ultrasound images, the potential of ELM in differentiating benign and malignant thyroid nodules was discussed. The proposed ELM-based method achieved an accuracy of 87.72%, a sensitivity of 86.72%, and a specificity of 94.55%. Xie et al. [13] proposed an automatic detection framework for pulmonary nodules based on a 2D-convolutional neural network. First, two regional networks and one deconvolutional layer were used to adjust the structure of the R-CNN to detect candidate nodules. Then, three different models were trained to obtain the subsequent fused results. Second, an enhanced structure based on 2D-CNN was designed to reduce false positives. Finally, these neural networks were fused to obtain the final classification results. The sensitivity of nodule detection was 86.42%. To reduce the false positives, the sensitivity was 73.4% and 74.4%, respectively, when 1/8 FPs/scan and 1/4 FPs/scan were scanned. Mei et al. [14] proposed using a random forest to predict the malignancy and invasiveness of pulmonary ground-glass nodules (GGNs). The results showed that the malignancy degree of GGNs was 95.1% and the invasive rate of GGNs was 83%, indicating that a random forest model could predict these qualities from lung images. Farahani et al. [15] developed a method to improve the quality of the original CT image based on a type II fuzzy algorithm and proposed a new opportunity-based fuzzy C-means clustering segmentation algorithm to detect nodules. They utilized three classifiers: a multi-layer perceptron (MLP), k-Nearest Neighbor (KNN), and an SVM. Pulmonary nodules were aggregated for benign and malignant diagnoses. Nibali et al. [16] classified pulmonary nodules based on deep residual networks. Based on the latest ResNet architecture, they explored the impacts of curriculum learning, migration learning, and network depth changes on the accuracy of malignant nodule classification. Zhu et al. [17] explored how to tap this vast, but currently unexplored data source to improve pulmonary nodule detection. We propose DeepEM, a novel deep 3D ConvNet framework augmented with expectation-maximization (EM), to mine weakly supervised labels in EMRs for pulmonary nodule detection, respectively, demonstrating the utility of incomplete information in EMRs for improving deep learning algorithms. Wu et al. [18] proposed a deep-neural-network-based detection system for lung nodule detection in CT. A primal-dual-type deep reconstruction network was applied first to convert the raw data to the image space, followed by a 3D-CNN for the nodule detection. With 144 multi-slice fan beam projections, the proposed end-to-end detector could achieve comparable sensitivity with the reference detector, which was trained and applied on the fully-sampled image data. It also demonstrated superior detection performance compared to detectors trained on the reconstructed images. Zhao et al. [19] presented a deep-learning based automatic all size pulmonary nodule detection system by cascading two artificial neural networks. They firstly use a U-net like 3D network to generate nodule candidates from CT images. Then, They use another 3D neural network to refine the locations of the nodule candidates generated from the previous subsystem. With the second sub-system, we bring the nodule candidates closer to the center of the ground truth nodule locations, their system has achieved high sensitivity. Khosravan and Bagci [20] propose S4ND, a new deep learning based method for lung nodule detection. Our approach uses a single feed forward pass of a single network for detection. The whole detection pipeline is designed as a single 3D CNN with dense connections, trained in an end-to-end manner, a high FROC-score have been achieved in the detection of pulmonary nodules. To reduce the incidence of false positive nodules, Zainudin et al. [21]used the deep learning structure of CNN to improve the detection rate of tumors through histopathological images. The performance evaluation of deep learning of CNN was been compared to other classifier algorithm and the result was better accuracy on detecting the mitosis and non-mitosis. Teramoto et al. [22] proposed an improved False Positive (FP)-dimensional reduction method to detect pulmonary nodules in PET/CT images using a CNN. Their FP-dimensional reduction method eliminated 93% of the false positive nodules, and their proposed method used CNN technology to eliminate approximately half of the false positive nodules compared to previous studies. Setio et al. [23] proposed a CADe system that used multi-view convolutional networks (CuNETs) to obtain nodule candidates by combining three candidate detectors specially designed for solid, subsolid, and large nodules. The false-positive nodules among them are then classified along with the false-negative nodules. The accuracy of the section scan reached 90.1%. Cao et al. [24] proposed a new method to compress FP nodules in CT images. First, 87 features were extracted from the candidate nodules, which were divided into five categories (shape, gray level, surface, gradient, and texture) by an SVM and then combined before selecting the most influential features for classification. Then, voxel values were obtained based on MTANNs, and the final result achieved 91.95% accuracy. Suzuki et al. [25]investigated a pattern-recognition technique based on an artificial neural network (ANN), which is called a MTANN, for reduction of false positives in computerized detection of lung nodules in low-dose CT images. By using the Multi-MTANN, the false-positive rate of our current scheme was improved from 0.98 to 0.18 false positives per section (from 27.4 to 4.8 per patient) at an overall sensitivity of 80.3% (57/71). Dou et al. [26] used a new 3D-CNN framework to automatically detect pulmonary nodules, reduce FPs and more effectively distinguish them from their hard simulators. Among the many methods that have been explored, CNNs are most commonly used to detect pulmonary nodules and pseudo positive pulmonary nodules because they can achieve a high sensitivity to reduce FPs.

In a CNN, the convolutional and pooling layers learn the local spatial structure of the input image, while a later fully connected layer learns at a more abstract level that includes more global information. However, because a traditional neural network uses the fullyconnected layer as the final classification layer, some details and spatial features of the image will be lost when the final feature is displayed. Therefore, the outline of the nodule cannot be judged based on this approach—true nodulesare highly similar to FP nodules (as shown in Fig 2). But in some details, there are many differences between positive pulmonary nodules and false positive pulmonary nodules. For example, the margin of positive pulmonary nodules is coronal radiation, with needle-like protuberances on the margin, non-smooth margins, transparent areas in the interior and ground glass elements; while the margin of false positive nodules is smooth, without ground glass elements. Therefore, it is difficult to identify FP pulmonary nodules using a traditional CNN, and traditional CNNs achieve low accuracy scores for malignant pulmonary nodule discovery. For this reason, our team proposed a new network framework, called A-CNN, to improve the sensitivity of lung nodule detection for nodules of 5–30 mm.

Fig 2. Comparison of true pulmonary nodules and false positive pulmonary nodules.

Fig 2

Materials and methods

A, Materials

Our experimental dataset was derived from the publicly available LUNA16 dataset and the Ali Tianchi dataset. The LUNA16 dataset comes from a larger dataset, the Lung Image Database Consortium and Image Database Resource Initiative (LIDC/IDRI) database. In both the LUNA16 dataset and the Ali Tianchi medical dataset the slice thickness is less than 2.5 mm. Our dataset included 1,888 CT scans, each of which contained a series of axial sections of the thoracic cavity. The number of slices included in each image varied according to the scanning machine, scanning thickness, and patient. The original images are three-dimensional. Each image contain a series of axial sections of the thorax. This three-dimensional image consists of different numbers of two-dimensional images. The number of two-dimensional images can vary based on factors such as the scanning machine used and the patients. The reference standard for these datasets is that 3 of 4 radiologists accept nodes larger than 3 mm. The nodules excluded from the reference criteria (which included non nodules, nodules smaller than 3mm, and nodules accepted by only 1–2 radiologists) were considered unrelated. We excluded nodules below 5 mm, and tested only pulmonary nodules with diameters ranging from 5 mm to 30 mm for false positives. Fig 3 shows the size distribution of the pulmonary nodules in the LUNA16 dataset, and Fig 4 shows the size distribution of the pulmonary nodules in the Ali Tianchi dataset.

Fig 3. Distribution of pulmonary nodule size in the LUNA16 dataset.

Fig 3

Fig 4. Size distribution map of pulmonary nodules in the Ali Tianchi dataset.

Fig 4

B, Preprocessing

We preprocessed the LUNA16 and Ali Tianchi datasets in three stages. First, we excluded nodules smaller than 5mmbased on the dataset labels, leaving only the nodes between 5 mm and 30 mm. Then, we segmented all the lung sections from the LUNA16 and Ali Tianchi dataset. Lung segmentation is challenging because no heterogeneity exists in the lung region. Additionally, the complex structure and uneven gray distribution of lung CT images makes it difficult to accurately segment and extract the lung parenchyma. Preprocessing lung CT images is a key step in a CADe system. Preprocessing involves removing the useless parts and redundant clutter to improve the image quality, allowing better results to be achieved to reduce FPs. For lung segmentation, the existing techniques can be divided into four categories: the threshold-based method, the variable model, and the shape or edge-based models [27]. In this paper, we use unsharp mask (UM) image sharpening technology to enhance the nodule signal in Charge eXchange Recombination Spectroscopy (CXRS). A technical flowchart is shown in Fig 5.

Fig 5. Flowchart of unsharp mask image sharpening technology.

Fig 5

The formula is as follows:

y(n,m)=x(n,m)+λz(n,m), (1)

where x(n,m) is the input image, y(n,m) is the output image, and z(n,m) is a correction signalusually obtained through high-pass filtering of x. Here, λ is a scaling factor used to control the enhancement effect. In the UM algorithm, z(n,m) can be obtained by Formula (2):

z(n,m)=4x(n,m)-x(n-1,m)-x(n+1,m)-x(n,m-1)-x(n,m+1). (2)

In the second pretreatment stage, the team rotated the LUNA16 dataset and the Ali Tianchi dataset clockwise by 90°, 180°, and 270°, respectively. In the third pretreatment stage, our team divided the rotated images from the original LUNA16 dataset and the CT images from the Ali Tianchi dataset into 96×96 images and down sampled them to generate 32×32, 64×64, and 96×96 images, resulting in 326,570 training samples. In these samples, there are 18,125 pulmonary nodules.

C, Training

The CNN is composed of several convolutional layers, pooling layers, and global average pooling layers. Small 3D images can be tailored around the markers in the whole CT image; these small 3D images correspond directly to the nodule labels. By learning these features, a neural network is trained to detect pulmonary nodules, evaluate the malignance degree of nodules, and predict the possibility of cancer in patients. During prediction, the neural network traverses the entire CT image and judges the possibility that malignant information occurs in each position of a sliding window. The above process is illustrated in Fig 6.

Fig 6. A neural network structure for detecting pseudo positive pulmonary nodules.

Fig 6

The activation function is very important during CNN calculations. Traditional neural networks usually use the rectified linear unit (ReLU) function as an activation function because compared with other linear functions, ReLU is more expressive, especially in deep networks. For nonlinear functions, ReLU does not have the vanishing gradient problem because the gradient of nonnegative intervals is constant, and the convergence rate of the model is stably maintained. The problem of gradient vanishing is that when the gradient becomes less than 1, the error between the predicted value and the real value will decay once per layer of propagation. This phenomenon is particularly obvious when a sigmoid function is used as the activation function in a deep model and will lead to stagnation during convergence. The mathematical expression of the ReLU function is shown in Formula (3):

f(x)={0,x0x,x0. (3)

When the input value to ReLU is negative, the output and first derivative are always 0. This causes the neurons to fail to update the parameters, and this phenomenon is called a Dead Neuron. To address the shortcomings of the ReLU function, our team used a new activation function—Random Translation ReLU (RT-ReLU) [28]as the activation function of our A-CNN fusion neural network. Cao et al. made full use of the input distribution of the nonlinear activation function and proposed the random translation nonlinear activation function for deep CNNs. During training, the nonlinear activation function is randomly converted from the offset of the Gaussian distribution, while during testing, a nonlinear activation with zero offset is used. Based on the method proposed by Cao et al., the input distribution of nonlinear activation is relatively decentralized, meaning that the acquired CNN is robust to small fluctuations of the nonlinear activation input. Thus, RT-ReLU is conducive to reducing over fitting of the neural network. The mathematical expression of RT-ReLU is shown in Formula (4):

yi={xi+ai,ifxi+ai>00,ifxi+ai0. (4)

The functional models of the ReLU and RT-ReLU functions are shown in Fig 7:

Fig 7. Contrast graph of ReLU function and RT-ReLU function.

Fig 7

A single CNN has a limited learning ability; therefore, while a single CNN may be able to identify all the basic features necessary to distinguish nodules from different types of nonnodular structures, using multiple different CNNs can distinguish more nonnodules. Therefore, our team proposed the A-CNN to detect pulmonary nodules.

We constructed three types of 3D-CNN structures for different input scales, namely, CNN-1, CNN-2, and CNN-3. CNN-1 used 32×32×32 CT segmentation images as input, CNN-2 used 64×64×64 images, and CNN-3 used 96×96×96 images as input. The structures of these CNNs are shown in Table 1. The structure of the proposed A-CNN is shown in Fig 8. The results of the three models are fused using an AdaBoost classifier.

Table 1. Single CNN structure table.

Layer Type Input Kernel Stride Pad Output
CNN-1 0 Input 32×32×32 N/A N/A N/A 32×32×32
1 Convolutional 32×32×32 5×5×5 1 0 20×32×32×32
2 Max pooling 20×32×32×32 2×2×2 2 0 20×14×14×14
3 Convolutional 20×14×14×14 5×5×5 1 0 50×10×10×10
4 Global Average Pooling
CNN-2 0 Input 64×64×64 N/A N/A N/A 64×64×64
1 Convolutional 64×64×64 5×5×5 1 0 20×60×60×60
2 Max pooling 20×60×60×60 2×2×2 2 0 20×30×30×30
3 Convolutional 20×30×30×30 5×5×5 1 0 50×26×26×26
4 Max pooling 50×26×26×26 2×2×2 2 0 50×13×13×13
5 Convolutional 50×13×13×13 4×4×4 1 0 100×10×10×10
6 Global Average Pooling
CNN-3 0 Input 96×96×96 N/A N/A N/A 96×96×96
1 Convolutional 96×96×96 5×5×5 1 0 20×92×92×92
2 Max pooling 20×92×92×92 2×2×2 2 0 20×46×46×46
3 Convolutional 20×46×46×46 5×5×5 1 0 50×42×42×42
4 Max pooling 50×42×42×42 2×2×2 2 0 50×21×21×21
5 Convolutional 50×21×21×21 5×5×5 1 0 100×17×17×17
6 Max pooling 100×17×17×17 2×2×2 2 0 100×9×9×9
7 Convolutional 100×9×9×9 5×5×5 1 0 100×5×5×5
8 Global Average Pooling

Fig 8. The structure of the neural network.

Fig 8

Different CNN architectures can process different input images and different scales of CT images. The A-CNN model proposed by our team combines the advantages of CNN-1, CNN-2, and CNN-3. We conducted a random 10-fold validation test to verify our algorithm and demonstrate its high robustness.

D, AdaBoost classifier reduces false positive nodules

The principle underlying A-CNN is shown in Fig 8. The boosting family of algorithms can upgrade weak learners to strong learners. The core idea of boosting is to train different classifiers (weak classifiers) using the same training set and then create an ensemble of those trained weak classifiers to form a stronger final classifier (strong classifier). The algorithm itself changes the data distribution by determining the weight of each sample based on whether the classification of each sample in each training set is correct, including the accuracy of the last overall classification. The new dataset with modified weights is sent to the individual classifiers for training. Finally, the classifiers obtained from each training operation are fused to form the final classifier.

The working approach to A-CNN is to train CNN-1 from the training set with the initial weights, update the weights of the training samples according to the learning error rate performance of CNN-1, make the weights of the training samples with high learning error rates by CNN-1 become higher to pay more attention to the points with high learning error rates in the subsequent CNN-2. Then, CNN-2 is trained based on the training set after weight adjustment. The points with high error rates in CNN-2 are treated similarly so that they will receive more attention in the subsequent CNN-3. Then, CNN-3 is trained based on the training set after the weight adjustment. This operation is repeated until CNN-1, CNN-2 and CNN-3 are eventually integrated through a set strategy to form the final A-CNN. The specific algorithm operation is as follows:

Recognition of pulmonary nodules is a two-class problem. The input training set is shown in Formula (5):

T={(x1,y1),(x2,y2),(xm,ym)}, (5)

whose output is {-1, +1}.

Initialize the training set weight as shown in Formula (6):

D(1)=(ω11,ω12ω1m);ω1i=1m;i=1,2m, (6)

For k = 1,2…K:

The weighted Dk training set is used to train the data, and the weak classifier Gk(x) is obtained.

The classification error rate of Gk(x) is calculated as shown in Formula (7):

ek=P(Gk(xi)yi)=i=1mωkiI(Gk(xi)yi), (7)

The coefficients of each weak classifier are calculated as shown in Formula (8):

αk=12log1-ekek, (8)

where ek represents the classification error rate, and αk represents the weight coefficient of the weak classifier. As the above formula shows, when the classification error rate ek is larger, the corresponding weak classifier weight coefficient αk is smaller; in other words, the smaller the error rate is, the larger the weight coefficient of the weak classifier is.

Update the sample weight D. Assuming that the sample set weight coefficient of the k th weak classifier is D(K)=(ωk1,ωk2ωkm), the corresponding sample set weight coefficient of the (k+1)th weak classifier is shown in formula (9):

ωk+1,i=ωkiZKexp(-αkyiGk(xi))i=1,2m, (9)

where Zk is the normalization factor, as shown in Formula (10):

Zk=i=1mωkiexp(-αkyiGk(xi)), (10)

From the formula for ωk+1,i, it can be seen that when the classification of the first sample is incorrect, yiGk(xi)<0, which leads to an increased weight of the sample in the k+1 weak classifier. Similarly, a decrease in the weight occurs in the k+1 weak classifier when the classification is correct.

The final classifier is constructed as shown in Formula (11):

f(x)=sign(k=1KαkGk(x)), (11)

E, Evaluation criterion

Finally, we validated the new detection method of pulmonary nodules with this new neural network. Our team uses metrics commonly used in CAD systems and widely used for performance analyses of image processing systems [29].

Sensitivity (SEN), also known as the true positive rate, is expressed by dividing the number of true positives by the total number of positive cases, as shown in Formula (12):

SEN=TPTP+FN, (12)

where TP denotes true positive, FN denotes false positives.

In the two-class problem of false positive reduction, we use thefree-responsereceiver operating characteristic (FROC) curve to evaluate our A-CNN model. The FROC curve is a coordinate map composed of the average number of false positives per scan as the horizontal axis, the sensitivity as the vertical axis, and the subject’s specific peak. Under excitation conditions, the curves drawn from different results are obtained by different criteria. Traditional diagnostic test evaluation methods have a common feature. The test results must be divided into two categories and then statistically analyzed. The evaluation method of the FROC curve is different from the traditional evaluation method. It does not require this restriction; instead, it allows the evaluation of multiple anomalies on an image based on the actual situation, making the FROC curve evaluation method more applicable. Therefore, we adopted the FROC curve as the evaluation criterion.

Experiment and results

This section presents and discusses the results obtained from the detection of pulmonary nodules from CT scans. We implemented the proposed neural network A-CNN to detect pulmonary nodules in the Python programming language on TensorFlow using the Keras deep learning library. We executed the model on a computer equipped with an Intel I7-8700 processor, 16 GB RAM, an Nvidia GeFrand GTX 1080Ti, and a Windows 10 operating system.

We preprocessed the LUNA16 dataset and the lung nodule slices from the Ali Tianchi dataset and obtained 326,570 slices. To test the effective detection of the new A-CNN model, we randomly divided the processed datasets into three groups: training, verification, and testing. The training set contains 266,570 slices, the verification set contains 30,000 slices and the test set contains 30,000 slices. Among these, the training set contains 14,674 nodules, the test set contains 1,795 nodules, and the verification set contains 1,656 nodules.

We trained the CNN-1, CNN-2, CNN-3, and the fusion neural network A-CNN; then, we verified and tested them using the same verification and test set to achieve a fair comparison between each neural network. Based on the cross-entropy loss function, we used stochastic gradient descent to supervise the training of each neural network, where the difference between the expected and current output of the network is continuously reduced by adjusting the weight. The weights were initialized to a standard value and updated by a standard back propagation algorithm for 30 epochs. The weights were attenuated to 0.0005, and the learning rate was set to 0.001.

We tested the false positive reduction performance of the fusion A-CNN on the LUNA16 and Ali Tianchi datasets. The results show that when the average number of false positives per scan is 0.125 FPs/scan, 0.25 FPs/scan, 0.5 FPs/scan, 1 FPs/scan, 2 FPs/scan, 4 FPs/scan, and 8 FPs/scan, the model achieves sensitivity values of 0.817, 0.851, 0.869, 0.883, 0.891, 0.907, and 0.914, respectively. We conducted a random 10-fold validation test to verify the robustness of the algorithm. As shown in Fig 9, A-CNN has high robustness.

Fig 9. FROC curve for pulmonary nodule detection.

Fig 9

FROC curves using the three individual CNNs and A-CNN are shown in Fig 10. When the average number of false positives per scan is 0.125 FPs/scan and 0.25 FPs/scan, the sensitivity of A-CNN reaches 81.7% and 85.1%, while the sensitivities of the other four single CNNs can only be up to 62.3% and 62.9%, respectively (as shown in Fig 10). These results demonstrate that A-CNN is significantly better than are the individual CNN-1, CNN-2, CNN-3and VGG16 models.

Fig 10. Performance comparison of A-CNN with CNN-1, CNN-2 and CNN-3.

Fig 10

The AdaBoost classifier is suitable for fusing multiple weak classifiers and it performs well with weak classifiers. Therefore, our team compared CNN-1, CNN-2, and CNN-3 with the VGG16 neural network. As Fig 10 shows, VGG16 performs better than the other three neural networks. And our team compared A-CNN with VGG16+AdaBoost neural network, A-CNN showed better performance. Therefore, the AdaBoost classifier is suitable for use in A-CNN in this paper.

In this study, we used UM image sharpening technology to enhance the nodule signals in CXRS. Therefore, as verification, we input the original images and the preprocessed images into A-CNN and compared the average sensitivity between the two types of input, as shown in Table 2. The results show that UM image processing can slightly increase the sensitivity of nodule detection. Although the average sensitivity difference before and after UM image processing is only 0.015, this result shows that UM processing of CT image is beneficial to improve the overall detection sensitivity.

Table 2. Comparison of average sensitivity between input of original images and those after image sharpening.

Input Average result
Original graph 0.876
Unsharp Mask 0.861

The relevant works cited in this paper do not provide their training, validation, or the test cases used, which makes it difficult to compare the experimental results reported in this paper to those of other related works. The only information the other works provided was the database used; consequently, we were unable to perform a rigorous methodological evaluation of other works, and are able to compare the FROC curve of pulmonary nodules only with the results of the known literature.

Table 3 compares the false positive results of the LUNA16 challenge between our team and those of other teams. We compare lung nodule test results only with other published test methods. Each team used a deep CNN to exclude false positive results. Two of the nine teams are based on 2D-CNN architectures, while seven are based on 3D-CNN architectures. In this study, we used a 3D-CNN framework [30] to detect pulmonary nodules. 3D-CNNsare conducive to the analysis and detection of pulmonary nodules from a stereoscopic perspective, and their pulmonary nodules detections are more sensitive than those of 2D-CNNs. It is worth noting that when the average number of false positives per scan is 0.125FPs/scan and 0.25FPs/scan, the sensitivity of A-CNN can reach 81.7% and 85.1%, respectively, which represent the best performances. Fig 11 clearly shows the sensitivity of A-CNN for the detection of pulmonary nodules compared to the best results by other teams.

Table 3. Sensitivity of pulmonary nodules compared with other groups.

Team CNN 0.125 0.25 0.5 1 2 4 8
Xie[19] 2D 0.734 0.744 0.763 0.796 0.824 0.832 0.834
RESNET 2D 0.675 0.745 0.813 0.879 0.903 0.939 0.948
MEDICAI 3D 0.744 0.802 0.847 0.881 0.901 0.927 0.928
Aidence 3D 0.747 0.801 0.834 0.872 0.922 0.963 0.976
3DCNN_NDET 3D 0.733 0.809 0.872 0.918 0.937 0.960 0.964
CASED 3D 0.775 0.835 0.869 0.903 0.928 0.953 0.965
iDST-VCT 3D 0.745 0.827 0.889 0.931 0.964 0.972 0.975
qfpxfd 3D 0.763 0.846 0.891 0.848 0.919 0.943 0.949
A-CNN 3D 0.817 0.851 0.869 0.883 0.891 0.907 0.914

Fig 11. A comparison of pulmonary nodules detection results between A-CNN and the best results of other methods.

Fig 11

Compared to other methods, A-CNN achieved highsensitivity and better detection results for pulmonary nodules. For a CAD system, sensitivity is a key measure because it reflects the performance of the model in correctly classifying nodules and achieving more convenient medical monitoring.

Conclusions

In this study, three CNN frameworks were establishedand then fused into a new A-CNN model to detect pulmonary nodules. The A-CNN network model proposed by our team was validated on the LUNA16 dataset, where it achieved sensitivity scores of 81.7% and 85.1% whenthe average false positives number per scan was 0.125 FPs/scan and 0.25 FPs/scan, respectively. These scores were 5.4% and 0.5% higher than those of the current optimal algorithm; thus, our A-CNN model can improve the efficiency of radiological analysis of CT.

Acknowledgments

This work was supported in part by the Social Sciences and Humanities of theMinistry of Education of Chinaunder Grant 18YJC88002, in part by the Guangdong Provincial Key Platform and Major Scientific Research Projects Featured Innovation Projects under Grant 2017GXJK136, and in part by the Guangzhou Innovation and Entrepreneurship Education Project under Grant 201709P14.

Data Availability

The Ali Tianchi data set used in this manuscript can be obtained on Aliyun official website (https://tianchi.aliyun.com/competition/entrance/231601/information), and the LUNA16 data set can be obtained on LUNA16 official website (https://luna16.grand-challenge.org/Download/).

Funding Statement

This work was supported in part by the Social Sciences and Humanities of the Ministry of Education of China under Grant 18YJC88002, in part by the Guangdong Provincial Key Platform and Major Scientific Research Projects Featured Innovation Projects under Grant 2017GXJK136, and in part by the Guangzhou Innovation and Entrepreneurship Education Project under Grant 201709P14.

References

  • 1.Hawkins S, Wang H, Liu Y, Garcia A, Stringfield O, Krewer H, et al. Predicting Malignant Nodules from Screening CT Scans. J Thorac Oncol. 2016;11(12):2120–8. 10.1016/j.jtho.2016.07.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Firmino M, Angelo G, Morais H, Dantas MR, Valentim R. Computer-aided detection (CADe) and diagnosis (CADx) system for lung cancer with likelihood of malignancy. Biomed Eng Online. 2016;15:17 10.1186/s12938-016-0134-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Orozco HM, Villegas OOV, Sanchez VGC, Domnguez HDO, Alfaro MDN. Automated system for lung nodules classification based on wavelet feature descriptor and support vector machine. Biomed Eng Online. 2015;14 10.1186/s12938-015-0003-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Kumar D, Wong A, Clausi DA, Ieee. Lung Nodule Classification Using Deep Features in CT Images2015. 133–8 p.
  • 5.Dehmeshki H, Ye XJ, Costello J, Ieee, Ieee. Shape based region growing using derivatives of 3D medical images: Application to semi-automated detection of pulmonary nodules. 2003 International Conference on Image Processing, Vol 1, Proceedings. IEEE International Conference on Image Processing (ICIP)2003. p. 1085–8.
  • 6.Li Q, Balagurunathan Y, Liu Y, Qi J, Schabath MB, Ye ZX, et al. Comparison Between Radiological Semantic Features and Lung-RADS in Predicting Malignancy of Screen-Detected Lung Nodules in the National Lung Screening Trial. Clinical Lung Cancer. 2018;19(2):148–U640. 10.1016/j.cllc.2017.10.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Arevalo J, Gonzalez FA, Ramos-Pollan R, Oliveira JL, Lopez MAG. Representation learning for mammography mass lesion classification with convolutional neural networks. Computer Methods and Programs in Biomedicine. 2016;127:248–57. 10.1016/j.cmpb.2015.12.014 [DOI] [PubMed] [Google Scholar]
  • 8.Gao XHW, Hui R, Tian ZM. Classification of CT brain images based on deep learning networks. Computer Methods and Programs in Biomedicine. 2017;138:49–56. 10.1016/j.cmpb.2016.10.007 [DOI] [PubMed] [Google Scholar]
  • 9.Zhao JJ, Ji GH, Qiang Y, Han XH, Pei B, Shi ZH. A New Method of Detecting Pulmonary Nodules with PET/CT Based on an Improved Watershed Algorithm. Plos One. 2015;10(4). 10.1371/journal.pone.0123694 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Shen W, Zhou M, Yang F, Yu DD, Dong D, Yang CY, et al. Multi-crop Convolutional Neural Networks for lung nodule malignancy suspiciousness classification. Pattern Recognition. 2017;61:663–73. 10.1016/j.patcog.2016.05.029 [DOI] [Google Scholar]
  • 11.Tajbakhsh N, Suzuki K. Comparing two classes of end-to-end machine-learning models in lung nodule detection and classification: MTANNs vs. CNNs. Pattern Recognition. 2017;63:476–86. 10.1016/j.patcog.2016.09.029 [DOI] [Google Scholar]
  • 12.Xia JF, Chen HL, Li Q, Zhou MD, Chen LM, Cai ZN, et al. Ultrasound-based differentiation of malignant and benign thyroid Nodules: An extreme learning machine approach. Computer Methods and Programs in Biomedicine. 2017;147:37–49. 10.1016/j.cmpb.2017.06.005 [DOI] [PubMed] [Google Scholar]
  • 13.Xie HT, Yang DB, Sun NN, Chen ZN, Zhang YD. Automated pulmonary nodule detection in CT images using deep convolutional neural networks. Pattern Recognition. 2019;85:109–19. 10.1016/j.patcog.2018.07.031 [DOI] [Google Scholar]
  • 14.Mei XY, Wang R, Yang WJ, Qian FF, Ye XD, Zhu L, et al. Predicting malignancy of pulmonary ground-glass nodules and their invasiveness by random forest. Journal of Thoracic Disease. 2018;10(1):458–63. 10.21037/jtd.2018.01.88 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Farahani FV, Ahmadi A, Zarandi MHF. Hybrid intelligent approach for diagnosis of the lung nodule from CT images using spatial kernelized fuzzy c-means and ensemble learning. Mathematics and Computers in Simulation. 2018;149:48–68. 10.1016/j.matcom.2018.02.001 [DOI] [Google Scholar]
  • 16.Nibali A, He Z, Wollersheim D. Pulmonary nodule classification with deep residual networks. International Journal of Computer Assisted Radiology and Surgery. 2017;12(10):1799–808. 10.1007/s11548-017-1605-6 [DOI] [PubMed] [Google Scholar]
  • 17.Zhu W, Vang YS, Huang Y, Xie X, editors. “DeepEM: Deep 3D ConvNets with EM for Weakly Supervised Pulmonary Nodule Detection.” Medical Image Computing and Computer Assisted Intervention—MICCAI 2018; 2018 2018//;
  • 18.Wu D, Kim K, Dong B, Fakhri GE, Li Q, editors. “End-to-End Lung Nodule Detection in Computed Tomography.” Machine Learning in Medical Imaging; 2018 2018//;
  • 19.Zhao YY, Zhao L, Yan ZN, Wolf M, Zhan YQ. “A deep-learning based automatic pulmonary nodule detection system”. In: Petrick N, Mori K, editors. Medical Imaging 2018: Computer-Aided Diagnosis. Proceedings of SPIE. 105752018.
  • 20.Khosravan N, Bagci U, editors. “S4ND: Single-Shot Single-Scale Lung Nodule Detection”. Medical Image Computing and Computer Assisted Intervention—MICCAI 2018; 2018 2018//;
  • 21.Zainudin Z, Shamsuddin SM, Hasan S, Ali A. Convolution Neural Network for Detecting Histopathological Cancer Detection. Advanced Science Letters. 2018;24(10):7494–500. 10.1166/asl.2018.12966 [DOI] [Google Scholar]
  • 22.Teramoto A., Fujita H., Yamamuro O., and Tamaki T.,“Automated detection of pulmonary nodules in PET/CT images: Ensemble false-positive reduction using a convolutional neural network technique,”MEDICAL PHYSICS,vol. 43, no. 6, pp. 2821–2827, June 2016. 10.1118/1.4948498 [DOI] [PubMed] [Google Scholar]
  • 23.Setio AAA, Ciompi F, Litjens G, Gerke P, Jacobs C, van Riel SJ, et al. Pulmonary Nodule Detection in CT Images: False Positive Reduction Using Multi-View Convolutional Networks. Ieee Transactions on Medical Imaging. 2016;35(5):1160–9. 10.1109/TMI.2016.2536809 [DOI] [PubMed] [Google Scholar]
  • 24.Cao G, Liu YZ, Suzuki K, Ieee. A new method for false-positive reduction in detection of lung nodules in CT images. 2014 19th International Conference on Digital Signal Processing. International Conference on Digital Signal Processing2014. p. 474–9.
  • 25.Suzuki K, Armato SG, Li F, Sone S, Doi K. Massive training artificial neural network (MTANN) for reduction of false positives in computerized detection of lung nodules in low-dose computed tomography. Medical Physics. 2003;30(7):1602–17. 10.1118/1.1580485 [DOI] [PubMed] [Google Scholar]
  • 26.Dou Q, Chen H, Yu LQ, Qin J, Heng PA. Multilevel Contextual 3-D CNNs for False Positive Reduction in Pulmonary Nodule Detection. Ieee Transactions on Biomedical Engineering. 2017;64(7):1558–67. 10.1109/TBME.2016.2613502 [DOI] [PubMed] [Google Scholar]
  • 27.Vital DA, Sais BT, Moraes MC. Robust pulmonary segmentation for chest radiography, combining enhancement, adaptive morphology and innovative active contours. Research on Biomedical Engineering. 2018;34(3):234–45. 10.1590/2446-4740.180035 [DOI] [Google Scholar]
  • 28.Krishnamurthy S, Narasimhan G, Rengasamy U. An Automatic Computerized Model for Cancerous Lung Nodule Detection from Computed Tomography Images with Reduced False Positives. In: Santosh KC, Hangarge M, Bevilacqua V, Negi A, editors. Recent Trends in Image Processing and Pattern Recognition. Communications in Computer and Information Science. 7092017. p. 343–55.
  • 29.Abe Y, Hanai K, Nakano M, Ohkubo Y, Hasizume T, Kakizaki T, et al. A computer-aided diagnosis (CAD) system in lung cancer screening with computed tomography. Anticancer Research. 2005;25(1B):483–8. [PubMed] [Google Scholar]
  • 30.Ji S. W., Xu W., Yang M., and Yu K., “3D convolutional neural networks for human action recognition,”Ieee Transactions on Pattern Analysis and Machine Intelligence, vol. 35, no. 1, pp. 221–231, January 2013. 10.1109/tpami.2012.59 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

The Ali Tianchi data set used in this manuscript can be obtained on Aliyun official website (https://tianchi.aliyun.com/competition/entrance/231601/information), and the LUNA16 data set can be obtained on LUNA16 official website (https://luna16.grand-challenge.org/Download/).


Articles from PLoS ONE are provided here courtesy of PLOS

RESOURCES