Skip to main content
Journal of Radiation Research logoLink to Journal of Radiation Research
. 2021 Jan 22;62(2):346–355. doi: 10.1093/jrr/rraa132

Automated approach for segmenting gross tumor volumes for lung cancer stereotactic body radiation therapy using CT-based dense V-networks

Yunhao Cui 1, Hidetaka Arimura 2,, Risa Nakano 3, Tadamasa Yoshitake 4, Yoshiyuki Shioyama 5, Hidetake Yabuuchi 6
PMCID: PMC7948852  PMID: 33480438

Abstract

The aim of this study was to develop an automated segmentation approach for small gross tumor volumes (GTVs) in 3D planning computed tomography (CT) images using dense V-networks (DVNs) that offer more advantages in segmenting smaller structures than conventional V-networks. Regions of interest (ROI) with dimensions of 50 × 50 × 6–72 pixels in the planning CT images were cropped based on the GTV centroids when applying stereotactic body radiotherapy (SBRT) to patients. Segmentation accuracy of GTV contours for 192 lung cancer patients [with the following tumor types: 118 solid, 53 part-solid types and 21 pure ground-glass opacity (pure GGO)], who underwent SBRT, were evaluated based on a 10-fold cross-validation test using Dice’s similarity coefficient (DSC) and Hausdorff distance (HD). For each case, 11 segmented GTVs consisting of three single outputs, four logical AND outputs, and four logical OR outputs from combinations of two or three outputs from DVNs were obtained by three runs with different initial weights. The AND output (combination of three outputs) achieved the highest values of average 3D-DSC (0.832 ± 0.074) and HD (4.57 ± 2.44 mm). The average 3D DSCs from the AND output for solid, part-solid and pure GGO types were 0.838 ± 0.074, 0.822 ± 0.078 and 0.819 ± 0.059, respectively. This study suggests that the proposed approach could be useful in segmenting GTVs for planning lung cancer SBRT.

Keywords: deep learning, segmentation, dense V-networks, lung stereotactic body radiation therapy

INTRODUCTION

Lung cancer, the most common fatal malignancy in the developed world [1], causes >1.3 million deaths worldwide each year, according to the world health organization (WHO) [2–3]. It is the most commonly occurring malignant cancer in men and the third most commonly occurring cancer in women [4–5]. Despite the development of multi-modality treatments over the past decade, lung cancer remains the leading cause of death from malignant cancers, accounting for ~25% of all cancer deaths [6]. The treatment of lung cancer should be selected appropriately according to the clinical stage identified by scientific evidence [7]. In general, there are four main treatments for lung cancer, namely surgery, radiation therapy, chemotherapy and immunotherapy. Stereotactic body radiotherapy (SBRT) refers to one of the treatment options for patients in the early stages (I or II) of non-small cell lung cancer (NSCLC) and who are medically inoperable or refuse surgery. Compared to conventional radiotherapy, SBRT can deliver higher doses (12, 20, 22 Gy/fraction and more) to small targets (Inline graphic50 mm) over 3–5 treatment fractions, using multiple conformal coplanar and non-coplanar beams [8–9]. Therefore, the contours of gross tumor volumes (GTVs), which are analyzed to produce estimates of clinical target volumes (CTVs) and planning target volumes (PTVs), should be as accurate as possible [10–11]. However, the GTV regions are manually delineated from treatment planning computed tomography (CT) images by treatment planners (e.g. radiation oncologists) with differing levels of experience and skills [12–15]. Thus, manual delineations can cause intra- and inter-observer variabilities in GTV contours, making them less repeatable and reproducible, leading to variability in treatment dose distributions. To address these issues, automated segmentation approaches are in high demanded in clinical SBRT practices for reducing observer variabilities.

A number of efforts have been made to develop more reproducible and more accurate GTV segmentation approaches using machine learning techniques such as deep learning in various imaging modalities (e.g. CT, positron emission tomography (PET), and magnetic resonance imaging) [16–23]. Automated approaches for segmenting GTV regions have been reported based on machine learning using PET- or PET/CT-based images [16–17]. Moreover, 18F-fluorodeoxyglucose (FDG)-PET directly shows relevant biomedical information, indicating the potential to improve the accuracy and achieve a more stable result of target volume delineation [13, 24]. If artificial intelligence, including machine learning or deep learning, is assumed to have the same ability to segment lung tumors for SBRT as a radiation oncologist with >6 years of experience, there would be some benefits for radiation oncologists, i.e. reducing the contour variability, and their time and labor for delineation of tumor contours.

Zhong et al. [25] studied deep-learning co-segmentation models of GTV regions in PET-CT images with 38 and 22 sets of training and test datasets, respectively. The models were based on deep fully convolutional networks (DFCN), which consisted of two coupled 3D U-nets [26] with an encoder–decoder architecture in lung SBRT. Additionally, Zhao et al. [27] studied deep-learning co-segmentation in PET/CT images with 48 and 36 sets of training and validation datasets, respectively, or segmentation in a single modality (PET or CT only) using multi-modality, fully convolutional neural networks (FCNs) in lung cancer. In contrast to past studies with machine learning [16, 17], deep learning techniques can automatically generate image features and segment the GTV regions of lung cancer.

However, there are three issues in the past studies. The use of the PET images has an issue of motion artifacts (e.g. blurring). Although using PET images may increase the performance of segmentation, it may also include ambiguous information in the datasets because of motion artifacts. Motion artifacts may cause non-negligible misregistration on tumor boundaries between PET and CT images. Reducing them in PET images is difficult due to the long duration of PET scans, which are taken under free respiration [28]. Furthermore, the number of PET scanners including PET/CT scanners is limited to <0.5 per 100 000 population compared with >8 CT scanners per 100 000 population in Japan [29]. Another benefit of using only CT images is that reducing PET examinations may not cause additional costs and radiation exposures to patients. Therefore, to diminish the risk of misregistration and extra costs, it is more appropriate to make use of only CT images for segmenting GTV regions, rather than using both PET and CT images. The number of cases investigated in previous deep learning-based studies [25, 27] is insufficient. Finally, the past studies did not employ the latest deep learning architectures and methods, like dense V-networks (DVNs) [30], which have the advantage of being able to segment smaller structures for small lung cancer in SBRT. Hence, our aim in this study was to develop an automated segmentation approach for small GTVs in lung cancer SBRT using CT-based densely connected V-Net. This approach has more advantages than conventional V-networks [31] in the segmentation of smaller structures.

MATERIALS AND METHODS

Figure 1 shows an overall scheme of the proposed approach for segmenting GTV regions. The procedure for the proposed approach is explained in the next section. An open source convolutional network platform [32] was applied to the CT-based DVNs to solve the GTV segmentation problem.

Fig. 1.

Fig. 1.

Overall scheme of the automated approach for segmenting GTV regions.

Clinical cases

This study was performed under the approval of an institutional review board of our university hospital. We selected 192 NSCLC patients (40–92 years old; mean age: 74 years; 70 females; 122 males; mean effective diameter of GTV: 11.039 mm; effective diameter range: 3.462–24.716 mm) who received SBRT at the university hospital. The patients were free to breathe when taking the planning CT images. The planning CT images were acquired from a 4-slice CT scanner (Mx 8000; Philips, Amsterdam, The Netherlands) with a matrix size from 512 × 512 × 103 to 512 × 512 × 235 pixels; pixel sizes of 0.781, 0.879 and 0.977 mm; and slice thicknesses of 2 or 3.2 mm. For the size of CT images used in this research, the first two values are the number of pixels in x and y directions, and the third value is the number of slices (z direction). The GTV contours were delineated on planning CT images using a commercially available radiation treatment planning (RTP) system (Eclipse version 6.5 or 8.1; Varian Medical System Inc., Palo Alto, USA), based on consensuses between two radiation oncologists with >6 years experience by referring to the PET and diagnostic CT images.

The database included three tumor types [118 solid, 53 part-solid and 21 pure ground-glass opacity (pure GGO) types]. Solid, part-solid and pure GGO types were defined based on consolidation-to-tumor ratio (CTR), which is calculated by dividing the solid portion size by the total tumor size in a lung window setting. The CTR for solid tumors is one, and that for pure GGO tumors is zero. The part-solid tumors show a CTR between zero and one, but usually around 0.5 [33]. Tumor types were determined by a radiologist (H.Y.) based on the planning CT images at a lung window level and width of −600 and 1500 Hounsfield units (HU), respectively.

Preprocessing of 3D planning CT images and GTV region datasets

The 3D planning CT and GTV binary images were converted into isovoxel images with a voxel size of 0.977 mm by a shape-based interpolation method [34]. The regions of interest (ROI; 50 × 50 × 6 to 50 × 50 × 72 pixels, hereinafter expressed as 50 × 50 × 6–72 pixels) were cropped based on the centroids of lung cancer regions from the original planning CT images; the dimensions are according to a guideline published in the Executive Summary of an ASTRO Evidence-Based Guideline [35] stating that the cancer regions should be smaller than 50 mm; and the maximum size of lung cancer in the database was 24.716 mm. The cropped ROI images were resized from 50 × 50 × 6–72 pixels to 40 × 40 × 5–58 pixels to meet input image sizes (as a multiple of 8 for the DVNs) in the NiftyNet platform [32].

Data augmentation

Augmentation techniques were used in the training step to avoid deep learning overfitting [36]. The images were randomly rotated from −10° to 10°, randomly flipped, and randomly scaled from −10 to 10%. Figure 2 shows examples of data augmentation applied in this study. Only the training images were augmented; the test images were not.

Fig. 2.

Fig. 2.

Examples of data augmentation applied in this study: (a) an original image, (b) flipped image, (c) rotated image, and (d) scaled image.

DVN segmentation

Gibson et al. [30] proposed DVNs in which the 2D segmentation model of a DenseNet [35] and a 3D medical segmentation model of a V-network [30] are combined for multi-organ segmentations. Figure 3 shows the architectures of the DVNs, which use a fully convolutional neural network based on convolutional units and enable high-resolution activation maps.

Fig. 3.

Fig. 3.

Architecture of dense V-networks used in this study.

DVNs have two main features, the dense connection and V-network. The densely connected network [37] is developed from a network called Residual network (ResNet) [38]. The main idea of dense connection is to input all the features extracted by previous layers, avoiding overfitting when using a deeper network and small dataset [37]. To increase memory efficiency, Gibson et al. [30] introduced a new batch-wise spatial dropout that can be combined with those of memory-efficient dense blocks by allocating only shared memory storage for the specific number of computed activation maps. Because dense feature stacks and batch-wise spatial dropouts are memory efficient, deeper networks are enabled, which have the advantage of being able to segment smaller structures [30]. V-network [39] is a network developed from U-network [40]. In V-network, the input images were changed to 3D images and residual learning [38] is employed in U-Net to improve performance. Residual learning has the potential to avoid gradient exploding and vanishing and reduce overfitting, enabling us to build a deeper network for difficult tasks. The dense connections and the V-network structure can significantly improve the performance, making the network more sensitive to small structures [30] and more effective on a small dataset [37]. Therefore, we have used DVNs to segment GTV regions.

Training datasets of the 3D planning CT images and contours of GTVs (reference contours) determined by radiation oncologists were fed into the DVNs as input and annotation data, respectively. Table 1 shows the hyperparameters to be optimized for ranges or candidates. Hyperparameters were optimized in a 10-fold cross-validation test with the 3D-planning CT volumes using a GPU of NVIDIA TITAN X Pascal (NVIDIA Corporation, Los Alamitos, CA).

Table 1.

Hyperparameters to be optimized for ranges or candidates

Hyperparameters Range or candidate for optimization
Loss function Dice loss [31] Cross entropy Generalised Dice overlap [41] Dice nosquare Generalised Wasserstein Dice loss [42]
Learning rate 0.0001 0.001 0.01
Activation function Parametric ReLU selu [43] Leacky ReLU ReLU
Gradient method RMSprop momentum adagrad Gradient discent Adam
Batch size 1 2 4 8 16 32
Iterations 5190 8650 17 300 25 950 86 500
Dropout 0.75 0.5 0.25

With three DVNs trained by the same training dataset and the same optimum hyper-parameters but initialed randomly, three output GTVs called DVN1, DVN2 and DVN3 were obtained each time we input a test image. To combine these results, these segmented GTVs produced single outputs, logical AND (common agreement) outputs, and logical OR (summation) outputs by combinations of two or three outputs from the DVNs. Therefore, for the three models DVN1, DVN2 and DVN3, we had four logical AND (AND DVN1–2, AND DVN1–3, AND DVN2–3, AND DVN1–2–3) and four logical OR (OR DVN1–2, OR DVN1–3, OR DVN2–3, OR DVN1–2–3) results. The four logical AND outputs were obtained by combinations of two or three outputs of DVN1, DVN2 or DVN3, and four logical OR outputs using the same combinations as those used in the logical AND outputs.

Evaluation of GTVs extracted by the proposed approach

The 192 cases were randomly split into 90% training cases and 10% test cases before data augmentation in a 10-fold cross-validation test. In the training step, 85% of the training cases were used for training the DVNs and 15% of the training cases for testing them. Following these settings, the numbers of cases in training, validation and testing datasets were 147, 26 and 19, respectively.

The segmentation accuracy of GTV contours for 192 lung cancer patients (118 solid, 53 part-solid and 21 pure GGO types) were evaluated using a 10-fold cross-validation test using Dice’s similarity coefficients (DSCs) [44] and the Hausdorff distance (HD) [45]. All DSCs and HDs were three-dimensionally derived from 3D volumes generated by DVNs.

The DSC denotes the similarity between the reference region determined by radiation oncologists and the GTV region estimated using the proposed approach; its values range from 0 to 1 and are calculated using the following equation:

graphic file with name M2.gif (1)

where T is the GTV ground truth region determined by experienced radiation oncologists, D is the GTV region estimated using the proposed approach, Inline graphic is the number of pixels in region Inline graphic, Inline graphic is the number of pixels in region Inline graphic and Inline graphic is the number of overlapping pixels between Inline graphic and Inline graphic.

The HD is the degree of misregistration between two sets of data measured by the distance within a point set Inline graphic that is farthest from any point within a point set Inline graphic and vice versa [45]. The HD is calculated using the following equation with the two finite point sets Inline graphic and Inline graphic(Inline graphic and Inline graphic are defined as the position vectors):

graphic file with name M16.gif (2)

where

graphic file with name M17.gif (3)

and Inline graphic is some underlying norm describing the points within Inline graphic and Inline graphic (e.g. the Inline graphic or Euclidean norm).

The HD is the maximum of Inline graphic and Inline graphic. The function Inline graphic identifies the point Inline graphic that is farthest from any point of Inline graphic and measures the distance from Inline graphic to its nearest neighbor in Inline graphic (using the given norm Inline graphic). If two images are identical, the score is zero, and the score is larger as the shapes of the two images are more different.

RESULTS

Figure 4 illustrates instances of each tumor type that shows reference GTVs (white line) and regions estimated by the proposed approach (green line; using the Dice loss function, a learning rate of 0.001, an activation function of selu, a batch size of 16, a gradient method of RMSprop, a dropout ratio of 0.75 (keep) and an iteration of 8650). The calculation time for delineation of a lung tumor was about 0.2 s on average. Table 2 shows DSCs and Table 3 shows HDs for 11 segmented GTVs obtained from three single outputs, four logical AND outputs, and four logical OR outputs using combinations of two or three outputs from DVNs based on a 10-fold cross-validation test. Figures 5 and 6 show 3D-DSCs and HDs obtained by the AND DNV 1–2–3 for solid, part-solid, and pure GGO tumor types, respectively. The average 3D DSCs for solid, part-solid, and pure GGO (Fig. 5) types were 0.838 ± 0.074, 0.822 ± 0.078, and 0.819 ± 0.059, respectively. The average HDs for solid, part-solid and pure GGO (Fig. 6) types were 4.03 ± 1.94 mm, 4.07 ± 2.02 mm and 4.70 ± 2.61 mm, respectively.

Fig. 4.

Fig. 4.

Reference GTVs (white line) and regions estimated with the dense V-networks (green line): (a) solid, (b) part-solid, and (c) pure GGO. 3-D DSCs and slice numbers (SNs) are included in these images. The isocenter slice number is 0.

Table 2.

DSCs and HDs for 11 segmented GTVs obtained from three single outputs, four logical AND outputs, or four logical OR outputs by combinations of two or three outputs from DVNs based on a 10-fold cross validation test

DVN1 DVN2 DVN3 AND DVN1–2 AND DVN1–3 AND DVN2–3 OR DVN1–2 OR DVN1–3 OR DVN2–3 AND DVN1–2–3 OR DVN1–2–3
DSC 0.822 0.822 0.820 0.831 0.829 0.829 0.814 0.814 0.813 0.832 0.808
HD 4.90 4.85 5.01 4.62 4.60 4.63 5.04 5.21 5.18 4.57 5.28

Table 3.

DSCs of different types of lung nodule for 11 segmented GTVs obtained from three single outputs, four logical AND outputs, or four logical OR outputs by combinations of two or three outputs from DVNs based on a 10-fold cross validation test

DVN1 DVN2 DVN3 AND DVN1–2 AND DVN1–3 AND DVN2–3 OR DVN1–2 OR DVN1–3 OR DVN2–3 AND DVN1–2–3 OR DVN1–2–3
All 0.822 0.822 0.820 0.831 0.829 0.829 0.814 0.814 0.813 0.832 0.808
Solid 0.826 0.825 0.825 0.836 0.834 0.835 0.816 0.817 0.816 0.838 0.810
Part solid 0.819 0.819 0.818 0.823 0.822 0.824 0.815 0.816 0.814 0.822 0.810
Pure GGO 0.810 0.814 0.795 0.823 0.815 0.813 0.801 0.790 0.796 0.819 0.788

Fig. 5.

Fig. 5.

3D-DSCs for solid, part-solid and pure GGO types.

Fig. 6.

Fig. 6.

HDs for solid, part-solid and pure GGO types.

The highest 3D DSC and lowest HD values obtained from AND DVN1–2–3 were 0.962 and 1.22 mm, respectively, whereas the lowest 3D DSC and highest HD values obtained similarly were 0.566 and 15.2 mm, respectively. These values derive from different cases, and the case with the highest 3D DSC value (Fig. 7a) has a distinct boundary owing to the high contrast between the lung tumor and its surrounding background tissue. On the other hand, the case with most accurate HD (Fig. 7b) is a small and solid tumor with a distinguishable outline. The case with the lowest 3D DSC is one in which the differences between lung cancer and blood vessels are difficult to identify, as shown Fig. 7c. The case with the least accurate HD (Fig. 7d) is a large solid tumor with a blurred contour.

Fig. 7.

Fig. 7.

Reference GTVs (white line) and regions estimated with the DVNs (green line): (a) highest 3D DSC, (b) highest HD, (c) lowest 3D DSC, and (d) lowest HD.

The relationship between the numbers of slices with reference and predicted contours is shown in Figure 8. In all, 2.9% (112 slices in 75 lung tumors) of 3823 slices in 192 lung tumors were underestimated in terms of the numbers of slices with predicted contours.

Fig. 8.

Fig. 8.

Relationship between the numbers of slices with reference and predicted contours of AND DNV1–2–3.

DISCUSSION

Table 4 compares the proposed approach with four automated approaches developed in past studies to delineate GTV regions for lung SBRT or NSCLC patients who have bigger GTVs than SBRT patients. Compared to machine learning-based studies with 2D images [16, 17], the accuracies of the proposed approach were higher for all three types of cancers viewed with 3D images, especially the part-solid and pure GGO types. This is because of the difficulty involved for machine learning to learn heterogeneity in the part-solid and pure GGO types, as shown in Figs 4 b and c. Besides, applying 3D images instead of 2D images can involve more information, e.g. the continuity of tumor region in several slices, making it possible for the segmentation system to learn more features among contiguous slices.

Table 4.

Comparison of the proposed approach with four automated approaches for delineation of GTV regions for lung SBRT or lung cancer patients

No. of patients Modality Machine learning Results
Our study 192 (Lung SBRT)
Solid: 118
Part solid: 53
Pure GGO:21
10-fold cross validation test (training: 174–175, test: 19–20)
3D-CT DVNs DSC: 0.832
HD: 4.57 mm
Solid: 0.838
Part solid: 0.822
Pure GGO: 0.819
Kawata et al. [16] 16 (Lung SBRT)
Solid: 6
Part solid: 6
Pure GGO: 4
2D-PET/CT Fuzzy-c-means clustering
method-based framework
DSC: 0.79
Solid: 0.83
Part solid: 0.76
Pure GGO: 0.79
Ikushima et al. [17] 14 (Lung SBRT)
Solid: 6
Part solid: 4
Pure GGO: 4
2D-PET/CT Support vector machine DSC: 0.777
Solid: 0.834
Part solid: 0.701
Pure GGO: 0.763
Zhong et al. [25] 60 (Lung SBRT)
Training: 38
Test: 22
3D-PET/CT
3D-CT
3D-PET
DFCN-based cosegmentation (DICELoss)
3D-Unet (DICELoss)
3D-Unet (DICELoss)
DSC: 0.861 (CT)
DSC: 0.828 (PET)
DSC: 0.811
DSC: 0.794
Zhao et al. [27] 84 (Lung cancer)
Training: 48
Test: 36
3D-PET/CT
3D-CT
3D-PET
Multi-modality FCN
CNN
CNN
DSC: 0.85
DSC: 0.76
DSC: 0.83

Our proposed approach failed in segmenting 75 lung tumors in 112 slices of 75 patients. These tumor regions were not segmented in the first, the last and central slices 70, 17 and 25, respectively. Therefore, the numbers of slices with predicted contours were underestimated as shown in Fig. 8. It is still uncertain why DVN failed to segment in those slices. Besides, when we combined three DVN outputs by using a logical AND method, there were more segmentation failures, because we successfully segmented the tumors on slices by AND DVN only if the two (or three) DVN models reach a common agreement. In further research, we will find some other method to avoid the failure that happened in DVNs and logical AND DVNs.

The proposed approach is a deep learning method based on supervised learning, which could not avoid the impact of the variability of the reference data on the performance, since the reference data were subjectively decided by human observers. One choice to reduce the variability is to use an unsupervised learning method, which does not need any reference data for training. However, it is hard to evaluate the results of unsupervised learning due to the lack of reference data (‘truths’). Therefore, further research is needed to develop new methods for reducing the reference variability and evaluating the segmentation accuracy at the same time.

Past studies on deep learning with SBRT or lung cancer patients [25, 27] (Table 4) achieved better performances than the proposed approach with both PET images and CT images. Because applying PET and CT images at the same time may cause a misregistration problem, it increases the difficulty of collecting data and additional work to register the tumor boundaries between PET and CT images; we employed only CT images, which can avoid misregistration on tumor boundaries between PET and CT images. Besides, a bigger database (192 cases in total) is employed than in past study, which reduces the risk of overfitting. Moreover, employing the common agreement (AND region) to delete outliers can increase the level of accuracy over those in past deep learning-based studies using a single modality. This also means that DVNs may increase the segmentation accuracy of GTVs if we used dual modalities (PET/CT images), like past studies.

There are three limitations in this study. First, the number of cases is not sufficient to improve segmentation performances, especially in pure GGO and part-solid tumor types. Second, we focused on GTVs instead of CTVs. Because radiation therapy requires CTVs to determine the appropriate PTVs (by adding internal and set-up margins), segmentation approaches for CTVs should be developed in future work, which we aim to accomplish. Third, we did not have a way to interpret the DSC to evaluate the impact on clinical contouring. Compared to the past study shown in Table 4, there is no significant improvement in the present study (in some cases results are even a little bit worse). But we do not know whether a small change in DSC or HD can affect the clinical contouring or not. The ‘good enough’ DSC is still unclear so we cannot evaluate the impact on clinical use.

CONCLUSIONS

We developed an automated approach for extracting GTVs of lung cancer patients using datasets of 3D planning CT images to segment tumors by deep learning-based DVNs. The 3D DSC and HD values obtained by the deep learning-based DVNs are 0.832 ± 0.074 and 4.57 ± 2.44 mm (AND DVN1–2–3), respectively. Therefore, the proposed approach has the potential to be useful in delineating GTVs for treatment planning of lung cancer SBRT and to assist radiation oncologists.

ACKNOWLEDGEMENTS

This study was partially supported by a grant from Center for Clinical and Translational Research of Kyushu University Hospital and JSPS KAKENHI Grant Number 20 K08084. The authors are grateful to all members of the Arimura Laboratory (http://web.shs.kyushu-u.ac.jp/∼arimura/) for their valuable comments and helpful discussion. The authors alone are responsible for the contents and writing of the paper.

Contributor Information

Yunhao Cui, Department of Health Sciences, Graduate School of Medical Sciences, Kyushu University, 3-1-1, Maidashi, Higashi-ku, Fukuoka 812-8582, Japan.

Hidetaka Arimura, Department of Health Sciences, Faculty of Medical Sciences, Kyushu University, 3-1-1, Maidashi, Higashi-ku, Fukuoka 812-8582, Japan.

Risa Nakano, Department of Health Sciences, Graduate School of Medical Sciences, Kyushu University, 3-1-1, Maidashi, Higashi-ku, Fukuoka 812-8582, Japan.

Tadamasa Yoshitake, Department of Clinical Radiology, Graduate School of Medical Sciences, Kyushu University, 3-1-1, Maidashi, Higashi-ku, Fukuoka 812-8582, Japan.

Yoshiyuki Shioyama, Saga International Heavy Ion Cancer Treatment Foundation, 3049 Harakogamachi, Tosu-shi, Saga 841-0071, Japan.

Hidetake Yabuuchi, Department of Health Sciences, Faculty of Medical Sciences, Kyushu University, 3-1-1, Maidashi, Higashi-ku, Fukuoka 812-8582, Japan.

CONFLICT OF INTEREST

None declared.

References

  • [1]. Sethi  T. Lung cancer. Introduction. Thorax  2002;57:992–3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2]. Sun  S, Bauer  C, Beichel  R. Automated 3-D segmentation of lungs with lung cancer in CT data using a novel robust active shape model approach. IEEE Trans Med Imaging  2012;31:449–60. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3]. Cheng  W, Ma  L, Yang  T  et al.  Joint lung CT image segmentation: A hierarchical Bayesian approach. PLoS One  2016;11:e0162211. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [4]. De Margerie-Mellon  C, De Bazelaire  C, De Kerviler  E. Image-guided biopsy in primary lung cancer: Why, when and how. Diagn Interv Imaging  2016;97:965–72. [DOI] [PubMed] [Google Scholar]
  • [5]. Worldwide cancer data. (2020). https://www.wcrf.org/dietandcancer/cancer-trends/worldwide-cancer-data (22 December 2020, date last accessed).
  • [6]. Lung Cancer Statistics | How Common is Lung Cancer. (2020). https://www.cancer.org/cancer/lung-cancer/about/key-statistics.html (22 December 2020, date last accessed).
  • [7]. Detterbeck  FC, Mazzone  PJ, Naidich  DP  et al.  Screening for lung cancer: Diagnosis and management of lung cancer, 3rd ed: American College of Chest Physicians evidence-based clinical practice guidelines. Chest  2013;143:e78S–92S. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [8]. Crabtree  TD, Denlinger  CE, Meyers  BF  et al.  Stereotactic body radiation therapy versus surgical resection for stageInon-small cell lung cancer. J Thorac Cardiovasc Surg  2010;140:377–86. [DOI] [PubMed] [Google Scholar]
  • [9]. Timmerman  R, McGarry  R, Yiannoutsos  C  et al.  Excessive toxicity when treating central Tumors in a phase II study of stereotactic body radiation therapy for medically inoperable early-stage lung cancer. J Clin Oncol  2006;24:4833–9. [DOI] [PubMed] [Google Scholar]
  • [10]. Weiss  E, Hess  C. The impact of gross tumor volume (GTV) and clinical target volume (CTV) definition on the total accuracy in radiotherapy theoretical aspects and practical experiments. Strahlenther Onkol  2003;179:21–30. [DOI] [PubMed] [Google Scholar]
  • [11]. Vinod  SK, Jameson  MG, Min  M  et al.  Uncertainties in volume delineation in radiation oncology: A systematic review and recommendations for future studies. Radiother Oncol J  2016;121:169–79. [DOI] [PubMed] [Google Scholar]
  • [12]. Van de Steene  J, Linthout  N, de Mey  J  et al.  Definition of gross tumor volume in lung cancer: Inter-observer variability. Radiother Oncol  2002;62:37–49. [DOI] [PubMed] [Google Scholar]
  • [13]. Caldwell  CB, Mah  K, Ung  YC  et al.  Observer variation in contouring gross tumor volume in patients with poorly defined non-small-cell lung tumors on CT: The impact of 18FDG-hybrid PET fusion. Int J Radiat Oncol Biol Phys  2001;51:923–31. [DOI] [PubMed] [Google Scholar]
  • [14]. Persson  GF, Nygaard  DE, Hollensen  C  et al.  Inetrobserver delineation variation in lung tumor stereotactic body radiation therapy. Br J Radiol  2012;85:e654–60. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [15]. Velazquez  ER, Aerts  HJWL, Gu  Y  et al.  A semiautomatic CT-based ensemble segmentation of lung tumors: Comparison with oncologist’s delineations and with the surgical specimen. Radiother Oncol  2012;105:167–73. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [16]. Kawata  Y, Arimura  H, Ikushima  K  et al.  Impact of pixel-based machine-learning techniques on automated frameworks for delineation of gross tumor volume regions for stereotactic body radiation therapy. Phys Med  2017;42:141–9. [DOI] [PubMed] [Google Scholar]
  • [17]. Ikushima  K, Arimura  H, Jin  Z  et al.  Computer-assisted framework for machine-learning-based delineation of GTV regions on datasets of planning CT and PET/CT images. J Radiat Res  2017;58:123–34. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [18]. Sharp  G, Fritscher  KD, Pekar  V  et al.  Vision 20/20: Perspectives on automated image segmentation for radiotherapy. Med Phys  2014;41:050902. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [19]. Van Baardwijk  A, Bosmans  G, Boersma  L  et al.  PET-CT-based auto-contouring in non-small-cell lung cancer correlates with pathology and reduces Interobserver variability in the delineation of the primary tumor and involved nodal volumes. Int J of Radiat Oncol Biol Phy  2018;68:771–8. [DOI] [PubMed] [Google Scholar]
  • [20]. Liu  Y, Stojadinovic  S, Hrycushko  B  et al.  A deep convolutional neural network-based automatic delineation strategy for multiple brain metastases stereotactic radiosurgery. PLoS One  2017;12:e0185844. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [21]. Men  K, Chen  X, Zhang  Y  et al.  Deep Deconvolutional neural network for target segmentation of nasopharyngeal cancer in planning computed tomography images. Front Oncol  2017;7:315. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [22]. Avendi  MR, Kheradvar  A, Jafarkhani  H. A combined deep-learning and deformable-model approach to fully automatic segmentation of the left ventricle in cardiac MRI. Med Image Anal  2016;30:108–19. [DOI] [PubMed] [Google Scholar]
  • [23]. Lustberg  T, van Soest  J, Gooding  M  et al.  Clinical evaluation of atlas and deep learning based automatic contouring for lung cancer. Radiother Oncol  2018;126:312–7. [DOI] [PubMed] [Google Scholar]
  • [24]. Ashamalla  H, Rafla  S, Parikh  K  et al.  The contribution of integrated PET/CT to the evolving definition of treatment volumes in radiation treatment planning in lung cancer. Int J Radiat Oncol Biol Phys  2005;63:1016–23. [DOI] [PubMed] [Google Scholar]
  • [25]. Zhong  Z, Kim  Y, Plichta  K  et al.  Simultaneous cosegmentation of tumors in PET-CT images using deep fully convolutional networks. Med Phys  2019;46:619–33. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [26]. Cicek  Ö, Abdulkadir  A, Lienkamp  SS  et al.  3D U-net: Learning dense volumetric segmentation from sparse annotation. Medical Image Computing and Computer-Assisted Intervention  2016;9901:424–32. [Google Scholar]
  • [27]. Zhao  X, Li  L, Lu  W  et al.  Tumor co-segmentation in PET/CT using multi-modality fully convolutional neural network. Phys Med Biol  2019;64:015011(15pp). doi: 10.1088/1361-6560/aaf44b. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [28]. Nehmeh  SA, Erdi  YE, Ling  CC  et al.  Effect of respiratory gating on reducing lung motion artifacts in PET imaging of lung cancer. Med Phys  2002;29:366–71. [DOI] [PubMed] [Google Scholar]
  • [29]. Matsumoto  M, Koike  S, Kashima  S  et al.  Geographic distribution of CT, MRI and PET devices in Japan: A longitudinal analysis based on national census data. PLoS One  2015;10:e0126036. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [30]. Gibson  E, Giganti  F, Hu  Y  et al.  Automatic multi-organ segmentation on abdominal CT with dense V-networks. IEEE Trans Med Imaging  2018;37:1822–34. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [31]. Fausto  M, Nassir  N, Seyed-Ahmad  A  et al.  V-net: Fully convolutional neural networks for volumetric medical image segmentation. 3D Vision  2016;565–71. [Google Scholar]
  • [32]. Gibson  E, Li  W, Sudre  C  et al.  NiffyNet: A deep-learning platform for medical imaging. Comput Methods Programs Biomed  2018;158:113–22. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [33]. Matsunaga  T, Suzuki  K, Takamochi  K  et al.  What is the radiological definition of part-solid tumour in lung cancer?  Eur J Cardiothorac Surg  2017;51:242–7. [DOI] [PubMed] [Google Scholar]
  • [34]. Herman  GT, Aheng  J, Bucholtz  C  et al.  Shape-based interpolation. IEEE Comput Graph Appl  1992;12:69–79. [Google Scholar]
  • [35]. Videtic  GM, Donington  J, Giuliani  M  et al.  Stereotactic body radiation therapy for early-stage non-small cell lung cancer: Executive summary of an ASTRO evidence-based guideline. Pract Radiat Oncol  2017;7:295–301. [DOI] [PubMed] [Google Scholar]
  • [36]. Hussain  Z, Gimenez  F, Yi  D  et al.  Differential data augmentation techniques for medical imaging classification tasks. AMIA Annu Symp Proc  2018;2017:979–84. [PMC free article] [PubMed] [Google Scholar]
  • [37]. Huang  G, Liu  Z, van der Maaten  L  et al.  Densely connected convolutional networks. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Honolulu, HI, 2017, 2261–9. doi: 10.1109/CVPR.2017.243. [DOI] [Google Scholar]
  • [38]. He  K, Zhang  X, Ren  S  et al.  Deep residual learning for image recognition. Proceedings of the IEEE conference on computer vision and pattern recognition.  2016;770–8. [Google Scholar]
  • [39]. Milletari  F, Navab  N, Ahmadi  SA. “V-net: Fully convolutional neural networks for volumetric medical image segmentation.” 2016 fourth international conference on 3D vision (3DV). IEEE;  2016;565–71. [Google Scholar]
  • [40]. Ronneberger  O, Fischer  P, Brox  T  et al.  U-net: Convolutional networks for biomedical image segmentation. International Conference on Medical image computing and computer-assisted intervention Springer, Cham, 2015, 234–41. [Google Scholar]
  • [41]. Carole  HS, Wenqi  L, Tom  V  et al.  Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations. Deep learning in medical image analysis and multimodal learning for clinical decision support, Springer, Cham, 2017, 240–8. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [42]. Lucas  F, Wenqi  L, Luis  C  et al.  Generalised Wasserstein dice score for imbalanced multi-class segmentation using holistic convolutional networks. International MICCAI Brainlesion Workshop Springer, Cham, 2017, 64–76. [Google Scholar]
  • [43]. Dabal  P. Comparison of non-linear activation functions for deep neural networks on MNIST classication task. arXiv:180402763v1 2018. Cs Stat. Available online: http://arxiv.org/abs/1804.02763 (18 December 2020, date last accessed). [Google Scholar]
  • [44]. Crum  WR, Camara  O, Hill  DL  et al.  Generalized overlap measures for evaluation and validation in medical image analysis. IEEE Trans Med Imaging  2006;25:1451–61. [DOI] [PubMed] [Google Scholar]
  • [45]. Huttenlocher  DP, Klanderman  GA, Rucklidge  WJ  et al.  Comparing images using the Hausdorff distance. IEEE Trans Pattern AnalMach Intell  1993;15:850–63. [Google Scholar]

Articles from Journal of Radiation Research are provided here courtesy of Oxford University Press

RESOURCES