Abstract
Background
Whole-body bone scan is the widely used tool for surveying bone metastases caused by various primary solid tumors including lung cancer. Scintigraphic images are characterized by low specificity, bringing a significant challenge to manual analysis of images by nuclear medicine physicians. Convolutional neural network can be used to develop automated classification of images by automatically extracting hierarchal features and classifying high-level features into classes.
Results
Using convolutional neural network, a multi-class classification model has been developed to detect skeletal metastasis caused by lung cancer using clinical whole-body scintigraphic images. The proposed method consisted of image aggregation, hierarchal feature extraction, and high-level feature classification. Experimental evaluations on a set of clinical scintigraphic images have shown that the proposed multi-class classification network is workable for automated detection of lung cancer-caused metastasis, with achieving average scores of 0.7782, 0.7799, 0.7823, 0.7764, and 0.8364 for accuracy, precision, recall, F-1 score, and AUC value, respectively.
Conclusions
The proposed multi-class classification model can not only predict whether an image contains lung cancer-caused metastasis, but also differentiate between subclasses of lung cancer (i.e., adenocarcinoma and non-adenocarcinoma). On the context of two-class (i.e., the metastatic and non-metastatic) classification, the proposed model obtained a higher score of 0.8310 for accuracy metric.
Keywords: Bone scan, Skeletal metastasis, Lung cancer, Image classification, Convolutional neural network
Key points
Automated detection of lung cancer-caused skeletal metastasis is first studied.
Convolutional neural network is exploited to develop automated classification method.
Clinical scintigraphic images are used to experimentally evaluate the proposed classification model.
Background
Skeletal metastasis is common in several of prevalent cancers including prostate, breast, and lung cancers [1], with 80% of all skeletal metastatic lesions originating from one of these primary sites [2]. The percentage of metastasis-related death reaches up to 90% for all lung cancer mortality [3]. Early detection of skeletal metastasis is extremely important not only for decreasing morbidity but also for disease staging, outcome prediction, and treatment planning [4].
Skeletal scintigraphy (bone scan) and positron emission tomography (PET) are commonly used for surveying bone metastasis [5, 6]. Compared to PET, bone scan is more affordable and available due to its low-cost equipment and radiopharmaceutical. Bone scan is typically characterized by high sensitivity but low specificity, bringing significant challenge to manual analysis of bone scan images by nuclear medicine physicians. The reasons of low specificity are multi-fold, mainly including low spatial resolution, accumulation of radiopharmaceutical in normal skeletal structures, soft tissues or viscera, and uptake in benign processes [7].
Automated analysis of bone scan images becomes therefore desired for accurate diagnosis of skeletal metastasis. There has been a substantial amount of works aimed at developing automated diagnosis approaches using conventional machine learning models to classify bone scan images into classes [5, 8–11], where the image features were manually extracted by researchers. The handcrafted features, however, often suffer from insufficient capability and unsatisfied performance for clinical tasks [6].
Convolutional neural network (CNN), a mainstream branch of deep learning techniques, has gained huge success in automated analysis of natural images [12–14] and medical images [14–17] due to their ability to automatically extracting hierarchical features from images in an optimal way. CNN-based automated classification methods have been proposed to detect metastasis caused by a variety of various primary tumors including prostate cancer [18–23], breast cancer [22–24], lung cancer [25, 26], and both of them [25–27]. The main purpose of existing works is to develop two-class classification models to determine whether or not an image contains metastasized lesion(s) by classifying this image (normal and metastatic). Differently, a series of CNN-based methods has been proposed to classify whole-body scintigraphic images for automated detection of skeletal metastases in our previous works [28, 29], in which we did not distinguish between the primary cancers.
Targeting at automated detection of skeletal metastasis caused by lung cancer, in this work, we propose a CNN-based multiclass classification network to classify whole-body scintigraphic images acquired from patients with clinically diagnosed lung cancer using a SPECT (single photon emission computed tomography) imaging device (i.e., GE SPECT Millennium MPR). The proposed network can not only determine whether an image contains lung cancer-caused skeletal metastasis, but also differentiate between subclasses of lung cancer (i.e., adenocarcinoma and non-adenocarcinoma).
The main contributions of this work can be summarized as: First, to the best of our knowledge, we are the first to attempt to solve the problem of automated detection of skeletal metastasis originated from various subclasses of lung cancer. Second, we convert the detection problem into the multiclass classification of low-resolution, large-size scintigraphic images using a CNN-based end-to-end network that first extracts hierarchal features from images, then aggregates these features, and finally classifies those high-level features into classes. Lastly, we use a group of scintigraphic images acquired from patients with clinically diagnosed lung cancer to evaluate the proposed method. Experimental results have shown that our CNN-based classification network performs well for distinguishing SPECT images between non-metastatic and metastatic as well as their sub-classes of metastasis.
The rest of this paper is organized as follows. We present in “Methods” section the proposed method. We report in “Results” section the experimental evaluation conducted on clinical SPECT images. In “Discussion” secton, we provide a brief discussion on the reasons that cause the misclassifications. In “Conclusions” section, we conclude this work and point out the future research directions.
Methods
To automatically detect metastasis of lung cancer in scintigraphic images, image fusion operation is first employed to enhance the lesion(s) in low-resolution whole-body scintigraphic images through aggregating the anterior- and posterior-view images of each bone scan. Parametric variation-based data augmentation is then applied to expand the size of the dataset used in this work to improve the performance of CNN-based network on classifying images as much as possible. A CNN-based end-to-end network is developed to classify the fused images by first extracting hierarchal features from images, then aggregating features, and finally classifying high-level features into classes of concerns, i.e., without metastasis (NoMet), adenocarcinoma metastasis (ADMet) and non-adenocarcinoma metastasis (nADMet). Figure 1 provides an overview of the proposed multiclass classification method, comprising of three main steps, i.e., view aggregation, data augmentation, and image classification.
View aggregation
During the SPECT imaging, two whole-body images were collected for each patient, corresponding to the anterior- and posteriorviews, respectively. When a primary tumor (e.g., lung cancer) invades into bone tissue, there will be an area of increased radionuclide’s uptake in the image. It is common, however, the metastatic areas have varied intensity of uptake in anterior- and posterior-view images. How to enhance the metastatic areas in images becomes crucial for accurate detection of metastasis. A pixel-wise view aggregation method is proposed to ‘excite’ those metastatic pixels, while ‘squeeze’ the normal pixels by fusing two views as shown in Fig. 2.
Let IAnt and IPost denote the anterior- and posterior-view image respectively, the pixel-wise view aggregation method works as follows.
Image flipping
The posterior-view image IPost is flipped horizontally around its central vertical line to obtain an image IT Post.
Pixel aligning
A horizontal line sweeps the image (i.e., IAnt and IT Post) line by line to find out the critical points PCeil and PFloor by examining the pixel value that represents uptake intensity. Similarly, we use a vertical line to sweep image line by line to find out the critical points PLeft and PRight. Two images IAnt and ITPost will be aligned according to these four critical points.
Pixel-wise image addition
The aligned images, IAnt and ITPost, will be aggregated to generate a composite image, IComp, according to Eq. 1.
1 |
where f is aggregation function, i.e., pixel-wise addition operation.
Data augmentation
It is widely accepted that the classification performance of CNN-based models depends on the size of dataset, with high classification accuracy always corresponding to the large dataset. Currently, a variety of various methods can be utilized to augment dataset including the parametric variation and adversarial learning techniques. In this work, we use the parametric variation technique to augment our dataset since the parametric variation-based data augmentation can obtain samples that have the same distribution as the original ones with the lower time complexity. Specifically, image translation and rotation are used, which are detailed as follows [30].
Image rotation
Given a constant r ∈ [0, rT], an image will be randomly rotated by ro in either the left or right direction around its geometric center. The parameter rT is experimentally determined according to the distribution of the radiotracer uptake of all images in the dataset. Figure 3d depicts the obtained image by rotating the image in Fig. 3a to the right direction by 3°.
Image translation
Given a constant r ∈ [0, tT], an image will be randomly translated by + t or −t pixels in either the horizontal or vertical direction. The parameter tT is experimentally determined according to the distribution of the radiotracer uptake of all images in the dataset. Figure 3c shows a resulting example by translating the given image in Fig. 3a + 3 pixels horizontally.
CNN-based classification network
Table 1 outlines the structure of the proposed 26-layer CNN-based classification network, consisting of one convolution layer (Conv), one normalization layer (Norm), one pooling layer (Pool), a set of residual convolution layer attached attention (RA-Conv) with varied kernel, one global average pooling layer (GAP), and 1 Softmax layer.
Table 1.
Layer | Configuration |
---|---|
Conv | 7 × 7, 64, Stride = 2 |
Norm | Batch normalization |
Pool | 3 × 3 Max pooling, Stride = 2 |
RA-Conv_2 | |
RA-Conv_3 | |
RA-Conv_5 | |
RA-Conv_2 | |
Global average pooling (GAP) | |
Softmax |
An input 256 × 1024 scintigraphic image is convolved by the Conv layer with filter of 7 × 7 to calculate a feature map made of neurons, followed by a batch normalization layer and a max pooling layer with kernel of 3 × 3. The subsequent convolutional layers are organized as residual convolution with hybrid attention inside or outside of the convolution. A global average pooling layer is used to alleviate the over-fitting problem while speeding up the training process. The Softmax layer points out the class of an image with a real number. The main layers will be detailed as follows.
Normalization layer
Batch normalization [31] is used to accelerate the network training by making normalization a part of the model architecture and performing the normalization for each training mini-batch. With batch normalization, we can thus use much higher learning rates and be less careful about initialization.
RA-Conv layer
Figure 4 demonstrates the structure of residual convolution with hybrid attention mechanism. We use residual connection to reduce the training parameters and training time. We also introduce hybrid attention mechanism to improve network focusing on those more important areas (i.e., lesions) on the feature maps by considering only the important information. Specifically, we use inRA-Conv (outRA-Conv) to indicate that a hybrid attention module is located inside (outside) the residual convolution. The classifiers are accordingly named as Classifer-inRAC and Classifer-outRAC, respectively.
The cascaded hybrid attention module in Fig. 4 using channel and spatial attention mechanism is capable of computing complementary attention by focusing on ‘what’ (channel attention) and ‘where’ (spatial attention), respectively [32]. Specifically, let F be the input of a 2D feature map to the channel attention sub-module. We can obtain a 1D output F, which will be further processed by the spatial attention sub-module to output a refined 2D feature map M according to Eq. 2.
2 |
where ⊗ is the element-wise multiplication, and fC and fS denotes the channel and spatial function, respectively, which are given in Eqs. 3 and 4.
3 |
4 |
where σ is the sigmoid function, MLP is the multi-layer perceptron, AvgPool (MaxPool) is the average (max) pooling, and fk×k is a convolutional operation with the kernel size of k × k.
Softmax layer
The network output nodes apply the Softmax function for the number of the unordered classes. A Softmax function is defined in Eq. 5 [33].
5 |
where f (xj) is the score of the j-th output node, xj is the network input to j-th output node, and n is the number of output nodes. In fact, all of the output values f (x) are a probability between 0 and 1, and their sum is 1.
Results
In this section, we provide an experimental evaluation of the proposed network using a set of clinical whole-body scintigraphic images.
Dataset
In this retrospective study, the whole-body scintigraphic images were collected from the Department of Nuclear Medicine, Gansu Provincial Tumor Hospital from Jan 2014 to Dec 2019 using a single-head gamma camera (GE SPECT Millennium MPR). SPECT imaging was performed between 2 and 3 h after intravenous injection of 99mTc-MDP (20–25 mCi) using a parallel-beam low-energy high-resolution (LEHR) collimator (energy peak = 140 keV, intrinsic energy resolution ≤ 9.5%, energy window = 20%, and intrinsic spatial resolution ≤ 6.9 mm). Each SPECT image was stored in a DICOM (Digital Imaging and Communications in Medicine) file with the imaging size of 256 × 1024. Every element in an image is represented by a 16-bit unsigned integer, differing from the natural images in which element ranges from 0 to 255.
A total of 506 patients who were clinically diagnosed with lung cancer were encompassed in this study. Figure 5 demonstrates the distribution of patients with respect to gender and age.
There are 1011 images collected from 506 patients due to the phenomenon of images being not successfully recorded. We categorize all these images into three subclasses, i.e., NoMet (n = 614, ≈ 60.73%), ADMet (n = 237, ≈ 23.44%), and nADMet (n = 160, ≈ 15.83%).
To keep the balance between samples in different subclasses, we randomly selected 226 images from NoMet class and group the original images into dataset D1 as shown in Table 2. Applying data augmentation technique on D1, we obtained an augmented dataset D2. The dataset D3 is achieved by aggregating images in D2.
Table 2.
Dataset | ADMet | nADMet | NoMet | Total |
---|---|---|---|---|
D1 | 237 | 160 | 226 | 623 |
D2 | 624 | 640 | 614 | 1878 |
D3 | 318 | 320 | 307 | 945 |
For supervised image classification problem, the CNN-based model is evaluated by comparing the automated classification results against ground truth (human performance) that is often obtained by manually labeling images. However, it is time-consuming, laborious, and subjective to manually label low-resolution, large-size SPECT images. To facilitate labeling SPECT image, in this work, we developed an annotation system based on the LabelMe (http://labelme.csail.mit.edu/Release3.0/) released by MIT.
With LabelMe-based annotation system, imaging findings including the DICOM file and the textual diagnostic report can be imported into the system in advance. In the labeling process, three nuclear medicine physicians from the Department of Nuclear Medicine, Gansu Provincial Tumor Hospital manually labeled areas on the visual presentation of DICOM file with a shape tool (e.g., polygon and rectangle). The labeled area will be annotated with a self-defined code combined with the name of disease or body part. The manually labeled results for all images act as ground truth in the experiments and form an annotation file together to feed into the classifiers.
Experimental setup
The evaluation metrics we use are accuracy, precision, recall, specificity, F-1 score, and AUC (Area Under ROC Curve), which are defined in Eqs. 6–10.
6 |
7 |
8 |
9 |
10 |
where the notations are TP = True Positive, TN = True Negative, FP = False Positive and FN = False Negative.
It is desirable that a classifier shows both a high true positive rate (TPR = Recall), and a low false positive rate (FPR = 1–Specificity) simultaneously. The ROC curve shows the true positive rate (y-axis) against the false positive rate (x-axis), and the AUC value is defined as the area under the ROC curve. As a statistical explanation, the AUC value is equal to the probability that a randomly chosen positive image is ranked higher than a randomly chosen negative image. Therefore, the closer to 1 the AUC value is, the higher performance the classifier achieves.
We divided every dataset (D1, D2 and D3) into two parts: training set and testing set, with the ratio of them being 7: 3. It means that we use 70% of samples in each dataset to train the classifiers, and the rest 30% for testing the classifiers. Images including the augmented ones from the same patient were not divided into the different subsets because they would show similarities. The parameters setting is shown in Table 3.
Table 3.
Parameter | Value |
---|---|
Learning rate | 0.01 |
Optimizer | Adam |
Batch size | 32 |
Epoch | 300 |
The experiments are run in Tensorflow 2.0 on an Intel Core i7-9700 PC with 32 GB RAM running Windows 10.
Experimental results
For the proposed multiclass classifiers Classifer-inRAC and Classifer-outRAC, Table 4 reports the scores of the defined evaluation metrics obtained on the testing samples in dataset D3.
Table 4.
Classifier | Accuracy | Precision | Recall | F-1 score |
---|---|---|---|---|
Classifer-inRAC | 0.7782 | 0.7799 | 0.7823 | 0.7764 |
Classifer-outRAC | 0.6725 | 0.7233 | 0.6831 | 0.6723 |
Best value in each column is highlighted in bold
Table 4 shows that the classifier Classifer-inRAC performs better than Classifer-outRAC. Results in Table 5 further show that Classifer-inRAC obtains the best performance on the aggregated samples in augmented dataset (i.e., D3).
Table 5.
Dataset | Accuracy | Precision | Recall | F-1 score |
---|---|---|---|---|
D1 | 0.6150 | 0.6324 | 0.6227 | 0.6058 |
D2 | 0.6968 | 0.7001 | 0.7024 | 0.6930 |
D3 | 0.7782 | 0.7799 | 0.7823 | 0.7764 |
Best value in each column is highlighted in bold
Figure 6 shows the ROC curve and AUC value obtained by Classifer-inRAC on classifying the testing samples in D3, where AUC value = 0.8364.
We further examine the ability of Classifer-inRAC on differentiating between the subclasses of images in dataset D3 by providing confusion matrix in Fig. 7 and scores of evaluation metrics in Fig. 8.
Experimental results in Figs. 7 and 8 show that differentiating subclasses between images with metastasis is more challenging than differentiating between metastatic and non-metastatic images. There are 22 ADMet images that have been incorrectly identified as nADMet ones.
With the testing samples in dataset D3, we show the impacts of network structure and depth on classification performance obtained by the proposed classifier Classifer-inRAC.
Table 6 reports the scores of evaluation metrics obtained by Classifer-inRAC after we remove the residual connection and hybrid attention module from Classifer-inRAC.
Table 6.
Residual | Attention | Accuracy | Precision | Recall | F-1 score |
---|---|---|---|---|---|
× | × | 0.6937 | 0.7032 | 0.7000 | 0.6940 |
× | √ | 0.7042 | 0.7416 | 0.7047 | 0.7031 |
√ | × | 0.7500 | 0.7614 | 0.7532 | 0.7497 |
√ | √ | 0.7782 | 0.7799 | 0.7823 | 0.7764 |
Best value in each column is highlighted in bold
It shows that the best performance can be obtained if Classifer-inRAC has residual connection and hybrid attention module simultaneously from the scores of evaluation metrics as shown in Table 6. Separately, residual connection has more positive impact than hybrid attention mechanism on the classification performance.
Following the architectural design of Classifer-inRAC, we define two classifiers with different network depth, which are given in Table 7.
Table 7.
Clasifier-18 | Clasifier-34 | Clasifier-inRAC | |
---|---|---|---|
Layer | Configuration | ||
Conv | 7 × 7, 64, Stride = 2 | ||
Norm | Batch normalization | ||
Pool | 3 × 3 Max pooling, Stride = 2 | ||
RA-Conv | |||
RA-Conv | |||
RA-Conv | |||
RA-Conv | |||
Global average pooling (GAP) | |||
Softmax |
Figure 9 reports the scores of evaluation metrics obtained by the classifiers defined in Table 7 and Classifer-inRAC, showing comparative advantage of the proposed classifier on classifying whole-body images.
We further test the performance of Classifer-inRAC on two-class classification by merging the metastatic subclasses (i.e., ADMet and nADMet) in the dataset D3. Specifically, the dataset for two-class classification is consisted of metastatic images (n = 638, ≈ 67.51%) and non-metastatic images (n = 307, ≈ 32.49%). Table 8 reports the scores of evaluation metrics on two-class classification of testing samples and Fig. 10 depicts the corresponding confusion matrix.
Table 8.
Accuracy | Precision | Recall | F-1 score | AUC |
---|---|---|---|---|
0.8310 | 0.8696 | 0.8696 | 0.8696 | 0.8147 |
The results of two-class classification show that our classifier performs better on differentiating between metastatic and non-metastatic images than classifying images in different subclasses.
A comparable analysis has also been performed between the proposed model and two classical deep models Inception-v1 [34] and VGG 11 [35], which are given in Table 9 by providing their network structures.
Table 9.
Model | Number of weight layers | Filter | Activation | Learning rate |
---|---|---|---|---|
Inception-v1 | 9 Inception blocks | 1 × 1, 3 × 3, 5 × 5 | ReLU | 10–2 |
VGG 11 | 11 | 3 × 3 | ReLU | 10–2 |
The scores of evaluation metrics obtained by three classifiers on the dataset D3 are reported in Table 10, showing that our model is more suitable for classifying lung cancer-caused images than the classical models. The possible reason is that the network structure of our model (i.e., residual convolution combined with hybrid attention) is capable of extracting more representative features of metastatic lesions.
Table 10.
Model | Accuracy | Precision | Recall | F-1 score |
---|---|---|---|---|
Inception v1 | 0.5387 | 0.6003 | 0.5490 | 0.5415 |
VGG 11 | 0.7324 | 0.7309 | 0.7333 | 0.7309 |
Classifer-inRAC | 0.7782 | 0.7799 | 0.7823 | 0.7764 |
Best value in each column is highlighted in bold
Discussion
In this section, we provide a brief discussion about the reasons that may cause the misclassifications by providing a group of examples in Fig. 11.
Now, we provide the reasons for misclassification explained by one nuclear medicine physician and one oncologist from Gansu Provincial Tumor Hospital.
Misclassification between the metastatic and non-metastatic
Uptake of 99mTc-MDP in benign processes (i.e., knee arthritis) is detected as metastatic lesions by the developed classifier due to the visually similar appearances to skeletal metastasis (see Fig. 11a). Furthermore, a normal bone would show a higher concentration of activity in trabecular bone with a large-mineralizing surface area like the spine. This brings huge challenge to the CNN-based automated classification of SPECT images, hence the metastatic images being misclassified as non-metastatic.
Misclassification between the diseased subclasses
It is very challenging to accurately classify metastatic images since skeletal metastases are often distributed irregularly in the axial skeleton and typically show variability in size, shape, and intensity [7]. The irregularly distributed radioactivity of ADMet can mimic nADMet, and vice versa, resulting in misclassification between ADMet and nADMet (see Fig. 11b).
Multiclass classification vs. two-class classification
Multiclass classification aims to not only determine whether an image contains lung cancer-caused skeletal metastasis, but also differentiate between subclasses of lung cancer (i.e., ADMet and nADMet). This is more difficult than to answer that an image whether contains metastasis (i.e., two-class classification). So, the proposed classifier Classifer-inRAC obtained score of 0.8310 and 0.7782 for accuracy metric for multiclass and tow-class classification problems, respectively.
Metastatic lesions are further examined by providing statistical analysis of shape, location (body region), and uptake intensity in Fig. 12. The mottling, patchy, punctate lesions dominate both the ADMet and nADMet metastasis as shown in Fig. 12a. The chest (vertebra and ribs) acts the main location (i.e., body region) where the lung cancer-caused metastasis is frequently present in as shown in Fig. 2b. As shown in Fig. 12c, the distribution of detected uptake intensity ranges widely, with 44% of lesions falling into [50, 100]; and much higher uptake can often be detected in the regions of urinary bladder and injection point. This further reveals that it is more difficult to develop an automated method for analyzing scintigraphic images than natural images in which the value of pixel ranges from 0 to 255.
To alleviate the issues mentioned above, technical solutions need to be developed in the future. With a large-scale dataset of SPECT images, representative image features can be extracted for each kind of subclasses by CNN-based end-to-end classifiers. This would contribute to improving the performance of distinguishing between metastatic and non-metastatic images. Moreover, statistical analysis conducted on large-scale SPECT images and pathologic findings would have the potential to develop a multi-modal fusion classifier, enabling to achieve higher classification performance between metastatic images caused by various subclasses of lung cancer.
Conclusions
Targeting the automated detection of lung cancer-caused metastasis with SPECT scintigraphy, we have developed a convolutional neural network with the hybrid attention mechanism in this work. Parametric variation was first conducted to augment the dataset of original images. An end-to-end CNN-based classification network has been proposed to automatically extract features from images, aggregate features, and classify high-level features into classes. Clinical whole-body scintigraphic images were utilized to evaluate the developed network. Experimental results have demonstrated that our self-defined network performs well in detecting lung cancer-caused metastasis as well as differentiating between subclasses of lung cancer. The analysis has also been conducted to compare the proposed model with other related models. The results reveal that our method can be used for determining whether an image contains lung cancer-caused skeletal metastasis and differentiating between subclasses of lung cancer.
In the future, we plan to extend our work in the following directions. First, we intend to collect more data of images and laboratory findings to improve the proposed multiclass classification model. Hopefully, a robust and effective computer-aided diagnosis system will be developed. Second, we attempt to develop deep learning-based approaches that can classify whole-body SPECT images with multiple lesions from various primary diseases that may present in a single image.
Abbreviations
- 99mTc-MDP
99MTc-methylene diphosphonate
- ADMet
Adenocarcinoma metastasis
- AUC
Area Under the Curve
- CNN(s)
Convolutional Neural Network(s)
- DICOM
Digital Imaging and Communications in Medicine
- nADMet
Non-adenocarcinoma metastasis
- NoMet
Without metastasis
- PET
Positron Emission Tomography
- ReLU
Rectified Linear Unit
- ROC
Receiver Operating Characteristic
- SPECT
Single Photon Emission Computed Tomography
Authors' contributions
Conceptualization was done by Qiang Lin and Ranru Guo; methodology was done by Qiang Lin and Ranru Guo; software was done by Yanru Guo, Tongtong Li, and Shaofang Zhao; validation was done by Qiang Lin, Yanru Guo, Tongtong Li, and Shaofang Zhao; formal analysis was done by Qiang Lin, Zhengxing Man, and Yongchun Cao; investigation was done by Qiang Lin and Ranru Guo; resources were done by Xianwu Zeng; data curation was done by Xianwu Zeng and Qiang Lin; writing—original draft preparation were done by Qiang Lin and Yanru Guo; writing—review and editing were done by Qiang Lin; visualization was done by Yanru Guo; supervision was done by Qiang Lin; project administration was done by Qiang Lin; funding acquisition was done by Qiang Lin. All authors have read and agreed to the published version of the manuscript. All authors read and approved the final manuscript.
Funding
The work was supported by the Youth Ph.D. Foundation of Education Department of Gansu Province (2021QB-063), the Fundamental Research Funds for the Central Universities (31920210013, 31920190029), Key R&D Plan of Gansu Province (21YF5GA063), the Natural Science Foundation of Gansu Province (20JR5RA511), the National Natural Science Foundation of China (61562075), the Gansu Provincial First-class Discipline Program of Northwest Minzu University (11080305), and the Program for Innovative Research Team of SEAC ([2018]98).
Availability of data and materials
Anyone can get the validation subset by emailing the corresponding author by stating that the data is used for research purposes only. The whole dataset will be publicly available in the future.
Declarations
Ethics approval and consent to participate
The study was approved by the Ethics Committee of Gansu Provincial Tumor Hospital (Approval No.: A202106100014). All procedures performed in this study involving human participants were in accordance with the ethical standards of the Gansu Provincial Tumor Hospital research committee and with the 1964 Helsinki declaration and its later amendments or comparable ethical standards. The used bone SPECT images were de-identified before the authors received the data. The fully anonymized image data were received by the authors on June 01, 2021. A requirement for informed consent was waived for this study because of the anonymous nature of the data.
Consent for publication
Not applicable.
Competing interests
Authors declare that they have no conflict of interest.
Footnotes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Contributor Information
Qiang Lin, Email: qiang.lin2010@hotmail.com.
Xianwu Zeng, Email: 1021619285@qq.com.
References
- 1.Valkenburg KC, Steensma MR, Williams BO, et al. Skeletal metastasis: treatments, mouse models, and the Wnt signaling. Chin J Cancer. 2013;32(7):380–396. doi: 10.5732/cjc.012.10218. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Hess KR, Varadhachary GR, Taylor SH, et al. Metastatic patterns in adenocarcinoma. Cancer. 2006;106:1624–1633. doi: 10.1002/cncr.21778. [DOI] [PubMed] [Google Scholar]
- 3.Mehlen P, Puisieux A. Metastasis: a question of life or death. Nat Rev Cancer. 2006;6(6):449–458. doi: 10.1038/nrc1886. [DOI] [PubMed] [Google Scholar]
- 4.Chang CY, Gill CM, Simeone FJ, et al. Comparison of the diagnostic accuracy of 99m-Tc-MDP bone scintigraphy and 18F-FDG PET/CT for the detection of skeletal metastases. Acta Radiol. 2014;58:1–8. doi: 10.1177/0284185114564438. [DOI] [PubMed] [Google Scholar]
- 5.Sadik M, Jakobsson D, Olofsson F, et al. A new computer-based decision-support system for the interpretation of bone scans. Nucl Med Commun. 2006;27(5):417–423. doi: 10.1097/00006231-200605000-00002. [DOI] [PubMed] [Google Scholar]
- 6.Shan H, Jia X, Yan P, LiY, Paganetti H, Wang G (2020) Synergizing medical imaging and radiotherapy with deep learning. Mach Learn: Sci Technol 1(2): 021001
- 7.Nathan M, Gnanasegaran G, Adamson K, Fogelman I (2013) Bone scintigraphy: patterns, variants, limitations and artefacts. Springer, Berlin
- 8.Sadik M, Hamadeh I, Nordblom P, Suurkula M, Hoglund P, Ohlsson M, Edenbrandt L. Computer–assisted interpretation of planar whole-body bone scans. J Nucl Med. 2008;49(12):1958–1965. doi: 10.2967/jnumed.108.055061. [DOI] [PubMed] [Google Scholar]
- 9.Aslanta A, Dandl E, Akrolu M. CADBOSS: a computer-aided diagnosis system for whole-body bone scintigraphy scans. J Cancer Res Ther. 2016;12(2):787–792. doi: 10.4103/0973-1482.150422. [DOI] [PubMed] [Google Scholar]
- 10.Mac A, Fgeb C, Svp D. Object-oriented classification approach for bone metastasis mapping from whole-body bone scintigraphy. Phys Med. 2021;84:141–148. doi: 10.1016/j.ejmp.2021.03.040. [DOI] [PubMed] [Google Scholar]
- 11.Elfarra FG, Calin MA, Parasca SV. Computer-aided detection of bone metastasis in bone scintigraphy images using parallelepiped classification method. Ann Nucl Med. 2019;33(11):866–874. doi: 10.1007/s12149-019-01399-w. [DOI] [PubMed] [Google Scholar]
- 12.Algan G, Ulusoy I (2021) Image classification with deep learning in the presence of noisy labels: a survey. Knowl-Based Syst 215(3): 106771
- 13.Minaee S, Boykov YY, Porikli F, Plaza AJ, Terzopoulos D. Image segmentation using deep learning: a survey. IEEE Trans Softw Eng. 2021;99:1–1. doi: 10.1109/TPAMI.2021.3059968. [DOI] [PubMed] [Google Scholar]
- 14.Taghanaki SA, Abhishek K, Cohen JP, Cohen-Adad J, Hamarneh G. Deep semantic segmentation of natural and medical images: a review. Artif Intell Rev. 2021;54:137–178. doi: 10.1007/s10462-020-09854-1. [DOI] [Google Scholar]
- 15.Lin Q, Man Z, Cao Y, et al. Classifying functional nuclear images with convolutional neural networks: a survey. IET Image Proc. 2020;14(14):3300–3313. doi: 10.1049/iet-ipr.2019.1690. [DOI] [Google Scholar]
- 16.Litjens G, Kooi T, Bejnordi BE, Setio AAA, Sánchez CI. A survey on deep learning in medical image analysis. Med Image Anal. 2017;42(9):60–88. doi: 10.1016/j.media.2017.07.005. [DOI] [PubMed] [Google Scholar]
- 17.Yin XH, Wang YC, Li DY. Survey of medical image segmentation technology based on U-Net structure improvement. J Softw. 2021;32(2):519–550. [Google Scholar]
- 18.Dang J (2016) Classification in none scintigraphy images using convolutional neural networks. Lund University
- 19.Papandrianos N, Papageorgiou E, Anagnostis A, Papageorgiou K (2020) Bone metastasis classification using whole body images from prostate cancer patients based on convolutional neural networks application. PLoS One 15(8):e0237213 [DOI] [PMC free article] [PubMed]
- 20.Papandrianos N, Papageorgiou E, Anagnostis A, Papageorgiou K. Efficient none metastasis diagnosis in bone scintigraphy using a fast convolutional neural network architecture. Diagnostics. 2020;10(8):532. doi: 10.3390/diagnostics10080532. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Papandrianos N, Papageorgiou E, Anagnostis A. Development of convolutional neural networks to identify bone metastasis for prostate cancer patients in bone scintigraphy. Ann Nucl Med. 2020;34:824–832. doi: 10.1007/s12149-020-01510-6. [DOI] [PubMed] [Google Scholar]
- 22.Cheng DC, Hsieh TC, Yen KY, Kao CH. Lesion-based bone metastasis detection in chest bone scintigraphy images of prostate cancer patients using pre-train, negative mining, and deep learning. Diagnostics. 2021;11(3):518. doi: 10.3390/diagnostics11030518. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Cheng DC, Liu CC, Hsieh TC, Yen KY, Kao CH. Bone metastasis detection in the chest and pelvis from a whole-body bone scan using deep learning and a small dataset. Electronics. 2021;10:1201. doi: 10.3390/electronics10101201. [DOI] [Google Scholar]
- 24.Papandrianos N, Papageorgiou E, Anagnostis A, Papageorgiou K. A deep-learning approach for diagnosis of metastatic breast cancer in bones from whole-body scans. Appl Sci. 2020;10(3):997. doi: 10.3390/app10030997. [DOI] [Google Scholar]
- 25.Pi Y, Zhao Z, Xiang Y, Li Y (2020) Automated diagnosis of bone metastasis based on multi-view bone scans using attention-augmented deep neural networks. Med Image Anal. 65: 101784 [DOI] [PubMed]
- 26.Zhao Z, Pi Y, Jiang L, Cai H. Deep neural network based artificial intelligence assisted diagnosis of bone scintigraphy for cancer bone metastasis. Sci Rep. 2020;10:17046. doi: 10.1038/s41598-020-74135-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Lin Q, Li T, Cao C, Cao Y, Man Z, Wang H. Deep learning based automated diagnosis of bone metastases with SPECT thoracic bone images. Sci Rep. 2021;11:4223. doi: 10.1038/s41598-021-83083-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Lin Q, Cao C, Li T, Cao Y, Man Z, Wang H. Multiclass classification of whole-body scintigraphic images using a self-defined convolutional neural network with attention modules. Med Phys. 2021;48:5782–5793. doi: 10.1002/mp.15196. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Lin Q, Cao C, Li T, Cao Y, Man Z, Wang H. dSPIC: a deep SPECT image classification network for automated multi-disease, multi-lesion diagnosis. BMC Med Imaging. 2021;21:122. doi: 10.1186/s12880-021-00653-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Shorten C, Khoshgoftaar TM. A survey on image data augmentation for deep learning. J Big Data. 2019;6(1):60. doi: 10.1186/s40537-019-0197-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. In: 32nd International conference on international conference on machine learning (ICML), Lille, France, July 6–11, 2015
- 32.Woo S, Park J, Lee JY, et al. (2018) CBAM: convolutional block attention module. arXiv:1807.06521v2
- 33.Chen B, Deng W, Du J (2017) Noisy Softmax: improving the generalization ability of DCNN via postponing the early softmax saturation. In: 2017 IEEE conference on computer vision and pattern recognition (CVPR), Honolulu, HI, USA, July 21–26, 2017
- 34.Szegedy C, Liu W, Jia Y, et al (2015) Going deeper with convolutions. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, USA, June 8–10, 2015
- 35.Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Anyone can get the validation subset by emailing the corresponding author by stating that the data is used for research purposes only. The whole dataset will be publicly available in the future.