Abstract
Background
The incidence rates of cervical cancer in developing countries have been steeply increasing while the medical resources for prevention, detection, and treatment are still quite limited. Computer-based deep learning methods can achieve high-accuracy fast cancer screening. Such methods can lead to early diagnosis, effective treatment, and hopefully successful prevention of cervical cancer. In this work, we seek to construct a robust deep convolutional neural network (DCNN) model that can assist pathologists in screening cervical cancer.
Methods
ThinPrep cytologic test (TCT) images diagnosed by pathologists from many collaborating hospitals in different regions were collected. The images were divided into a training dataset (13,775 images), validation dataset (2301 images), and test dataset (408,030 images from 290 scanned copies) for training and effect evaluation of a faster region convolutional neural network (Faster R-CNN) system.
Results
The sensitivity and specificity of the proposed cervical cancer screening system was 99.4 and 34.8%, respectively, with an area under the curve (AUC) of 0.67. The model could also distinguish between negative and positive cells. The sensitivity values of the atypical squamous cells of undetermined significance (ASCUS), the low-grade squamous intraepithelial lesion (LSIL), and the high-grade squamous intraepithelial lesions (HSIL) were 89.3, 71.5, and 73.9%, respectively. This system could quickly classify the images and generate a test report in about 3 minutes. Hence, the system can reduce the burden on the pathologists and saves them valuable time to analyze more complex cases.
Conclusions
In our study, a CNN-based TCT cervical-cancer screening model was established through a retrospective study of multicenter TCT images. This model shows improved speed and accuracy for cervical cancer screening, and helps overcome the shortage of medical resources required for cervical cancer screening.
Keywords: Cervical cancer, ThinPrep cytologic test (TCT), Deep leaning, Convolutional neural network (CNN)
Background
Cervical cancer is one of the most common malignant tumors in the world, and it is the fourth leading cause of cancer in women [1–3]. The morbidity and mortality of cervical cancer in the developing countries are distinctly higher than those in the developed countries [1, 4]. About four out of five medical cases occur in developing countries, and especially in China and India. Human papillomavirus (HPV) infections have been established as a main cause of cervical cancer [5–7]. Based on a survey of 30,207 cases at 13 Chinese medical centers, the HPV infection rate in China was found to be about 17.7% [8].
The incidence of cervical cancer can be reduced through early screening [9, 10]. For severe cervical cancer types, screening can help in avoiding cancer-related deaths [11]. In 2012, the National Comprehensive Cancer Network (NCCN) published clinical practice guidelines for cervical cancer screening. These guidelines show that the combination of HPV testing and cytology has been used as a primary option for cervical cancer screening in women [12].
In the United States, the number of deaths from cervical cancer has decreased since the implementation of widespread cervical cancer screening from 2.8 to 2.3 deaths per 100,000 women between 2000 and 2015 [13]. However, cervical cancer screening is still a problem with low diagnostic sensitivity and specificity, especially in developing countries. While cervical cancer goes undetected in some patients, it is overtreated in others [14]. In short, it is important to improve the efficiency and accuracy of cervical cancer screening, especially in China and other developing countries.
In the 1990 s, ThinPrep cytologic test (TCT) technology was approved by the Food and Drug Administration (FDA) for clinical applications. Compared with the pap smear, this technique reduced the effects of mucus, blood, and inflammation in cervical cancer screening. It can also effectively reduce the dissatisfaction rate of cell screening by maintaining or improving the sensitivity of disease screening [15–17]. According to Chinese guidelines for diagnosis and treatment of cervical cancer published in 2018, the recommended preliminary screening method is TCT combined with HPV screening [18].
The diagnostic results of the TCT were classified using the Bethesda system [19]. This system represents an artificial screening method for determining the precancerous cervical lesions by microscopic assessment of the changes in cytoplasm, nuclear shape, and the fluid base color [18]. According to the Bethesda system, the smears are mainly divided into the following categories: Intraepithelial Lesion/Malignant Lesion cell (NILM), atypical squamous cells of undetermined significance (ASCUS), low-grade squamous intraepithelial lesion (LSIL), and high-grade squamous intraepithelial lesion (HSIL). Above all, this method requires trained pathologists to make a correct diagnosis at the cellular recognition stage.
In developing countries with a large population, like China, the following three problems are mostly observed in cervical screening that reduces its diagnostic accuracy: (1) identification of cells on TCT smear images is subjective and mostly relies on the experience and technology of the pathologists. It has been reported that the increase in pathologists’ workload also leads to a reduced diagnostic sensitivity [20]. (2) The correct diagnosis of TCT images requires much of the pathologists’ time. However, there is a significant shortage of pathologists in China. Indeed, 68.4% of the pathologists who provide cervical cancer screening services have only primary technical qualifications or no qualifications at all. (3)The gap between urban and rural medically trained pathologists has been increasing [21]. Moreover, it takes more time and effort for human eyes to read images, the number of readings per day is also limited [22].
With the rapid development of science and technology, the emergence of artificial intelligence has provided more sensitive and effective solutions to many medical imaging problems [23, 24]. In recent years, deep learning algorithms have helped to identify patterns in classifying and quantifying medical images [25]. Significant progress has been made in the computer-aided diagnosis of medical images, such as the skin cancer diagnosis from melanoma images [25, 26] and the detection of pulmonary nodules via computerized tomography (CT) images [27, 28] and image-base detection of hepatocellular carcinoma using multiphasic magnetic resonance imaging (MRI) [29]. The concept of deep learning originated from the area of artificial neural network (ANN), which was helpful to extracting appropriate features from medical images. Among them, one of the leading neural networks for image recognition is a convolutional neural network (CNN) [30]. At present, this method is applied in most deep learning applications to achieve good detection results[31, 32].
We have developed a target recognition system based on a CNN model to assist pathologists in the diagnosis of early cervical cancer screening. This method is based on faster region convolutional neural networks (Faster R-CNN)[33] to locate and identify lesion cells in the cervical TCT images, extract features, and automatically optimize them through updating the convolutional kernel parameters in the cyclic training process [34]. Convolutional neural networks have the advantages of automatically learning features during training, and possessing strong inferential capabilities in comparison with conventional image processing methods in complex scenes [35, 36].
The trained faster RCNN model can automatically identify target regions and reduce the false detections of other non-target regions. As well, this model can generate region proposals on the feature map of the target region and its adjacent areas [33, 37]. The region proposal features can distinguish between target and non-target regions through merging layers and classifiers. It can learn the common characteristics of cells of the same lesion grade and the characteristic differences among different lesion grade cells. The identification of generalized performance was reported to be better than performance of conventional image processing methods under the conditions of cell deformation and cell superposition.
In this research project, we have successfully developed an auxiliary diagnosis model for the diagnosis of early cervical cancer. This model separated the positive or negative images effectively, minimizing the rate of missed diagnosis. It also reduced pathologists’ workload by saving their time and improved the accuracy of cervical cancer screening.
Methods
The analysis pipeline for the detection of cervical cancer from the TCT smear using the Faster R-CNN is depicted in Fig. 1.
Study design and participants
Images of TCT were obtained from multiple collaborating hospitals and research institutes in China. The collaborating entities are: (1) The First Affiliated Hospital, Sun Yat-sen University, (2) The First Affiliated Hospital, Xi’an Jiaotong University, (3) Shanghai General Hospital, Shanghai Jiaotong University, School of Medicine, (4) Shanghai First Maternity and Infant Hospital, (5) The First Affiliated Hospital of Soochow University, (6) Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, (7) The Central Hospital of Wuhan, Tongji Medical College, Huazhong University of Science and Technology, (8) Xiangyang Hospital Affiliated to Hubei University of Medicine, and (9) the Molecular Department, Kindstar Global, Wuhan. About 16,000 TCT images of cervical brush smears were randomly selected and collected from outpatients at the above nine hospitals. All TCT images were histopathologically diagnosed as with different grades of a lesion by at least two local experienced pathologists according to the Bethesda system. This diagnosis was used as the gold standard for subsequent experiments. The images were sent to Tongji Hospital affiliated to Tongji Medical College of Huazhong University of Science and Technology, and the abnormal cells were examined and labeled by a pathologist.
Image preparation and preprocessing
A pathologist from Tongji Hospital examined and preprocessed all images. High-quality images that met the requirements were selected. The inclusion criteria were as follows: The cells were evenly spread on the slide, the field of vision was clear, and the number of cells in the slide were moderate, with fewer overlapping cells.
Images of non-JPG formats (such as scanned copies from slides) couldn’t be directly marked by pathologists because of the extra-large image size. So, these images were sent to Zhejiang University Rui Medical Artificial Intelligence Research Center for further processing and segmentation. One slide image was divided into 1407 single-field images with enlarged cervical cytology images using seamless slider technology. Figure 2 displays the process of the seamless slider technique, with a scanned image size of 83,712 × 86,784 × 3. In this work, the eyepiece has a 10X magnification, while the objective lens has magnifications of 20X or 40X. Hence, images with combined magnification factors of 200X and 400X were obtained. Then, those images were sent to the Tongji Hospital for creating test data through labeling irregular cells in images. The LabelImg software (Version 1.5.0.), an open-source graphical image annotation tool, was used to annotate the lesion areas in the eligible images. The primary purpose of lesion labeling was to train a model to recognize the abnormal cells. After that, all images with abnormal cell labeling were sent to Zhejiang University Rui Medical Artificial Intelligence Research Center for model training.
Image processing
Image patching
Cervical cytology has a more diverse form of color expression compared with the standard red and blue. Because of the differences in the staining methods and the severity of cervical cancer, the cells show varied and complicated cell morphology patterns. The image analysis flow chart for the Faster R-CNN system for the detection of cervical cancer from the TCT smear is depicted in Fig. 1. In this paper, we created three different datasets. The first one was named as the training dataset with 13,775 images, of which 5414, 4292, and 4069 were ASCUS, LSIL, and HSIL images, respectively. These images showed 200X and 400X magnification levels. We used this dataset to train the Faster R-CNN model. The second dataset was named as the validation dataset with 2301 images, of which 1000 images had a 200X magnification and 1301 images had a 400X magnification. The detailed statistics of each kind of picture scan piece or field of view are presented in Table 1.
Table 1.
A: Training dataset | ||||
ASCUS | LSIL | HSIL | Total | |
Images | 5414 | 4292 | 4069 | 13775 |
B: Validation dataset | ||||
ASCUS | LSIL | HSIL | Total | |
Images | 801 | 780 | 720 | 2301 |
C: Test dataset | ||||
Normal | ASCUS | LSIL | HSIL | |
Scanning copy | 112 | 47 | 70 | 61 |
Images | 157584 | 250,446 |
Image enhancement
Because the data samples in this study are limited, we employ image processing methods to simulate and transform the color and shape of the sample images in order to increase the data diversity and better simulate the real data variability. The image processing methods were implemented by an image rollover, changing the brightness, saturation, or adjusting the color deviation within the original images. Figure 3 illustrates some of the images that are simulated or transformed during the enhancement process.
Network model
The processed cervical cytological images, 13,775 pictures, were used as training samples, and the labeled boxes and the pathological changes in irregular cells within the images were used as training labels. The deep CNN architecture contains a convolution kernel, an activation function and a pooling function of 3 × 3 and 1 × 1 sizes. During training, this architecture is used to extract image features and get feature maps. The feature map is fed into the region proposal network, and the IoU metric is used to select the region proposals for real and interfering targets. The proposal box was used to select the feature map in the feature graph of the previous step and process it through the grid pooling layer to obtain a pooled feature map. Then, the real target and interference target features were identified contained in the pooled feature graph by classification network. The deviation of the proposal box and ground-truth box was calculated. The network was optimized by backpropagation combining classification error and proposal box deviation, to achieve higher classification accuracy and more accurate localization of region proposals.
Loss function
During training, an objective loss function was used to measure the model localization and detection performance. This function is defined as:
1 |
where, i is the subscript of an anchor in a mini-batch, is the prediction possibility of an object in anchor i, and is the ground-truth tag. If the anchor is positive, the ground-truth label is 1, otherwise, it is 0. represents four parameterized coordinates of bounding-box of the predicted target region, (anchor-based transformation), is the ground-truth box corresponding to this positive anchor (transformation based on anchor) represents classification loss. stands for regression loss.
Regression loss
Regression loss is the robust smooth-L1 loss function. indicates that regression loss is only activated when the value of positive anchor () is equal to unity. Therefore, the regression adopted the calculation stable smooth-L1 method and used it to calculate the deviation between the predicted target coordinates and ground-truth. Then it minimized the error to make the predicted target position closer to ground-truth to achieve an accurate model positioning. The smooth-L1 function is used to calculate the distance between the coordinates of the predicted and ground-truth targets. The smooth-L1 loss is defined as:
2 |
Classification loss
There were two ways to classify the training samples. One was to divide the samples into normal or abnormal ones based on the presence of diseased cells. The classification loss uses cross-entropy, which is a binary classifier (only in the case of yes and no). For each category, the two probabilities predicted by the model were p and 1-p, respectively, and the classification loss is defined as:
3 |
The other way was to divide the microscopic field of view into LSIL, HSIL, and ASCUS images based on the lesion degree. For the second classification method, the algorithm encoded multiple categories as a continuous sequence with 0 starting and 1 interval. The classification loss is as follows:
4 |
where, N represents the number of categories, i.e., the types of lesions, values 0 or 1, if the sample category is the same as category C, the value is 1, otherwise, it is 0. represents the probability that the prediction samples belong to category C.
Model identification
For the feature identification of the same area, the voting method was used to select the category. The number of probability values that each model could output was the same as the number of categories to be identified, i.e., each category corresponded to a probability value. The probability value is directly proportional to the possibility. The higher the probability value is, the higher the possibility value is. The probability outputs of the two models were averaged, and then the category corresponding to the maximum probability value was selected as the output category. The average value calculated by the voting algorithm is as follows:
5 |
where, i is the serial number of the category, N is the number of categories to be identified, is the probability value of category i in the N probability values output by the M th model, and is the average probability value of the model in category i, and m is the number of models. The category corresponding to the maximum probability value was selected as the final recognition category.
Statistical analysis
The receiver operating characteristic (ROC) curve was used to evaluate the ability of the auxiliary diagnosis model of cervical cancer to distinguish between the negative and positive images and to calculate the Area Under ROC Curve (AUC) values. AUC is a more objective evaluation index to measure the advantages and disadvantages of the two-classification model. The ROC curve and AUC value are calculated by Python (version 3.6) and scikit -learn (version 0.20.0).
Results
The samples were divided into normal, ASCUS, LSIL, and HSIL cells as well as squamous cell carcinoma (SqCa). Performance indicators included sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV), defined as follows:
6 |
7 |
8 |
9 |
where, TN, TP, FN, and FP represent the numbers of true-negative, true-positive, false- negative, and false- positive images, respectively.
Evaluation of positive and negative classification
Table 2 shows the confusion matrix of the classification results of the TCT technique used for the test data composed of 290 scans of complete slides. Of the 112 normal slides, 39 were correctly classified as normal and 73 were incorrectly classified as abnormal. Out of 178 abnormal slides, a total of 177 slides were correctly classified. This classifier resulted in a sensitivity of 99.4% and a specificity of 34.8%. The diagnostic test performance was better quantified using the positive predictive value (PPV) and the negative predictive value (NPV). The predictive value represented the probability that an experiment could make a correct judgment and the actual clinical significance of the results. In our system, the PPV and NPV were 70.8 and 97.5%, respectively. Furthermore, we constructed the ROC curve which is a comprehensive indicator of the continuous variables of sensitivity and specificity. The proposed system had an AUC of 0.67 for the ROC curve in Fig. 4.
Table 2.
CNN system | Turth | Total | |
---|---|---|---|
Normal | Abnormal | ||
Normal | 177 (TP) | 73 (FP) | 242 |
Abnormal | 1 (FN) | 39 (TN) | 48 |
Total | 178 | 112 | 290 |
Evaluation of lesiondegree classification
Our tests were performed on 1975 images generated after slide segmentation as shown in Fig. 2. This system didn’t provide a direct diagnosis but provided a reference for the pathologists according to the cells with the highest degree of the lesion. Table 3 shows the confusion matrix used to test the results of TCT classification according to lesion degree. Of the 606 ASCUS images, 541 were correctly classified. Also, 422 of the 590 LSIL images and 576 of the 799 HSIL images were classified correctly. Therefore, the sensitivities for the ASCUS, LSIL, and HSIL were 89.3, 71.5, and 73.9%, respectively.
Table 3.
ASCUS | LISL | HSIL | |
---|---|---|---|
True-positive | 541 | 422 | 576 |
False-negative | 65 | 168 | 193 |
Total | 606 | 590 | 779 |
Sensitivity | 89.3% | 71.5% | 73.9% |
Discussion
In this paper, we have proposed a cervical cancer screening method based on CNN, which could successfully extract features from TCT images. The pathological condition of glass slides was diagnosed by dividing the images scanned by glass slides into a large number of single visual field images to establish the model, and by integrating the results of the visual field processing. In general, this model could distinguish between negative and positive cells, with a recognition sensitivity of 99.4%. Proper diversion of negative and positive marker images was also the focus of developing the cervical cancer-assisted screening system. This model achieved a 99.6% accuracy to exclude negative images (i.e. normal or healthy populations), and the remaining suspicious images were analyzed and confirmed by pathologists.
The specificity of the system was only 34.8%, and hence many normal cells were misclassified as abnormal ones. However, the key goal of this model was to ensure that there were no false-negative images as much as possible, and thus reduce the rate of missed diagnosis. The PPV was 70.8%, which meant that the test-positive patients had a 70.8% chance of actually having the disease. The NPV was 97.5%, which meant that the test-negative patients had a 97.5% chance of actually not having the disease. Although the low specificity means that more than half of the normal women may be classified as suspicious cases for further diagnosis, the system had a low false-negative rate of 0.06%, which indicated that the rate of missed diagnosis greatly decreased and it could effectively identify suspicious cases. This also helps pathologists improve their disease diagnosis efficiency, so they can spend more time on analyzing complex cases and ensuring that diseases are diagnosed quickly and accurately. Although cervical cancer screening has become very popular, there is an imbalance in medical resources and untrained pathologists in some parts of China. Our deep learning model may help solve the above problems, and lead to a more accurate and effective diagnosis.
Moreover, this model can be re-classified according to the abnormal cell details. The sensitivity of ASCUS, LSIL, and HSIL was 89.3, 71.5, and 73.9%, respectively. According to the degree of a lesion, the classification of cervical cancer cells indicated that the model has good sensitivity to identify the degree of a lesion in a single field. The diagnosis of the TCT smear is a synthesis of the results of a large number of the single-field images, and our system was able to assist the specialists in locating and identifying different degrees of diseased cells quickly, allowing them to devote more time on more complex cases.
To our knowledge, only few studies on automatic diagnosis of cervical cancer were based on ThinPrep cytologic test (TCT) images instead of pap smears. This is mainly because TCT images are relatively difficult to analyze compared to the pap smear images. More specifically: (1) the TCT images do not currently have a good database, and the collection of such an image database is more difficult compared to the Harlve database of pap-smear images [38, 39]; (2) TCT images have many overlapping cells, which are not as easy to analyze as the single cells in pap-smear images [40, 41]; (3) the color and quality of TCT images obtained from different medical institutions may vary greatly. Therefore, these systems are poorly adapted for the proposed system, we trained it with data from hospitals at different levels in different regions of north and south China, which ensured the high adaptability of the system. Theoretically, the CNN system has good robustness and low dependence on smears staining and preparation. The system can adapt to the images from different hospitals and reduce the secondary examination caused by low-quality images. Future research is needed for further verification in other clinical applications.
A sensitivity screening system may help pathologists to quickly diagnose some positive lesions with a computational speed of 70 − 10 milliseconds per single field. Therefore, this system can relieve the burden of the pathologists to some extent. In addition, the system not only reduces the influence of appellate factors, and ensures consistency in diagnosing images of different pathologists from different hospitals, but also increases the possibility of remote diagnosis. Meanwhile, because the automatic cervical cancer diagnosis system is mainly based on CNN models, the construction and diagnosis costs are very low.
However, our proposed system also has several major limitations. Firstly, Due to a high number of overlapping and adhesion cells, the differences between diseased cells and normal cells were not distinct. It was difficult for the system to learn the characteristics of cells, and hence its accuracy was relatively low. Secondly, the collected dataset didn’t have enough samples of some pathological types such as atypical squamous cells-cannot exclude high-grade squamous intraepithelial lesion (ASC-H) or atypical glandular cells (AGC). This scarcity can’t lead to effective training and accurate classification. Therefore, we plan to collect more samples and enlarge our dataset in order to arrive at more valuable findings. Ultimately, it is difficult to obtain complete TCT image information from outpatients. Future research should focus on more relevant patient information, such as age, HPV infection, or vaginal inflammation. Similarly, large amounts of data will be needed to design prospective follow-up studies to enhance the stability and accuracy of the model. In addition, with the discovery of new biomarkers, tumor diagnosis based on molecular imaging may be explored as a viable path [42–44].
Conclusions
The CNN-based TCT cervical cancer cell classification system proposed in this paper can effectively exclude negative smear samples, and identify the suspicious population in cervical cancer screening. It also showed high sensitivity and excellent performance in the identification of cervical cancer screening, which can save time for pathologists and provide an excellent secondary prevention effect. In comparison to the conventional diagnostic methods, this system has good robustness, objectivity, and small computational cost. Meanwhile, our system provides a possibility for online diagnosis of TCT images and is expected to contribute to the construction of primary medical care.
Acknowledgements
We would like to thank all the partner hospitals which helped collect the TCT smear images.
Abbreviations
- HPV
Human papillomavirus
- TCT
Thinprep Cytologic Test
- Faster R-CNN
Faster region- convolution neural network
- NCCN
National Comprehensive Cancer Network
- DCNN
Deep convolutional neural network
- FDA
Food and Drug Administration
- ANN
Artificial neural network
- ASCUS
Atypical squamous cells of undetermined significance
- NILM
Intraepithelial Lesion/Malignant Lesion cell
- ASC-H
Atypical squamous cells-cannot exclude high-grade squamous intraepithelial lesion
- AGC
Atypical glandular cells
- LSIL
Low-grade squamous intraepithelial lesion
- HSIL
High grade squamous intraepithelial lesions
- SqCa
Squamous cell carcinoma
- CT
Computerized tomography
- MRI
Magnetic Resonance Imaging
- AUC
Area under the curve
- ROC
Receiver operating characteristic
- TN
Ture negative
- TP
Ture positive
- FN
False negative
- FP
False positive
- PPV
Positive predictive value
- NPV:
Negative predictive value
Authors’ contributions
XYH and JW made contribution to the conception and design of this study. XPL labeled all the smear images with abnormal cells using the LabelImg software. XYT and KXL participated in the data collection, assisting with the analysis of data, editing the original manuscript. JCZ, WZW and BW performed scanned image processing, annotation data pre-processing, code implementation. All authors read and approved the final manuscript. We faithfully thank all participators of this study for their time and effort.
Funding
This study was supported by Technical Innovation Special Project of Hubei Province, Grant/Award Number: 2018ACA138.
Availability of data and materials
All the data used in this article is available from the corresponding authors on reasonable request.
Ethics approval and consent to participate
This retrospective study was approved by the Medical Ethics Committee of Tongji Medical College of Huazhong University of Science and Technology (Hu Bei, China), which waived the requirement for patients’ written informed consent referring to the guidelines of the Council for International Organizations of Medical Sciences (CIOMS).
Consent for publication
The publication was approved by all the authors.
Competing interests
All authors declare no conflict on interest.
Footnotes
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Xiangyu Tan and Kexin Li contributed equally to this work
Contributor Information
Xiaoping Li, Email: xiaopingli11@163.com.
Xiaoyuan Huang, Email: huangxy@tjh.tjmu.edu.cn.
References
- 1.Siegel RL, Miller KD, Jemal A. Cancer statistics, 2019. CA Cancer J Clin. 2019;69(1):7–34. doi: 10.3322/caac.21551. [DOI] [PubMed] [Google Scholar]
- 2.Ghasemi F, Shafiee M, Banikazemi Z, Pourhanifeh MH, Khanbabaei H, Shamshirian A, Amiri Moghadam S, ArefNezhad R, Sahebkar A, Avan A, et al. Curcumin inhibits NF-kB and Wnt/beta-catenin pathways in cervical cancer cells. Pathol Res Pract. 2019;215(10):152556. doi: 10.1016/j.prp.2019.152556. [DOI] [PubMed] [Google Scholar]
- 3.Sadri Nahand J, Moghoofei M, Salmaninejad A, Bahmanpour Z, Karimzadeh M, Nasiri M, Mirzaei HR, Pourhanifeh MH, Bokharaei-Salim F, Mirzaei H, et al. Pathogenic role of exosomes and microRNAs in HPV-mediated inflammation and cervical cancer: a review. Int J Cancer. 2020;146(2):305–20. doi: 10.1002/ijc.32688. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Torre LA, Bray F, Siegel RL, Ferlay J, Lortet-Tieulent J, Jemal A. Global cancer statistics, 2012. CA Cancer J Clin. 2015;65(2):87–108. doi: 10.3322/caac.21262. [DOI] [PubMed] [Google Scholar]
- 5.Goodman A. HPV testing as a screen for cervical cancer. BMJ. 2015;350:h2372. doi: 10.1136/bmj.h2372. [DOI] [PubMed] [Google Scholar]
- 6.Shafabakhsh R, Reiter RJ, Mirzaei H, Teymoordash SN, Asemi Z. Melatonin: a new inhibitor agent for cervical cancer treatment. J Cell Physiol. 2019;234(12):21670–82. doi: 10.1002/jcp.28865. [DOI] [PubMed] [Google Scholar]
- 7.Nahand JS, Taghizadeh-Boroujeni S, Karimzadeh M, Borran S, Pourhanifeh MH, Moghoofei M, Bokharaei-Salim F, Karampoor S, Jafari A, Asemi Z, et al. microRNAs: New prognostic, diagnostic, and therapeutic biomarkers in cervical cancer. J Cell Physiol. 2019;234(10):17064–99. doi: 10.1002/jcp.28457. [DOI] [PubMed] [Google Scholar]
- 8.Chen L, Song Y, Ruan G, Zhang Q, Lin F, Zhang J, Wu T, An J, Dong B, Sun P. Knowledge and attitudes regarding HPV and vaccination among Chinese women aged 20 to 35 years in Fujian Province: a cross-sectional study. Cancer Control. 2018;25(1):1073274818775356. doi: 10.1177/1073274818775356. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Sasieni P, Adams J. Effect of screening on cervical cancer mortality in England and Wales: analysis of trends with an age period cohort model. BMJ. 1999;318(7193):1244–5. doi: 10.1136/bmj.318.7193.1244. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Herbert A, Stein K, Bryant TN, Breen C, Old P. Relation between the incidence of invasive cervical cancer and the screening interval: is a five year interval too long? Journal of medical screening. 1996;3(3):140. doi: 10.1177/096914139600300307. [DOI] [PubMed] [Google Scholar]
- 11.Landy R, Pesola F, Castanon A, Sasieni P. Impact of cervical screening on cervical cancer mortality: estimation using stage-specific results from a nested case-control study. Br J Cancer. 2016;115(9):1140–6. doi: 10.1038/bjc.2016.290. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Saslow D, Solomon D, Lawson HW, Killackey M, Kulasingam SL, Cain J, Garcia FA, Moriarty AT, Waxman AG, Wilbur DC, et al. American Cancer Society, American Society for Colposcopy and Cervical Pathology, and American Society for Clinical Pathology screening guidelines for the prevention and early detection of cervical cancer. CA Cancer J Clin. 2012;62(3):147–72. doi: 10.3322/caac.21139. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Force USPST, Curry SJ, Krist AH, Owens DK, Barry MJ, Caughey AB, Davidson KW, Doubeni CA, Epling JW, Jr, Kemper AR, et al. Screening for cervical cancer: US preventive services task force recommendation statement. JAMA. 2018;320(7):674–86. doi: 10.1001/jama.2018.10897. [DOI] [PubMed] [Google Scholar]
- 14.Schiffman M, Kinney WK, Cheung LC, Gage JC, Fetterman B, Poitras NE, Lorey TS, Wentzensen N, Befano B, Schussler J, et al. Relative performance of HPV and cytology components of cotesting in cervical screening. J Natl Cancer Inst. 2018;110(5):501–8. doi: 10.1093/jnci/djx225. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Costa MO, Heraclio SA, Coelho AV, Acioly VL, Souza PR, Correia MT. Comparison of conventional Papanicolaou cytology samples with liquid-based cervical cytology samples from women in Pernambuco, Brazil. Braz J Med Biol Res. 2015;48(9):831–8. doi: 10.1590/1414-431X20154252. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Pankaj S, Nazneen S, Kumari S, Kumari A, Kumari A, Kumari J, Choudhary V, Kumar S. Comparison of conventional Pap smear and liquid-based cytology: a study of cervical cancer screening at a tertiary care center in Bihar. Indian J Cancer. 2018;55(1):80–3. doi: 10.4103/ijc.IJC_352_17. [DOI] [PubMed] [Google Scholar]
- 17.Jeong H, Hong SR, Chae SW, Jin SY, Yoon HK, Lee J, Kim EK, Ha ST, Kim SN, Park EJ, et al. Comparison of unsatisfactory samples from conventional smear versus liquid-based cytology in uterine cervical cancer screening test. J Pathol Transl Med. 2017;51(3):314–9. doi: 10.4132/jptm.2017.03.17. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.National Health Commission of the Peoples Republic of China Chinese guidelines for diagnosis and treatment of cervical cancer 2018 (English version) Chin J Cancer Res. 2019;31(2):295–305. doi: 10.21147/j.issn.1000-9604.2019.02.04. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Davey DD, Souers RJ, Goodrich K, Mody DR, Tabbara SO, Booth CN. Bethesda 2014 implementation and human papillomavirus primary screening: practices of laboratories participating in the College of American Pathologists PAP education program. Arch Pathol Lab Med. 2019;143(10):1196–202. doi: 10.5858/arpa.2018-0603-CP. [DOI] [PubMed] [Google Scholar]
- 20.Renshaw AA, Elsheikh TM. Sensitivity and workload for manual and automated gynecologic screening: best current estimates. Diagn Cytopathol. 2011;39(9):647–50. doi: 10.1002/dc.21439. [DOI] [PubMed] [Google Scholar]
- 21.Di J, Rutherford S, Chu C. Review of the cervical cancer burden and population-based cervical cancer screening in China. Asian Pac J Cancer Prev. 2015;16(17):7401–7. doi: 10.7314/apjcp.2015.16.17.7401. [DOI] [PubMed] [Google Scholar]
- 22.Ellis K, Renshaw AA, Dudding N. Individual estimated sensitivity and workload for manual screening of SurePath gynecologic cytology. Diagn Cytopathol. 2012;40(2):95–7. doi: 10.1002/dc.21495. [DOI] [PubMed] [Google Scholar]
- 23.Ahmad OF, Soares AS, Mazomenos E, Brandao P, Vega R, Seward E, Stoyanov D, Chand M, Lovat LB. Artificial intelligence and computer-aided diagnosis in colonoscopy: current evidence and future directions. The Lancet Gastroenterology Hepatology. 2019;4(1):71–80. doi: 10.1016/S2468-1253(18)30282-6. [DOI] [PubMed] [Google Scholar]
- 24.Hosny A, Parmar C, Quackenbush J, Schwartz LH, Aerts H. Artificial intelligence in radiology. Nat Rev Cancer. 2018;18(8):500–10. doi: 10.1038/s41568-018-0016-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Hekler A, Utikal JS, Enk AH, Berking C, Klode J, Schadendorf D, Jansen P, Franklin C, Holland-Letz T, Krahl D, et al. Pathologist-level classification of histopathological melanoma images with deep neural networks. Eur J Cancer. 2019;115:79–83. doi: 10.1016/j.ejca.2019.04.021. [DOI] [PubMed] [Google Scholar]
- 26.Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, Thrun S. Dermatologist-level classification of skin cancer with deep neural networks. Nature. 2017;542(7639):115–8. doi: 10.1038/nature21056. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Zhao W, Yang J, Sun Y, Li C, Wu W, Jin L, Yang Z, Ni B, Gao P, Wang P, et al. 3D deep learning from CT scans predicts tumor invasiveness of subcentimeter pulmonary adenocarcinomas. Cancer Res. 2018;78(24):6881–9. doi: 10.1158/0008-5472.CAN-18-0696. [DOI] [PubMed] [Google Scholar]
- 28.Yu KH, Zhang C, Berry GJ, Altman RB, Re C, Rubin DL, Snyder M. Predicting non-small cell lung cancer prognosis by fully automated microscopic pathology image features. Nat Commun. 2016;7:12474. doi: 10.1038/ncomms12474. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Wang CJ, Hamm CA, Savic LJ, Ferrante M, Schobert I, Schlachter T, Lin M, Weinreb JC, Duncan JS, Chapiro J, et al. Deep learning for liver tumor diagnosis part II: convolutional neural network interpretation using radiologic imaging features. Eur Radiol. 2019;29(7):3348–57. doi: 10.1007/s00330-019-06214-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Shin HC, Roth HR, Gao M, Lu L, Xu Z, Nogues I, Yao J, Mollura D, Summers RM. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans Med Imaging. 2016;35(5):1285–98. doi: 10.1109/TMI.2016.2528162. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Araujo FHD, Silva RRV, Ushizima DM, Rezende MT, Carneiro CM, Campos Bianchi AG, Medeiros FNS. Deep learning for cell image segmentation and ranking. Comput Med Imaging Graph. 2019;72:13–21. doi: 10.1016/j.compmedimag.2019.01.003. [DOI] [PubMed] [Google Scholar]
- 32.Toratani M, Konno M, Asai A, Koseki J, Kawamoto K, Tamari K, Li Z, Sakai D, Kudo T, Satoh T, et al. A convolutional neural network uses microscopic images to differentiate between mouse and human cell lines and their radioresistant clones. Cancer Res. 2018;78(23):6703–7. doi: 10.1158/0008-5472.CAN-18-0653. [DOI] [PubMed] [Google Scholar]
- 33.Ren S, He K, Girshick R, Sun J. Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell. 2017;39(6):1137–49. doi: 10.1109/TPAMI.2016.2577031. [DOI] [PubMed] [Google Scholar]
- 34.Krizhevsky A, Sutskever I, Hinton GJ. ImageNet classification with deep convolutional neural networks. Commun ACM. 2017;60(6):84–90. [Google Scholar]
- 35.Chen Y , Fan H , Xu B , et al. Drop an octave: reducing spatial redundancy in convolutional neural networks with octave convolution[C]//2019 IEEE/CVF International conference on computer vision (ICCV). IEEE, 2020
- 36.Lecun Y, Kavukcuoglu K, Farabet CM: Convolutional networks and applications in vision. In: Proceedings of 2010 IEEE International Symposium on Circuits and Systems: 2010; 2010.
- 37.Girshick R: Fast R-CNN. In: 2015 IEE International Conference on Computer Vision (ICCV): 2016; 2016.
- 38.Iliyasu AM, Fatichah C. A quantum hybrid PSO combined with fuzzy k-NN approach to feature selection and cell classification in cervical cancer detection. Sensors (Basel) 2017;17(12):2935. doi: 10.3390/s17122935. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Kar S, Majumder DD. A novel approach of mathematical theory of shape and neuro-fuzzy based diagnostic analysis of cervical cancer. Pathol Oncol Res. 2019;25(2):777–90. doi: 10.1007/s12253-019-00582-8. [DOI] [PubMed] [Google Scholar]
- 40.Song Y, Tan EL, Jiang X, Cheng JZ, Ni D, Chen S, Lei B, Wang T. Accurate cervical cell segmentation from overlapping clumps in pap smear images. IEEE Trans Med Imaging. 2017;36(1):288–300. doi: 10.1109/TMI.2016.2606380. [DOI] [PubMed] [Google Scholar]
- 41.William W, Ware A, Basaza-Ejiri AH, Obungoloch J. A pap-smear analysis tool (PAT) for detection of cervical cancer from pap-smear images. Biomed Eng Online. 2019;18(1):16. doi: 10.1186/s12938-019-0634-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Jafari SH, Saadatpour Z, Salmaninejad A, Momeni F, Mokhtari M, Nahand JS, Rahmati M, Mirzaei H, Kianmehr M. Breast cancer diagnosis: Imaging techniques and biochemical markers. J Cell Physiol. 2018;233(7):5200–13. doi: 10.1002/jcp.26379. [DOI] [PubMed] [Google Scholar]
- 43.Saadatpour Z, Bjorklund G, Chirumbolo S, Alimohammadi M, Ehsani H, Ebrahiminejad H, Pourghadamyari H, Baghaei B, Mirzaei HR, Sahebkar A, et al. Molecular imaging and cancer gene therapy. Cancer Gene Ther. 2016;18:1–5. doi: 10.1038/cgt.2016.62. [DOI] [PubMed] [Google Scholar]
- 44.Keshavarzi M, Darijani M, Momeni F, Moradi P, Ebrahimnejad H, Masoudifar A, Mirzaei H. Molecular imaging and oral cancer diagnosis and therapy. J Cell Biochem. 2017;118(10):3055–60. doi: 10.1002/jcb.26042. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
All the data used in this article is available from the corresponding authors on reasonable request.