Abstract
Cervical cancer continues to pose a significant global health challenge, highlighting the urgent need for accurate and efficient diagnostic techniques. Recent progress in deep learning has demonstrated considerable potential in improving the detection and classification of cervical cancer. This review presents a thorough analysis of deep learning methods utilized for cervical cancer diagnosis, with an emphasis on critical approaches, evaluation metrics, and the ongoing challenges faced in the field. We explore various deep learning architectures, particularly convolutional neural networks (CNNs), and their applications in the segmentation and classification of cervical cytology images. Key performance indicators, such as accuracy, sensitivity, specificity, and the area under the curve (AUC), are reviewed to assess the effectiveness of these models. Despite advancements, challenges like limited annotated datasets, inconsistencies in medical imaging, and the demand for more resilient models remain. Strategies like data augmentation, transfer learning, and semi-supervised learning are examined as potential solutions. This review synthesizes current research to guide future studies and clinical implementations, ultimately advancing early detection and treatment of cervical cancer through cutting-edge deep learning technologies.
Keywords: cervical cancer, machine learning, deep learning, segmentation
Introduction
According to the Global Cancer Observatory (GCO), 661,000 new cases of cervical cancer were identified worldwide in 2022, with 348,000 lakh associated deaths [1]. Within India, cervical cancer ranks as the second most prevalent cancer among women, with approximately 75,209 new cases with a crude rate of 10.9 and cumulative risk of 1 in 75 reported in 2020 [2]. Annually, India sees between 2000 and 3000 hospitalizations due to cervical cancer, with a majority of cases being diagnosed at advanced stages. India’s cervical cancer mortality rate is more than double that of countries like the United Kingdom, the Netherlands, and Finland, despite efforts such as screening programs and vaccination campaigns. Malignancy treatment continues to be the most expensive illness in India, with the average expense for cancer treatment varying from 100,000 INR to 1,000,000 INR or more (approximately 1,200 USD to 12,000+ USD) depending on several factors, including the stage of the disease, treatment methods, and the healthcare facility [3]. The majority of these costs, amounting to three quarters, are incurred as out-of-pocket expenses [4].
Screening, recommended for women aged 25 to 74, includes methods like colposcopy, where the cervix and vagina are examined under magnification [5]. Early detection is crucial, as delays reduce survival rates. Research indicates that inadequate detection and treatment lead to high mortality rates. Early screening is vital to reducing the global burden of cervical cancer, yet barriers such as lack of awareness and costly procedures limit access in developing nations [6].
Figure 1.
A., C. Cross sectional view of cervix [8]; B. Detection of pre-cervical cancer
Deep learning-based convolutional neural networks, relying on robust feature extraction capabilities, can autonomously identify key features in an image through gradient descent [7]. This advancement significantly enhances object recognition accuracy, establishing deep learning as a prominent tool in computer image processing. Implementing this technology in cervical cancer diagnosis can assist clinicians in making precise auxiliary assessments, decreasing their workload, and enhancing diagnostic accuracy. The growth in computational power, availability of large datasets, and proven efficacy of deep learning in fields like radiology have been key drivers behind the growing favour towards deep learning algorithms for cervical cancer screening. Moreover, the superior accuracy and robustness of deep learning models compared to traditional machine learning methods have made them the preferred choice for researchers working in this domain.
Traditional cervical cell screening depends on pathologist experience, leading to inefficiencies and low accuracy due to the similarity between cancerous and normal cells. Recent advancements include novel human papilloma virus (HPV) detection methods and the use of nanotechnology, biosensors, and deep learning techniques to enhance the analysis of Pap smear images and improve cervical cancer detection.
Despite progress in artificial intelligence (AI) healthcare, its use in cervical cancer remains limited. Our research focuses on developing and validating a portable AI-based cervical cancer screening system. We aim to address two key questions: Which AI algorithms are effective for detecting cervical cancer using pelvic images? Which algorithms show the most promise for diagnosis? To explore these, we conducted a preliminary investigation to synthesize existing research on machine learning algorithms’ accuracy in detecting cervical cancer, paving the way for a potential systematic review.
Materials and methods
Search strategy
Search renowned datasets for studies using computerized or automated methods for cancer diagnosis, following the Arksey and O’Malley approach. The search strategy, developed with an information specialist and adhering to PRISMA criteria, covered publications from database inception to November 2023, with an update in July 2024.
Criteria for approval and rejection
Articles between 2007 to April 2024 have been included as part of the Calman-Hine report, a milestone for cancer management document. Table 2 contains a collection of all Approval and Exclusion Standards.
Table 2.
Approval and rejection standards
| Approval standards | Rejection standards |
|---|---|
| Studies on how cervical cancer affects patients’ life Research that incorporates cervical cancer patient images as a component of a broader study Knowledge of linguistics in English |
Professional and scientific papers on the optimal attributes for cervical cancer The study outlines the way pros are observed Viewpoints, judgments, and remarks |
Evaluation of studies
Two authors conducted a comprehensive assessment of search results, focusing on screening titles and abstracts to determine relevance. Duplicate entries were removed, and additional manual searches were conducted by examining reference lists. A member of the research team obtained full copies of potentially relevant studies and reviewed them independently. Any discrepancies regarding the inclusion of studies in the final evaluation were resolved through team discussion.
Table 1.
Population-Exposure Framework (PEO) investigation plans
| PEO element | Title/Focus | Descriptive terms | Improved pursuit terms |
|---|---|---|---|
| Population | Cervical cancer patients | Malignant cervical cancer images, patient conditions | Patients with cervical cancer, women undergoing diagnostic imaging (cervicography, colposcopy), cancer screening |
| Combined with AND | |||
| Exposure | Deep learning and AI in cervical cancer detection | Deep learning, machine learning, cervical cancer detection methodologies | Convolutional neural networks (CNNs), transfer learning, colposcopy image analysis, AI-enhanced diagnostics |
| Combined with AND | |||
| Outcome | Performance and diagnostic accuracy | Diagnostic skills, performance evaluation, outcome accuracy | Sensitivity, specificity, precision, F1 score, clinical applicability, AI in clinical workflow |
| Related concepts: clinical validation, interpretability of AI models, patient outcomes, diagnostic impact | |||
AI — artificial intelligence
Data extraction
Data was collected through research in various categories, including research objectives (whether focused on cervical cancer specifically or more broadly universal), research methods, sample size, tumor characteristics, objectives, and main results.
Quality appraisal
To evaluate academic quality, the Critical Appraisal Skills Programme (CASP) qualitative checklist was used by two independent authors, with discrepancies resolved through discussion [9]. Pilot studies were included if they reported results and impacts on patient care, with no restrictions on publication date or language. Studies on cervical cancer treatment were excluded due to differences in research settings, as were retrospective studies and those focusing only on patient symptom screening without physician involvement. Relevant articles were selected based on full-text references.
Evaluation, synthesis, assessment, and analytical review
After conducting the de-duplication process outlined in the Bramer method, studies were organized in EndNote and reviewed in Covidence for titles and abstracts, noting reasons for removal. Relevant citations were included in the full-text review. Key factors for data synthesis were extracted from the analyzed articles.
Figure 2 shows the PRISMA flow diagram, which visually represents the stages of study identification, screening, eligibility assessment, and inclusion in systematic reviews and meta-analyses.
Figure 2.
PRISMA flow diagram
Cervical cancer detection and diagnosis approaches
Ongoing screening is essential for cervical cancer prevention and therapy since early detection can improve treatment results significantly. Figure 2 depicts the steps for cervical screening and diagnosis.
Visual inspection with acetic acid (VIA) is a quick screening technique for cervical cancer that employs a 3–5% acetic acid solution to identify pre-cancerous cells. VIA is cost-effective and can be performed by mid-level providers in resource-limited settings, though it has lower sensitivity compared to cytology and colposcopy. Cytology involves pap smears or liquid-based tests to examine cervical cells, while HPV testing uses polymerase chain reaction (PCR) or hybridization techniques to detect HPV DNA. Confirmatory diagnosis of cervical intraepithelial neoplasia (CIN) involves colposcopy and biopsy, where a gynecologist or physician examines tissue for cancer cells. Recent progress in deep learning has notably advanced image analysis tasks, including semantic segmentation, classification, and object detection, leading to better screening and diagnostic outcomes [10].
Multiple deep learning-based approaches for detecting cervical cancer are currently available that have shown promising results in research studies. Some of the notable approaches include:
Utilizing convolutional neural networks (CNNs) to analyze cervical cell images obtained from Pap smears or colposcopy exams for the presence of abnormal cells or lesions [11].
Developing deep learning models that can accurately classify different stages of cervical cancer based on histopathological images [12].
Implementing automated image analysis algorithms to assist pathologists in identifying and diagnosing cervical cancer more efficiently [13].
Integrating deep learning techniques with other like magnetic resonance imaging (MRI) or ultrasound, to improve the accuracy of cervical cancer detection [14].
Figure 3.
Key procedures in cervical screening and diagnosis. HPV — human papilloma virus; PCR — polymerase chain reaction
These methods can greatly enhance the accuracy, efficiency, and cost-effectiveness of cervical cancer screening and diagnosis. However, it is essential to continue validating and refining these deep learning-based approaches through rigorous clinical trials and real-world implementation to ensure their effectiveness in clinical practice.
Deep learning in cervical cancer classification
Medical imaging, essential for diagnosis, is often hindered by individual interpretation and resource limitations, especially in low- and developing countries. The request for diagnostic images exceeds practitioner capacity in these regions. Artificial intelligence, particularly deep learning, offers a potential solution by improving computerized diagnosis [15]. Despite initial excitement, thorough studies are needed to address concerns about technology bias, research generality, and real-world applicability. Deep learning has shown exceptional speed and accuracy in medical imaging tasks such as segmentation, classification, and record-keeping. Its multilayer neural networks can uncover complex patterns and enhance diagnostic precision, potentially enabling less experienced healthcare professionals to achieve expert-level evaluations [16]. A major challenge in deep learning models for cervical cancer detection is their ability to generalize across different datasets. Models trained on specific datasets may perform poorly when applied to new populations or imaging conditions due to variations in imaging devices, population demographics, and dataset characteristics. Generalization issues are particularly acute in medical imaging, where high variability exists between datasets.
Various deep learning approaches for cervical cancer classification is shown in Figure 4.
Figure 4.
Various artificial intelligence (AI) algorithms utilized in medical studies
Self-supervision
Self-supervised learning is an advanced method that generates supervisory signals from unlabeled data, reducing the need for human annotation. This approach extracts semantic features from the data itself, which are then used for tasks with limited annotated data. It merges unsupervised learning’s ability to avoid manual labeling with supervised learning’s use of self-generated labels.
Research has shown that applying self-supervised learning to Pap smear images, which are abundant and unlabeled, can outperform standard pre-trained models. The Cervical Cell Copy-Pasting technique exemplifies this by enhancing performance beyond simple transfer methods and benefiting from integration with multiple instance learning techniques.
Unsupervised learning, which lacks direct supervision, relies on deriving signals without explicit guidance. Table 4 summarizes recent methods in unsupervised learning. Semi-supervised learning (SSL) is a key technique where models learn representations through auxiliary pretext tasks applied later to target tasks. SSL’s effectiveness depends on the design of these tasks, as they introduce biases. It includes four main types: predictive, generative, contrastive, and multi-self-supervision. A conceptual view of generative and contrastive learning in shown in Figure 5.
Table 4.
Advancements in using unsupervised techniques for cervical cancer classification
| Reference | Method | Task performed | Description | Dataset used | Result obtained |
|---|---|---|---|---|---|
| [37] | 3D CNN and ViT | Classification | 3D CNN: spatiotemporal features ViT: sequence modeling | Not specified | Accuracy: 98.6% |
| [14] | Exemplar pyramid deep feature extraction model | Image classification | Uses deep feature extraction based on exemplar pyramids | SIPaKMeD Pap-smear images Training: 70%; validation: 15%; testing: 15% | Accuracy: 92.95%, precision: 93.89%, recall: 92.71%, F1 score: 93.30% |
| [38] | FSOD-GAN | Object detection and classification | Uses GANs for detecting small cervical spots | Open-source cervical cancer datasets (Herlev database) | Accuracy: 99%, sensitivity: 76.92%, specificity: 90.91% |
| [39] | 3cDe-Net (dilated convolution ResNet and feature pyramid network) | Multiscale feature extraction and classification | Combines dilated convolutions and multiscale feature fusion | Tian-chi initiative and Herlev | Achieved superior performance with a MAP of 50.4% |
| [40] | Decision Tree with RFE and SMOTETomek | Risk prediction and classification | Uses decision trees with feature selection and class imbalance techniques | Patient medical records | Accuracy of 98.72% and sensitivity of 100% |
| [41] | Convolutional Neural Networks (CNN) for Histological Images | Image classification | Uses CNNs for classifying histological images of cervical cancer. | Histological images | Improved classification accuracy and AUC: 0.966 |
| [42] | SIPaKMeD and Herlev Dataset Classification | Feature and image classification | Uses a new dataset for classifying normal and pathological cervical cells. | SIPaKMeD and Herlev datasets | Enhanced recognition ability Accuracy: 99.81%, precision: 89.4% |
| [43] | Hybrid DL Pre-Trained Models with Fuzzy Min–Max NN | Diagnosis and classification | Combines deep learning models with fuzzy neural networks | SIPaKMeD and Herlev dataset | Accuracy of 95.33% is obtained using Resnet-50 |
| [44] | CNN with ELM | Classification | Integrates CNN with ELM for better performance | Herlev database | Accuracy: detection (2-class) — 99.5%, classification (7-class) — 91.2% |
| [11] | Ensemble Deep Learning Network | Classification | Combines multiple deep learning models for better accuracy | Colposcopy images | Sensitivity: 92.4%, specificity: 962%, Kappa scores: 88% |
| [45] | RFE and LASSO | Feature selection and classification | Uses RFE and LASSO for feature selection in risk prediction | Cervical cancer risk datasets | Accuracy: RFE — 98.0% and LASSO — 96.0% |
| [39] | Multiscale Feature Fusion Network and Channel-Wise Cross-Attention | Feature extraction and classification | Uses channel-wise cross-attention and feature fusion for better feature extraction | SIPaKMeD, Herlev and Motic dataset | Higher MAP scores compared to baseline methods |
| [46] | Deep Learning with Depthwise Separable Convolutions | Image recognition and classification | Uses shuffle net for efficient deep learning | Cervigram dataset | Accuracy: 81.38% AUC: 0.99 |
| [47] | Generative Adversarial Networks (GANs) for Cervical Cell Detection | Detection and classification | Uses GANs for detecting cervical cells in images | Cervical cell images | Accuracy of almost 97% and a loss of less than 1% |
| [48] | Hybrid Convolutional Neural Network and Transfer Learning | Image classification | Combines CNN with transfer learning for better performance | Pap smear images | Accuracy of 91.46% |
| [49] | Multiscale Feature Extraction and Classification Network | Feature extraction and classification | Uses multiscale feature extraction techniques for better accuracy | SIPaKMeD dataset | Accuracy of 98.08% |
| [50] | Deep Learning-for ADSC | Screening and classification | Automates dual stain cytology using deep learning | Cervical screening data | Enhanced accuracy and efficiency in screening |
CNN — convolutional neural networks;
Figure 5.
Conceptual view of generative and contrastive learning [17]
Table 3.
Frequently utilized datasets for achieving data-efficient deep learning in cervical cancer analysis
| Dataset | Dataset type | Task performed | Description | Access link |
|---|---|---|---|---|
| UCI Machine Learning Repository [27] | Tabular (clinical data) | Classification (predicting cancer presence based on clinical features) | Contains clinical and demographic data related to cervical cancer, including attributes like age, number of sexual partners, and number of pregnancies | https://archive.ics.uci.edu/ml/datasets/Cervical |
| Cervix Cell Dataset (Kaggle) [28] | Image data (cell images) | Image classification (classifying images of cervix cells into different categories) | Includes images of cervix cells that are useful for training deep learning models for cell classification and anomaly detection | https://www.kaggle.com/datasets/prahladmehandiratta/cervical-cancer-largest-dataset-sipakmed |
| The Cancer Imaging Archive (TCIA) [29] | Medical imaging data (MRI, CT scans) | Image analysis (segmentation and classification of cervical cancer in medical imaging) | Provides imaging data including MRI and CT scans relevant for cervical cancer research | https://www.cancerimagingarchive.net/collection/tcga-hnsc/#citations |
| SEER Cancer Statistics [30] | Statistical data (cancer incidence and survival data) | Statistical analysis (analyzing cancer incidence, survival rates, and trends) | Provides a comprehensive range of cancer-related statistics, including detailed data on cervical cancer incidence and survival | https://seer.cancer.gov/statistics/ |
| Cervical Cancer Screening Dataset [31] | Tabular (Screening Test Data) | Classification (predicting the risk of cervical cancer based on screening test results) | Contains data related to cervical cancer screening, including results from various tests and demographic information | https://doi.org/10.1016/j.sxmr.2019.09.005 |
| Pap Smear Data [32] | Tabular (Pap smear test results) | Classification (classifying Pap smear test results to detect cervical cancer) | Includes information from Pap smear tests used for diagnosing cervical cancer, with attributes related to cell morphology | https://doi.org/10.24432/C5HS5C |
| Cervical Cancer Dataset (Kaggle Competitions) [33] | Tabular and/or image data | Various (depending on competition objectives; often classification or prediction tasks) | Data used in Kaggle competitions to develop predictive models for cervical cancer. The exact nature of the dataset can vary between competitions. | https://www.kaggle.com/datasets/ranzeet013/cervical-cancer-dataset |
| Genomic Data Commons (GDC) Cervical Cancer Dataset [34] | Genomic data | Genomic analysis (studying genetic mutations and expression related to cervical cancer) | Offers genomic and clinical data relevant to cervical cancer, useful for studying genetic factors and molecular mechanisms. | https://portal.gdc.cancer.gov |
| NCI Genomic Data [35] | Genomic data | Genomic and molecular analysis (analyzing genetic and molecular data across various cancers including cervical cancer) | Provides access to a broad range of genomic data for various cancers, including cervical cancer, for research and analysis | https://gdc.cancer.gov |
| Papanicolaou Smear Dataset (UCI) [36] | Tabular (Pap smear test data) | Classification (classifying Pap smear test results to detect cervical cancer) | Contains data from Pap smear tests used for detecting cervical cancer, including attributes related to cell characteristics | https://doi.org/10.1016/j.dib.2020.105589 |
MRI — magnetic resonance imaging; CT — computed tomography
In SSL, the pretext task trains the model on unlabeled data by generating self-created labels, allowing the model to learn representations. These representations are then used as initial weights for the downstream task, which fine-tunes them for specific goals. This workflow is illustrated in Figure 6.
Figure 6.
Overview of the semi-supervised learning (SSL) process. Top: in SSL, an auxiliary task is trained using pseudo-labels derived from a large, unlabeled dataset. Bottom: the features learned from this initial task are then applied to the main task, enabling training with a limited set of data containing true labels. CNN — convolutional neural networks
As the field evolves, the choice and complexity of pretext tasks become increasingly important, as they significantly influence the quality of the learned features. Tasks such as predicting rotations, filling in missing parts of an image, or distinguishing between augmented versions of the same data help the model capture various aspects of the data’s structure. In downstream tasks, these pre-learned features can greatly reduce the need for large annotated datasets, making SSL a powerful tool in areas where labeled data is scarce or expensive to obtain. Researchers continue to develop more sophisticated pretext tasks and refine the transfer process of learned representations to improve the performance of downstream applications.
Predictive self-supervision
In this a pretext task is framed as an identification or analysis problem. Common predictive tasks include relative position calculation, jigsaw completion, and rotational orientation identification. These methods are beneficial in medical image analysis (MIA).
For example, one method for cardiac magnetic resonance imaging (MRI) segmentation uses a pretext task to predict anatomical positions, employing various cardiac perspectives like short-axis and long-axis views. A similar approach for cervical cancer could involve segmenting cervical images to predict anatomical positions, using different views such as colposcopic images and Pap smear slides. This could improve the detection and diagnosis of cervical abnormalities by training networks to predict specific anatomical landmarks.
Contrastive learning
Contrastive learning has shown promising results in improving cervical cancer classification. By leveraging large amounts of unlabeled cervical images, contrastive learning trains a model to distinguish between similar and dissimilar image pairs. This approach enhances the model’s ability to learn robust and discriminative features, leading to better performance in identifying cervical abnormalities.
In contrastive learning as shown in Figure 7, input images are augmented and classified into positive or negative pairs based on labels. A pre-trained CNN encoder extracts features, which are projected into a space to enhance similarity in positive pairs and reduce it in negative ones. This process helps in accurate classification by improving feature representation.
Figure 7.
Steps for the execution of contrastive learning
Generative self-supervision
Generative self-supervision uses generative models, like variational auto-encoder (VAEs) and generative adversarial networks (GANs), to enhance cervical cancer classification by creating meaningful data representations. These models utilize large volumes of unlabeled images to improve accuracy and robustness. VAEs help in learning latent representations, while GANs augment datasets and detect anomalies. Techniques include generating synthetic cervical images to increase data diversity and identifying anomalies by comparing original and reconstructed images. A novel approach as shown in Figure 8 involves swapping patches in an image and using a generative model to restore it, aiding in early detection of cervical abnormalities [18].
Figure 8.
In the convolutional neural networks (CNN) computational structure for self-supervised reconstruction studying, blue, green, and orange represent layered elements, downsampling elements, and upsampling elements, respectively. The CNN’s structure throughout the reconstruction stage can be modified depending on the next assignment [19]
Inexact supervision
Inexact supervision involves using weak or imprecise labels for training models, which is particularly relevant in medical fields like cervical cancer detection, where precise annotations are challenging. This approach reduces the annotation burden by employing partially labeled data or heuristic methods. Techniques like Multiple Instance Learning (MIL), which labels groups of instances rather than individuals, and SSL, which uses both labeled and unlabeled data, are effective in handling the ambiguity in medical images. These methods improve the scalability and performance of cervical cancer diagnostic systems, but ensuring model reliability under inexact supervision remains a challenge [20].
Multiple instance learning (MIL) simplifies the labeling process by assigning labels to groups of instances, making it suitable for heterogeneous medical images where precise annotation is difficult. MIL has improved accuracy in cervical cancer screening by managing variability in cytological images. Strategies include MIL, which uses broad labels for sets of instances, and transfer learning, where models are fine-tuned on specific medical tasks, also helps in developing scalable and cost-effective screening tools, enhancing early detection despite annotation challenges [21].
Incomplete supervision
Incomplete supervision involves limited labeled data and abundant unlabeled data, and is divided into semi-supervised learning, active learning, and domain-adaptive learning. SSL improves model performance by using both labeled and unlabeled data, with techniques like consistency regularization, pseudo-labeling, and generative models to enhance cervical cancer detection.
Active learning, a type of incomplete supervision, improves labeling efficiency by focusing on the most informative samples, especially in resource-intensive tasks like cervical cancer detection (Fig. 9). It selects key samples from large unlabeled datasets for expert annotation, using strategies like uncertainty sampling, where the model queries labels for uncertain samples, and query-by-committee, where models select samples with high disagreement. This approach reduces annotation effort while improving model accuracy and generalizability.
Figure 9.
This map illustrates the abstract approach for self-teaching in semi-supervised learning and active learning. In self-training, a classifier is first trained on a small labeled set and then updated with additional pseudo-labeled data. Active learning involves retraining the segmentation model using original labeled data and new labels obtained through human interaction [22]. A. Self-training; B. Active learning
Domain adaptive learning (DAL) enhances cervical cancer detection by addressing domain shifts between labeled and unlabeled data, which can degrade model performance due to differences in data sources and conditions. DAL techniques and correlation alignment (CORAL), align feature distributions across domains to reduce discrepancies. Self-ensembling methods refine a model’s predictions on unlabeled data to adapt to new domains [23]. These strategies improve the robustness and generalizability of cervical cancer screening models, making them more effective in diverse clinical settings and aiding early detection. A typical DAL framework that involves adapting a pre-trained segmentation model to a new target domain, using a semantic discriminator to maintain consistent image-to-label mappings and ensuring realistic segmentations is shown in Figure 10.
Figure 10.
The designed optimal unsupervised context adaption system is made up of a segmentation framework and an intuitive detector. This system uses adversarial learning to correlate the statistical representations of the source and domains of concern. The segmentation systems for both regions have the same weights, with the initial domain model trained supervised and the intended domain model trained adversarially. The linguistic detector learns disease-specific spatial patterns explicitly by processing primary images, lexical masks, edge maps, and inverted maps [24]
Inaccurate supervision
Inaccurate supervision occurs when discrepancies in label information arise due to human error, inter-observer variability. This can harm deep neural networks’ ability to generalize, making it crucial to develop strategies to mitigate these effects, especially in medical image analysis, where precision is vital. Modern methods to address faulty supervision fall into three categories: robust loss design, data re-weighting, and training procedures.
Robust loss design involves creating loss functions that reduce the impact of mislabeled data during training. Techniques like using noise-tolerant loss functions, such as mean absolute error (MAE) over mean squared error (MSE), or loss correction approaches that adjust for label noise are common. For cervical cancer detection, enhanced loss functions, like an adapted dice loss or adaptive cross entropy (ACE), improve model accuracy despite noisy labels, aiding in the correct identification of cancerous cells [25].
Data re-weighting helps minimize the effects of noisy labels by assigning weights to training samples based on their likelihood of being correct. This technique uses validation sets or iterative updates to adjust sample weights, prioritizing reliable data. In cervical cancer screening, data re-weighting ensures more robust models, improving automated detection and patient outcomes by reducing the impact of incorrect labels [26].
Limited supervision
Limited supervision occurs when only a small amount of labeled data is available, a common challenge in medical imaging due to the high cost and expertise required for annotation. To overcome this, several strategies have been developed to make efficient use of both labeled and unlabeled data. Semi-supervised learning leverages the combined datasets, with techniques like pseudo-labeling, where model predictions on unlabeled data serve as additional labels, and transfer learning, which pre-trains models on large datasets and fine-tunes them on specific tasks like cervical cancer detection. Data augmentation further addresses the challenge of limited annotations by artificially increasing the size and diversity of the training set. This approach is particularly valuable in cervical cancer screening, where the cost of acquiring labeled medical images is high.
Transfer learning plays a crucial role in enhancing cervical cancer detection by using pre-trained models from large datasets. Since obtaining extensive labeled data in medical imaging is often difficult, this method allows models to utilize previously learned features and representations, improving accuracy even with limited annotated samples. Pre-trained models, such as those trained on ImageNet, can be fine-tuned for cervical cancer detection, enhancing the detection of precancerous lesions and other abnormalities [21].
Few-shot learning is another promising technique that addresses the scarcity of labeled data by enabling models to learn from only a few labeled examples. Techniques like meta-learning, which helps models quickly adapt to new tasks, and metric learning, which allows models to classify new examples based on their similarity to reference samples, have shown significant potential in improving cervical cancer detection with minimal data. Together, these strategies enhance early diagnosis and improve patient outcomes.
Results
Numerous works highlight significant advancements in the use of DL algorithms for the accurate diagnosis of various cancers, particularly in analyzing cervical cancer images, where these algorithms have achieved high precision rates. Central to these advancements are several frequently utilized datasets that enable data-efficient deep learning for cervical cancer analysis. Other datasets, such as the Cervix Cell Dataset from Kaggle [28], aid in classifying cervix cells into different categories, essential for training DL models for cell classification and anomaly detection. The Pap Smear Data [32] includes results from Pap smear tests critical for detecting cervical cancer, detailing attributes related to cell morphology. Lastly, the Cervical Cancer Dataset used in Kaggle Competitions [33] encompasses both tabular and image data for various tasks, focusing on classification or prediction based on competition objectives. Together, these datasets are invaluable for advancing cervical cancer research through deep learning techniques.
Numerous innovative models have emerged that leverage these datasets. For instance, Jia et al. [66] developed a Strong Feature CNN-SVM network that achieved an accuracy of 99.3% using the Herlev dataset. Similarly, Rahaman et al. [67] introduced the Deep Cervix Framework, which attained high accuracies on the SIPaKMeD and Herlev datasets. Cao et al. [68] utilized an Attention-Guided Convolutional Network, achieving notable sensitivity and specificity in detecting abnormal cells. Furthermore, Pacal et al. [14] proposed the Exemplar Pyramid Deep Feature Extraction Model, which also demonstrated strong performance metrics. Tan et al. [38] advanced the field further with the FSOD-GAN, achieving impressive accuracy and specificity in detecting small cervical spots.
Recent advancements in inexact supervised techniques for cervical cancer classification have yielded significant results. Ali et al [53] employed an ensemble learning approach using SHapley Additive exPlanations (SHAP) and achieved accuracy rates of 98.06% and 95.45% on two datasets. Wang et al. [69] introduced a Multi-Task Learning AI model that effectively predicts various clinical features of cervical cancer.
Chen and his team developed a compact CNN framework designed for use in embedded devices, emphasizing minimal parameters and low processing needs. Their study focuses on how information distillation can enhance the performance of this small-scale CNN model [70]. This framework integrates a CNN module to extract local features, a vision transformer to capture global attributes, and a multi-layer perceptron to combine these elements for precise classification. The goal of this system is to quickly and accurately evaluate pap images. Figure 11 provides a comparison of the results from these new deep learning approaches with those from earlier studies on similar topics.
Figure 11.
A. Illustration of number of articles published in a particular machine learning technique; B. Illustration of number of articles published year wise rages from 2008 to 2024
Evaluation metrics
When assessing how well a deep learning model performs in classifying cervical cancer, several key metrics are used to measure its effectiveness. These metrics include accuracy, precision, recall, and overall classification performance, as detailed in Table 6.
Table 6.
Comparative analysis of the evaluation metrics
| Parameter | Mathematical equation | Remarks |
|---|---|---|
| Dice [71] | It is a metric used to assess the similarity between two sets | |
| Accuracy [72] | Accuracy evaluates the percentage of samples that are classified correctly | |
| Recall [71] | Recall assesses the percentage of actual positive cases that were correctly identified by the model | |
| Precision [72] | Precision assesses the fraction of true positive predictions compared to the overall number of positive predictions generated by the model | |
| Specificity [71] | Specificity assesses the fraction of true negatives out of all negative cases | |
| Score [71] | Score is the average of dice and recall | |
| F1-Score [72] | It is the harmonic mean of precision and recall, which is viewed as a more effective performance indicator compared to the standard accuracy measure | |
| Kappa score [73] | (Observed agreement – Expected agreement)/(1 – Expected agreement) | Cohen’s Kappa, also known as the Kappa Statistic, is a metric used to assess the agreement between raters for categorical data |
| Area under the receiver operating characteristic curve (AUC-ROC) [74] | AUC is a measure of segregation, with values close to 1 indicating an algorithm’s ability to accurately classify patients with and without cancer | |
| Mathew correlation coefficient (MCC) [75] | It varies between +1 to −1, with +1 indicating perfect agreement and 0 showing no better than random predictions | |
| Confusion Matrix [76] | – | It serves as a valuable method for assessing the model’s accuracy across various classes |
Discussion
This review examined research articles covering a period of 20 years (2005–2024) that focused on the application of machine learning, including deep learning algorithms, for image classification in cervical cancer screening. The findings suggest a growing favouritism towards deep learning algorithms over time. The majority of these studies appeared in non-clinical publications, emphasizing the importance of sharing these algorithms in healthcare journals to enhance AI’s role as a diagnostic tool in clinical research.
The integration of deep learning algorithms into cervical cancer diagnosis has marked a transformative phase in oncological imaging, significantly enhancing detection and classification accuracy. Notably, recent architectures, like the 3D CNN and Vision Transformer, achieved an impressive accuracy of 0.986, reflecting a trend towards more sophisticated models that leverage advanced feature extraction techniques. These architectures are well-suited for processing the high-dimensional and complex data typical of medical imaging. The emergence of frameworks such as the CVM-Cervix, which combines CNNs with vision transformers, signifies a shift towards hybrid approaches that capitalize on both local and global feature representations. This adaptability may lead to more nuanced assessments of cervical cytology images, enabling quicker and more accurate evaluations of Pap smear results.
While these advancements are encouraging, challenges persist, including the need for larger, annotated datasets and the management of inconsistencies in medical imaging. For example, models that performed well on specific datasets may not generalize effectively across different populations or imaging conditions. Furthermore, the reliance on high-quality annotations remains a barrier, as many datasets are limited in their breadth and depth.
Table 5.
Advancements in using inexact supervised techniques for cervical cancer classification
| Reference | Method | Task performed | Algorithm | Description | Dataset used | Result obtained |
|---|---|---|---|---|---|---|
| [37] | 3D CNN and Vision Transformer (ViT) | Classification | 3D CNN, ViT | Combines spatiotemporal features from 3D CNN with global context from ViT | SIPaKMeD dataset | Accuracy of 98.6% |
| (51) | Deep Learning with Coarse Labels | Classification with sparse labels | Two-stage deep learning | Efficient training with coarse labels on gigapixel images | WSIs | High accuracy despite sparse annotations |
| [52] | Feature Fusion with ShuffleNet | Classification | Deep learning, ShuffleNet | Combines deep learning with ShuffleNet for enhanced feature extraction | Pap smear images, SIPaKMeD datasets | Classification accuracy: 99.1% |
| [53] | Ensemble Learning-Based Classification using behavioral risk factors | Classification | Ensemble learning, SHapley Additive exPlanations (SHAP) | Utilize a five-fold cross-validation technique | Cervical cytology images | Accuracy: 98.06% for Dataset 1and 95.45% for Dataset 2 AUC: 0.95 for Dataset 1 and 0.97 for Dataset 2 |
| [54] | Deep Learning with Generative Adversarial Networks (GANs) | Small object detection | GANs | Detects small objects in cervical images | Custom dataset | High detection accuracy and AUC value is 0.984 |
| [55] | CNNs with RELMs | Classification | CNNs, RELMs | Combines CNNs for feature extraction with RELMs for classification | Cervical cytology images | High classification accuracy |
| [56] | EfficientNet-B0 and GRU- | Colposcopy diagnosis | EfficientNet-B0, GRU | Combines EfficientNet-B0 for feature extraction and GRU for sequence modeling | Colposcopy images | Accuracy: 91.18%, sensitivity: 93.6%, specificity: 87.6% |
| [57] | Ensemble of FDBD Model | Identification | Fuzzy distance metrics, deep learning | Ensemble of deep models using fuzzy distance metrics | PaP smear images | Performance of 96.96% |
| [58] | Improving Classification with Imbalanced Data Handling Techniques | Classification | Various imbalance handling techniques | Methods to handle data imbalance in cervical cancer classification | LBCPS dataset, SIPAKMeD, and Herlev dataset | Accuracy: LBCPS dataset (4-class) — 98.79%, SIPAKMeD (5-class) — 99.58%, Herlev (7-class) — 99.88%, IS: 3.75, FID: 0.71, Recall: 0.32, Precision: 0.65 |
| [8] | Cervical Transformation Zone Segmentation | Segmentation and classification | Multi-scale feature fusion framework | 2-phase method that includes segmenting and augmentation | Colposcopy images | Accuracy: 81.24%, sensitivity: 81.24%, specificity: 90.62%, precision: 87.52%, FPR: 9.38%, F1 score: 81.68%, MCC: 75.27%, Kappa coefficient: 57.79% |
| [59] | Self-Supervised Learning Approaches | Feature extraction | ConvNeXt | Leverages unlabeled data for feature extraction. | Deep Cervical Cytological Levels (DCCL) dataset | Classification accuracy 8.85% higher than that of previous advanced models |
| [60] | Weakly Supervised Learning with Partial Annotations | Detection | Weakly supervised learning (LD-WSCCD) | Weakly supervised model for cervical cell detection, based on a local distillation mechanism | Publicly accessible cervical cell dataset | mAP value of 73.6% |
| [61] | Graph Neural Networks for Cytology Image Analysis | Image analysis | Graph Convolutional Network (GCN) | Examined how cervical cell images might improve classification accuracy | SIPaKMeD and Motic liquid-based cytology image dataset | Better classification performance |
| [62] | Meta-Learning for Adaptation to New Data | Adaptation | Optimized meta-learning (OML) algorithm | Quickly adapts models to new data | Custom dataset | Accuracy: 99.39% |
| [63] | Semi-Supervised Learning Using Unlabeled Data | Classification | SSL | Uses labeled and unlabeled data | Tianchi dataset and the Comparison Detector dataset | High classification accuracy |
| [64] | Federated Learning for Distributed Data | Classification | Federated learning | Uses distributed data without centralizing it | UCI cervical cancer risk factors data | Accuracy: 99.26%, misprediction rate: 0.74% |
| [65] | Comparison detector for cervical cell/clumps detection | Detector | Faster R-CNN with FPN for proposal-based object detection | Proposed a dataset for comparison to tackle limited data issues | Comparison Detector dataset | mAP: 48.8% AR: 64% |
CNN — convolutional neural networks
Strategies such as data augmentation and transfer learning have shown promise in mitigating these issues, facilitating the development of resilient models capable of generalizing better to unseen data. The exploration of inexact supervised techniques and ensemble learning methods further illustrates the ongoing innovation within this domain, pushing the boundaries of what is possible in cervical cancer detection.
Conclusion and future scope
This review has highlighted the critical role of diverse datasets in facilitating these developments and underscored the importance of selecting appropriate models that integrate both local and global features for optimal performance. As illustrated by the comparative performance metrics, various innovative architectures, including CNNs and hybrid models, have achieved impressive results, indicating a promising direction for future research.
Despite these advancements, challenges such as limited annotated datasets and the need for more robust models remain. Addressing these issues through strategies like data augmentation, transfer learning, and the incorporation of explainable AI techniques will be essential for further improving diagnostic accuracy and generalizability across diverse populations.
Ultimately, ongoing research in this field must continue to prioritize the translation of these technological advancements into clinical practice, ensuring that deep learning approaches not only enhance diagnostic capabilities but also lead to better patient outcomes in cervical cancer care. By synthesizing current findings and exploring new methodologies, we can pave the way for more effective and accessible diagnostic solutions in the fight against cervical cancer.
Table 7.
Comparative analysis of the performance metrics
| Model | No. of images | Precision | Accuracy | Recall | Dice | Specificity | F1-Score | AUC |
|---|---|---|---|---|---|---|---|---|
| Convolutional Neural Network [49] | Training — 7,498 images, validation — 1,884, and testing — 1,884 images | 0.924 | – | – | – | 0.962 | – | – |
| RCNN [77] | 906 | 0.7387 | 0.9291 | 0.7912 | 0.7431 | 0.9589 | – | – |
| Gradient boosting classifier [78] | 199 patients images | 0.80 | 0.938 | – | – | 0.964 | 0.85 | |
| IDCNN-CDC [79] | – | 0.9784 | 0.9796 | 0.9820 | – | – | 0.9781 | – |
| GBT [79] | – | 0.9618 | 0.9565 | 0.9699 | – | – | 0.9636 | – |
| XGBoost [79] | – | 0.9421 | 0.9515 | 0.9663 | – | – | 0.9577 | – |
| ELM [79] | – | 0.9367 | 0.9407 | 0.9599 | – | – | 0.9520 | – |
| DCNN [80] | 488 images: 117 cancer patients. 509 images: 181 non-cancer patients | 0.883 | 0.908 | – | – | 0.933 | – | 0.932 |
| DCNN [81] | 13775 images | 0.994 | – | – | – | 0.348 | – | 0.67 |
| ResNet and combining ResNet with clinical features [82] | 15,276 | ResNet + clinical features: 0.932 Only ResNet: 0.901 | ResNet + clinical features: 0.886 Only ResNet: 0.882 | – | – | ResNet + clinical features: 0.846 Only ResNet: 0.867 | – | ResNet + clinical features: 0.953 Only ResNet: 0.945 |
| Compare KNN, SVM, and DT [83] | 2000 | 0.989 | 0.993 | – | – | 0.994 | – | – |
| Assessment of VIA, VILI, and GIVI contrast combinations [84] | 1426 | Resnet-v2: 0.833 Resnet-152: 0.852 |
Resnet-v2: 0.877 Resnet−152: 0.878 |
– | – | Resnet-v2: 0.886 Resnet−152: 0.882 |
– | Resnet-v2: 0.932 Resnet-152: 0.947 |
| CNN: AlexNet and VGG-16 with modified filters [48] | 4753 | 0.9822 | 0.9613 | – | – | 0.9509 | – | – |
| Convolutional Neural Network CAIADS [85] | 240 | KNN: 0.75 SVM: 0.727 DT: 0.636 |
KNN: 0.783 SVM: 0.742 DT: 0.758 |
– | – | KNN: 0.803 SVM: 0.75 DT: 0.829 |
– | KNN: 0.807 SVM: 0.744 DT: 0.767 |
| SVM based on radiomics and deep learning [73] | 4753 | 0.9568 | 0.9687 | – | – | 0.9872 | – | – |
| CNN-SVM [86] | 5679 | VGG-19: 0.33 CYENET: 0.924 |
VGG-19: 0.733 CYENET: 0.923 |
– | – | VGG-19: 0.0.79 CYENET: 0.962 |
– | – |
| Convolutional Neural Network [87] | 253 | 0.9560 | 0.9410 | – | – | 0.8330 | – | 0.963 |
| Convolutional Neural Network [88] | Training — 111 (44 + LVSI and 67 — LVSI) Validation — 56 (23 + LVSI and 33 – LVSI) | Training cohort: 0.773 Validation cohort: 0.739 |
– | – | – | Training cohort: 0.776 Validation cohort: 0.667 |
– | Training cohort: 0.842 Validation cohort: 0.775 |
| CCNet [89] | – | 0.7264 | 0.9191 | 0.7179 | 0.6849 | 0.9560 | – | – |
| DarkNet53 [90] | – | 0.8541 | 0.8477 | 0.8401 | – | – | 0.8470 | – |
| SVM using VIA and VILI [91] | 200 | 0.813 | 0.80 | – | – | 0.786 | – | – |
| SVM and DT [92] | 102 | For SVM: 0.981 For DT: 0.95 |
For SVM: 0.9833 For DT: 0.97 |
– | – | For SVM: 0.985 For DT: 0.9867 |
– | – |
| Xception [93] | – | 0.8678 | 0.8511 | 0.8695 | – | – | 0.8686 | – |
| ResNeXt50 [94] | – | 0.8908 | 0.8633 | 0.8658 | – | – | 0.8781 | – |
| SegNet [95] | – | 0.6867 | 0.9097 | 0.7057 | 0.6600 | 0.9517 | – | – |
| DeepLabV3+ [96] | – | 0.6889 | 0.9083 | 0.6828 | 0.6416 | 0.9545 | – | – |
| ResNet18 [97] | – | 0.8637 | 0.8254 | 0.8287 | – | – | 0.8458 | – |
| FCN8x [71] | – | 0.7102 | 0.9094 | 0.6434 | 0.6311 | 0.9522 | – | – |
| UNet [98] | – | 0.6941 | 0.9073 | 0.6593 | 0.6307 | 0.9575 | – | – |
| Multilayer perceptron Networks [99] | 170 | 0.6978 | – | – | – | 0.68 | – | 0.73 |
AUC — ???; KNN — ???; SVM — ???; DT — ???; VGG-19 — ???; CYENET — ???
Footnotes
Author contributions: P.P.: conceptualization and research design, literature review and theoretical framework, data collection and analysis, manuscript drafting and final revisions; D.V.: methodology developement and implementation, data interpretation and validation, formatting, editing, and proofreading.
Conflict of interests: The authors declare no conflict of interests.
Ethical approval: Not required.
Funding: None declared.
References
- 1.Bray F, Laversanne M, Sung H, et al. Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. 2024;74(3):229–263. doi: 10.3322/caac.21834. [DOI] [PubMed] [Google Scholar]
- 2.Singh P, Singh S, Mishra A, et al. Multimodality treatment planning using the Markov decision process: a comprehensive study of applications and challenges. Res Biomed Eng. 2024;40(2):435–450. doi: 10.1007/s42600-024-00349-4. [DOI] [Google Scholar]
- 3.Singh P, Tripathi S, Tamrakar R. Fluence map optimisation for prostate cancer intensity modulated radiotherapy planning using iterative solution method. Pol J Med Phys Eng. 2020;26(4):201–209. doi: 10.2478/pjmpe-2020-0024. [DOI] [Google Scholar]
- 4.Singh P, Tripathi S, Gupta S. A unified approach for optimal dose delivery and trajectory optimization for the treatment of prostate cancer. Biomed Signal Process Control. 2021;69:102884. doi: 10.1016/j.bspc.2021.102884. [DOI] [Google Scholar]
- 5.Basera A, Hull R, Demetriou D, et al. Cervical cancer in low and middle-income countries. Oncol Lett. 2020;20(3):2058–2074. doi: 10.3892/ol.2020.11754. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Singh P, Mishra A, Mishra S. A comprehensive analysis of the challenges and potential side effects of radiation therapy for palliative cancer treatment. Méd Palliat. 2024;23(2):75–91. doi: 10.1016/j.medpal.2023.12.002. [DOI] [Google Scholar]
- 7.Levine AB, Schlosser C, Grewal J, et al. Rise of the Machines: Advances in Deep Learning for Cancer Diagnosis. Trends Cancer. 2019;5(3):157–169. doi: 10.1016/j.trecan.2019.02.002. [DOI] [PubMed] [Google Scholar]
- 8.Dash S, Sethy PK, Behera SK. Cervical Transformation Zone Segmentation and Classification based on Improved Inception-ResNet-V2 Using Colposcopy Images. Cancer Inform. 2023;22:11769351231161477. doi: 10.1177/11769351231161477. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Singh P, Singh S, Mishra A, et al. Multimodality treatment planning using the Markov decision process: a comprehensive study of applications and challenges. Res Biomed Eng. 2024;40(2):435–450. doi: 10.1007/s42600-024-00349-4. [DOI] [Google Scholar]
- 10.Singh P, Tripathi S, Tamrakar R, et al. Dose-Volume Constraints Based Inverse Treatment Planning For Optimizing the Delivery of Radiation Therapy. Gedrag Organ Rev. 2020;33(03) doi: 10.37896/gor33.03/489. [DOI] [Google Scholar]
- 11.Chandran V, Sumithra MG, Karthick A, et al. Diagnosis of Cervical Cancer based on Ensemble Deep Learning Network using Colposcopy Images. Biomed Res Int. 2021;2021:5584004. doi: 10.1155/2021/5584004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Alsalatie M, Alquran H, Mustafa WA, et al. Analysis of Cytology Pap Smear Images Based on Ensemble Deep Learning Approach. Diagnostics (Basel) 2022;12(11) doi: 10.3390/diagnostics12112756. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Li C, Xue D, Kong F, Hu Z, Chen H, Yao Y. Information Technology in Biomedicine. Springer; 2019. Cervical histoathology image classification using ensembled transfer learning; pp. 26–37. [Google Scholar]
- 14.Pacal I, Kılıcarslan S. Deep learning-based approaches for robust classification of cervical cancer. Neural Comput Appl. 2023;35(25):18813–18828. doi: 10.1007/s00521-023-08757-w. [DOI] [Google Scholar]
- 15.Mishra A, Bhusnur S, Mishra SK, et al. Exploring a new frontier in cardiac diagnosis: ECG analysis enhanced by machine learning and parametric quartic spline modeling. J Electrocardiol. 2024;85:19–24. doi: 10.1016/j.jelectrocard.2024.05.086. [DOI] [PubMed] [Google Scholar]
- 16.Mishra A, Bhusnur S, Mishra SK, et al. Comparative analysis of parametric B-spline and Hermite cubic spline based methods for accurate ECG signal modeling. J Electrocardiol. 2024;86:153783. doi: 10.1016/j.jelectrocard.2024.153783. [DOI] [PubMed] [Google Scholar]
- 17.Liu X, Zhang F, Hou Z, et al. Self-supervised Learning: Generative or Contrastive. IEEE Trans Knowl Data Eng. 2021:1–1. doi: 10.1109/tkde.2021.3090866. [DOI] [Google Scholar]
- 18.Schlegl T, Seeböck P, Waldstein S, et al. Unsupervised Anomaly Detection with Generative Adversarial Networks to Guide Marker Discovery. Lect Notes Comp Sci. 2017:146–157. doi: 10.1007/978-3-319-59050-9_12. [DOI] [Google Scholar]
- 19.Chen L, Bentley P, Mori K, et al. Self-supervised learning for medical image analysis using image context restoration. Med Image Anal. 2019;58:101539. doi: 10.1016/j.media.2019.101539. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Freitas NP, da SM. Classification of Cervical Cancer Risk Using Deep Learning. 2022 https://repositorio-aberto.up.pt/bitstream/10216/143003/2/573067.pdf . [Google Scholar]
- 21.Tajbakhsh N, Shin JY, Gurudu SR, et al. Convolutional Neural Networks for Medical Image Analysis: Full Training or Fine Tuning? IEEE Trans Med Imaging. 2016;35(5):1299–1312. doi: 10.1109/TMI.2016.2535302. [DOI] [PubMed] [Google Scholar]
- 22.Peng J, Wang Ye. Medical image segmentation with limited supervision: a review of deep network models. IEEE Access. 2021;9:36827–36851. doi: 10.1109/access.2021.3062380. [DOI] [Google Scholar]
- 23.Perone C, Cohen-Adad J. Deep Semi-supervised Segmentation with Weight-Averaged Consistency Targets. Lect Notes Comp Sci. 2018:12–19. doi: 10.1007/978-3-030-00889-5_2. [DOI] [Google Scholar]
- 24.Li H, Loehr T, Sekuboyina A, et al. Domain adaptive medical image segmentation via adversarial learning of disease-specific spatial patterns. arXiv Prepr. 2020 arXiv200109313. [Google Scholar]
- 25.Matuszewski D, Sintorn IM. Minimal annotation training for segmentation of microscopy images. 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018) 2018:387–390. doi: 10.1109/isbi.2018.8363599. [DOI] [Google Scholar]
- 26.Zhu H, Shi J, Wu Ji. Pick-and-Learn: Automatic Quality Evaluation for Noisy-Labeled Image Segmentation. Lect Notes Comp Sci. 2019:576–584. doi: 10.1007/978-3-030-32226-7_64. [DOI] [Google Scholar]
- 27.Dua D, Graff CU. Machine learning repository. University of California, School of Information and Computer Science; Irvine, CA: 2019. [Google Scholar]
- 28.Ali MdM, Ahmed K, Bui FM, et al. Machine learning-based statistical analysis for early stage detection of cervical cancer. Comput Biol Med. 2021;139:104985. doi: 10.1016/j.compbiomed.2021.104985. [DOI] [PubMed] [Google Scholar]
- 29.Lucchesi FR, Aredes ND. Radiology data from the cancer genome atlas cervical squamous cell carcinoma and endocervical adenocarcinoma (tcga-cesc) collection. the cancer imaging archive. Cancer Imaging Arch. 2016;10:k9. [Google Scholar]
- 30.Lin J, Rodriguez S, Zhou H, et al. Racial disparities in late-stage prostate cancer: A SEER database analysis 2005–2015. J Clin Oncol. 2019;37(7_suppl):115–115. doi: 10.1200/jco.2019.37.7_suppl.115. [DOI] [PubMed] [Google Scholar]
- 31.Bedell SL, Goldstein LS, Goldstein AR, et al. Cervical Cancer Screening: Past, Present, and Future. Sex Med Rev. 2020;8(1):28–37. doi: 10.1016/j.sxmr.2019.09.005. [DOI] [PubMed] [Google Scholar]
- 32.Kelly M, Longjohn R, Nottingham K. The UCI machine learning repository. 2023. https//archiveicsuciedu .
- 33.Hemalatha B, Karthik B, Reddy C, et al. Cervical Cancer Classification: Optimizing Accuracy, Precision, and Recall using SMOTE Preprocessing and t-SNE Feature Extraction. 2024 International Conference on Intelligent and Innovative Technologies in Computing, Electrical and Electronics (IITCEE); 2024; pp. 1–7. [DOI] [Google Scholar]
- 34.Cappelli E, Cumbo F, Bernasconi A, et al. OpenGDC: Unifying, Modeling, Integrating Cancer Genomic Data and Clinical Metadata. Appl Sci. 2020;10(18):6367. doi: 10.3390/app10186367. [DOI] [Google Scholar]
- 35.Heath AP, Ferretti V, Agrawal S, et al. The NCI Genomic Data Commons. Nat Genet. 2021;53(3):257–262. doi: 10.1038/s41588-021-00791-5. [DOI] [PubMed] [Google Scholar]
- 36.Kitchen FL, Cox CM. Papanicolaou smear. StatPearls Publishing; Treasure Island: 2022. [Google Scholar]
- 37.KA, BS A Deep Learning-Based Approach for Cervical Cancer Classification Using 3D CNN and Vision Transformer. J Imaging Inform Med. 2024;37(1):280–296. doi: 10.1007/s10278-023-00911-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Tan SL, Selvachandran G, Ding W, et al. Cervical Cancer Classification From Pap Smear Images Using Deep Convolutional Neural Network Models. Interdiscip Sci. 2024;16(1):16–38. doi: 10.1007/s12539-023-00589-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Wang W, Tian Y, Xu Y, et al. 3cDe-Net: a cervical cancer cell detection network based on an improved backbone network and multiscale feature fusion. BMC Med Imaging. 2022;22(1):130. doi: 10.1186/s12880-022-00852-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Tanimu J, Hamada M, Hassan M, et al. A Machine Learning Method for Classification of Cervical Cancer. Electronics. 2022;11(3):463. doi: 10.3390/electronics11030463. [DOI] [Google Scholar]
- 41.Li YX, Chen F, Shi JJ, et al. Convolutional Neural Networks for Classifying Cervical Cancer Types Using Histological Images. J Digit Imaging. 2023;36(2):441–449. doi: 10.1007/s10278-022-00722-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Xu L, Cai F, Fu Y, et al. Cervical cell classification with deep-learning algorithms. Med Biol Eng Comput. 2023;61(3):821–833. doi: 10.1007/s11517-022-02745-3. [DOI] [PubMed] [Google Scholar]
- 43.Kalbhor M, Shinde S, Popescu DE, et al. Hybridization of Deep Learning Pre-Trained Models with Machine Learning Classifiers and Fuzzy Min-Max Neural Network for Cervical Cancer Diagnosis. Diagnostics (Basel) 2023;13(7) doi: 10.3390/diagnostics13071363. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Ghoneim A, Muhammad G, Hossain MS. Cervical cancer classification using convolutional neural networks and extreme learning machines. Futur Gener Comput Syst. 2020;102:643–649. https://www.sciencedirect.com/science/article/pii/S0167739X19306065. [Google Scholar]
- 45.Hamada M, Tanimu JJ, Hassan M, Kakudi HA, Robert P. Evaluation of Recursive Feature Elimination and LASSO Regularization-based optimized feature selection approaches for cervical cancer prediction. 2021 IEEE 14th International Symposium on Embedded Multicore/Many-core Systems-on-Chip (MCSoC); 2021; pp. 333–339. [Google Scholar]
- 46.Fang S, Yang J, Wang M, et al. An Improved Image Classification Method for Cervical Precancerous Lesions Based on ShuffleNet. Comput Intell Neurosci. 2022;2022:9675628. doi: 10.1155/2022/9675628. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.Elakkiya R, Teja K, Deborah LJ, et al. Imaging based cervical cancer diagnostics using small object detection - generative adversarial networks. Multimed Tools Applic. 2021;81(1):191–207. doi: 10.1007/s11042-021-10627-3. [DOI] [Google Scholar]
- 48.Kudva V, Prasad K, Guruvare S. Hybrid Transfer Learning for Classification of Uterine Cervix Images for Cervical Cancer Screening. J Digit Imaging. 2020;33(3):619–631. doi: 10.1007/s10278-019-00269-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.Mathivanan SK, Francis D, Srinivasan S, et al. Enhancing cervical cancer detection and robust classification through a fusion of deep learning models. Sci Rep. 2024;14(1):10812. doi: 10.1038/s41598-024-61063-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Goodman A. Deep-Learning-Based Evaluation of Dual Stain Cytology for Cervical Cancer Screening: A New Paradigm. J Natl Cancer Inst. 2021;113(11):1451–1452. doi: 10.1093/jnci/djaa067. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Chai S, Xin J, Wu J, et al. Cervical Cytology Classification with Coarse Labels Based on Two-Stage Weakly Supervised Contrastive Learning Framework. 2023 IEEE International Conference on Bioinformatics and Biomedicine (BIBM); 2023; pp. 825–830. [DOI] [Google Scholar]
- 52.Alquran H, Alsalatie M, Mustafa WA, et al. Cervical Net: A Novel Cervical Cancer Classification Using Feature Fusion. Bioengineering (Basel) 2022;9(10) doi: 10.3390/bioengineering9100578. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Ali M, Hossain M, Kona M, et al. An ensemble classification approach for cervical cancer prediction using behavioral risk factors. Healthcare Anal. 2024;5:100324. doi: 10.1016/j.health.2024.100324. [DOI] [Google Scholar]
- 54.Yu S, Zhang S, Wang B, et al. Generative adversarial network based data augmentation to improve cervical cell classification model. Math Biosci Eng. 2021;18(2):1740–1752. doi: 10.3934/mbe.2021090. [DOI] [PubMed] [Google Scholar]
- 55.He CM, Kang H, Yu Yao T, et al. An effective classifier based on convolutional neural network and regularized extreme learning machine. Math Biosci Eng. 2019;16(6):8309–8321. doi: 10.3934/mbe.2019420. [DOI] [PubMed] [Google Scholar]
- 56.Chen X, Pu X, Chen Z, et al. Application of EfficientNet-B0 and GRU-based deep learning on classifying the colposcopy diagnosis of precancerous cervical lesions. Cancer Med. 2023;12(7):8690–8699. doi: 10.1002/cam4.5581. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Pramanik R, Biswas M, Sen S, et al. A fuzzy distance-based ensemble of deep models for cervical cancer detection. Comput Methods Programs Biomed. 2022;219:106776. doi: 10.1016/j.cmpb.2022.106776. [DOI] [PubMed] [Google Scholar]
- 58.Zhao C, Shuai R, Ma Li, et al. Improving cervical cancer classification with imbalanced datasets combining taming transformers with T2T-ViT. Multimed Tools Appl. 2022;81(17):24265–24300. doi: 10.1007/s11042-022-12670-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Tang J, Zhang T, Gong Z, et al. High Precision Cervical Precancerous Lesion Classification Method Based on ConvNeXt. Bioengineering (Basel) 2023;10(12) doi: 10.3390/bioengineering10121424. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Yin J, Zhang Q, Xi X, et al. Enhancing Cervical Cell Detection Through Weakly Supervised Learning With Local Distillation Mechanism. IEEE Access. 2024;12:77104–77113. doi: 10.1109/access.2024.3407066. [DOI] [Google Scholar]
- 61.Shi J, Wang R, Zheng Y, et al. Cervical cell classification with graph convolutional network. Comput Methods Programs Biomed. 2021;198:105807. doi: 10.1016/j.cmpb.2020.105807. [DOI] [PubMed] [Google Scholar]
- 62.Dhivya P, Karthiga M, Indirani A, Nagamani T. Cervical Cancer Prediction Using Otimized Meta-Learning BT. In: Saini HS, Sayal R, Govardhan A, Buyya R, editors. Innovations in Computer Science and Engineering Proceedings of 8th ICICSE. Springer; Nature Singapore, Singapore: 2023. pp. 393–401. [Google Scholar]
- 63.Zhang Z, Yao P, Chen M, et al. SCAC: A Semi-Supervised Learning Approach for Cervical Abnormal Cell Detection. IEEE J Biomed Health Inform. 2024;28(6):3501–3512. doi: 10.1109/JBHI.2024.3375889. [DOI] [PubMed] [Google Scholar]
- 64.Nasir M, Khalil O, Ateeq K, et al. Cervical Cancer Prediction Empowered with Federated Machine Learning. Comput Mater Contin. 2024;79(1):963–981. doi: 10.32604/cmc.2024.047874. [DOI] [Google Scholar]
- 65.Liang Y, Tang Z, Yan M, et al. Comparison detector for cervical cell/clumps detection in the limited data scenario. Neurocomputing. 2021;437:195–205. doi: 10.1016/j.neucom.2021.01.006. [DOI] [Google Scholar]
- 66.Jia AD, Li BZ, Zhang CC. Detection of cervical cancer cells based on strong feature CNN-SVM network. Neurocomputing. 2020;411:112–127. doi: 10.1016/j.neucom.2020.06.006. [DOI] [Google Scholar]
- 67.Rahaman MdM, Li C, Yao Y, et al. DeepCervix: A deep learning-based framework for the classification of cervical cells using hybrid deep feature fusion techniques. Comput Biol Med. 2021;136:104649. doi: 10.1016/j.compbiomed.2021.104649. [DOI] [PubMed] [Google Scholar]
- 68.Cao L, Yang J, Rong Z, et al. A novel attention-guided convolutional network for the detection of abnormal cervical cells in cervical cancer screening. Med Image Anal. 2021:102197. doi: 10.1016/j.media.2021.102197. [DOI] [PubMed] [Google Scholar]
- 69.Wang Z, Gao H, Wang X, et al. A multi-Task Learning based applicable AI model simultaneously predicts stage, histology, grade and LNM for cervical cancer before surgery. BMC Womens Health. 2024;24(1):425. doi: 10.1186/s12905-024-03270-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70.Chen W, Gao L, Li X, et al. Lightweight convolutional neural network with knowledge distillation for cervical cells classification. Biomed Signal Process Control. 2022;71:103177. doi: 10.1016/j.bspc.2021.103177. [DOI] [Google Scholar]
- 71.Long J, Shelhamer E, Darrell T. Fully convolutional networks for semantic segmentation. IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2015; 2015. [DOI] [PubMed] [Google Scholar]
- 72.Dhawan S, Singh K, Arora M. Cervix Image Classification for Prognosis of Cervical Cancer using Deep Neural Network with Transfer Learning. EAI Endorsed Trans Pervasive Heal Technol. 2021;7(27):e5. doi: 10.4108/eai.12-4-2021.169183. [DOI] [Google Scholar]
- 73.Hua W, Xiao T, Jiang X, et al. Lymph-vascular space invasion prediction in cervical cancer: Exploring radiomics and deep learning multilevel features of tumor and peritumor tissue on multiparametric MRI. Biomed Signal Process Control. 2020;58:101869. doi: 10.1016/j.bspc.2020.101869. [DOI] [Google Scholar]
- 74.Chaudhuri A, Ray A, Banerjee D, et al. A Multi-Stage Approach Combining Feature Selection with Machine Learning Techniques for Higher Prediction Reliability and Accuracy in Cervical Cancer Diagnosis. Int J Intell Syst Appl. 2021;13(5):46–63. doi: 10.5815/ijisa.2021.05.05. [DOI] [Google Scholar]
- 75.Attallah O. Cervical Cancer Diagnosis Based on Multi-Domain Features Using Deep Learning Enhanced by Handcrafted Descriptors. App Sci. 2023;13(3):1916. doi: 10.3390/app13031916. [DOI] [Google Scholar]
- 76.Kumawat G, Vishwakarma S, Chakrabarti P, et al. Prognosis of Cervical Cancer Disease by Applying Machine Learning Techniques. J Circuits Syst Comp. 2022;32(01) doi: 10.1142/s0218126623500196. [DOI] [Google Scholar]
- 77.Mukku L, Thomas J. Deep learning-based cervical lesion segmentation in colposcopic images. Appl Eng Technol. 2024;3(1):16–25. doi: 10.31763/aet.v3i1.1345. [DOI] [Google Scholar]
- 78.Harsono AB, Susiarno H, Suardi D, et al. Cervical pre-cancerous lesion detection: development of smartphone-based VIA application using artificial intelligence. BMC Res Notes. 2022;15(1):356. doi: 10.1186/s13104-022-06250-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 79.Waly MI, Sikkandar MY, Aboamer MA, et al. Optimal Deep Convolution Neural Network for Cervical Cancer Diagnosis Model. Comput Mater Contin. 2022;70(2):3295–3309. doi: 10.32604/cmc.2022.020713. [DOI] [Google Scholar]
- 80.Urushibara A, Saida T, Mori K, et al. Diagnosing uterine cervical cancer on a single T2-weighted image: Comparison between deep learning versus radiologists. Eur J Radiol. 2021;135:109471. doi: 10.1016/j.ejrad.2020.109471. [DOI] [PubMed] [Google Scholar]
- 81.Tan X, Li K, Zhang J, et al. Automatic model for cervical cancer screening based on convolutional neural network: a retrospective, multicohort, multicenter study. Cancer Cell Int. 2021;21(1):35. doi: 10.1186/s12935-020-01742-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82.Liu Lu, Wang Y, Liu X, et al. Computer-aided diagnostic system based on deep learning for classifying colposcopy images. Ann Transl Med. 2021;9(13):1045. doi: 10.21037/atm-21-885. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 83.Bae JK, Roh HJ, You JS, et al. Quantitative Screening of Cervical Cancers for Low-Resource Settings: Pilot Study of Smartphone-Based Endoscopic Visual Inspection After Acetic Acid Using Machine Learning Techniques. JMIR Mhealth Uhealth. 2020;8(3):e16467. doi: 10.2196/16467. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 84.Asiedu MN, Skerrett E, Sairo G, Ramanujam N. In: Skerrett E, Sairo G, Ramanujam N, editors. Combining multile contrasts for imroving machine learning-based classification of cervical cancers with a low-cost oint-of-care Pocket coloscoe; 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC) IEEE; 2020; pp. 1148–51. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85.Xue P, Tang C, Li Q, et al. Development and validation of an artificial intelligence system for grading colposcopic impressions and guiding biopsies. BMC Med. 2020;18(1):406. doi: 10.1186/s12916-020-01860-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86.Jia AD, Li BZ, Zhang CC. Detection of cervical cancer cells based on strong feature CNN-SVM network. Neurocomputing. 2020;411:112–127. doi: 10.1016/j.neucom.2020.06.006. [DOI] [Google Scholar]
- 87.Miyagi Y, Takehara K, Nagayasu Y, et al. Application of deep learning to the classification of uterine cervical squamous epithelial lesion from colposcopy images combined with HPV types. Oncol Lett. 2020;19(2):1602–1610. doi: 10.3892/ol.2019.11214. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 88.Yue Z, Ding S, Zhao W, et al. Automatic CIN Grades Prediction of Sequential Cervigram Image Using LSTM With Multistate CNN Features. IEEE J Biomed Health Inform. 2020;24(3):844–854. doi: 10.1109/JBHI.2019.2922682. [DOI] [PubMed] [Google Scholar]
- 89.Huang Z, Wang X, Huang L, et al. CCNet: Criss-Cross Attention for Semantic Segmentation. 2019 IEEE/CVF International Conference on Computer Vision (ICCV); 2019; pp. 603–612. [DOI] [Google Scholar]
- 90.Redmon J, Farhadi A. Yolov3: An incremental improvement. 2018 arXiv Prepr arXiv18040. 2767. [Google Scholar]
- 91.Asiedu MN, Simhal A, Chaudhary U, et al. Development of algorithms for automated detection of cervical pre-cancers with a low-cost, point-of-care, pocket colposcope. IEEE Trans Biomed Eng. 2019;66(8):2306–2318. doi: 10.1109/TBME.2018.2887208. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 92.Kudva V, Prasad K, Guruvare S. Andriod Device-Based Cervical Cancer Screening for Resource-Poor Settings. J Digit Imaging. 2018;31(5):646–654. doi: 10.1007/s10278-018-0083-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 93.Chollet F. Xcetion: Dee learning with dethwise searable convolutions. Proceedings of the IEEE conference on comuter vision and attern recognition; 2017; pp. 1251–1258. [Google Scholar]
- 94.Xie S, Girshick R, Dollar P, et al. Aggregated Residual Transformations for Deep Neural Networks. IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2017; 2017. [DOI] [Google Scholar]
- 95.Badrinarayanan V, Kendall A, Cipolla R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans Pattern Anal Mach Intell. 2017;39(12):2481–2495. doi: 10.1109/TPAMI.2016.2644615. [DOI] [PubMed] [Google Scholar]
- 96.Chen LC, Papandreou G, Schroff F, et al. Rethinking atrous convolution for semantic image segmentation. arXiv Prepr arXiv17060. 5587. 2017 [Google Scholar]
- 97.He K, Zhang X, Ren S, et al. Deep Residual Learning for Image Recognition. IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2016; 2016. [DOI] [Google Scholar]
- 98.Ronneberger O, Fischer P, Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In: Navab N, Hornegger J, Wells WM, Frangi AF, editors. International Conference on Medical Image Computing and Computer-Assisted Intervention. Springer International Publishing; Cham: 2015. pp. 234–241. [Google Scholar]
- 99.Simões PW, Izumi NB, Casagrande RS, et al. Classification of images acquired with colposcopy using artificial neural networks. Cancer Inform. 2014;13:119–124. doi: 10.4137/CIN.S17948. [DOI] [PMC free article] [PubMed] [Google Scholar]











