Abstract
Simple Summary
The rapid development of technology has enabled numerous applications of artificial intelligence (AI), especially in medical science. Histopathological assessment of tissues remains the gold standard for diagnosis of gastrointestinal (GI) cancers for subsequent management. In a traditional anatomical pathology (AP) laboratory, histopathologists are required to manually assess, quantify and classify diseases under a microscope in a semiquantitative or qualitative manner. The conversion of analogue-to-digital pathology is on the rise, conforming with trends toward digitalization. The rapid adoption of digital pathology (DP) is driven by factors such as the worldwide shortage of pathologists and medical technologists, the increasing incidence of cancer and the critical need to improve laboratory efficiency. In this review, we aim to provide a comprehensive summary of algorithms for AI detection and classification of GI cancer. We will also provide critical insight into the application of algorithms for routine care in clinical practice.
Abstract
The implementation of DP will revolutionize current practice by providing pathologists with additional tools and algorithms to improve workflow. Furthermore, DP will open up opportunities for development of AI-based tools for more precise and reproducible diagnosis through computational pathology. One of the key features of AI is its capability to generate perceptions and recognize patterns beyond the human senses. Thus, the incorporation of AI into DP can reveal additional morphological features and information. At the current rate of AI development and adoption of DP, the interest in computational pathology is expected to rise in tandem. There have already been promising developments related to AI-based solutions in prostate cancer detection; however, in the GI tract, development of more sophisticated algorithms is required to facilitate histological assessment of GI specimens for early and accurate diagnosis. In this review, we aim to provide an overview of the current histological practices in AP laboratories with respect to challenges faced in image preprocessing, present the existing AI-based algorithms, discuss their limitations and present clinical insight with respect to the application of AI in early detection and diagnosis of GI cancer.
Keywords: digital pathology, computational pathology, whole-slide imaging, artificial intelligence, histopathology, gastrointestinal tract, algorithms, machine learning, deep learning, cancer diagnosis
1. Introduction
GI cancer is a collective term encompassing various cancers related to the GI system, which comprises the organs from the oral cavity down to the anal canal. Malignancy can develop in any part of the GI tract, as it is constantly and directly exposed to carcinogens in the environment through the ingestion of food. Given the numerous hotspots for malignant transformation and the constant exposure to carcinogens, GI cancer accounts for 26% of global cancer incidence and 35% of global cancer-related mortality [1]. The recent surge in incidence and mortality are linked to the increasing prevalence of modifiable risk factors, such as sedentary lifestyle, obesity, unhealthy diet and other metabolic abnormalities [2]. Recent analysis revealed a significant increase in GI cancer incidence in young adults aged 25–49 years, alerting the public to emerging medical burdens [3]. The five major GI cancers causing considerable global burdens are oesophageal squamous cell carcinoma, gastric adenocarcinoma, colorectal cancer (CRC), hepatocellular carcinoma and pancreatic cancer. In order to identify and classify GI cancers, histopathological analysis of endoscopic biopsies or resected tumour specimens remains the gold standard for disease diagnosis. Unfortunately, early diagnosis of GI-related cancers is often missed due to a lack of specific symptoms. Medical attention is often sought only when non-specific symptoms become unbearable [4].
In addition to the existing lack of medical professionals, there are various reasons for the growing demand for histopathological diagnosis, including an increase in morbidity within the aging population, the increasing incidence of malignancies among young adults and the increase in cancer screening programs, resulting in an increased number of annual laboratory cases [5,6,7,8]. Furthermore, the complexity of cases, as well arduous criteria for case reporting, constitute burdens to histopathologists and may prolong the turnaround time (TAT) associated with generating reports in AP laboratories [9]. Standardization among AP laboratories is crucial for pathologists to make precise diagnoses. Medical laboratory technologists are not only responsible for preparing tissue samples for diagnoses but also for ensuring that the laboratory performance is satisfactory through quality assurance programs while operating in compliance with laboratory accreditation [10,11]. With the lack of manpower and restricted laboratory budget, the adoption of more advanced technology is imminent with respect to addressing the overwhelming workload in an AP laboratory.
Whole-slide imaging (WSI) refers to the digitization of traditional glass slides into a digital image. Technological advancement has led to the development of high-throughput WSI scanners, which allow for large amounts of slides to be digitized in a short period of time. The digitized image of stained histological specimens can be used for a variety of purposes. Digital pathology (DP) is a term that incorporates the acquisition, management and interpretation of pathology information [12]. Currently, DP is used in both academic and clinical settings for teaching, diagnostic and archival purposes [13,14,15,16]. More importantly, the worldwide adoption of DP is a trend that could help AP laboratories improve productivity and reduce costs while improving patient care. The establishment of DP workflow in an AP laboratory opens opportunities for computational pathology; image analysis techniques, such as quantification and measurement; and the application of AI in computer-aided diagnosis [17]. The potential of AI in a DP workflow may alleviate burdens faced by histopathologists by reducing the time spent on tedious activities, such as counting cells and measuring tumour parameters, and, likewise, aid in standardizing immunohistochemical (IHC) staining for companion diagnostic testing [18,19,20]. AI-assisted endoscopy can also facilitate the detection and characterization of suspicious polyps, as well as the use of skin surface microscopy to classify skin diseases [21], which may allow for early identification and intervention in cancer. The adoption of AI will be of mutual benefit to healthcare institutes, medical practitioners and patients by increasing laboratory throughput, shortening TAT and reducing operating and patient costs.
The potential of AI in DP and GI-related malignancy is encouraging. However, the process from algorithm development to clinical application is complicated by various obstacles. An example of such a problem is the preprocessing of specimen for WSI generation. This could create variability within datasets [22,23], thus increasing the difficulty and cost of model training in a hospital setting. Machine learning [24,25,26,27] and deep learning [28,29,30,31,32,33,34,35,36,37,38,39,40] are common methods adopted for image analysis and both have unique advantages and disadvantages. Currently, there is a variety of existing models that can be selected and applied easily. Using existing models is convenient but may not be suitable for specific tasks, leading to poor performance. The design and features of models should be carefully considered during algorithm design to complete specific tasks to minimize unnecessary computational time while maintaining high output performance. Models with high performance, as measured by area under the receiver operating characteristic (AUROC), may not necessary be clinically applicable [16,22]. It is important to understand the clinical needs and current acceptability of AI as part of the clinical workflow. As the end users of these algorithm are medical professionals, it is vital to discuss and develop a model specific to their needs and workflow, while also balancing with generalizability for use in other hospital or clinical settings [16,41]. There is a real challenge associated with developing clinically applicable AI-based tools to facilitate pathologists in their workflow. In this regard, we will critically review the processes from preparation of tissue samples to current challenges in applying AI in diagnosis of GI cancers. We will also explore how the development and implementation of AI in DP may improve existing practice in screening and diagnosis of GI malignancies.
2. Current Histopathology Practices and Opportunities in Digital Pathology
In modern clinical practice, the advancement of technology enables the development of highly efficient, high-throughput and high-resolution (up to 0.23–0.25 μm/pixel) WSI devices that can digitize traditional glass slides into whole-slide images (WSIs) that can be remotely viewed on a display monitor for remote diagnosis of cases [42,43]. This digital process is termed “digital pathology” and has additional benefits compared to traditional pathology, including the ease of quickly and securely sharing pathology WSIs worldwide with other pathologists for second opinions, as well as providing a better view of the tissue with additional annotation and measurement tools while allowing for rapid access to archived cases without loss of quality [44]. Previous literature reports support the safe transition to digital pathology, showing high concordance of diagnosis using digital pathology with WSIs as compared to traditional slides with light microscopy [45,46,47,48,49]. The quality of WSIs is a major factor that affects concordance rates and diagnostic accuracy. In order for high-quality WSIs to be generated, specific tasks must be performed by medical laboratory technologists. The tissues of interest extracted from esophagogastroduodenoscopy or surgery must be preserved in fixatives, such as 10% neutral buffered formalin, to prevent autolysis of biological tissue while maintaining tissue structure and integrity. The tissue is then embedded, sectioned, stained and mounted onto a glass slide, ensuring the quality of the tissue and glass slide will. This includes checking for artefacts on the tissue, including possible scores, tears, floating contamination, thick and thin sections and ensuring that the glass slide is intact and free from dust before and after digitization [50,51] (Figure 1).
Considering the forthcoming transition from traditional pathology to digital pathology, there is a demand for the development of new tools to facilitate the reporting process for pathologists. In recent years, powerful WSI analysis software tools that are user-friendly yet packed with clinically relevant tools have been developed. The majority of such software applications are open-source and freely available; these include ImageJ 1.53s (Madison, United States), QuPath 0.3.0 (Belfast, United Kingdom), SlideRunner 2.2.0 (Erlangen, Germany) and Cytomine (Liège, Belgium) [52,53,54]. Such software applications are capable of handling large WSIs and metadata generated from different hardware brands and contain interactive drawing tools for annotation. They also include features that can perform cellular detection and feature extraction. Furthermore, to complement software development, the cost of hardware required for high-performance computation, including high-speed network infrastructure and data storage, has become increasingly affordable. This has led to increased adoption of digital pathology in major hospitals worldwide. However, more investment is required to expand the roles of digital pathology in most hospitals [12]. With an increasing number of digital pathology centres, the generation of large and high-quality WSI databases will slowly emerge [17,47,55,56]. This growth increases the feasibility of obtaining large datasets and designing algorithms for analysis of WSIs using computer software through a broad range of methods for the study of diseases. This concept is termed computational pathology [57]. The combination of computational pathology with digital pathology opens new opportunities for case diagnosis. Algorithms developed for the early detection of cancer are important to improve patients’ chance of survival [58]. Automated screening of a large number of specimens may provide improved accessibility and make diagnosis and treatment more affordable [59]. Furthermore, the escalating amount of cancer specimens with increasingly complex classification and the urgent need for shorter TAT have resulted in an upsurge in workload in diagnostic pathology services [60]. With these challenges in mind, more research has been focused on the use of AI to perform diagnostic tasks for histopathological examination. Some models have been proposed to use computational methods to triage patients and present urgent cases to pathologists with prioritizations. Despite the excellent results and potential application in histopathological diagnosis prediction using AI [61,62,63], there are still significant challenges to overcome prior to the implementation of these AI-based tools and algorithms. In the following section, we will explore the critical areas that hinder the current development and deployment of AI-based tools in clinical practice.
3. Current Challenges of Algorithm Development in Computational Pathology
AI has become a popular tool in the field of medical image processing and analysis [64]. AI is capable of extracting meaningful information from images that cannot be obtained with the naked eye. Machine learning is a subset of AI that utilizes algorithms developed to process these data and perform tasks without explicit programming. Deep learning is a subset of machine learning that uses highly complex structures of algorithms based on the neural network of the human brain [65] (Figure 2). In recent years, AI has become increasingly popular in histological imaging analysis, with clinical applications that include primary tumour detection, grading and subtyping [66,67,68]. Several well-known research challenge competitions, such as CAMELYON16 and CAMELYON17, encourage the development of novel and automated solutions that are clinically applicable to improve diagnostic accuracy and efficiency [69,70]. This is achieved through various combinations of machine learning and deep learning techniques, which will be discussed further. Large annotated datasets are also indispensable for the development of successful deep learning algorithms. Through the adoption of digital pathology for diagnosis, it is possible for pathologists to annotate regions of interest during the case reporting process, which may indirectly facilitate the generation of such valuable datasets for future algorithm development [16].
Furthermore, the inevitable integration of these AI tools is necessary due to the aging population worldwide and the shortage of pathologists [71]. The training of new pathologists requires a long period of time in order to ensure competency [72]. Thus, there is an urgent need to develop clinically applicable AI-based tools to relieve the high workload of pathologists, producing more precise and reproducible diagnoses while reducing the TAT of cases. However, there remain obstacles and challenges that are hinder the development process, which will be discussed below.
3.1. Colour Normalization
For routine histopathological diagnosis, haematoxylin and eosin (H&E) staining is the most preferred method for visualizing cellular morphology and tissue structure. Haematoxylin stains cell nuclei a purplish-blue, whereas eosin stains the extracellular matrix and cytoplasm pink. The patterns of coloration play a central role in the differentiation between cells and structures. Colour and staining variations can be affected by, but are not limited to, the thickness of the specimen, non-standardized staining protocols and slide scanner variations [73]. It is important to perform colour normalization to ensure consistency across WSI databases, as it can affect the robustness of deep learning models. Common methods for colour normalization include histogram matching [74,75], colour transfer [76,77,78] and spectral matching [79,80,81]. However, these methods of colour normalization heavily rely on the expertise of pathologists, which makes WSI colour normalization difficult for general researchers with limited knowledge of the colour profile of histological staining. Moreover, this process of manually assessing and adjusting each WSI is impractical. Several algorithms have been proposed by researchers that are capable of performing colour normalization by deep learning using a template glass slide image for reference, such as StainGAN [82] and SA-GAN [83]. These models have shown promising results with respect to ensuring consistent representation of colour and texture [84,85,86]. Nevertheless, future work is needed to explore the clinical performance of AI-based tools developed using such colour normalization systems.
3.2. Pathologist Interpretation for Model Training
For histopathologists, understanding normal histology is essential for investigating abnormal cells and tissue. Final interpretation comprises several processes, including visualization, spatial awareness, perception and empirical experience. When developing new algorithms for the classification of diseases, the wide variation in the interpretation or classification of diseases cannot be categorized or easily represented in a cardinal manner. If so, the classification task is oversimplified, leading to a lack of detail. Physiologically, cancers are characterized by an increase in cellular proliferation, invasive growth, evasion of apoptosis, and altered genome and expression [87]. With H&E staining, histopathologists can identify the changes in cellular characteristics by several broad features of malignant cells, which include anaplasia, loss of polarity, hyperchromatism, nuclear pleomorphism and irregularity, abnormal distribution of chromatin, prominent nucleoli and atypical mitotic figures. Diagnostic reports are often used as the ground truth to label datasets. However, pathologists may use descriptive terminology to explain pathology concepts that are difficult to interpret and categorize [88]. Hence, it is important for researchers to obtain diagnostic reports with standardized reporting, including the specimen type, site, histologic type, grading and staging, in accordance with the American Joint Committee on Cancer, which represents different categories of histopathological knowledge under fuzzy ontology [89], and streamline the process of algorithm development to generate more accurate disease classifications [90].
3.3. Model Transparency and Interpretability for Deployment of AI-Based Tools in Clinical Practice
The incorporation of digital pathology, let alone computational pathology, in the clinical workflow requires solid validation of WSI analysis. However, various studies have reported that there is high discordance between diagnoses made using WSI and those made with light microscopy in GI malignancies [46,91,92,93]. More evidence is required to demonstrate that diagnoses made through digital pathology are as performant as conventional light microscopy diagnoses to ensure reliable and safe application in digital pathology practice [91]. According to practical recommendations by the Royal College of Pathologists regarding the implementation of digital pathology [94], validation and verification processes should be performed in the image analysis system to demonstrate its clinical utility before integrating into clinical workflows. Thus, it is also of considerable importance for AI-based tools to be transparent and interpretable in order to address possible moral and fairness biases by providing evidence in making a specific decision [95]. However, there is a dilemma when designing new algorithms, as most algorithm developments focus on using sophisticated deep learning and ensemble methods—so-called “black-box” models—to tackle multidimensional problems. On the other hand, much simpler methods, such as linear regression or decision-tree-based algorithms, may not be sophisticated and powerful enough to achieve the desired outcome [96,97,98,99]. Nonetheless, the demand for more explainable AI is on the rise [100]. Developers of new AI-based tools targeting applications in digital pathology should continuously assess the model explainability of their tool to meet the expectations of various stakeholders. They should also consider the ethical concerns and possible regulatory barriers imposed by governments and professional bodies regulating practice [101,102].
4. Machine Learning in GI Cancer Diagnosis
Machine learning algorithms constitute a method that requires the input of large datasets to study patterns or relationships to predict outcomes. The most common machine learning algorithms include artificial neural networks, decision tree algorithms, support vector machine, regression analysis, Bayesian networks, etc. In computational pathology, due to the large image size of each WSI (>3 billion pixels, >1 GB), the majority of these algorithms are designed based on manually extracted or handcrafted features from WSIs to improve their performance and minimise the computational time [103]. Such algorithm design processes can be summarised as feature engineering and can be divided in a two-step process. First, the regions of interest related to a specific clinical problem are identified in images. Then, the images are translated into numerical or categorical values using several predictor variables to create an accurate machine learning algorithm [104]. In digital pathology, there are two kinds of hand-crafted features, namely domain-inspired and domain-agnostic features. Domain-inspired features require the intrinsic domain knowledge of pathologists and oncologists, whereas domain-agnostic features include the general computational features used in the machine learning algorithms [15]. In practice, the generation of handcrafted features is often achieved by software, such as wndchrm, which was developed by Shamir et al. [105], and QuPath by Bankhead et al. [53]. These handcrafted, feature-based machine learning algorithms have demonstrated GI cancer classification, detection and prognostication abilities. The main researches related to machine learning models and algorithms in GI cancer diagnosis are presented in Table 1.
Table 1.
Author | Task | Cancer Type | Type of WSI | Dataset | Algorithm/ Model |
Performance | Clinical Application |
---|---|---|---|---|---|---|---|
Yoshida et al. [24] | Classification | Gastric cancer | H&E | Training and testing: 3062 WSIs | e-Pathologist | Positive for carcinoma or suspicion of carcinoma vs. caution for adenoma or suspicion of a neoplastic lesion vs. negative for a neoplastic lesion Overall concordance rate: 55.6% Kappa coefficient: 0.28 (95% CI: 0.26–0.30) Negative vs. non-negative Sensitivity: 89.5% (95% CI: 87.5–91.4%) Specificity: 50.7% (95% CI: 48.5–52.9%) Positive predictive value: 47.7% (95% CI: 45.4–49.9%) Negative predictive value: 90.6% (95% CI, 88.8–92.2%) |
Differentiation and diagnosis gastric cancer grade |
Yasuda et al. [25] | Classification | Gastric cancer | H&E | Training and testing: 66 WSIs | wndchrm | Noncancer vs. well-differentiated gastric cancer AUC: 0.99 Noncancer vs. moderately differentiated gastric cancer AUC: 0.98 Noncancer vs. poorly differentiated gastric cancer AUC: 0.99 |
Differentiation and diagnosis gastric cancer grade |
Jiang et al. [106] | Classification and prognosis | Gastric cancer | H&E | Training: 251 patients Internal validation: 248 patients External validation: 287 patients |
Support vector machine | Patients might benefit more from postoperative adjuvant chemotherapy vs. patient might not postoperative adjuvant chemotherapy training cohort: 5-year overall survival AUC: 0.796 5-year disease-free survival AUC: 0.805 Internal validation cohort: 5-year overall survival AUC: 0.809 5-year disease-free survival AUC: 0.813 External validation cohort: 5-year overall survival AUC: 0.834 5-year disease-free survival AUC: 0.828 |
Prognosis of gastric cancer patients and identification of patients who might benefit from adjuvant chemotherapy |
Cosatto et al. [26] | Detection | Gastric cancer | H&E | Training set: 8558 patients Test set: 4168 patients |
Semi-supervised multi-instance learning framework | Positive vs. negative AUC: 0.96 |
Detection of gastric cancer |
Jiang et al. [27] | Classification | Colon cancer | H&E | Training: 101 patients Internal validation: 67 patients External validation: 47 patients |
InceptionResNetV2 + gradient-boosting decision tree machine classifier | High-risk recurrence vs. low-risk recurrence Internal validation hazard ratio: 8.9766 (95% CI: 2.824–28.528) External validation hazard ratio: 10.273 (95% CI: 2.177–48.472) Poor vs. good prognosis groups: Internal validation hazard ratio: 10.687 (95% CI: 2.908–39.272) External validation hazard ratio: 5.033 (95% CI: 1.792–14.132) |
Prognosis of stage III colon cancer |
WSI = whole-slide imaging; H&E = haematoxylin and eosin; CI= confidence interval; AUC = area under the curve; wndchrm = weighted neighbour distance using compound hierarchy of algorithms representing morphology.
Yoshida et al. [24] pioneers in the application of machine learning algorithms to classify the different grades of gastric cancer (GC) using “e-Pathologist” software. Their system performance showed high sensitivity (>89%) and low specificity (<51%). The authors demonstrated that their software, “e-Pathologist”, is not sensitive enough to detect gastric carcinoma using clinical samples. Similar work has been done by Yasuda et al. [25], with “wndchrm” software; their system exhibited improved performance relative to “e-Pathologist” and showed a superior AUC of >0.98. However, with a limited validation dataset used (<100 WSIs), a larger dataset is required to verify the clinical applicability of wndchrm. Furthermore, Cosatto et al. reported the use of a machine learning algorithm integrated with a multi-instance learning framework to identify the presence of GC [26]. Their framework achieved promising performance (AUC: 0.96), and their validation using a large dataset (4168 patients) illustrates the effectiveness of integration of machine learning with a multi-instance learning architecture. Support vector machine is one of the most popular algorithms in machine learning used for classification and prognostication for cancer research. Jiang et al. [106] used support vector machine to predict the prognosis of GC patients and identify patients who might benefit from adjuvant chemotherapy. Sharma et al. [107] also used support vector machine for detection of gastric tumour necrosis. The use of support vector machine will likely remain popular. Its adoption in computational pathology should be further explored.
To achieve shorter computational time and higher accuracy, machine learning is commonly synergised with deep learning in the field of computational pathology [27,108]. Jiang et al. proposed the combination of the InceptionResNetV2 deep learning model and a gradient-boosting decision tree machine classifier for prognosis prediction of stage III colon cancer. The hybrid model calculated a hazard ratio of 10.273 in the prediction of high-risk or low-risk recurrence and a hazard ratio of 5.033 for patients with poor and good prognoses. [27]
Machine learning algorithms are also capable of performing tasks related to the colour of histology images, as the colour of biomarkers is important for colour normalization of various histology images, as well as disease prediction. Kothari et al. [74] used a four-class linear discriminant classifier to develop a new colour segmentation algorithm for standardization of WSI colour to improve machine learning model accuracy in classifying different colour components from histological cancer images. Furthermore, as aggressive tumour progression is associated with high cell proliferation rates, immunostaining of tumour nuclei is performed by pathologists for estimation of tumour growth. Rokshana et al. [109] proposed the use of an IHC colour histogram to perform colour separation for the quantification of the Ki67 proliferation index for breast cancer. Such a colour-thresholding algorithm can assist pathologists in obtaining automatic and robust prognosis predictions for cancer patients.
5. Deep Learning in GI Cancer Diagnosis
Deep learning is a subcategory of machine learning and is used to directly study image patterns. It has also gained popularity in the world of computational pathology. Convolution neural network (CNN) is a widely used deep learning algorithm applied in pathological image analysis. It is a specific type of neural network that uses a convolution operation to process each pixel’s data for image recognition and processing through image processing with different convolution and pooling layers [110]. CNN is able to compress and extract problem-specific information directly from images to train algorithms for the classification, segmentation, detection and prognosis of cancers. Prior to model training, image preprocessing is vital, as most WSIs are excessively large (approx. 40,000 × 50,000 pixels), and the input requirement of CNN allows for small images only (≤512 × 512 pixels). Hence, it is normal practice to tile WSIs prior to algorithm training and development. Herein, we will review a selection of deep learning algorithms developed for GI cancer diagnosis. The main researches related to deep learning models and algorithms in GI cancer diagnosis are presented in Table 2.
Table 2.
Author | Degree of Supervision |
Task | Cancer Type | Type of WSI | Dataset | Algorithm/ Model |
Performance | Clinical Application |
---|---|---|---|---|---|---|---|---|
Shen et al. [112] | Fully supervised |
Classification | Gastric cancer | H&E | Training, validation and testing: 432 WSIs (TCGA-STAD cohort) + 460 WSIs (TCGA-COAD) + 171 WSIs (TCGA-READ) + 400 WSIs (Camelyon16) |
DenseNet + Deformable Conditional Random Field model |
Accuracy: 0.9398 (TCGA-STAD), 0.9337 (TCGA-COAD), 0.9294 (TCGA-READ), 0.9468 (Camelyon16) | Identification of suspected cancer area from histological imaging |
Song et al. [30] | Fully supervised |
Classification | Gastric cancer | H&E | Training: 2123 WSIs Validation: 300 WSIs Internal testing: 100 WSIs External validation: 3212 WSIs (daily gastric dataset) + 595 WSIs (PUMCH) + 987 WSIs (CHCAMS and Peking Union Medical College) |
DeepLab v3 | Malignant vs. benign training AUC: 0.923 Internal testing AUC: 0.931 AUC: 0.995 (daily gastric dataset) AUC: 0.990 (PUMCH) AUC: 0.996 (CHCAMS and Peking Union Medical College) |
Diagnosis of gastric cancer |
Su et al. [28] | Fully supervised |
Classification and detection | Gastric cancer | H&E | Training: 348 WSIs Testing: 88 WSIs External Validation: 31 WSIs |
ResNet-18 | Poorly differentiated adenocarcinoma vs. well-differentiated adenocarcinoma and other normal tissue F1 score: 0.8615 Well-differentiated adenocarcinoma vs. poorly differentiated adenocarcinoma and other normal tissue F1 score: 0.8977 Patients with MSI vs. without MSI Accuracy: 0.7727 (95% CI 0.6857–0.8636) |
Differentiation of cancer grade and diagnosis of MSI |
Song et al. [29] | Fully supervised |
Classification | Colorectal cancer | H&E | Training: 177 WSIs Validation: 40 WSIs Internal test: 194 WSIs External validation: 168 WSIs |
Deep Lab v2 with ResNet34 | Adenomatous vs. normal AUC: 0.92 Accuracy: 0.904 |
Diagnosis of colorectal adenomas |
Sirinukunwattana et al. [114] | Fully supervised |
Classification | Colorectal cancer | H&E | Training: 510 WSIs External validation: 431 WSIs (TCGA cohort) + 265 WSIs (GRAMPIAN cohort) |
Inception V3 | Colorectal cancer consensus molecular subtypes 1 vs. 2 vs. 3 vs. 4 Training average accuracy: 70% Training AUC: 0.9 External validation accuracy: 0.64 (TCGA cohort) + 0.72 (GRAMPIAN cohort)External validation AUC:0.84 (TCGA cohort) + 0.85 (GRAMPIAN cohort) |
Prediction of colorectal cancer molecular subtype |
Popovici et al. [115] | Fully supervised |
Classification | Colorectal cancer | H&E | Training: 100 WSIs Test: 200 WSIs |
VGG-F | Molecular subtype A vs. B vs. C vs. D vs. EOverall accuracy: 0.84 (95% CI: 0.79−0.88)Overall recall: 0.85 (95% CI: 0.80−0.89)Overall precision: 0.84 (95% CI: 0.80−0.88) | Prediction of colorectal cancer molecular subtype |
Korbar et al. [116] | Fully supervised |
Classification | Colorectal cancer | H&E | Training: 458 WSIs Testing: 239 WSIs |
ResNet-152 | Hyperplastic polyp vs. sessile serrated polyp vs. traditional serrated adenoma vs. tubular adenoma vs. tubulovillous/villous adenoma vs. normal Accuracy: 0.930 (95% CI: 0.890−0.959) Precision: 0.897 (95% CI: 0.852−0.932) Recall: 0.883 (95% CI: 0.836−0.921) F1 score: 0.888 (95% CI: 0.841−0.925) |
Characterization of colorectal polyps |
Wei et al. [118] | Fully supervised |
Classification | Colorectal cancer | H&E | Training: 326 WSIs Validation: 25 WSIs Internet test: 157 WSIs External validation: 238 WSIs |
Ensemble ResNet×5 | Hyperplastic polyp vs. sessile serrated adenoma vs. tubular adenoma vs. tubulovillous or villous adenoma. Internal test mean accuracy: 0.935 (95% CI: 0.896–0.974) External validation mean accuracy: 0.870 (95% CI: 0.827–0.913) |
Colorectal polyp classification |
Gupta et al. [119] | Fully supervised |
Classification | Colorectal cancer | H&E | Training and testing: 303,012 normal WSI patches and approximately 1,000,000 abnormal WSI patches |
Customized Inception-ResNet-v2 Type 5 (IR-v2 Type 5) model. |
Abnormal region vs. normal region F-score: 0.99 AUC: 0.99 |
Identification of suspected cancer area from histological imaging |
Kather et al. [117] | Fully supervised |
Classification and prognosis | Colorectal cancer | H&E | Training: 86 WSIs Testing: 25 WSIs External validation: 862 WSIs (TCGA cohort) + 409 WSIs (DACHS cohort) |
VGG19 | Adipose tissue vs. background vs. lymphocytes vs. mucus vs. smooth muscle vs. normal colon mucosa vs. cancer-associated stroma vs. colorectal adenocarcinoma epithelium Internal testing Overall Accuracy: 0.99 External testing Overall accuracy: 0.943 High deep stroma score predicts shorter survival Hazard ratio: 1.99 (95% CI: 1.27–3.12) |
Colorectal cancer detection and prediction of patient survival outcome |
Zhu et al. [31] | Fully supervised |
Classification and segmentation | Gastric and colorectal cancer | H&E | Training: 750 WSIs Testing: 250 WSIs |
Adversarial CAC-UNet | Malignant region vs. benign region DSC: 0.8749 Recall: 0.9362 Precision: 0.9027 Accuracy: 0.8935 |
Identification of suspected cancer area from histological imaging |
Xu et al. [33] | Fully supervised |
Segmentation | Colorectal cancer | H&E | Training: 750 WSIs Testing: 250 WSIs |
CoUNet | Malignant region vs. benign region Dice: 0.746 AUC: 0.980 |
Identification of suspected cancer area from histological imaging |
Feng et al. [32] | Fully supervised |
Segmentation | Colorectal cancer | H&E | Training: 750 WSIs Testing: 250 WSIs |
U-Net-16 | Malignant region vs. benign region DSC: 0.7789 AUC:1 |
Identification of suspected cancer area from histological imaging |
Mahendra et al. [120] | Fully supervised |
Segmentation | Colorectal cancer | H&E | Training: 270 WSIs (CAMELYON16) + 500 WSIs (CAMELYON17) + 660 WSIs (DigestPath) + 50 WSIs (PAIP) Testing: 129 WSIs (CAMELYON16) + 500 WSIs (CAMELYON17) + 212 WSIs (DigestPath) + 40 WSIs (PAIP) |
DenseNet-121 + Inception-ResNet-V2 + DeeplabV3Plus |
Malignant region vs. benign region Cohen kappa score: 0.9090 (CAMELYON17) DSC: 0.782 (DigestPath) |
Identification of suspected cancer area from histological imaging |
Gehrung et al. [34] | Fully supervised |
Detection | Oesophageal cancer | H&E and TFF3 pathology slides | Training: 100 + 187 patients Validation: 187 patients External validation: 1519 patients |
VGG-16 | Patients with Barrett’s oesophagus vs. no Barrett’s oesophagus AUC: 0.88 (95% CI: 0.85–0.91) Sensitivity: 0.7262 (95% CI: 0.6742–0.7821) Specificity: 0.9313 (95% CI: 0.9004–0.9613) Simulated realistic cohort workload reduction: 57% External validation cohort reduction: 57.41% |
Detection of Barrett’s oesophagus |
Kather et al. [122] | Fully supervised |
Detection | Gastric and colorectal cancer | H&E | Training: 81 patients (UMM and NCT tissue bank) + 216 patients (TCGA-STAD) + 278 patients (TCGA-CRC-KR) + 260 patients (TCGA-CRC-DX) + 382 patients (UCEC) External validation: 99 patients (TCGA-STAD) + 109 patients (TCGA-CRC-KR) +100 patients (TCGA-CRC-DX) +110 patients (UCEC) + 185 patients (KCCH) |
Resnet18 | Patients with MSI vs. no MSI Training AUC: >0.99 (UMM and NCT tissue bank) AUC: 0.81 (CI: 0.69–0.90) (TCGA-STAD) AUC: 0.84 (CI: 0.73–0.91) (TCGA-CRC-KR) AUC: 0.77 (CI: 0.62–0.87) (TCGA-CRC-DX) AUC: 0.75 (CI: 0.63–0.83) (UCEC) AUC: 0.69 (CI: 0.52–0.82) (KCCH) |
Detection of MSI |
Echle et al. [36] | Fully supervised |
Detection | Colorectal cancer | H&E | Training: 6406 WSIs External validation: 771 WSIs |
Shufflenet | Colorectal tumour sample with dMMR or MSI vs. no dMMR or MSI Mean AUC: 0.92 AUPRC: 0.93 Specificity: 0.67 Sensitivity: 0.95 External validation AUC without colour normalisation: 0.95 External validation AUC with colour normalisation: 0.96 |
Detection of MSI |
Cao et al. [121] | Fully supervised |
Detection | Colorectal cancer | H&E | Training: 429 WSIs External validation: 785 WSIs |
ResNet-18 | Colorectal cancer patients with MSI vs. no MSI AUC: 0.8848 (95% CI: 0.8185–0.9512) External validation AUC: 0.8504 (95% CI: 0.7591–0.9323) |
Detection of MSI |
Meier et al. [124] | Fully supervised |
Prognosis | Gastric cancer | H&E IHC staining, including CD8, CD20, CD68 and Ki67 |
Training and testing: 248 patients | GoogLeNet | Risk of the presence of Ki67&CD20 Hazard ratio = 1.47 (95% CI: 1.15–1.89) Risk of the presence of CD20&CD68 Hazard ratio = 1.33 (95% CI: 1.07–1.67) |
Cancer prognosis based on various IHC markers to predict patient survival outcome |
Bychkov et al. [123] | Fully supervised |
Prognosis | Colorectal cancer | H&E | Training: 220 WSIs Validation: 60 WSIs Testing: 140 WSIs |
VGG-16 | High-risk patients vs. low-risk patients Prediction with small tissue hazard ratio: 2.3 (95% CI: 1.79–3.03) |
Survival analysis of colorectal cancer |
Wang et al. [127] | Weakly supervised | Classification | Gastric cancer | H&E | Training: 408 WSIs Testing: 200 WSIs |
recalibrated multi-instance deep learning |
Cancer vs. dysplasia vs. normal Accuracy: 0.865 |
Diagnosis of gastric cancer |
Xu et al. [37] | Weakly supervised | Classification | Gastric cancer | H&E | Training, validation and testing: 185 WSIs (SRS dataset) + 2032 WSIs (Mars dataset) |
multiple instance classification framework based on graph convolutional networks |
Tumour vs. normal Recall: 0.904 (SRS dataset), 0.9824 (Mars dataset) Precision: 0.9116 (SRS dataset), 0.9826 (Mars dataset) F1-score: 0.9075 (SRS dataset), 0.9824 (Mars dataset) |
Diagnosis of gastric cancer |
Huang et al. [39] | Weakly supervised | Classification | Gastric cancer | H&E | Training and testing: 2333 WSIs External validation: 175 WSIs |
GastroMIL | Gastric cancer vs. normal External validation accuracy: 0.92 GastroMIL risk score associated with patient overall survival Hazard ratio: 2.414 |
Diagnosis of gastric cancer and prediction of patient survival outcome |
Li et al. [131] | Weakly supervised | Classification | Gastric cancer | H&E | Training and testing: 10,894 WSIs | DLA34 + Otsu’s method | Tumour vs. normal Sensitivity: 1.0000 Specificity: 0.8932 AUC: 0.9906 |
Diagnosis of gastric cancer |
Chen et al. [133] | Weakly supervised | Classification | Colorectal cancer | H&E | Training and testing: 400 WSIs | CNN classifier | Normal (including hyperplastic polyp) vs. adenoma vs. adenocarcinoma vs. mucinous adenocarcinoma vs. signet ring cell carcinoma Overall accuracy: 0.76 |
Prediction of colorectal cancer tumour grade |
Ye et al. [38] | Weakly supervised | Classification | Colorectal cancer | H&E | Training and testing: 100 WSIs | Multiple-instance CNN | With epithelial cell nuclei vs. no epithelial cell nuclei Accuracy: 0.936 Precision: 0.922 Recall: 0.960 |
Detection of colon cancer |
Sharma et al. [129] | Weakly supervised | Classification | Gastrointestinal cancer | H&E | Training and testing: 413 WSIs | Cluster-to-Conquer framework |
Celiac cancer vs. normal Accuracy: 0.862 Precision: 0.855 Recall: 0.922 F1-score: 0.887 |
Detection of gastrointestinal cancer |
Klein et al. [130] | Weakly supervised | Detection | Gastric cancer | H&E + Giemsa staining | Training: 191 H&E WSIs and 286 Giemsa-stained WSIs Validation: 71 H&E WSIs and 87 Giemsa-stained WSIs External validation: 364 H&E WSIs and 347 Giemsa-stained WSIs |
VGG+ + active learning |
H. pylori vs. no H. pylori
External validation AUC: 0.81 (H&E) + 0.92 (Giemsa-stained) |
Detection of H. pylori |
WSI = whole-slide imaging; H&E = haematoxylin and eosin; AUC = area under the curve; CI = confidence interval; TCGA = The Cancer Genome Atlas; STAD = stomach adenocarcinoma; DACHS = Darmkrebs: Chancen der Verhütung durch Screening; MSI = microsatellite instability; dMMR = deficient mismatch repair; TFF3 = trefoil factor 3; DSC = Dice similarity coefficient; UMM = University Medical Centre Mannheim, Heidelberg University; NCT = National Centre for Tumour Diseases; CRC = colorectal cancer; PUMCH = Peking Union Medical College Hospital; CHCAMS = Chinese Academy of Medical Sciences; H. pylori = Helicobacter pylori; IHC = immunohistochemistry; CNN = convolutional neural network.
5.1. Fully Supervised Approach
Fully supervised learning is defined by its use of labelled data to predict outcomes. Computational pathology-related research requires large amounts of pathologist-labelled data with pixel-wise annotation for each WSI for accurate model training.
5.1.1. Classification of GI Cancer
Several deep learning models have been proposed to differentiate between benign and malignant tissue for GI cancer classification. Staging of cancers is also an important parameter that affects the patient’s treatment strategy and prognosis. Recently, research has been focused on the development of high-performance deep learning models to classify normal and cancerous cases with grading. For instance, a ResNet-18 model was reported by Su et al., with an F1 score of >0.86 in terms of distinguishing between poorly differentiated adenocarcinoma and well-differentiated adenocarcinoma [28]. DeepLab is another type of CNN architecture utilized by Song et al. to classify different types of GI cancers and showed promising performance [29,30]. In an earlier publication, Song et al. reported that DeepLab v2 achieved a high AUC (0.92) in recognizing different types of colorectal adenomas (tubular, villous, tubulovillous, etc.) [29]. Later, they utilized an updated DeepLab v3 model for diagnosis of GC with clinical applicability. Their model showed outstanding performance, with an AUC > 0.99 on multiple large external validation datasets, including various difficult cases [30].
Despite research reporting a reasonably reliable model, there is a demand for higher accuracy in differentiating between grades of GI cancer. Alternative novel network structures applied in other types of cancer classification can act as a reference for future model training for GI cancers. Graph transformer network is an example that has demonstrated a high accuracy of 0.97 in the grading of lung carcinoma [111]. Moreover, as WSIs are characterized by high morphological heterogeneity in the shape and scale of tissues, simple tiling of the WSI may compromise spatial information, which may be of benefit in accurately predicting outcomes. A deformable conditional random field (DCRF) model was developed to study the interrelationship by obtaining offsets and weights of neighbouring tiled images. Integration with well-established models, such as ResNet-18, ResNet34 and DenseNet, resulted in improvements in classification accuracy of gastric adenocarcinoma and CRC [112].
Moreover, during the peer-review process of this research, the issue of missed diagnosis by pathologists was identified and discussed. We found that two cases of malignancy were initially missed by pathologists in the initial diagnostic report. However, reassessment by a trained DeepLab v3 AI assistance system identified the malignancy and successfully flagged the tumour regions. This demonstrates that diagnoses by pathologists are not 100% accurate, are subject to interobserver variability and are prone to error when pathologists are facing large workloads and stress. Given the compromised accuracy of diagnoses made solely by pathologists, concerns have been raised about the reliability of using pathologists to train deep learning systems. Interestingly, several studies have shown that deep learning algorithms trained by pathologists exceeded the average performance of pathologists [30,113]. This means that AI systems can analyse and integrate training sets with satisfactory performance, even if all of the annotations by pathologists designated for training might not be fully accurate.
DL-based algorithms are also used for the differentiation of molecular subtypes. Sirinukunwattana et al. [114] developed a deep learning model (imCMS) for the differentiation of four consensus molecular subtype groups with distinct clinical behaviours and underpinning biology from standard H&E-stained WSIs. A similar concept was implemented by Popovici et al. for the classification of low-risk subtypes and high-risk subtypes [115].
The recognition of different types of colorectal polyps (hyperplastic, sessile serrated, traditional serrated adenoma, etc.), cancer-associated tissues (lymphocytes, cancer-associated stroma, colorectal adenocarcinoma epithelium, etc.) and normal tissues is critical to determine the risk of CRC and future rates of surveillance for patients. This has been demonstrated by various research groups using ResNet-152 [116], VGG19 [117] and the ensembled ResNet [118]. Among these models, VGG19 [117] showed the highest model accuracy of 0.943, followed by ResNet-152 (0.93) [116] and ensembled ResNet [118] (0.87). VGG19 (143 million parameters) had a much larger network size than that of ResNet-152 (60 million parameters). It is clear that larger and more complex models with more parameters play an important role in boosting model performance. As a comparable example, the Inception V3 model used by Popovici et al. [115] obtained an average accuracy of 0.72, whereas the VGG-F model used by Sirinukunwattana et al. [114] achieved a greater accuracy of 0.84. This can be explained by the fact that VGG-F (61 million parameters) is more complex than Inception V3 (23 million parameters) with respect to model enhancement. Moreover, a large dataset for model training is also essential for improved prediction results. For example, Gupta et al. [119] used the Inception-ResNet-v2 model to obtain superior performance (F-Score, 0.99; AUC, 0.99) in classifying and locating abnormal and normal tissue regions based on a model trained with more than 1,000,000 WSI patches. Despite the relatively low network complexity of the Inception-ResNet-v2 model (56 million parameters), network performance can be compensated with the support of a large amount of training data.
5.1.2. Segmentation in GI Cancer
Determining the spatial location of cancerous regions in WSI is an important step for accurate diagnosis by pathologists using digital pathology. Various networks have been applied to delineate malignant and benign regions, such as adversarial CAC-UNet [31], U-Net-16 [32] and CoUNet [33]. Remarkably, these networks are trained (750 WSIs) and tested (250 WSIs) with the same DigestPath dataset for detection of the location of gastric cancerous lesions. The adversarial CAC-UNet [31] achieved the highest segmentation performance, with a Dice coefficient score of 0.8749 as compared with U-Net-16 [32], with a Dice coefficient score of 0.7789, and CoUNet [33], with a Dice coefficient score of 0.746. U-Net-16 and CoUNet sequentially classified the corresponding colorectal malignant and benign regions successfully (AUC: 1.00 vs. 0.980), demonstrating that pipelines following segmentation and subsequent classification can potentially achieve high performance in the classification of specific histopathological regions. Ensembling different networks together is another method for performing segmentation tasks in CRC. One such example was reported by Mahendra et al. [120], who developed a general framework combining DenseNet-121, Inception-ResNet-V2 and DeeplabV3Plus models to specifically segment malignant and benign regions.
5.1.3. Detection of Prognostic Markers and Prognosis in GI Cancer
Detecting and identifying prognostic markers is important for the development of the most suitable treatment plan for patients. However, histology WSIs, by nature, contain highly heterogenous cell types, inevitably increasing the dataset difficulty and hindering the model development process of new algorithms to predict outcomes with high prognostic value. Despite these challenges, several models have been proposed to aid in the early detection of prognostic markers that may influence disease progression. The presence of goblet cells with overexpression of TFF3 is a key feature of Barrett’s oesophagus. The identification and quantification of these cells may be indicative of intestinal metaplasia of Barrett’s oesophagus. The VGG16 model showed strong adaptability, with an AUC = 0.88 with respect to detection of patients with Barrett’s oesophagus in TFF3-stained biopsies. Moreover, this approach allowed biopsies to be defined into eight classes with varying priority for pathologists to review. This method can be applied to prioritize important cases for early case review in a busy AP laboratory [34].
Infection with H. pylori can lead to chronic inflammation, which is a risk factor for GC. However, H. pylori is not easily identified, as the H&E section requires examination under high magnification. This significantly increases the cost and TAT of diagnosis. Hence, early identification of H. pylori is a valuable prognostic marker that allows for early and cost-effective treatment to eradicate H. pylori and prevent progression to GC. Several DL-based algorithms were tested by combining DenseNet-121 and ResNet-18 models [35]. Overall, this ensemble method achieved satisfactory performance, with accuracy >0.90 in detecting H. pylori in human gastric WSIs.
Similarly, identifying microsatellite instability (MSI) in patients with CRC can affect the treatment strategy. However, not all laboratories have the resources to perform MSI testing for each patient. The development of algorithms to directly predict MSI from H&E-stained slides may allow for prediction of MSI status without additional genetic or IHC tests. Shufflenet [36] and ResNet18 with a single dataset [121] and ResNet18 [122] with multiple source datasets were trained and tested for their performance. The models were trained with 6406, 429 and 1230 WSIs, achieving AUCs of 0.92, 0.8848 and 0.69–0.84, respectively. Overall, Shufflenet demonstrated higher performance than the ResNet18 models. Despite the higher WSI number of ResNet18 with multiple source data, its performance was lower than that of ResNet18 trained with a smaller and single database. Overall, the intention of developing a deep-learning-based algorithm is to predict MSI status, with the aim of providing immunotherapy to more patients without the additional cost of MSI testing. It is worth noting that large and reliable WSI databases are essential for the development of highly accurate and performant algorisms.
Similar tactics have been applied to identify MSI in GC. Su et al. reported the use of ResNet-18 to detect MSI status in GC [28]. Their model performance showed poor accuracy of 0.7727 in reporting patient-level MSI status. This is limited by the lack of well-defined regions of MSI status at the tile level. Thus, their study reveals the necessity to obtain accurate pixel-level annotated MSI labels from tumour tissue regions for better algorithm development for improved MSI detection.
Additionally, deep learning algorithms are capable of predicting prognosis in GI cancer by estimating patient survival outcome. Bychkov et al. used the VGG16 model trained with H&E slides only to predict the 5-year disease-specific survival of patients with CRC into either low- or high-risk groups [123]. Their results demonstrated the superior performance of DL-based algorithms compared to manual visual inspection in the stratification of low- and high-risk patients, with a hazard ratio of 2.3 (95% CI: 1.79–3.03). Moreover, the use of different IHC markers to assess the risk for cancer-specific death was demonstrated by Meier et al. Tissue slides were stained with a panel of markers, including Ki67, CD8, CD20 and CD68, and were used as input to train GoogLeNet. Their results showed that a combination of these markers can support the prediction of 5-year survival classification in patients (Ki67 and CD20 vs. CD20 and CD68; hazard ratio: 1.47 vs. 1.33). These works demonstrate the potential of using DL-based algorithms in a clinical setting by adding value to field-of-precision oncology [124].
5.2. Weakly Supervised Learning
Weakly supervised learning refers to the use of data labelled with a noisy or imprecise source to predict outcome. Due to increasing challenges associated with algorithm development in computational pathology, such conditions will continue to increase the burden of production and access to larger datasets with time-consuming annotation for model training. To maximize algorithm learning ability from data from limited annotations, multi-instance learning strategies are widely adopted in weakly supervised learning algorithms, especially in the field of computation pathology [125]. The general assumption of a multi-instance learning strategy is that the same histopathological information exists between a WSI and all its patch images. For instance, in a WSI with MSI, all generated tiled images are labelled as “with MSI” [126].
By implying such a straightforward assumption, considerable success has been achieved with a variety of techniques, such as recalibrated multi-instance learning [127] for classification of cancer and dysplasia in GC, graph convolutional network-based multi-instance learning [37] for identification of the presence of tumours in gastric biopsies, S3TA multi-instance learning method [38] for prediction of the presence of epithelial cell nuclei in CRC and GastroMIL [39] for GC diagnosis and risk prediction. Other than GI cancer, multi-instance learning also shows its strong adaptability for cancer tissue classification, such as dual-stream multiple instance learning network [40] and TransMIL [128] for recognition of tumours in lung or breast cancer and cluster-to-conquer [129] for identification of celiac cancer. It is anticipated that multi-instance learning will remain popular for the classification of GI cancer. Beyond the widely used multi-instance learning strategy, other novel learning strategies, network structures and loss function are proposed to perform classification tasks in GI cancer. Klein et al. [130] proposed the use of a VGG architectural active learning network to detect patients with H. pylori and obtained an outstanding AUC (>0.80). Li et al. [131] proposed a definition of a new hybrid supervised learning method combining pixel-level and image-level annotations and achieved exceptional performance (sensitivity, 1.0000; AUC, 0.9906) for the recognition of GC using the ResNet-34 and Otsu methods. Kwabena et al. [132] proposed the use of dual horizontal squash capsule networks for the identification of malignant and benign regions in CRC and obtained a high AUC (0.998). Chen et al. [133] proposed a new loss function, namely rectified cross-entropy and upper transition loss for the prediction of CRC tumour grade and obtained an average accuracy of 0.76. The study showed that the new learning strategy and network structure were more effective than the optimization of loss function in weakly supervised learning. Furthermore, reduced performance would occur when applying the trained model to multiple datasets, especially for weakly supervised learning algorithms. To overcome such challenges, colour normalization and adversarial learning [134] can be applied to train weakly supervised algorithms from different datasets for cancer grading.
6. Clinical Insight for Selected AI Applications in Early Diagnosis and Monitoring of the Progression of GI Cancer
The application of AI in GI cancer is an important and rapidly growing area of research. Early diagnosis and monitoring of GI cancer can minimize cancer incidence, lower mortality rates and reduce the cost of care. Effective clinical management is a prerequisite to achieve these goals, and the incorporation of AI in healthcare systems is indispensable. Many AI-based software applications and algorithms have been developed to assist in the detection of GI premalignant lesions. In addition, several investigations have shown impact through advanced AI algorithms for tumour subtyping by providing valuable information that may affect subsequent treatment planning and prediction of patient outcomes. Finally, promising results have been demonstrated by various intelligential WSI analytical neural networks for the screening of GI cancer, ranging from AI monitoring to providing rapid analysis and diagnosis of GI cancer, optimizing the workflow in an AP laboratory. Based on these promising results, the possible practical applications of AI for histopathological analysis will also be highlighted below.
6.1. Detection of Early GI Premalignant Lesions
Early detection of precancerous lesions is crucial to identify certain GI cancers, especially in GC and CRC [135], which are associated with high incidence and mortality. As chronic H. pylori infection (Figure 3A) is a major risk factor for GC, evaluation of H. pylori is critical in gastric biopsies. Identification of these organisms requires examination of the slide under high magnification (400×). Although this bacterium can be easily identified in H&E-stained samples if present in large quantities, identifying cases with scanty numbers of H. pylori can be extremely time-consuming. Although ancillary staining (e.g., Warthin–Starry staining, IHC) is possible, these methods require extra cost and time for diagnosis. Deep learning assistance [35,130] can further speed up GC diagnosis and provide alternative solutions to low-resource settings, where ancillary stains are not readily available. Klein et al. [130] designed VGG-based algorithms under an active learning scheme. By utilizing algorithms that are able to proactively learn and train from standard GC histology images with H. pylori, the required amount of pixelwise labelling annotation by pathologists can be significantly reduced, eliminating the need for large, annotated image datasets established by pathologists. Therefore, only a small portion of manual annotation will be required to develop AI applications for detection of H. pylori in order to self-generate annotation of regions with H. pylori.
Unlike GC, differentiating high-risk colorectal polyps from low-risk polyps (e.g., innocuous hyperplastic polyps) is an important task in CRC screening (Figure 3J). As identifying high-risk polyps is based on characterization of specific types of polyps (e.g., sessile serrated polyps) [136] and considerable interobserver variability exists among pathologists [137,138,139], accurate diagnosis of high-risk polyps is needed for effective and early detection of CRC. As all US adults aged 50–75 years old are recommended by the US Preventive Services Task Force [140] to screen for CRC yearly, AI applications [29,116,118] for classification of high-risk colorectal polyps could demonstrate its clinical utility in prioritising slides with higher likelihood of malignancy. Wei et al. [118] demonstrated that their AI algorithms for colorectal polyp detection can further highlight regions of high-risk polyps on WSIs to provide effective, consistent and accurate diagnoses, demonstrating that AI systems, such as computer-aided diagnosis, can possibly be developed to assist pathologists in the interpretation of CRC WSIs. However, as polyp annotations were provided by several pathologists for AI model training, similar errors were found in both pathologist and AI applications with respect to classifying high-risk polyps and low-risk polyps. This suggests that future AI systems for colorectal polyp detection may require additional development with extra manual annotations by different experienced GI pathologists to reduce errors in classification.
Currently, diagnostic procedures still rely heavily on pattern recognition by pathologists to identify regions of interests. Future AI algorithms will guide pathologist by identifying areas of interest by highlighting H. pylori and high-risk polyps for confirmation. Given the promising performance of AI algorithms in accurate delineation of premalignant lesion regions, pathologists might integrate computational pathology to adopt AI-based tools in clinical workflows with increased confidence. Eventually, this will reduce the time spent per case and allow for early detection and diagnosis of patients with GI premalignant lesions.
6.2. Tumour Subtyping Characterization and Estimation of Patient Outcome
In histopathology, the identification and classification of tumour subtypes is important for personalized medicine [141]. As certain gene mutations and molecular alterations are associated with specific morphological changes [142], deep-learning-based image analysis has the potential to uncover molecular tumour subtyping and build robust classifiers to enhance treatment response in cancer patients. MSI refers to DNA mismatch repair deficiency resulting in accumulation of mutations within short repetitive sequences of DNA (microsatellites). MSI testing is critical for treatment of GI cancer [143], as CRCs with high MSI are associated with poor response to conventional chemotherapy but improved response to immunotherapy, [144,145] leading to improved median overall survival [146]. However, due to the high cost of PCR-based MSI testing [147], many studies have focused on developing deep learning methods based on H&E imaging to provide fast and accurate MSI detection for GC [28,122] and CRC [36,121,122] in countries with limited resources, especially in Latin America and the Caribbean [148]. Due to the lack of exact 3D quantitative MSI information, training deep learning models without accurate information may limit the prognostic performance in GI cancer patients with MSI. Despite imperfect specificity, the increasing amount of MSI related-AI research indicates that AI will be rapidly applied to screen for MSI in GI cancers to lower laboratory operation costs.
Assessment of survival outcomes in GI cancer is another ongoing research area using AI. Many studies have explored the question of whether deep learning applications can be used as prognostic tools for GC [124] (Figure 3C) and CRC [27,117,123] (Figure 3K,L) based on H&E-stained WSIs. Using deep learning, suspected GI tumour tissue (e.g., tissue microarrays [27,123,124] and stroma [117]) can be assessed quantitatively using H&E-stained WSIs to assist pathologists in identifying patients at high risk of mortality for immediate advanced treatment. Furthermore, some pioneering studies [114,115] have attempted to explore the classification of consensus molecular subtype (CMS) classification of CRC and patient survival outcomes using deep learning. With the capacity to identify tumour subtypes and perform risk assessment of cancer patients, AI applications can provide a cost-effective and time-saving solution for clinical decision making. Guided by AI with high reproducibility and objectivity, more clinically relevant information can be generated from WSIs to allow for improved clinical management to enhance the survival of GI cancer patients.
Performing supplementary laboratory tests, including PCR and IHC tests, remains the gold standard to determine the molecular tumour subtype profile of GI cancer. Advancements in AI algorithms allow for application in clinical settings with limited resources to detect specific biomarkers, such as MSI, using commonly available H&E staining images, facilitating early personalized treatment. Additionally, AI application will allow for improved allocation of medical resource by shortening the time required to identify GI cancer patients at high risk of mortality for earlier treatment.
6.3. Screening GI Cancer in Daily Clinical Operation
As clinical follow up is essential to reduce cancer mortality [59], effective pathology service must start by providing early detection and accurate cancer diagnosis. The shortage of pathologists continues to worsen globally, resulting in delayed cancer diagnosis and treatment, especially during the recent COVID-19 pandemic health crisis [149,150,151]. The pressure to provide fast and accurate diagnoses, with overloaded histopathology workforces, has forced the transformation of practice from conventional light microscopy to digital pathology with AI analytical applications. Studies have focused on improving clinical workflow by diagnosing various GI cancers, including oesophageal cancer [34] (Figure 3E–H), GC [30,39] (Figure 3C,I), etc. Pertinently, Gehrung et al. [34] established an H&E-based trained deep learning system to identify and triage oesophageal cases with Barrett’s oesophagus (BO) for early detection of oesophageal adenocarcinoma. The AI application resulted in a workload reduction of more than 55% of for BO identification after triage and further demonstrated that the developed AI system can improve operational efficiency and support clinical decision making in a realistic primary care setting. However, the deep learning system developed by Gehrung et al. [34] did not tackle the problem of successful training algorithms requiring large annotated WSI datasets established by pathologists [16]. Campanella et al. [152] demonstrated a weakly supervised system trained by diagnostic data that can be automatically retrieved from pathology reports to triage prostate, skin and breast cancer cases. A workload reduction of more than 65% [152] was observed. As weakly supervised algorithms have proven their value in cancer screening, efforts should be made to develop AI applications to triage and identify GI cancers for daily operation without further overloading the histopathology workforce.
At present, histological cases are reviewed in chronological order in traditional AP laboratories. Clinical needs and the current acceptability of AI have not been well-studied and discussed. One of the ways to demonstrate the clinical usefulness of AI is to measure the potential workload reduction associated with automatic review of benign/disease-free cases. With the utilization of AI for case screening, cases susceptible to GI cancer can be flagged for rapid review, and cases with no potential indication of GI cancer can be semiautomatically reviewed (Figure 3B,D). By measuring the change in pathologist workload in prospective trials [153], the clinical usefulness of AI can be assessed, enabling earlier and improved treatment planning. Furthermore, using slide-level diagnostics in a weakly supervised setting is also recommended for AI algorithm development. This could also alleviate the overwhelming workload issue in AP laboratories by freeing pathologists from the generation of pixel-level annotation for future algorithmic development used in clinical settings.
7. Conclusions
In this review, we outlined the potential of AI for the histological diagnosis of GI cancers. The recent increase in demand and complexity of GI cancer diagnosis has prompted the incorporation of digital pathology into the diagnostic workflow. Digital pathology allows pathology information to be acquired, managed and interpreted in a digital environment, opening up opportunities for computational analysis using AI. We presented the applications of the two major AI methodologies, i.e., machine learning and deep learning, for the diagnosis of GI cancers. Various machine learning algorithms show promising sensitivity and specificity in diagnosing GC. In particular, the most popular machine learning algorithm, support vector machine, demonstrates high robustness and generalization ability. Deep learning, which uses highly complex algorithms to simulate the human neural network, is also a powerful tool in digital pathology. We summarized the performance characteristics and limitations of fully supervised and weakly supervised approaches in terms of classification, segmentation, detection and prognostication of GI cancer. We also presented clinical insights with respect to how AI could facilitate the early detection of lesions, tumour subtyping and widespread screening for cancers. Despite the merits offered by AI, the major challenges of algorithm development, such as variation in histological image data, interpathologist variability with respect to interpretations, model transparency and interpretability in clinical settings, have to be overcome. Although there is an urgent need for AI development due to the overwhelming workload of pathologists, it requires pathologists to spend extra time to provide annotations. This dilemma is remains unsolved. From a clinical perspective, solid external validation and quality controls with reference to a large dataset are important in ensuring acceptable standards of AI. This warrants continuous studies on model design by relying on patch/pixel-level annotation, explainability and generalizability of AI algorithms with the inclusion of variability of the multiethnic populations. Through rigid comparison of the current AI model for GI cancer, our studies have highlighted the current challenges associated with developing AI for histological analysis, providing a detailed summary of current AI algorithms developed for GI cancer and spotlighting clinical insights for future development of GI cancer AI algorithms. As AI algorithms continue to advance, we believe that the transparency of such application will also improve. We also believed that the clinical usefulness of AI could be demonstrated though prospective trials. Thus, the adoption of such conceptual AI will be transformed into effective applications of computational pathology for clinical practice.
Author Contributions
Writing—original draft preparation, A.N.N.W., Z.H., K.L.L., C.C.K.T., C.Y.W., A.Z.C., M.D.L. and M.H.Y.Y.; writing—review and editing, A.N.N.W., Z.H., S.C.C.W., J.S.Y., C.K.R.C., A.Z.C., M.D.L. and M.H.Y.Y.; supervision, M.H.Y.Y. All authors have read and agreed to the published version of the manuscript.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
Data sharing not applicable.
Conflicts of Interest
The authors declare no conflict of interest.
Funding Statement
This study was supported by Departmental Seed Fund for M.H.Y.Y. from The Hong Kong Polytechnic University. Grant Number: DDP.
Footnotes
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Arnold M., Abnet C.C., Neale R.E., Vignat J., Giovannucci E.L., McGlynn K.A., Bray F. Global Burden of 5 Major Types of Gastrointestinal Cancer. Gastroenterology. 2020;159:335–349.e15. doi: 10.1053/j.gastro.2020.02.068. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Islami F., Goding Sauer A., Miller K.D., Siegel R.L., Fedewa S.A., Jacobs E.J., McCullough M.L., Patel A.V., Ma J., Soerjomataram I., et al. Proportion and number of cancer cases and deaths attributable to potentially modifiable risk factors in the United States. CA Cancer J. Clin. 2018;68:31–54. doi: 10.3322/caac.21440. [DOI] [PubMed] [Google Scholar]
- 3.Sung H., Siegel R.L., Rosenberg P.S., Jemal A. Emerging cancer trends among young adults in the USA: Analysis of a population-based cancer registry. Lancet Public Health. 2019;4:e137–e147. doi: 10.1016/S2468-2667(18)30267-6. [DOI] [PubMed] [Google Scholar]
- 4.Yan H., Sellick K. Symptoms, psychological distress, social support, and quality of life of Chinese patients newly diagnosed with gastrointestinal cancer. Cancer Nurs. 2004;27:389–399. doi: 10.1097/00002820-200409000-00009. [DOI] [PubMed] [Google Scholar]
- 5.Kamel H.M. Trends and challenges in pathology practice: Choices and necessities. Sultan Qaboos Univ. Med. J. 2011;11:38. [PMC free article] [PubMed] [Google Scholar]
- 6.Rao G.G., Crook M., Tillyer M. Pathology tests: Is the time for demand management ripe at last? J. Clin. Pathol. 2003;56:243–248. doi: 10.1136/jcp.56.4.243. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Hassell L.A., Parwani A.V., Weiss L., Jones M.A., Ye J. Challenges and opportunities in the adoption of College of American Pathologists checklists in electronic format: Perspectives and experience of Reporting Pathology Protocols Project (RPP2) participant laboratories. Arch. Pathol. Lab. Med. 2010;134:1152–1159. doi: 10.5858/2009-0386-OA.1. [DOI] [PubMed] [Google Scholar]
- 8.Hewitt S.M., Lewis F.A., Cao Y., Conrad R.C., Cronin M., Danenberg K.D., Goralski T.J., Langmore J.P., Raja R.G., Williams P.M. Tissue handling and specimen preparation in surgical pathology: Issues concerning the recovery of nucleic acids from formalin-fixed, paraffin-embedded tissue. Arch. Pathol. Lab. Med. 2008;132:1929–1935. doi: 10.5858/132.12.1929. [DOI] [PubMed] [Google Scholar]
- 9.Ribé A., Ribalta T., Lledó R., Torras G., Asenjo M.A., Cardesa A. Evaluation of turnaround times as a component of quality assurance in surgical pathology. Int. J. Qual. Health Care. 1998;10:241–245. doi: 10.1093/intqhc/10.3.241. [DOI] [PubMed] [Google Scholar]
- 10.Tamil S.M., Srinivas A. Evaluation of quality management systems implementation in medical diagnostic laboratories benchmarked for accreditation. J. Med. Lab. Diagn. 2015;6:27–35. doi: 10.5897/JMLD2015.0104. [DOI] [Google Scholar]
- 11.Peter T.F., Rotz P.D., Blair D.H., Khine A.-A., Freeman R.R., Murtagh M.M. Impact of laboratory accreditation on patient care and the health system. Am. J. Clin. Pathol. 2010;134:550–555. doi: 10.1309/AJCPH1SKQ1HNWGHF. [DOI] [PubMed] [Google Scholar]
- 12.The Royal College of Pathologists Digital Pathology. [(accessed on 20 March 2022)]. Available online: https://www.rcpath.org/profession/digital-pathology.html.
- 13.Williams B.J., Hanby A., Millican-Slater R., Nijhawan A., Verghese E., Treanor D. Digital pathology for the primary diagnosis of breast histopathological specimens: An innovative validation and concordance study on digital pathology validation and training. Histopathology. 2018;72:662–671. doi: 10.1111/his.13403. [DOI] [PubMed] [Google Scholar]
- 14.Williams B.J., Jayewardene D., Treanor D. Digital immunohistochemistry implementation, training and validation: Experience and technical notes from a large clinical laboratory. J. Clin. Pathol. 2019;72:373–378. doi: 10.1136/jclinpath-2018-205628. [DOI] [PubMed] [Google Scholar]
- 15.Bera K., Schalper K.A., Rimm D.L., Velcheti V., Madabhushi A. Artificial intelligence in digital pathology—new tools for diagnosis and precision oncology. Nat. Rev. Clin. Oncol. 2019;16:703–715. doi: 10.1038/s41571-019-0252-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Van der Laak J., Litjens G., Ciompi F. Deep learning in histopathology: The path to the clinic. Nat. Med. 2021;27:775–784. doi: 10.1038/s41591-021-01343-4. [DOI] [PubMed] [Google Scholar]
- 17.Tizhoosh H.R., Pantanowitz L. Artificial Intelligence and Digital Pathology: Challenges and Opportunities. J. Pathol. Inform. 2018;9:38. doi: 10.4103/jpi.jpi_53_18. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Kashyap A., Khartchenko A.F., Pati P., Gabrani M., Schraml P., Kaigala G.V. Quantitative microimmunohistochemistry for the grading of immunostains on tumour tissues. Nat. Biomed. Eng. 2019;3:478–490. doi: 10.1038/s41551-019-0386-3. [DOI] [PubMed] [Google Scholar]
- 19.Arar N.M., Pati P., Kashyap A., Khartchenko A.F., Goksel O., Kaigala G.V., Gabrani M. Computational immunohistochemistry: Recipes for standardization of immunostaining. In: Descoteaux M., Maier-Hein L., Franz A., Jannin P., Collins D., Duchesne S., editors. Medical Image Computing and Computer-Assisted Intervention−MICCAI 2017. Springer; Cham, Switzerland: 2017. pp. 48–55. (Lecture Notes in Computer Science). [Google Scholar]
- 20.Arar N.M., Pati P., Kashyap A., Khartchenko A.F., Goksel O., Kaigala G.V., Gabrani M. High-quality immunohistochemical stains through computational assay parameter optimization. IEEE Trans. Biomed. Eng. 2019;66:2952–2963. doi: 10.1109/TBME.2019.2899156. [DOI] [PubMed] [Google Scholar]
- 21.Srinivasu P.N., SivaSai J.G., Ijaz M.F., Bhoi A.K., Kim W., Kang J.J. Classification of Skin Disease Using Deep Learning Neural Networks with MobileNet V2 and LSTM. Sensors. 2021;21:2852. doi: 10.3390/s21082852. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Kelly C.J., Karthikesalingam A., Suleyman M., Corrado G., King D. Key challenges for delivering clinical impact with artificial intelligence. BMC Med. 2019;17:195. doi: 10.1186/s12916-019-1426-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Vulli A., Srinivasu P.N., Sashank M.S.K., Shafi J., Choi J., Ijaz M.F. Fine-Tuned DenseNet-169 for Breast Cancer Metastasis Prediction Using FastAI and 1-Cycle Policy. Sensors. 2022;22:2988. doi: 10.3390/s22082988. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Yoshida H., Shimazu T., Kiyuna T., Marugame A., Yamashita Y., Cosatto E., Taniguchi H., Sekine S., Ochiai A. Automated histological classification of whole-slide images of gastric biopsy specimens. Gastric Cancer. 2017;21:249–257. doi: 10.1007/s10120-017-0731-8. [DOI] [PubMed] [Google Scholar]
- 25.Yasuda Y., Tokunaga K., Koga T., Sakamoto C., Goldberg I.G., Saitoh N., Nakao M. Computational analysis of morphological and molecular features in gastric cancer tissues. Cancer Med. 2020;9:2223–2234. doi: 10.1002/cam4.2885. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Cosatto E., Laquerre P.-F., Malon C., Graf H.P., Saito A., Kiyuna T., Marugame A., Kamijo K. Automated gastric cancer diagnosis on H&E-stained sections; ltraining a classifier on a large scale with multiple instance machine learning. Proc. SPIE. 2013;8676:867605. doi: 10.1117/12.2007047. [DOI] [Google Scholar]
- 27.Jiang D., Liao J., Duan H., Wu Q., Owen G., Shu C., Chen L., He Y., Wu Z., He D., et al. A machine learning-based prognostic predictor for stage III colon cancer. Sci. Rep. 2020;10:10333. doi: 10.1038/s41598-020-67178-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Su F., Li J., Zhao X., Wang B., Hu Y., Sun Y., Ji J. Interpretable tumor differentiation grade and microsatellite instability recognition in gastric cancer using deep learning. Lab. Investig. 2022;102:641–649. doi: 10.1038/s41374-022-00742-6. [DOI] [PubMed] [Google Scholar]
- 29.Song Z., Yu C., Zou S., Wang W., Huang Y., Ding X., Liu J., Shao L., Yuan J., Gou X., et al. Automatic deep learning-based colorectal adenoma detection system and its similarities with pathologists. BMJ Open. 2020;10:e036423. doi: 10.1136/bmjopen-2019-036423. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Song Z., Zou S., Zhou W., Huang Y., Shao L., Yuan J., Gou X., Jin W., Wang Z., Chen X., et al. Clinically applicable histopathological diagnosis system for gastric cancer detection using deep learning. Nat. Commun. 2020;11:1–9. doi: 10.1038/s41467-020-18147-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Zhu C., Mei K., Peng T., Luo Y., Liu J., Wang Y., Jin M. Multi-level colonoscopy malignant tissue detection with adversarial CAC-UNet. Neurocomputing. 2021;438:165–183. doi: 10.1016/j.neucom.2020.04.154. [DOI] [Google Scholar]
- 32.Feng R., Liu X., Chen J., Chen D.Z., Gao H., Wu J. A Deep Learning Approach for Colonoscopy Pathology WSI Analysis: Accurate Segmentation and Classification. IEEE J. Biomed. Health Inform. 2021;25:3700–3708. doi: 10.1109/JBHI.2020.3040269. [DOI] [PubMed] [Google Scholar]
- 33.Xu W., Liu H., Wang X., Ouyang H., Qian Y. ICVIP 2020: 2020 The 4th International Conference on Video and Image Processing, 25–27 December 2020, Xi’an China. Association for Computing Machinery; New York, NY, USA: 2020. CoUNet: An End-to-End Colonoscopy Lesion Image Segmentation and Classification Framework; pp. 81–87. [Google Scholar]
- 34.Gehrung M., Crispin-Ortuzar M., Berman A.G., O’Donovan M., Fitzgerald R.C., Markowetz F. Triage-driven diagnosis of Barrett’s esophagus for early detection of esophageal adenocarcinoma using deep learning. Nat. Med. 2021;27:833–841. doi: 10.1038/s41591-021-01287-9. [DOI] [PubMed] [Google Scholar]
- 35.Zhou S., Marklund H., Bláha O., Desai M., Martin B.A., Bingham D.B., Berry G.J., Gomulia E., Ng A., Shen J. Deep learning assistance for the histopathologic diagnosis of Helicobacter pylori. Intell. Med. 2020;1–2:100004. doi: 10.1016/j.ibmed.2020.100004. [DOI] [Google Scholar]
- 36.Echle A., Grabsch H.I., Quirke P., van den Brandt P.A., West N.P., Hutchins G.G.A., Heij L.R., Tan X., Richman S.D., Krause J., et al. Clinical-Grade Detection of Microsatellite Instability in Colorectal Tumors by Deep Learning. Gastroenterology. 2020;159:1406–1416.e11. doi: 10.1053/j.gastro.2020.06.021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Xiang X., Wu X. Multiple Instance Classification for Gastric Cancer Pathological Images Based on Implicit Spatial Topological Structure Representation. Appl. Sci. 2021;11:10368. doi: 10.3390/app112110368. [DOI] [Google Scholar]
- 38.Ye T., Lan R., Luo X. Multiple-instance CNN Improved by S3TA for Colon Cancer Classification with Unannotated Histopathological Images; Proceedings of the 2021 11th International Conference on Intelligent Control and Information Processing (ICICIP); Dali, China. 3–7 December 2021; pp. 444–448. [Google Scholar]
- 39.Huang B., Tian S., Zhan N., Ma J., Huang Z., Zhang C., Zhang H., Ming F., Liao F., Ji M., et al. Accurate diagnosis and prognosis prediction of gastric cancer using deep learning on digital pathological images: A retrospective multicentre study. EBioMedicine. 2021;73:103631. doi: 10.1016/j.ebiom.2021.103631. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Li B., Li Y., Eliceiri K.W. Dual-stream Multiple Instance Learning Network for Whole Slide Image Classification with Self-supervised Contrastive Learning; Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); Nashville, TN, USA. 20–25 June 2021; pp. 14313–14323. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41.Tellez D., Litjens G., Bándi P., Bulten W., Bokhorst J.-M., Ciompi F., van der Laak J. Quantifying the effects of data augmentation and stain color normalization in convolutional neural networks for computational pathology. Med. Image Anal. 2019;58:101544. doi: 10.1016/j.media.2019.101544. [DOI] [PubMed] [Google Scholar]
- 42.Kim D., Pantanowitz L., Schuffler P., Yarlagadda D.V.K., Ardon O., Reuter V.E., Hameed M., Klimstra D.S., Hanna M.G. (Re) Defining the High-Power Field for Digital Pathology. J. Pathol. Inform. 2020;11:33. doi: 10.4103/jpi.jpi_48_20. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Dun X.-p., Parkinson D.B. Visualizing peripheral nerve regeneration by whole mount staining. PLoS ONE. 2015;10:e0119168. doi: 10.1371/journal.pone.0119168. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Heffner S., Colgan O., Doolan C. Digital Pathology. [(accessed on 30 April 2022)]. Available online: https://www.leicabiosystems.com/en-br/knowledge-pathway/digital-pathology/
- 45.Borowsky A.D., Glassy E.F., Wallace W.D., Kallichanda N.S., Behling C.A., Miller D.V., Oswal H.N., Feddersen R.M., Bakhtar O.R., Mendoza A.E., et al. Digital Whole Slide Imaging Compared with Light Microscopy for Primary Diagnosis in Surgical Pathology. Arch. Pathol. Lab. Med. 2020;144:1245–1253. doi: 10.5858/arpa.2019-0569-OA. [DOI] [PubMed] [Google Scholar]
- 46.Snead D.R., Tsang Y.W., Meskiri A., Kimani P.K., Crossman R., Rajpoot N.M., Blessing E., Chen K., Gopalakrishnan K., Matthews P., et al. Validation of digital pathology imaging for primary histopathological diagnosis. Histopathology. 2016;68:1063–1072. doi: 10.1111/his.12879. [DOI] [PubMed] [Google Scholar]
- 47.Hanna M.G., Reuter V.E., Ardon O., Kim D., Sirintrapun S.J., Schüffler P.J., Busam K.J., Sauter J.L., Brogi E., Tan L.K., et al. Validation of a digital pathology system including remote review during the COVID-19 pandemic. Mod. Pathol. 2020;33:2115–2127. doi: 10.1038/s41379-020-0601-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Cheng C.L., Azhar R., Sng S.H., Chua Y.Q., Hwang J.S., Chin J.P., Seah W.K., Loke J.C., Ang R.H., Tan P.H. Enabling digital pathology in the diagnostic setting: Navigating through the implementation journey in an academic medical centre. J. Clin. Pathol. 2016;69:784–792. doi: 10.1136/jclinpath-2015-203600. [DOI] [PubMed] [Google Scholar]
- 49.Aloqaily A., Polonia A., Campelos S., Alrefae N., Vale J., Caramelo A., Eloy C. Digital Versus Optical Diagnosis of Follicular Patterned Thyroid Lesions. Head Neck Pathol. 2021;15:537–543. doi: 10.1007/s12105-020-01243-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 50.Salvi M., Acharya U.R., Molinari F., Meiburger K.M. The impact of pre- and post-image processing techniques on deep learning frameworks: A comprehensive review for digital pathology image analysis. Comput. Biol. Med. 2021;128:104129. doi: 10.1016/j.compbiomed.2020.104129. [DOI] [PubMed] [Google Scholar]
- 51.Taqi S.A., Sami S.A., Sami L.B., Zaki S.A. A review of artifacts in histopathology. J. Oral Maxillofac. Pathol. 2018;22:279. doi: 10.4103/jomfp.JOMFP_125_15. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 52.Schneider C.A., Rasband W.S., Eliceiri K.W. NIH Image to ImageJ: 25 years of image analysis. Nat. Methods. 2012;9:671–675. doi: 10.1038/nmeth.2089. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Bankhead P., Loughrey M.B., Fernández J.A., Dombrowski Y., McArt D.G., Dunne P.D., McQuaid S., Gray R.T., Murray L.J., Coleman H.G., et al. QuPath: Open source software for digital pathology image analysis. Sci. Rep. 2017;7:16878. doi: 10.1038/s41598-017-17204-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 54.Aubreville M., Bertram C., Klopfleisch R., Maier A. Bildverarbeitung für die Medizin 2018. Springer; Berlin/Heidelberg, Germany: 2018. SlideRunner; pp. 309–314. [Google Scholar]
- 55.Williams B., Hanby A., Millican-Slater R., Verghese E., Nijhawan A., Wilson I., Besusparis J., Clark D., Snead D., Rakha E., et al. Digital pathology for primary diagnosis of screen-detected breast lesions—Experimental data, validation and experience from four centres. Histopathology. 2020;76:968–975. doi: 10.1111/his.14079. [DOI] [PubMed] [Google Scholar]
- 56.Eloy C., Vale J., Curado M., Polonia A., Campelos S., Caramelo A., Sousa R., Sobrinho-Simoes M. Digital Pathology Workflow Implementation at IPATIMUP. Diagnostics. 2021;11:2111. doi: 10.3390/diagnostics11112111. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.Abels E., Pantanowitz L., Aeffner F., Zarella M.D., van der Laak J., Bui M.M., Vemuri V.N., Parwani A.V., Gibbs J., Agosto-Arroyo E., et al. Computational pathology definitions, best practices, and recommendations for regulatory guidance: A white paper from the Digital Pathology Association. J. Pathol. 2019;249:286–294. doi: 10.1002/path.5331. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Hawkes N. Cancer survival data emphasise importance of early diagnosis. BMJ. 2019;364:l408. doi: 10.1136/bmj.l408. [DOI] [PubMed] [Google Scholar]
- 59.Schiffman J.D., Fisher P.G., Gibbs P. Early detection of cancer: Past, present, and future. Am. Soc. Clin. Oncol. Educ. Book. 2015;35:57–65. doi: 10.14694/EdBook_AM.2015.35.57. [DOI] [PubMed] [Google Scholar]
- 60.Williams B.J., Bottoms D., Treanor D. Future-proofing pathology: The case for clinical adoption of digital pathology. J. Clin. Pathol. 2017;70:1010–1018. doi: 10.1136/jclinpath-2017-204644. [DOI] [PubMed] [Google Scholar]
- 61.Maung R. Pathologists’ workload and patient safety. Diagn. Histopathol. 2016;22:283–287. doi: 10.1016/j.mpdhp.2016.07.004. [DOI] [Google Scholar]
- 62.Esteva A., Robicquet A., Ramsundar B., Kuleshov V., DePristo M., Chou K., Cui C., Corrado G., Thrun S., Dean J. A guide to deep learning in healthcare. Nat. Med. 2019;25:24–29. doi: 10.1038/s41591-018-0316-z. [DOI] [PubMed] [Google Scholar]
- 63.Miotto R., Wang F., Wang S., Jiang X., Dudley J.T. Deep learning for healthcare: Review, opportunities and challenges. Brief. Bioinform. 2018;19:1236–1246. doi: 10.1093/bib/bbx044. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Jiang F., Jiang Y., Zhi H., Dong Y., Li H., Ma S., Wang Y., Dong Q., Shen H., Wang Y. Artificial intelligence in healthcare: Past, present and future. Stroke Vasc. Neurol. 2017;2:230–243. doi: 10.1136/svn-2017-000101. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Holzinger A., Langs G., Denk H., Zatloukal K., Muller H. Causability and explainability of artificial intelligence in medicine. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2019;9:e1312. doi: 10.1002/widm.1312. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Whitney J., Corredor G., Janowczyk A., Ganesan S., Doyle S., Tomaszewski J., Feldman M., Gilmore H., Madabhushi A. Quantitative nuclear histomorphometry predicts oncotype DX risk categories for early stage ER+ breast cancer. BMC Cancer. 2018;18:610. doi: 10.1186/s12885-018-4448-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Hinata M., Ushiku T. Detecting immunotherapy-sensitive subtype in gastric cancer using histologic image-based deep learning. Sci. Rep. 2021;11:22636. doi: 10.1038/s41598-021-02168-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 68.Rathore S., Iftikhar M.A., Chaddad A., Niazi T., Karasic T., Bilello M. Segmentation and Grade Prediction of Colon Cancer Digital Pathology Images Across Multiple Institutions. Cancers. 2019;11:1700. doi: 10.3390/cancers11111700. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 69.Ehteshami Bejnordi B., Veta M., Johannes van Diest P., van Ginneken B., Karssemeijer N., Litjens G., van der Laak J., Hermsen M., Manson Q.F., Balkenhol M., et al. Diagnostic Assessment of Deep Learning Algorithms for Detection of Lymph Node Metastases in Women with Breast Cancer. JAMA. 2017;318:2199–2210. doi: 10.1001/jama.2017.14585. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 70.Bandi P., Geessink O., Manson Q., Van Dijk M., Balkenhol M., Hermsen M., Ehteshami Bejnordi B., Lee B., Paeng K., Zhong A., et al. From Detection of Individual Metastases to Classification of Lymph Node Status at the Patient Level: The CAMELYON17 Challenge. IEEE Trans. Med. Imaging. 2019;38:550–560. doi: 10.1109/TMI.2018.2867350. [DOI] [PubMed] [Google Scholar]
- 71.The Royal College of Pathologists of Australasia Ageing Pathologists. [(accessed on 1 May 2022)]. Available online: https://www.rcpa.edu.au/getattachment/95c190e1-bdbe-4ab1-83e1-0a218c69ad82/Ageing-Pathologists.aspx.
- 72.The Royal College of Pathologists of Australia Becoming a Pathologist. [(accessed on 30 April 2022)]. Available online: https://www.rcpa.edu.au/Pathology-Careers/Becoming-a-Pathologist.
- 73.Yagi Y. Color standardization and optimization in whole slide imaging. Diagn. Pathol. 2011;6((Suppl. S1)):S15. doi: 10.1186/1746-1596-6-S1-S15. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.Kothari S., Phan J.H., Moffitt R.A., Stokes T.H., Hassberger S.E., Chaudry Q., Young A.N., Wang M.D. Automatic batch-invariant color segmentation of histological cancer images; Proceedings of the 2011 IEEE International Symposium on Biomedical Imaging: From Nano to Macro; Chicago, IL, USA. 30 March–2 April 2011; pp. 657–660. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 75.Tabesh A., Teverovskiy M., Pang H.-Y., Kumar V.P., Verbel D., Kotsianti A., Saidi O. Multifeature prostate cancer diagnosis and Gleason grading of histological images. IEEE Trans. Med. Imaging. 2007;26:1366–1378. doi: 10.1109/TMI.2007.898536. [DOI] [PubMed] [Google Scholar]
- 76.Reinhard E., Adhikhmin M., Gooch B., Shirley P. Color transfer between images. IEEE Comput. Graph. Appl. 2001;21:34–41. doi: 10.1109/38.946629. [DOI] [Google Scholar]
- 77.Abe T., Murakami Y., Yamaguchi M., Ohyama N., Yagi Y. Color correction of pathological images based on dye amount quantification. Opt. Rev. 2005;12:293–300. doi: 10.1007/s10043-005-0293-6. [DOI] [Google Scholar]
- 78.Magee D., Treanor D., Crellin D., Shires M., Smith K., Mohee K., Quirke P. Colour normalisation in digital histopathology images; Proceedings of the Optical Tissue Image analysis in Microscopy, Histopathology and Endoscopy (MICCAI Workshop); London, UK. 24 September 2009; London, UK: Daniel Elson; 2009. pp. 100–111. [Google Scholar]
- 79.Macenko M., Niethammer M., Marron J.S., Borland D., Woosley J.T., Guan X., Schmitt C., Thomas N.E. A method for normalizing histology slides for quantitative analysis; Proceedings of the 2009 IEEE International Symposium on Biomedical Imaging: From Nano to Macro; Boston, MA, USA. 28 June–1 July 2009; pp. 1107–1110. [Google Scholar]
- 80.Tani S., Fukunaga Y., Shimizu S., Fukunishi M., Ishii K., Tamiya K. Color standardization method and system for whole slide imaging based on spectral sensing. Anal. Cell. Pathol. 2012;35:107–115. doi: 10.1155/2012/154735. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Niethammer M., Borland D., Marron J.S., Woosley J., Thomas N.E. Appearance normalization of histology slides. In: Wang F., Yan P., Suzuki K., Shen D., editors. MLMI 2010: Machine Learning in Medical Imaging. Volume 6357. Springer; Berlin/Heidelberg, Germany: 2010. pp. 58–66. (Lecture Notes in Computer Science). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 82.Shaban M., Baur C., Navab N., Albarqouni S. Staingan: Stain Style Transfer for Digital Histological Images; Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019); Venice, Italy. 8–11 April 2019; pp. 953–956. [Google Scholar]
- 83.Kausar T., Kausar A., Ashraf M.A., Siddique M.F., Wang M., Sajid M., Siddique M.Z., Haq A.U., Riaz I. SA-GAN: Stain Acclimation Generative Adversarial Network for Histopathology Image Analysis. Appl. Sci. 2022;12:288. doi: 10.3390/app12010288. [DOI] [Google Scholar]
- 84.Cong C., Liu S., Di Ieva A., Pagnucco M., Berkovsky S., Song Y. Texture Enhanced Generative Adversarial Network for Stain Normalisation in Histopathology Images; Proceedings of the 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI); Nice, France. 13–16 April 2021; pp. 1949–1952. [Google Scholar]
- 85.Patil A., Talha M., Bhatia A., Kurian N.C., Mangale S., Patel S., Sethi A. Fast, Self Supervised, Fully Convolutional Color Normalization of H&E Stained Images; Proceedings of the 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI); Nice, France. 13–16 April 2021; pp. 1563–1567. [Google Scholar]
- 86.Bug D., Schneider S., Grote A., Oswald E., Feuerhake F., Schüler J., Merhof D. Context-Based Normalization of Histological Stains Using Deep Convolutional Features; Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Cham, Switzerland. 9 September 2017; pp. 135–142. [Google Scholar]
- 87.Baba A.I., Câtoi C. Comparative Oncology. The Publishing House of the Romanian Academy; Bucharest, Romania: 2007. [PubMed] [Google Scholar]
- 88.Elmore J.G., Longton G.M., Carney P.A., Geller B.M., Onega T., Tosteson A.N., Nelson H.D., Pepe M.S., Allison K.H., Schnitt S.J., et al. Diagnostic concordance among pathologists interpreting breast biopsy specimens. JAMA. 2015;313:1122–1132. doi: 10.1001/jama.2015.1405. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 89.Ali F., Khan P., Riaz K., Kwak D., Abuhmed T., Park D., Kwak K.S. A fuzzy ontology and SVM–based Web content classification system. IEEE Access. 2017;5:25781–25797. doi: 10.1109/ACCESS.2017.2768564. [DOI] [Google Scholar]
- 90.Amin M.B., Greene F.L., Edge S.B., Compton C.C., Gershenwald J.E., Brookland R.K., Meyer L., Gress D.M., Byrd D.R., Winchester D.P. The Eighth Edition AJCC Cancer Staging Manual: Continuing to build a bridge from a population-based to a more “personalized” approach to cancer staging. CA Cancer J. Clin. 2017;67:93–99. doi: 10.3322/caac.21388. [DOI] [PubMed] [Google Scholar]
- 91.Azam A.S., Miligy I.M., Kimani P.K.U., Maqbool H., Hewitt K., Rajpoot N.M., Snead D.R.J. Diagnostic concordance and discordance in digital pathology: A systematic review and meta-analysis. J. Clin. Pathol. 2021;74:448. doi: 10.1136/jclinpath-2020-206764. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 92.Buck T.P., Dilorio R., Havrilla L., O’Neill D.G. Validation of a whole slide imaging system for primary diagnosis in surgical pathology: A community hospital experience. J. Pathol. Inform. 2014;5:43. doi: 10.4103/2153-3539.145731. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 93.Tabata K., Mori I., Sasaki T., Itoh T., Shiraishi T., Yoshimi N., Maeda I., Harada O., Taniyama K., Taniyama D. Whole-slide imaging at primary pathological diagnosis: Validation of whole-slide imaging-based primary pathological diagnosis at twelve Japanese academic institutes. Pathol. Int. 2017;67:547–554. doi: 10.1111/pin.12590. [DOI] [PubMed] [Google Scholar]
- 94.Cross S., Furness P., Igali L., Snead D., Treanor D. Best Practice Recommendations for Implementing Digital Pathology January 2018. The Royal College of Pathologists; London, UK: 2018. [(accessed on 16 June 2022)]. Available online: https://www.rcpath.org/uploads/assets/f465d1b3-797b-4297-b7fedc00b4d77e51/Best-practice-recommendations-for-implementing-digital-pathology.pdf. [Google Scholar]
- 95.Linardatos P., Papastefanopoulos V., Kotsiantis S. Explainable AI: A Review of Machine Learning Interpretability Methods. Entropy. 2020;23:18. doi: 10.3390/e23010018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 96.LeCun Y., Bengio Y., Hinton G. Deep learning. Nature. 2015;521:436–444. doi: 10.1038/nature14539. [DOI] [PubMed] [Google Scholar]
- 97.Ertosun M.G., Rubin D.L. Automated Grading of Gliomas using Deep Learning in Digital Pathology Images: A modular approach with ensemble of convolutional neural networks. AMIA Annu. Symp. Proc. 2015;2015:1899–1908. [PMC free article] [PubMed] [Google Scholar]
- 98.Barker J., Hoogi A., Depeursinge A., Rubin D.L. Automated classification of brain tumor type in whole-slide digital pathology images using local representative tiles. Med. Image Anal. 2016;30:60–71. doi: 10.1016/j.media.2015.12.002. [DOI] [PubMed] [Google Scholar]
- 99.Langer L., Binenbaum Y., Gugel L., Amit M., Gil Z., Dekel S. Computer-aided diagnostics in digital pathology: Automated evaluation of early-phase pancreatic cancer in mice. Int. J. Comput. Assist. Radiol. Surg. 2015;10:1043–1054. doi: 10.1007/s11548-014-1122-9. [DOI] [PubMed] [Google Scholar]
- 100.Amann J., Blasimme A., Vayena E., Frey D., Madai V.I. Explainability for artificial intelligence in healthcare: A multidisciplinary perspective. BMC Med. Inform. Decis. Mak. 2020;20:310. doi: 10.1186/s12911-020-01332-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101.Goldfarb A., Teodoridis F. Why Is AI Adoption in Health Care Lagging? [(accessed on 30 April 2022)]. Available online: https://www.brookings.edu/research/why-is-ai-adoption-in-health-care-lagging/
- 102.McKay F., Williams B.J., Prestwich G., Bansal D., Hallowell N., Treanor D. The ethical challenges of artificial intelligence-driven digital pathology. J. Pathol. Clin. Res. 2022;8:209–216. doi: 10.1002/cjp2.263. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 103.Madabhushi A., Lee G. Image analysis and machine learning in digital pathology: Challenges and opportunities. Med. Image Anal. 2016;33:170–175. doi: 10.1016/j.media.2016.06.037. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 104.Tizhoosh H.R., Babaie M. Representing Medical Images With Encoded Local Projections. IEEE Trans. Biomed. Eng. 2018;65:2267–2277. doi: 10.1109/TBME.2018.2791567. [DOI] [PubMed] [Google Scholar]
- 105.Shamir L., Orlov N.V., Eckley D.M., Macura T.J., Johnston J., Goldberg I.G. Wndchrm—An open source utility for biological image analysis-0. Source Code Biol. Med. 2008;3:13. doi: 10.1186/1751-0473-3-13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 106.Jiang Y., Xie J.-j., Han Z., Liu W., Xi S., Huang L., Huang W., Lin T., Zhao L.-y., Hu Y., et al. Immunomarker Support Vector Machine Classifier for Prediction of Gastric Cancer Survival and Adjuvant Chemotherapeutic Benefit. Clin. Cancer Res. 2018;24:5574–5584. doi: 10.1158/1078-0432.CCR-18-0848. [DOI] [PubMed] [Google Scholar]
- 107.Sharma H., Zerbe N., Klempert I., Lohmann S., Lindequist B., Hellwich O., Hufnagl P. Appearance-based necrosis detection using textural features and SVM with discriminative thresholding in histopathological whole slide images; Proceedings of the 2015 IEEE 15th International Conference on Bioinformatics and Bioengineering (BIBE); Belgrade, Serbia. 2–4 November 2015; pp. 1–6. [Google Scholar]
- 108.Wang H., Cruz-Roa A., Basavanhally A., Gilmore H., Shih N., Feldman M.D., Tomaszewski J.E., González F.A., Madabhushi A. Mitosis detection in breast cancer pathology images by combining handcrafted and convolutional neural network features. J. Med. Imaging. 2014;1:034003. doi: 10.1117/1.JMI.1.3.034003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 109.Geread R.S., Morreale P., Dony R.D., Brouwer E., Wood G.A., Androutsos D., Khademi A. IHC Color Histograms for Unsupervised Ki67 Proliferation Index Calculation. Front. Bioeng. Biotechnol. 2019;7:226. doi: 10.3389/fbioe.2019.00226. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 110.Hou L., Samaras D., Kurç T.M., Gao Y., Davis J.E., Saltz J.H. Patch-Based Convolutional Neural Network for Whole Slide Tissue Image Classification; Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Las Vegas, NV, USA. 27–30 June 2016; pp. 2424–2433. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 111.Zheng Y., Gindra R.H., Green E.J., Burks E.J., Betke M., Beane J.E., Kolachalama V.B. A graph-transformer for whole slide image classification. IEEE Trans. Med. Imag. 2022. Online ahead of print . [DOI] [PMC free article] [PubMed]
- 112.Shen Y., Ke J. MICCAI 2020: Medical Image Computing and Computer Assisted Intervention—MICCAI 2020. Volume 12265. Springer; Cham, Switzerland: 2020. A Deformable CRF Model for Histopathology Whole-Slide Image Classification; pp. 500–508. (Lecture Notes in Computer Science). [Google Scholar]
- 113.Iizuka O., Kanavati F., Kato K., Rambeau M., Arihiro K., Tsuneki M. Deep Learning Models for Histopathological Classification of Gastric and Colonic Epithelial Tumours. Sci. Rep. 2020;10:1–11. doi: 10.1038/s41598-020-58467-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 114.Sirinukunwattana K., Domingo E., Richman S.D., Redmond K.L., Blake A., Verrill C., Leedham S.J., Chatzipli A., Hardy C.W., Whalley C.M., et al. Image-based consensus molecular subtype (imCMS) classification of colorectal cancer using deep learning. Gut. 2020;70:544–554. doi: 10.1136/gutjnl-2019-319866. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 115.Popovici V., Budinská E., Dušek L., Kozubek M., Bosman F.T. Image-based surrogate biomarkers for molecular subtypes of colorectal cancer. Bioinformatics. 2017;33:2002–2009. doi: 10.1093/bioinformatics/btx027. [DOI] [PubMed] [Google Scholar]
- 116.Korbar B., Olofson A.M., Miraflor A.P., Nicka K.M., Suriawinata M.A., Torresani L., Suriawinata A.A., Hassanpour S. Deep Learning for Classification of Colorectal Polyps on Whole-slide Images. J. Pathol. Inform. 2017;8:30. doi: 10.4103/jpi.jpi_34_17. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 117.Kather J.N., Krisam J., Charoentong P., Luedde T., Herpel E., Weis C.-A., Gaiser T., Marx A., Valous N.A., Ferber D., et al. Predicting survival from colorectal cancer histology slides using deep learning: A retrospective multicenter study. PLoS Med. 2019;16:e1002730. doi: 10.1371/journal.pmed.1002730. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 118.Wei J., Suriawinata A.A., Vaickus L.J., Ren B., Liu X., Lisovsky M., Tomita N., Abdollahi B., Kim A.S., Snover D.C., et al. Evaluation of a Deep Neural Network for Automated Classification of Colorectal Polyps on Histopathologic Slides. JAMA Netw. Open. 2020;3:e203398. doi: 10.1001/jamanetworkopen.2020.3398. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 119.Gupta P., Huang Y., Sahoo P.K., You J.-F., Chiang S.-F., Onthoni D.D., Chern Y.-J., Chao K.-Y., Chiang J.-M., Yeh C.-Y., et al. Colon Tissues Classification and Localization in Whole Slide Images Using Deep Learning. Diagnostics. 2021;11:1398. doi: 10.3390/diagnostics11081398. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 120.Khened M., Kori A., Rajkumar H., Srinivasan B., Krishnamurthi G. A generalized deep learning framework for whole-slide image segmentation and analysis. Sci. Rep. 2021;11:11579. doi: 10.1038/s41598-021-90444-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 121.Cao R., Yang F., Ma S.-C., Liu L., Zhao Y., Li Y., Wu D.-H., Wang T., Lu W.-J., Cai W.-J., et al. Development and interpretation of a pathomics-based model for the prediction of microsatellite instability in Colorectal Cancer. Theranostics. 2020;10:11080–11091. doi: 10.7150/thno.49864. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 122.Kather J.N., Pearson A.T., Halama N., Jäger D., Krause J., Loosen S.H., Marx A., Boor P., Tacke F., Neumann U.P., et al. Deep learning can predict microsatellite instability directly from histology in gastrointestinal cancer. Nat. Med. 2019;25:1054–1056. doi: 10.1038/s41591-019-0462-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 123.Bychkov D., Linder N., Turkki R., Nordling S., Kovanen P.E., Verrill C., Walliander M., Lundin M., Haglund C., Lundin J. Deep learning based tissue analysis predicts outcome in colorectal cancer. Sci. Rep. 2018;8:1–11. doi: 10.1038/s41598-018-21758-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 124.Meier A., Nekolla K., Hewitt L.C., Earle S., Yoshikawa T., Oshima T., Miyagi Y., Huss R., Schmidt G., Grabsch H.I. Hypothesis-free deep survival learning applied to the tumour microenvironment in gastric cancer. J. Pathol. Clin. Res. 2020;6:273–282. doi: 10.1002/cjp2.170. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 125.Lu M.Y., Williamson D.F.K., Chen T.Y., Chen R.J., Barbieri M., Mahmood F. Data-efficient and weakly supervised computational pathology on whole-slide images. Nat. Biomed. Eng. 2021;5:555–570. doi: 10.1038/s41551-020-00682-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 126.Joshi R., Kruger A.J., Sha L., Kannan M., Khan A.A., Stumpe M.C. Learning relevant H&E slide morphologies for prediction of colorectal cancer tumor mutation burden using weakly supervised deep learning. J. Clin. Oncol. 2020;38:e15244. doi: 10.1200/jco.2020.38.15_suppl.e15244. [DOI] [Google Scholar]
- 127.Wang S., Zhu Y., Yu L., Chen H., Lin H., Wan X.-B., Fan X., Heng P.-A. RMDL: Recalibrated multi-instance deep learning for whole slide gastric image classification. Med. Image Anal. 2019;58:101549. doi: 10.1016/j.media.2019.101549. [DOI] [PubMed] [Google Scholar]
- 128.Shao Z., Bian H., Chen Y., Wang Y., Zhang J., Ji X., Zhang Y. TransMIL: Transformer based Correlated Multiple Instance Learning for Whole Slide Image Classication. Adv. Neural Inf. Process. Syst. 2021;34:2136–2147. [Google Scholar]
- 129.Sharma Y., Shrivastava A., Ehsan L., Moskaluk C.A., Syed S., Brown D.E. Cluster-to-Conquer: A Framework for End-to-End Multi-Instance Learning for Whole Slide Image Classification. Proc. Mach. Learn. Res. 2021;143:682–698. [Google Scholar]
- 130.Klein S., Gildenblat J., Ihle M.A., Merkelbach-Bruse S., Noh K.-W., Peifer M., Quaas A.M., Büttner R. Deep learning for sensitive detection of Helicobacter Pylori in gastric biopsies. BMC Gastroenterol. 2020;20:1–11. doi: 10.1186/s12876-020-01494-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 131.Li J., Chen W., Huang X., Yang S., Hu Z., Duan Q., Metaxas D.N., Li H., Zhang S. Medical Image Computing and Computer Assisted Intervention—MICCAI 2021. Volume 12908. Springer; Cham, Switzerland: 2021. Hybrid Supervision Learning for Pathology Whole Slide Image Classification; pp. 309–318. (Lecture Notes in Computer Science). [Google Scholar]
- 132.Adu K., Yu Y., Cai J., Owusu-Agyemang K., Twumasi B.A., Wang X. DHS-CapsNet: Dual horizontal squash capsule networks for lung and colon cancer classification from whole slide histopathological images. Int. J. Imaging Syst. Technol. 2021;31:2075–2092. doi: 10.1002/ima.22569. [DOI] [Google Scholar]
- 133.Chen H., Han X., Fan X., Lou X., Liu H., Huang J., Yao J. Medical Image Computing and Computer Assisted Intervention—MICCAI 2019. Volume 11764. Springer; Cham, Switzerland: 2019. Rectified Cross-Entropy and Upper Transition Loss for Weakly Supervised Whole Slide Image Classifier; pp. 351–359. (Lecture Notes in Computer Science). [Google Scholar]
- 134.Ren J., Hacihaliloglu I., Singer E.A., Foran D.J., Qi X. Unsupervised Domain Adaptation for Classification of Histopathology Whole-Slide Images. Front. Bioeng. Biotechnol. 2019;7:102. doi: 10.3389/fbioe.2019.00102. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 135.Bray F., Ferlay J., Soerjomataram I., Siegel R.L., Torre L.A., Jemal A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2018;68:394–424. doi: 10.3322/caac.21492. [DOI] [PubMed] [Google Scholar]
- 136.Biscotti C.V., Dawson A.E., Dziura B., Galup L., Darragh T., Rahemtulla A., Wills-Frank L. Assisted primary screening using the automated ThinPrep Imaging System. Am. J. Clin. Pathol. 2005;123:281–287. doi: 10.1309/AGB1MJ9H5N43MEGX. [DOI] [PubMed] [Google Scholar]
- 137.Vu H.T., Lopez R., Bennett A., Burke C.A. Individuals with sessile serrated polyps express an aggressive colorectal phenotype. Dis. Colon Rectum. 2011;54:1216–1223. doi: 10.1097/DCR.0b013e318228f8a9. [DOI] [PubMed] [Google Scholar]
- 138.Osmond A., Li-Chang H., Kirsch R., Divaris D., Falck V., Liu D.F., Marginean C., Newell K., Parfitt J., Rudrick B., et al. Interobserver variability in assessing dysplasia and architecture in colorectal adenomas: A multicentre Canadian study. J. Clin. Pathol. 2014;67:781–786. doi: 10.1136/jclinpath-2014-202177. [DOI] [PubMed] [Google Scholar]
- 139.Foss F.A., Milkins S., McGregor A.H. Inter-observer variability in the histological assessment of colorectal polyps detected through the NHS Bowel Cancer Screening Programme. Histopathology. 2012;61:47–52. doi: 10.1111/j.1365-2559.2011.04154.x. [DOI] [PubMed] [Google Scholar]
- 140.Davidson K.W., Barry M.J., Mangione C.M., Cabana M., Caughey A.B., Davis E.M., Donahue K.E., Doubeni C.A., Krist A.H., Kubik M.J.J. Screening for colorectal cancer: US Preventive Services Task Force recommendation statement. JAMA. 2016;315:2564–2575. doi: 10.1001/jama.2016.5989. [DOI] [PubMed] [Google Scholar]
- 141.Zhao L., Lee V., Ng M.K., Yan H., Bijlsma M.F. Molecular subtyping of cancer: Current status and moving toward clinical applications. Briefings Bioinform. 2018;20:572–584. doi: 10.1093/bib/bby026. [DOI] [PubMed] [Google Scholar]
- 142.De Smedt L., Lemahieu J., Palmans S., Govaere O., Tousseyn T., Van Cutsem E., Prenen H., Tejpar S., Spaepen M., Matthijs G., et al. Microsatellite instable vs stable colon carcinomas: Analysis of tumour heterogeneity, inflammation and angiogenesis. Br. J. Cancer. 2015;113:500–509. doi: 10.1038/bjc.2015.213. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 143.Baretti M., Le D.T. DNA mismatch repair in cancer. Pharmacol. Ther. 2018;189:45–62. doi: 10.1016/j.pharmthera.2018.04.004. [DOI] [PubMed] [Google Scholar]
- 144.Li K., Luo H., Huang L., Luo H., Zhu X. Microsatellite instability: A review of what the oncologist should know. Cancer Cell Int. 2020;20:1–13. doi: 10.1186/s12935-019-1091-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 145.André T., Shiu K.-K., Kim T.W., Jensen B.V., Jensen L.H., Punt C., Smith D., Garcia-Carbonero R., Benavides M., Gibbs P., et al. Pembrolizumab in Microsatellite-Instability–High Advanced Colorectal Cancer. N. Engl. J. Med. 2020;383:2207–2218. doi: 10.1056/NEJMoa2017699. [DOI] [PubMed] [Google Scholar]
- 146.Fan J., Qian J., Zhao Y.J.N. The loss of PTEN expression and microsatellite stability (MSS) were predictors of unfavorable prognosis in gastric cancer (GC) Neoplasma. 2020;67:1359–1366. doi: 10.4149/neo_2020_200422N427. [DOI] [PubMed] [Google Scholar]
- 147.Snowsill T., Coelho H., Huxley N., Jones-Hughes T., Briscoe S., Frayling I.M., Hyde C. Molecular testing for Lynch syndrome in people with colorectal cancer: Systematic reviews and economic evaluation. Health Technol. Assess. 2017;21:1–238. doi: 10.3310/hta21510. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 148.Goss P.E., Lee B.L., Badovinac-Crnjevic T., Strasser-Weippl K., Chavarri-Guerra Y., St Louis J., Villarreal-Garza C., Unger-Saldaña K., Ferreyra M., Debiasi M., et al. Planning cancer control in Latin America and the Caribbean. Lancet Oncol. 2013;14:391–436. doi: 10.1016/S1470-2045(13)70048-2. [DOI] [PubMed] [Google Scholar]
- 149.Banatvala J.J.L. COVID-19 testing delays and pathology services in the UK. Lancet. 2020;395:1831. doi: 10.1016/S0140-6736(20)31037-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 150.Al-Shamsi H.O., Abu-Gheida I., Rana S.K., Nijhawan N., Abdulsamad A.S., Alrawi S., Abuhaleeqa M., Almansoori T.M., Alkasab T., Aleassa E.M., et al. Challenges for cancer patients returning home during SARS-COV-19 pandemic after medical tourism - a consensus report by the emirates oncology task force. BMC Cancer. 2020;20:1–10. doi: 10.1186/s12885-020-07115-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 151.Balasubramani B., Newsom K.J., Martinez K.A., Starostik P., Clare-Salzler M., Chamala S. Pathology informatics and robotics strategies for improving efficiency of COVID-19 pooled testing. Acad. Pathol. 2021;8:23742895211020485. doi: 10.1177/23742895211020485. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 152.Campanella G., Hanna M.G., Geneslaw L., Miraflor A., Silva V.W.K., Busam K.J., Brogi E., Reuter V.E., Klimstra D.S., Fuchs T.J. Clinical-grade computational pathology using weakly supervised deep learning on whole slide images. Nat. Med. 2019;25:1301–1309. doi: 10.1038/s41591-019-0508-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 153.Ianni J.D., Soans R.E., Sankarapandian S., Chamarthi R.V., Ayyagari D., Olsen T.G., Bonham M.J., Stavish C.C., Motaparthi K., Cockerell C.J., et al. Tailored for Real-World: A Whole Slide Image Classification System Validated on Uncurated Multi-Site Data Emulating the Prospective Pathology Workload. Sci. Rep. 2020;10:1–12. doi: 10.1038/s41598-020-59985-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
Data sharing not applicable.