Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2021 Dec 1.
Published in final edited form as: Dig Dis Sci. 2020 Oct 15;65(12):3448–3455. doi: 10.1007/s10620-020-06643-2

Artificial Intelligence and Its Role in Identifying Esophageal Neoplasia

Taseen Syed 1,2, Akash Doshi 3, Shan Guleria 4, Sana Syed 5, Tilak Shah 1,2
PMCID: PMC8139616  NIHMSID: NIHMS1695740  PMID: 33057945

Abstract

Randomized trials have demonstrated that ablation of dysplastic Barrett’s esophagus can reduce the risk of progression to cancer. Endoscopic resection for early stage esophageal adenocarcinoma and squamous cell carcinoma can significantly reduce postoperative morbidity compared to esophagectomy. Unfortunately, current endoscopic surveillance technologies (e.g., high-definition white light, electronic, and dye-based chromoendoscopy) lack sensitivity at identifying subtle areas of dysplasia and cancer. Random biopsies sample only approximately 5% of the esophageal mucosa at risk, and there is poor agreement among pathologists in identifying low-grade dysplasia. Machine-based deep learning medical image and video assessment technologies have progressed significantly in recent years, enabled in large part by advances in computer processing capabilities. In deep learning, sequential layers allow models to transform input data (e.g., pixels for imaging data) into a composite representation that allows for classification and feature identification. Several publications have attempted to use this technology to help identify dysplasia and early esophageal cancer. The aims of this reviews are as follows: (a) discussing limitations in our current strategies to identify esophageal dysplasia and cancer, (b) explaining the concepts behind deep learning and convolutional neural networks using language appropriate for clinicians without an engineering background, (c) systematically reviewing the literature for studies that have used deep learning to identify esophageal neoplasia, and (d) based on the systemic review, outlining strategies on further work necessary before these technologies are ready for “prime-time,” i.e., use in routine clinical care.

Keywords: Computer assisted diagnosis, Esophageal cancer, Barrett’s esophagus, Convolutional neural network, Artificial intelligence, Deep learning

Rationale for Endoscopic Surveillance

Esophageal cancer is the 8th most common cancer and the 6th leading cause of cancer death worldwide [1]. Primary malignancies of the esophagus include esophageal adenocarcinoma (EAC) and squamous cell carcinoma (SCC). The overall prognosis for both EAC and SCC is poor because most of these cancers are identified at an advanced stage. Randomized trials have demonstrated that identification and endoscopic eradication of dysplasia via ablation or resection reduce the risk of progression to cancer [2, 3]. Endoscopic resection of esophageal cancer at an early stage (i.e., limited to the mucosa or muscularis mucosa) can achieve > 90% cure rates without the need for esophagectomy, which carries substantial morbidity [4].

Barrett’s esophagus (BE) occurs in 1–2% of the population in Western countries and is characterized by replacement of the squamous lining of the esophagus by columnar mucosa (i.e., intestinal metaplasia) [5]. Because BE patients have an increased risk of EAC compared to the general population, endoscopic surveillance is generally recommended in Western nations to identify dysplasia and early stage cancer. Similarly, endoscopic surveillance is performed in many areas with a high prevalence of SCC (e.g., China, Eastern Asia, and Africa) to identify dysplasia and early stage SCC [6].

Endoscopic Surveillance Techniques and Limitations

Visual identification of dysplasia and early esophageal cancer using high-definition white light endoscopy (HD-WLE) is a highly subjective task that requires significant effort and experience. Even in expert hands, the sensitivity of HD-WLE for identifying dysplasia and cancer in BE is a mere 48% [7] likely due to lack of specific training for gastroenterologists to recognize subtle mucosal changes visualized by this modality. Guidelines therefore recommend random 4-quadrant biopsies at 1–2 cm intervals, a practice dubbed the Seattle protocol [5]. While the Seattle protocol does improve yield for dysplasia, adherence to the protocol is as low in clinical practice, likely due to the time-consuming nature of the protocol [8]. Furthermore, only 5% of the esophageal mucosa at risk is sampled [11], and there is poor agreement among pathologists at identifying low-grade dysplasia (LGD) [12, 13], both of which can lead to missed diagnoses and progression of disease.

The limitations in our current approach in identifying esophageal neoplasia have spurred development of several novel technologies that may enhance diagnostic capabilities. Commercially available imaging-based technologies such as probe-based confocal endomicroscopy (pCLE) and optical coherence tomography (OCT) are expensive and operator-dependent, requiring specialized training to perform and interpret. Chromoendoscopy technologies are less costly, highlighting the mucosal surface pattern and include dye-based (e.g., acetic acid, Lugol’s iodine) and electronic techniques (e.g., narrow-band imaging, Fuji intelligent chromoendoscopy, and i-scan) [9]. However, chromoendoscopy also requires specialized training. Wide area trans-epithelial sampling (WATS) utilizes an abrasive brush that can sample tissue from a wide area of the esophagus [10]. A disadvantage of this technology is that cytology rather than histologic specimens with preserved tissue architecture is used for analysis.

In 2016, the American Society for Gastrointestinal Endoscopy (ASGE) published thresholds for a technology to replace random biopsies in clinical practice. The ASGE recommends an imaging technology along with targeted biopsy must demonstrate a per-patient sensitivity of ≥ 90%, negative predictive value (NPV) ≥ 98%, and specificity > 80% for detecting dysplasia or early EAC when compared to random biopsies [11]. Unfortunately, none of the aforementioned technological advances met these criteria for replacing random biopsies [12, 13]. There exists a clear need for technological improvements to the current random biopsy protocol, and recent developments in computing power have brought artificial intelligence to the forefront of this discussion.

Artificial Intelligence (AI): A Primer

Computer-aided detection is not an entirely new concept in esophageal neoplasia identification; in fact, WATS samples are analyzed using a computerized imaging processing system. Yet, only recently progressive increases in computer storage and processing speed, along with the availability of large datasets (“big data”), converged to markedly expand the capabilities of AI. The vocabulary of AI studies may seem daunting to those without a background in computer science and may dissuade clinicians from critically appraising AI-based gastroenterology studies. Lay persons’ definitions of terms commonly utilized in AI-based studies are summarized in Table 1. Beginning with the basics, let us first define “machine learning (ML),” which refers to the development of algorithms that can use structured data in order to perform tasks with minimal human input [14]. For example, if one feeds a ML algorithm thousands of images of sedans and trucks, the algorithm will eventually be able to classify a brand-new image as being either a sedan or a truck. Deep learning (DL), by contrast, is a subset of ML in which the algorithm uses layers of data transformations to identify features of the given dataset in order to learn how to classify the data [14]. To use our prior example, one could once again provide the DL algorithm with images of sedans and trucks, and this time it would learn to identify the features (this is the key distinction in DL) of each class of vehicle in order to classify new images. This ability to identify features and learn with minimal human supervision is part of what makes DL so exciting in the field of medicine, which (between CT scans, endoscopy video data, surgical biopsies, and beyond) has ample visual data in need of analysis and interpretation.

Table 1.

Artificial intelligence terminology

Term Abbreviation Definition
Activation Map The output of a convolutional layer that highlights features detected by the algorithm
Convolutional neural network CNN An algorithm that can classify images by identifying features in the image using mathematical filters
Deep learning DL A subset of machine learning in which the data provided to the algorithm do not need to be structured in order for the algorithm to learn
Feature An aspect of an image that can be identified by a CNN. Can be simple (e.g., lines, edges) or complex (e.g., specific faces).
Filter The mathematical “lens” that the CNN hovers over (i.e., convolves over) the image in order to highlight features
Layers Aspects of the CNN algorithm that refine the feature identification and image classification.
Convolutional layers refer to the filters described above. Other layers can perform complex mathematical tasks in the algorithm.
Machine learning ML The development of algorithms that can use structured data in order to perform tasks without human input

The second key term to define in the world of AI is “convolutional neural network” or CNN, which represents one of the most common ways in which modern DL algorithms accomplish their feature recognition. The term is actually based on the way in which the neurons in the visual cortex in the brain respond specifically in the presence of certain visual stimuli (e.g., horizontal vs. vertical lines) [15]. In similar fashion, the goal of a CNN is to develop filters that hover over an image (i.e., convolve over them) in order to identify features. Each of these filters produces an “activation map,” which essentially highlights the feature identified by a given filter. These filters start out very simply, with the initial filters or convolutional “layers” identifying simple features such as curves or edges. Each subsequent layer of the CNN works on the activation maps from the previous layer; the result is the identification of increasingly complex features as the network moves deeper and deeper. Instead of simple lines and edges, deeper layers may light up in response to complex shapes, faces, words, etc. [16]. In the end, the CNN then uses these activation maps to classify the image as belonging to one of the multiple classes by correlating specific features with each class. To effectively make such correlations and ascribe appropriate weights to given features, a CNN requires data to train itself. After many iterations of training on labeled images, a CNN should have fine-tuned its layers and is ready to be tested on a fresh dataset. The larger and higher quality the dataset for the training phase, the more accurate the CNN is likely to be for the testing phase. As mentioned, we as clinicians have access to massive quantities of data; CNNs and DL represent means by which we can help make sense of this treasure trove.

Exciting though this new technology undoubtedly is, there are significant pitfalls of which to be aware. For example, it can be difficult for humans to distinguish between esophageal dysplasia and non-dysplastic BE with inflammation [17], meaning that a classification algorithm will need to identify large numbers of features in order to separate the two classes. In general, complex models with many features are needed to tackle difficult datasets; to achieve this level of complexity, extremely large training datasets are necessary to generalize the model to new data. [18]. Unfortunately, datasets in medicine tend to be fragmented (e.g., rather than having access to one million biopsies across the nation, researchers may have access to only a few hundred at one or two centers), and these piecemeal datasets make applications of DL more difficult. An additional and related pitfall involves the quality of the data, specifically the accuracy of “ground truth” to which the model is being compared. To once again use our example of esophageal dysplasia, human pathologists have relatively poor inter-observer agreement at identifying LGD in BE [1921]. A DL model that attempts to distinguish LGD from non-dysplastic BE would be difficult to train and test because the certainty of the underlying data is suspect. Unlike our relatively simple example of sedans versus trucks, medical image analysis is much murkier territory. One potential solution to these pitfalls would be to collect and annotate large quantities of medical data from across multiple centers and to have these data reviewed by multiple experts to ensure quality, as has been done with breast cancer data in the CAMELYON dataset, among others [22].

Identifying Esophageal Neoplasia with AI

In order to summarize the literature on using DL to identify esophageal neoplasia, we searched Pubmed/Medline, Embase, Cochrane library, and Web of Science for studies on the diagnostic performance of AI in the detection of esophageal neoplasia (BE, squamous dysplasia, EAC, and SCC) as per the guidelines outlined by the Preferred Reporting Items for Systematic Reviews and Meta-analyses statement (PRISMA) [23]. The literature search was performed in close collaboration with a health sciences librarian at Virginia Commonwealth University. When key information regarding accuracy of the DL models at identifying neoplasia was missing, we attempted to contact authors to obtain this information. We identified 21 relevant studies, 6 of which were conference abstracts (Fig. 1). All studies were published after 2017. Characteristics of all studies including methodology and sample sizes are outlined in Table 2. Study outcomes are listed in Table 3. The primary outcome was the identification of squamous dysplasia or SCC in 11 studies, and BE with dysplasia or EAC in 15 studies. All studies used endoscopic images for training. One study used live endoscopy videos for training and validation [24]. In 17 out of 21 studies, data were available for both the training and the test/validation datasets. Among these studies, the number of unique patients in the training set ranged from 17 to 1530. (The number of specimens ranged from 60 to 196,200.) The number of unique patients in the test/validation set ranged from 9 to 880. (The number of specimens ranged from 62 to 21,800.) In three studies the, test/validation was performed prospectively [2426]. Sensitivity across the studies ranged from 70 to 100%, specificity ranged from 42 to 100%, and overall accuracy ranged from 66 to 98% (Table 3). The models achieved ASGE recommended sensitivity and specificity thresholds in 15 and 14 studies, respectively. In most of these studies, ground truth was established based on expert pathologist review of histopathology.

Fig. 1.

Fig. 1

Flow diagram of studies identified in the systematic review

Table 2.

Characteristics of clinical studies involving computer-aided diagnosis for early esophageal cancer

References Study design Target esophageal disease* Type of specimens for training Number of specimens for training set Number of unique patients for training set Number of specimens for test/validation Number of unique patients for test/validation
Zhang et al. [28] Retrospective SCC, EAC Endoscopic images* 196,200 21,800
Mendel et al.[29] Retrospective EAC Endoscopic images 100 39 100 39
Ebigbo et al.[30] (Augsburg data) Retrospective EAC Endoscopic images 148 62
Ebigbo et al.[30] (MICCAI data) Retrospective EAC Endoscopic images 100 78
Everson et al.[31] Retrospective SCC Endoscopic images/video 2045 17 812
Guo et al. [32] Retrospective SCC Endoscopic images 2085 293 1478 59
Horie et al. [33] Retrospective SCC, EAC Endoscopic images 8428 387 1128 97
Cai et al. [34] Retrospective SCC Endoscopic images 2428 746 187 52
de Groof et al.[26] Prospective EAC, BE Endoscopic images 60 60
Ghatwary et al.[35] Retrospective EAC Endoscopic images 100 21 9
Horie et al. [36] Retrospective SCC, EAC Endoscopic images 8428 384 1118 97
Liu et al. [37] Retrospective SCC Endoscopic images 10,950 503 500
Shiroma et al.[38] Retrospective SCC, SD, EAC Endoscopic images 8428 40¥ 20
Struyvenberg et al. [39] Retrospective EAC Endoscopic images 1247 183 100
Tang et al. [40] Retrospective SCC, SD Endoscopic images 5174 1530 946 280
de Groof et al.[25] Prospective EAC Endoscopic images 297 95 80 80
Ebigbo et al.[41] Retrospective EAC Endoscopic images 129 62 14
Guo et al. [42] Retrospective SCC Endoscopic images 6473 549 6671 880
Ohmori et al.[43] Retrospective SCC, SD Endoscopic images 22,562 727 135
de Groof et al.[24] Prospective EAC Endoscopic images (application in live endoscopy) 1544 160
Hashimoto et al.[44] Retrospective EAC Endoscopic images 1832 100 458 39
Iwagami et al.[45] Retrospective EAC-EJ Endoscopic images 3443 159 275 36
*

1000 patches were made from each endoscopic image, and total number of endoscopic images were 218.

¥

Test/validation was done on 40 test movies of 20 patients.

The CAD system was pre-trained on a unique dataset of 494,364 labeled endoscopic images named GastroNet. SCC squamous cell cancer, SD squamous dysplasia, EAC esophageal adenocarcinoma, BE Barrett’s esophagus, EAC-EJ esophageal adenocarcinoma at esophagogastric junction

Table 3.

Performance data on artificial intelligence algorithms used for computer-aided diagnosis for early esophageal cancer

References Deep learning model? Model Sensitivity (%) Specificity (%) PPV (%) NPV (%) Accuracy (%)
Zhang et al. [28]* No Supervised learning via CNN. 218,000 patches generated from 218 endoscopic images 73.41 83.54 72.09 84.44 79.38
Mendel et al. [29] Yes Deep CNN with a residual net (ResNet) architecture 94 88
Ebigbo et al. [30] Yes Deep CNN with a residual net (ResNet) architecture 97 (WLE) 94 (NBI) 88 (WLE) 80 (NBI)
Ebigbo et al. [30] Yes Deep CNN with a residual net (ResNet) architecture 92 100
Everson et al. [31] Yes Deep CNN 89.3 98 93.7
Guo et al. [32]* NA NA 97.8 76.2
Horie et al. [33]* Deep CNN 54 98
Cai et al. [34] Yes CNN 97.8 85.4 86.4 97.6 91.4
de Groof et al. [26] No Supervised learning technique 95 85 87.2 91.7
Ghatwary et al. [35] Yes Regional-based convolutional neural network (R-CNN), Fast R-CNN, Faster R-CNN and Single Shot MultiBox Detector (SSD) 96 92
Horie et al. [36] Yes Deep CNN (Single Shot MultiBox Detector) 98 79 39 95 98
Liu et al. [37] Yes CNN 77.58 78.06
Shiroma et al. [38]* Yes CNN 70 42.1 80
Struyvenberg et al. [39]* Yes CNN 88 78 84
Tang et al. [40]* Yes CNN 97 94 97
de Groof et al. [25] Yes ResNet U-net 93 83 88
Ebigbo et al. [41] Yes Encoder–decoder artificial neural network 83.7 100 89.9
Guo et al. [42] Yes Detection neural network 98.04 95.03
Ohmori et al. [43] Yes CNN 100 (NME-NBI) 90 (WLE) 63(NME-NBI) 76(WLE) 63(NME-NBI) 70(WLE) 100(NME-NBI 93(WLE) 77(NME-NBI) 81 (WLE)
de Groof et al. [24] Yes Custom-made, fully residual, hybrid ResNet/U-Net model 91 89 90
Hashimoto et al.[44] Yes CNN: inception-ResNet-v2 algorithm 96.4 94.2 89.2 95.4
Iwagami et al. [45] Yes CNN: single shot multiBox detector 94 42 58 90 66
*

Conference abstracts. WLE; White light endoscopy, NBI; narrow-band imaging, NME; non-magnified endoscopy

Summary

As described earlier, random esophageal biopsies to identify dysplasia and cancer during endoscopic surveillance are fraught with limitations. Technologies that have attempted to replace random biopsies to date have not been able to achieve sufficient accuracy and require substantial expertise to interpret, which limits widespread utilization. Additionally, the high cost of imaging technologies such as pCLE and OCT represents a significant barrier to adoption in routine clinical practice. As with other technologies, in order to replace the Seattle protocol, DL-based AI must not only meet recommended accuracy thresholds in prospective randomized trials but should also be inexpensive and easy to use and interpret. In an ideal scenario, manufacturers would incorporate AI as a toggle on the endoscope, similar to electronic chromoendoscopy technologies like NBI. Once AI is turned “on,” the algorithm would rapidly and accurately discriminate non-dysplastic areas from segments with a very high probability of dysplasia for targeted biopsy or mucosal resection. Alternatively, if the technology is sufficiently accurate, the endoscopist could proceed with ablation during the same session, thereby eliminating the time, cost, and risk of an additional endoscopic procedure. DL also has the potential to improve uncertainty in histologic diagnosis, particularly for LGD. For instance, our group recently demonstrated that DL algorithms are capable of identifying dysplasia in histologic BE specimens with high accuracy, suggesting that future algorithms may be able to distinguish low-grade from high-grade dysplasia [27]. The most recent published studies using this technology have used real-time endoscopic videos to target areas of concern [24]. Preliminary results suggest that some algorithms achieve high levels of accuracy. Rather than working in silos, researchers must conduct head-to-head comparisons in order to converge on the optimal iterative algorithm and, importantly, compare the results of DL models to human accuracy. Finally, well-designed randomized trials comparing DL to Seattle protocol biopsies are necessary before endoscopists can utilize AI in routine clinical practice.

Acknowledgments

John Cyrus (Research and Education Librarian): Assisted in literature search of following databases: Pubmed/Medline, Embase, Cochrane library and Web of Science for studies. Tilak Shah: ASGE, McGuire Research Institute. Research contract with Allergan. Sana Syed: K23DK117061-01A1 (SS) National Institute of Diabetes and Digestive and Kidney Diseases (NIDDK) of the National Institutes of Health.

References

  • 1.Arnold M, Soerjomataram I, Ferlay J, Forman D. Global incidence of oesophageal cancer by histological subtype in 2012. Gut. 2015;64:381–387. [DOI] [PubMed] [Google Scholar]
  • 2.Cotton CC, Wolf WA, Overholt BF, et al. Late recurrence of Barrett’s esophagus after complete eradication of intestinal metaplasia is rare: final report from ablation in intestinal metaplasia containing dysplasia trial. Gastroenterology. 2017;153:681–688. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Phoa KN, Rosmolen WD, Weusten B, et al. The cost-effectiveness of radiofrequency ablation for Barrett’s esophagus with low-grade dysplasia: results from a randomized controlled trial (SURF trial). Gastrointest Endosc. 2017;86:120–129. [DOI] [PubMed] [Google Scholar]
  • 4.Naveed M, Kubiliun N. Endoscopic treatment of early-stage esophageal cancer. Curr Oncol Rep. 2018;20:71. [DOI] [PubMed] [Google Scholar]
  • 5.Shaheen NJ, Falk GW, Iyer PG, Gerson LB, American College of G. ACG clinical guideline: diagnosis and management of Barrett’s esophagus. Am J Gastroenterol. 2016;111:30–50. (quiz 1). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Domper Arnal MJ, Ferrandez Arenas A, Lanas Arbeloa A. Esophageal cancer: Risk factors, screening and endoscopic treatment in Western and Eastern countries. World J Gastroenterol. 2015;21:7933–7943. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Sharma N, Ho KY. Recent updates in the endoscopic diagnosis of Barrett’s oesophagus. Gastrointest Tumors. 2016;3:109–113. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Wani S, Williams JL, Komanduri S, Muthusamy VR, Shaheen NJ. Endoscopists systematically undersample patients with long-segment Barrett’s esophagus: an analysis of biopsy sampling practices from a quality improvement registry. Gastrointest Endosc. 2019;90:732–741. [DOI] [PubMed] [Google Scholar]
  • 9.Davila RE. Chromoendoscopy. Gastrointest Endosc Clin N Am. 2009;19:193–208. [DOI] [PubMed] [Google Scholar]
  • 10.Goldblum JR, Shaheen NJ, Vennalaganti PR, Sharma P, Light-dale CJ. WATS for Barrett’s surveillance. Gastrointest Endosc. 2018;88:201–202. [DOI] [PubMed] [Google Scholar]
  • 11.Committee AT, Thosani N, Abu Dayyeh BK, et al. ASGE Technology Committee systematic review and meta-analysis assessing the ASGE preservation and incorporation of valuable endoscopic innovations thresholds for adopting real-time imaging-assisted endoscopic targeted biopsy during endoscopic surveillance of Barrett’s esophagus. Gastrointest Endosc. 2016;83:684–698. [DOI] [PubMed] [Google Scholar]
  • 12.Ferguson DD, DeVault KR, Krishna M, Loeb DS, Wolfsen HC, Wallace MB. Enhanced magnification-directed biopsies do not increase the detection of intestinal metaplasia in patients with GERD. Am J Gastroenterol. 2006;101:1611–1616. [DOI] [PubMed] [Google Scholar]
  • 13.Sanghi V, Thota PN. Barrett’s esophagus: novel strategies for screening and surveillance. Ther Adv Chronic Dis. 2019;10:2040622319837851. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Patel V, Khan MN, Shrivastava A, et al. Artificial intelligence applied to gastrointestinal diagnostics: a review. J Pediatr Gastroenterol Nutr. 2020;70:4–11. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Hubel DH, Wiesel TN. Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J Physiol. 1962;160:106–154. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Zeiler MD, Fergus R. Visualizing and understanding convolutional networks. Eur Conf Comput Vis. 2014;8689:818–833. [Google Scholar]
  • 17.Goldblum JR. Controversies in the diagnosis of Barrett esophagus and Barrett-related dysplasia: one pathologist’s perspective. Arch Pathol Lab Med. 2010;134:1479–1484. [DOI] [PubMed] [Google Scholar]
  • 18.Deo RC. Machine learning in medicine. Circulation. 2015;132:1920–1930. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Wani S, Rubenstein JH, Vieth M, Bergman J. Diagnosis and management of low-grade dysplasia in Barrett’s esophagus: expert review from the clinical practice updates committee of the american gastroenterological association. Gastroenterology. 2016;151:822–835. [DOI] [PubMed] [Google Scholar]
  • 20.Shah T, Lippman R, Kohli D, Mutha P, Solomon S, Zfass A. Accuracy of probe-based confocal laser endomicroscopy (pCLE) compared to random biopsies during endoscopic surveillance of Barrett’s esophagus. Endosc Int Open. 2018;6:E414–E420. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Vennalaganti P, Kanakadandi V, Goldblum JR, et al. Discordance among pathologists in the United States and Europe in diagnosis of low-grade dysplasia for patients with Barrett’s Esophagus. Gastroenterology. 2017;152:564–570. [DOI] [PubMed] [Google Scholar]
  • 22.Litjens G, Bandi P, Ehteshami Bejnordi B, et al. 1399 H&E-stained sentinel lymph node sections of breast cancer patients: the CAMELYON dataset. Gigascience. 2018;7:giy065. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Moher D, Liberati A, Tetzlaff J, Altman DG, Group P. Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. Ann Intern Med. 2009;151:264–269. [DOI] [PubMed] [Google Scholar]
  • 24.de Groof AJ, Struyvenberg MR, Fockens KN, et al. Deep learning algorithm detection of Barrett’s neoplasia with high accuracy during live endoscopic procedures: a pilot study (with video). Gastrointest Endosc. 2020;91:1242–1250. [DOI] [PubMed] [Google Scholar]
  • 25.de Groof AJ, Struyvenberg MR, van der Putten J, et al. Deep-learning system detects neoplasia in patients with Barrett’s esophagus with higher accuracy than endoscopists in a multistep training and validation study with benchmarking. Gastroenterology. 2020;158:915–929. [DOI] [PubMed] [Google Scholar]
  • 26.de Groof J, van der Sommen F, van der Putten J, et al. The Argos project: the development of a computer-aided detection system to improve detection of Barrett’s neoplasia on white light endoscopy. United Eur Gastroenterol J. 2019;7:538–547. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Sali R, Moradinasab N, Guleria S, Ehsan L, Fernandes P, Shah T, Syed S, Brown DE. Deep learning for whole slide tissue histopathology classification: a comparative study in identification of dysplastic and non-dysplastic Barrett’s esophagus. J Pers Med. Accepted manuscript. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Zhang CML, Uedo N, Matsuura N, Tam P, Teoh AY. The use of convolutional neural artificial intelligence network to aid the diagnosis and classification of early esophageal neoplasia. A feasibility study. Gastrointestinal Endosc. 2017;85:AB581–AB582. [Google Scholar]
  • 29.Mendel R, Ebigbo A, Probst A, Messmann H, Palm C. Barrett’s Esophagus analysis using convolutional neural networks. Bildver-arbeitung für die Medizin. 2017:80–5. [Google Scholar]
  • 30.Ebigbo A, Mendel R, Probst A, et al. Computer-aided diagnosis using deep learning in the evaluation of early oesophageal adenocarcinoma. Gut. 2019;68:1143–1145. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Everson M, Herrera L, Li W, et al. Artificial intelligence for the real-time classification of intrapapillary capillary loop patterns in the endoscopic diagnosis of early oesophageal squamous cell carcinoma: a proof-of-concept study. United Eur Gastroenterol J. 2019;7:297–306. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Guo L, Wu C, Xiao X, et al. Automated detection of precancerous lesion and early esophageal squamous cancer using a deep learning model. Am J Gastroenterol. 2018;113:S173–S174. [Google Scholar]
  • 33.Horie Y, Yoshio T, Aoyama K, Fujisaki J, Tada T. Application of artificial intelligence using convolutional neural networks in the detection of esophageal cancer. Gastrointestinal Endosc. 2018;87:AB538. [Google Scholar]
  • 34.Cai SL, Li B, Tan WM, et al. Using a deep learning system in endoscopy for screening of early esophageal squamous cell carcinoma (with video). Gastrointest Endosc. 2019;90:745–753. [DOI] [PubMed] [Google Scholar]
  • 35.Ghatwary N, Zolgharni M, Ye X. Early esophageal adenocarcinoma detection using deep learning methods. Int J Comput Assist Radiol Surg. 2019;14:611–621. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Horie Y, Yoshio T, Aoyama K, et al. Diagnostic outcomes of esophageal cancer by artificial intelligence using convolutional neural networks. Gastrointest Endosc. 2019;89:25–32. [DOI] [PubMed] [Google Scholar]
  • 37.Liu DY, Jiang H, Rao N, Luo C, Du W, Li Z, et al. Computer aided annotation of early esophageal cancer in gastroscopic images based on Deeplabv3+ network. ICBIP ‘19: Proceedings of the 2019 4th International Conference on Biomedical Signal and Image Processing (ICBIP 2019). 2019: pp 56–61. [Google Scholar]
  • 38.Shiroma S, Yoshio T, Aoyama K, Tsuchida T, Fujisaki J, Tada T. The application of artificial intelligence to detect esophageal squamous cell carcinoma in movies using convoluntional neural networks. Gastrointest Endosc. 2019;89:AB89. [Google Scholar]
  • 39.Struyvenberg MRDG, Van Der Putten J, Van Der Sommen F, et al. Deep learning algorithm for characterization of barrett’s neoplasia demonstrates high accuracy on NBI-zoom images. Gastroenterology. 2019;156:S-58. [Google Scholar]
  • 40.Tang DW, Wang L, He G, et al. Artificial intelligence network to aid the diagnosis of early esophageal squamous cell carcinoma and esophageal inflammations in white light endoscopic images. Gastrointestinal Endosc. 2019;89:AB654. [Google Scholar]
  • 41.Ebigbo A, Mendel R, Probst A, et al. Real-time use of artificial intelligence in the evaluation of cancer in Barrett’s oesophagus. Gut. 2020;69:615–616. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Guo L, Xiao X, Wu C, et al. Real-time automated diagnosis of precancerous lesions and early esophageal squamous cell carcinoma using a deep learning model (with videos). Gastrointest Endosc. 2020;91:41–51. [DOI] [PubMed] [Google Scholar]
  • 43.Ohmori M, Ishihara R, Aoyama K, et al. Endoscopic detection and differentiation of esophageal lesions using a deep neural network. Gastrointest Endosc. 2020;91:301–309. [DOI] [PubMed] [Google Scholar]
  • 44.Hashimoto R, Requa J, Dao T, et al. Artificial intelligence using convolutional neural networks for real-time detection of early esophageal neoplasia in Barrett’s esophagus (with video). Gastrointest Endosc. 2020;91:1264–1271. [DOI] [PubMed] [Google Scholar]
  • 45.Iwagami H, Ishihara R, Aoyama K, Fukuda H, Shimamoto Y, Kono M, et al. Artificial intelligence for the detection of esophageal and esophagogastric junctional adenocarcinoma. J Gastroenterol Hepatol. 2020. [DOI] [PubMed] [Google Scholar]

RESOURCES