Skip to main content
Lippincott Open Access logoLink to Lippincott Open Access
. 2021 Nov 4;56(1):23–35. doi: 10.1097/MCG.0000000000001629

Artificial Intelligence in the Diagnosis of Upper Gastrointestinal Diseases

Pierfrancesco Visaggi *, Nicola de Bortoli *, Brigida Barberio , Vincenzo Savarino , Roberto Oleas §, Emma M Rosi *, Santino Marchi *, Mentore Ribolsi , Edoardo Savarino †,
PMCID: PMC9988236  PMID: 34739406

Abstract

Artificial intelligence (AI) has enormous potential to support clinical routine workflows and therefore is gaining increasing popularity among medical professionals. In the field of gastroenterology, investigations on AI and computer-aided diagnosis (CAD) systems have mainly focused on the lower gastrointestinal (GI) tract. However, numerous CAD tools have been tested also in upper GI disorders showing encouraging results. The main application of AI in the upper GI tract is endoscopy; however, the need to analyze increasing loads of numerical and categorical data in short times has pushed researchers to investigate applications of AI systems in other upper GI settings, including gastroesophageal reflux disease, eosinophilic esophagitis, and motility disorders. AI and CAD systems will be increasingly incorporated into daily clinical practice in the coming years, thus at least basic notions will be soon required among physicians. For noninsiders, the working principles and potential of AI may be as fascinating as obscure. Accordingly, we reviewed systematic reviews, meta-analyses, randomized controlled trials, and original research articles regarding the performance of AI in the diagnosis of both malignant and benign esophageal and gastric diseases, also discussing essential characteristics of AI.

Key Words: artificial intelligence, gastrointestinal endoscopy, upper gastrointestinal cancer, GERD, Helicobacter pylori infection

WHAT IS ARTIFICIAL INTELLIGENCE (AI) AND HOW IT WORKS

The term AI generically refers to complex computer algorithms that mimic human cognitive functions, including learning and problem-solving.1 Machine learning (ML) is a field of AI that involves computer-based methods to elaborate data through algorithms. Conventional ML is based on hand-crafted algorithms in which researchers, based on clinical knowledge, manually indicate features of interest of an input dataset to train the system to recognize discriminative features, and provides appropriate outputs (ie, solving the problem of interpreting given data).1 The aim of ML is to find a generalizable model applicable to new data, which were not included in the training dataset, so that the computer can learn to interpret previously unknown information and provides reliable outputs.2 Learning techniques are divided into supervised, unsupervised, and reinforcement methods. A supervised learning model learns from known patterns,2 and requires the training dataset to contain input-output pairs to map new input to output.3 Unsupervised models are designed to classify subgroups of data according to commonalities without an a priori knowledge of groups significance.3 In reinforcement learning, the computer learns from its previous errors, adjusting the output over time.2

Recently, a derivative of ML referred to as deep learning (DL) has enthusiastically broken into the scene (Fig. 1). In contrast to ML, DL is more powerful as it autonomously extracts discriminative attributes of input data through an artificial neural network (ANN), often organized as convolutional neural networks (CNNs), which are constituted of multiple layers of nonlinear functions (Fig. 2).1,4 AI, ML, and DL are increasingly being integrated into computer-aided diagnosis (CAD) systems that can be applied to gastrointestinal (GI) diseases to improve recognition and characterization of pathology. The main application of AI in the upper GI tract is endoscopy. The ability to recognize endoscopic images depends on individual expertise, being interobserver and intraobserver variability a limit of endoscopic procedures. CAD tools have the potential to successfully assist both trainee and expert physicians to reduce variability in the detection of upper GI pathology, thus increasing the diagnostic accuracy regardless of individual expertise, and virtually overcoming current limitations of esophagogastroduodenoscopies (EGDS).5 Besides luminal imaging, AI has been applied to numerical and categorical data describing upper GI pathology to automate and optimize the assessment of diseases, including gastroesophageal reflux disease (GERD), eosinophilic esophagitis (EoE), and primary esophageal motility disorders.

FIGURE 1.

FIGURE 1

Relationship between artificial intelligence, machine learning, and deep learning. ANN indicates artificial neural network.

FIGURE 2.

FIGURE 2

Structure of convolutional neural networks.

As promising as it is, DL has its own limitations as models cannot apply reason throughout the decision process, and this may be counterproductive.5 DL models are black boxes in which the input data and the output (diagnosis) are known, but the processes by which the diagnosis is achieved are not, thus it is difficult to investigate the rationale for the diagnosis made by DL. In this regard, science not only is required to provide answers but also to explain why those are the answers for academic, legal, and ethical implications.6 Research is already heading to understand how DL models make decisions to solve interpretability gaps, and methods to understand the process of CNN-based choices are being developed. For example, an option is the use of “heat maps” that indicate what parts of an image the CNN has analyzed or altering model input data to appreciate how the outputs change.7 Of note, these highly sophisticated computational systems are not cost-effective at present and could lead physicians to rely on machines more than their clinical judgment while still retaining responsibility for decisions nonetheless.4 Despite some limitations, AI ex vivo and in vivo real-time support in decision-making is a fascinating hot topic in our time. Accordingly, we reviewed current knowledge regarding the application and performance of AI in the diagnosis of several esophageal and gastric diseases.

METHODS

We searched MEDLINE (PubMed), EMBASE, EMBASE Classic, and the Cochrane Library (from inception to April 2021) to identify systematic reviews, meta-analyses, randomized controlled trials, and original research articles reporting the performance of AI systems in the instrumental or clinical diagnosis of several esophageal and gastric diseases. The following terms were searched: AI or machine learning, deep learning. We combined these using the set operator AND with studies identified with the following terms: esophageal and gastroduodenal endoscopy, cancer, carcinoma, neoplasia, Barrett’s esophagus, esophagitis, gastro-esophageal reflux disease, GERD, eosinophilic esophagitis, EoE, motility disorder, varices, gastric cancer, atrophic gastritis, Helicobacter pylori infection. All terms were used as MeSH terms. Restriction to English language was applied. We screened titles and abstracts of all citations identified by our search for potential suitability and retrieved those that appeared relevant to examine them in more detail. The “snowball strategy” (ie, a manual search of the references listed in online databases publications) was performed to increase sources of information.

AI in Barrett’s Esophagus (BE) and Early Esophageal Adenocarcinoma (EAC)

The replacement of squamous esophageal epithelium with intestinal metaplasia containing goblet cells defines BE, which represents a well-known preneoplastic lesion for the development of EAC. BE has a predictable course through intermediate stages, namely BE with low-grade dysplasia, high-grade dysplasia (HGD), intramucosal carcinoma, and eventually invasive EAC.8 Thus, BE-derived EAC is preventable and amenable to surveillance strategies.

In 2018, esophageal cancer (EC) was estimated to account for 508,000 deaths, being the seventh most common cancer and the sixth cause of cancer death worldwide. Histologically, EAC accounts for 20% of all ECs, and its main risk factors include GERD and high body mass index.9 When diagnosed in advanced stages, EAC has a poor prognosis, with a 5-year survival rate of < 20%.3,9 However, when early detection and management are possible, the outcome improves significantly.10 At present, dysplasia and cancer surveillance in BE follows the Seattle protocol with random 4-quadrant biopsies every 2 cm, which is expensive, time-consuming, and has a sensitivity ranging from 28% to 85% for the detection of HGD/EAC.11 When in the hands of expert endoscopists, advanced endoscopic imaging techniques as narrow-band imaging (NBI) and confocal laser endomicroscopy (CLE) can meet optical diagnosis performance thresholds required by the Preservation and Incorporation of Valuable Endoscopic Innovations (PIVI) initiative by the American Society for Gastrointestinal Endoscopy12,13 [ie, per-patient sensitivity of 90% or greater, a negative predictive value (NPV)] of 98% or greater, and a specificity of at least 80% for detecting HGD or EAC).11 However, improving the diagnostic performance of EGDS in the detection of EC regardless of individual expertise is highly desirable, and AI is demonstrating enormous potential in this matter (Table 1).

TABLE 1.

AI in the Diagnosis of Esophageal Adenocarcinoma

Performance
References AI Model Study Type Aim Endoscopic Technique Accuracy Sensitivity Specificity
de Groof et al14 ML—SVM Retrospective Detection of early Barrett’s neoplasia WLE 92% 95% 85%
de Groof et al15 DL Prospective Detection of early Barrett’s neoplasia WLE 90% 91% 89%
Ebigbo et al16 DL Prospective Detection of early Barrett’s neoplasia WLE 89.9% 83.7% 100%
de Groof et al17 DL Retrospective AI vs. endoscopists in detection of early Barrett’s neoplasia WLE AI EE 88%  73% 93%  72% 83%  74%
Swager et al18 ML—SVM Retrospective AI vs. endoscopists in detection of early Barrett’s neoplasia VLE AI EE AUC=0.95  AUC=0.81 90%  85% 93%  68%
van der Sommen et al19 ML—SVM Retrospective AI vs. endoscopists in detection of early Barrett’s neoplasia WLE AI Best endoscopist —  — 86%  90% 87%  91%
Ebigbo et al20 DL Retrospective AI vs. endoscopists in predicting invasion in Barrett’s cancer WLE AI EE 77%  63% 64%  78% 71%  70%

AI indicates artificial intelligence; AUC, area under the curve; DL, deep learning; EE, expert endoscopists; ML, machine learning; SVM, support vector machine; VLE, volumetric laser endomicroscopy; WLE: white-light endoscopy.

A recent meta-analysis21 revealed that AI systems detected BE-related EC with white-light endoscopy (WLE)14,15,17,19 or volumetric laser endomicroscopy (VLE)18 with a pooled sensitivity of 88% [95% confidence interval (CI), 82.0%-92.1%], pooled specificity of 90.4% (95% CI, 85.6%-94.5%), and an area under the curve (AUC) of 0.96 (95% CI, 0.93-0.99). When compared with general endoscopists operating with standard WLE17,19 or VLE,18 AI performed better than physicians on the detection of neoplastic lesions in BE.21 Specifically, AI systems had an AUC of 0.96 (95% CI, 0.94-9.97) versus 0.82, P<0.001; sensitivity 90.7% (95% CI, 89.8%-91.5%) versus 72.3% (95% CI, 70.2%-74.3%), P<0.001; and specificity 88.0% (95% CI, 87.1%-88.9%) versus 74.0% (95% CI, 72.2%-75.7%), P<0.001. However, AI was tested on optimal endoscopic images but not during live EGDS. The retrospective study of Van Riel et al22 showed consistent results on still endoscopic images. The AI system detected early BE neoplasms from the public MICCAI 2015 image dataset with AUC of 0.92. Another study applied image data augmentation through Generative Adversarial Networks (GANs) to increase the identification of BE and EAC compared with standard endoscopic images.23 The combination of CNNs and GANs allowed to achieve 85% accuracy in the task. Liu et al24 recently tested a DL-SVM combined CAD tool to automate the classification of esophageal findings on WLE images. The proposed network achieved accuracy for the classification of cancer, premalignant lesion, and normal esophagus of 77.14%, 82.5%, and 94.23%, respectively. Iwagami et al25 trained a DL model to recognize esophageal junctional cancers under WLE and compared its performance with that of expert physicians. The AI system showed a favorable sensitivity of 94%, and specificity of 42% for noncancerous lesions (esophagitis, polyp), which was comparable to that of expert endoscopists (43%). Another DL algorithm was trained to differentiate EC from BE, inflammation, and normal mucosa.26 The CAD tool classified the above lesions with overall accuracy of 96%.

AI has shown good performance also in lesion characterization. A pilot study demonstrated that an AI-based system performed as good as international expert physicians in the prediction of submucosal invasion (ie, differentiating stage T1a vs. T1b) in endoscopic images of Barrett’s cancer, having a sensitivity of 77%, a specificity of 64% and an accuracy of 71%.20

To enable the integration of CAD systems into clinical practice, research is now concentrating on the real-time use of CAD tools that instantly provide feedbacks to the endoscopist. A preceding to this is the application of AI diagnosis to video clips. Accordingly, in a recent study, a DL model was trained on images and tested on NBI zoom video clips of EAC and NDBE.27 The CAD system showed good performance with 83% accuracy, 85% sensitivity, and 83% specificity. Further, de Groof et al15 assessed the accuracy of a CAD system for the detection of Barrett’s neoplasia within endoscopic images systematically taken every 2 cm in Barrett’s areas during live endoscopic procedures. The system was tested in real-time on 10 patients with NDBE and 10 patients with BE-related EAC. Standard WLE images were obtained and analyzed live by a DL algorithm that met PIVI threshold with an accuracy of 90%, a sensitivity of 91%, and a specificity of 89%. A different real-time approach was designed by Ebigbo et al.16 In their AI system, endoscopic images were randomly captured from the camera livestream during endoscopic procedures. The immediate AI analysis differentiated between NDBE and EAC with a sensitivity of 83.7%, a specificity of 100%, and an accuracy of 89.9%.

AI in Squamous Dysplasia and Early Esophageal Squamous Cell Carcinoma (ESCC)

ESCC accounts for up to 90% of ECs in lower income countries.9 It has a poor prognosis with an overall 5-year survival rate of 18%, which decreases to < 5% when distant metastases are present at diagnosis.28 Early detection may potentially improve the outcome of the disease. Endoscopic recognition of early ESCC is challenging, as lesions often pass unrecognized with standard WLE. Lugol’s dye spray chromoendoscopy, virtual chromoendoscopy with NBI and blue-laser imaging (BLI) bright have shown accuracy in the detection of ESCC,2931 although nonexpert endoscopists may not perform as good as experts,31 limiting their applicability.

To fill the gap, AI has been explored (Table 2). Regarding the diagnosis by NBI or BLI-bright, the performance of AI was compared with that of endoscopy specialists of the Japan Gastroenterological Endoscopy Society.35 The sensitivity of the AI system was greater than that of experienced physicians (100% vs. 92%), and the specificity for noncancerous lesions was not significantly lower (63% vs. 69%). DL-based CAD tools have also been challenged to recognize early ESCC under WLE, proving higher accuracy than nonexpert endoscopists and comparable accuracy to expert endoscopists (97.6% vs. 88.8%, and 77.2%, respectively).32 Wang et al41 developed a single shot multibox detector with a CNN algorithm that performed well in the detection of ESCC using WLE and NBI. The system diagnosed ESCC with 90.9% accuracy. A meta-analysis confirmed that the accuracy of AI in the detection of ESCC was significantly higher when images were analyzed with NBI33,35,42 than WLE,33,42 being the AUC 0.92 (95% CI, 0.86-1.00) versus 0.83 (95% CI, 0.82-0.84).21 When pooled together, the studies that used AI with NBI, WLE, endocytoscopy,34 or optical magnifying endoscopy (ME)35 to recognize ESCCs, had an AUC of 0.88 (95% CI, 0.82-0.96), a specificity of 92.5% (95% CI, 66.8%-99.5%), and a sensitivity of 75.6% (95% CI, 48.3%-92.5%).21

TABLE 2.

AI in the Diagnosis of ESCC

Performance
References AI Model Study Type Aim Endoscopic Technique Accuracy Sensitivity Specificity
Cai et al32 DL Retrospective Detection of ESCC WLE 91.4% 98% 85%
Guo et al33 DL Retrospective Detection of ESCC NBI images  NBI videos AUC=0.989  100% 98%  100% 95%  100%
Kumagai et al34 DL Prospective Detection of ESCC ECS AUC=0.85 39% 98%
Ohmori et al35 DL Retrospective Detection of ESCC WLE  NBI/BLI  ME—BLI/NBI 81%  77%  77% 90%  100%  98% 76%  63%  56%
Tokai et al36 DL Retrospective Detection of ESCC WLE/NBI 96%
Estimating invasion depth of ESCC WLE SM1=93%  SM2=97% —  — —  —
NBI SM1=97%  SM2=100% —  — —  —
Nakagawa et al37 DL Retrospective Estimating invasion depth of ESCC WLE/NBI/BLI SM1=93%  SM2=90% 95%  94% 79%  75%
ME—WLE/NBI/BLI SM1=90  SM2=92 92%  94% 79%  86%
Shimamoto et al38 DL Retrospective Estimating invasion depth of ESCC WLE/NBI/BLI  ME—WLE/NBI/BLI 87%  89% 50%  71% 99%  95%
Everson et al39 DL Retrospective Detection of abnormal IPCL ME-NBI 98% 99% 97%
Zhao et al40 DL Retrospective Classification of IPCL ME-NBI 89%

AI indicates artificial intelligence; AUC, area under the curve; BLI, blue-laser imaging; DL, deep learning; EAC, esophageal adenocarcinoma; ECS, endocytoscopic system; ESCC, esophageal squamous cell carcinoma; IPCL, interpapillary capillary loop; ME, magnified endoscopy; NBI, narrow-band imaging; SM, submucosal; WLE, white-light endoscopy.

AI has also been tested on the characterization of mucosal invasion of ESCC through the analysis of esophageal intrapapillary capillary loops (IPCLs), which are microvascular structures on the surface of the esophagus. IPCLs appear as brown loops on ME with NBI and show morphologic changes that strictly correlate with neoplastic invasion depth of ESCC, allowing intraprocedural decisions for endoscopic resections.43,44 However, optical classification of IPCL requires experience and is mastered by experts only. Accordingly, it was developed an AI-based automated IPCL classification whose accuracy was significantly higher than that of endoscopists with <15 years of experience.40 A CNN-based AI system39 was trained with sequential high-definition ME-NBI images from 17 patients (10 ESCN, 7 normal), and distinguished abnormal IPCL patterns with 93.7% accuracy. The sensitivity and specificity to classify abnormal IPCL patterns were 89.3% and 98%, respectively. In another study, AI estimated the invasion depth of ESCC from NBI/WLE images better than 13 expert endoscopists, showing a sensitivity of 84.1% versus 78.8%, a specificity of 73.3% versus 61.7%, and an accuracy of 80.9% versus 73.5%.36 Another AI system showed good performance in differentiating mucosal and submucosal microinvasive (SM1) cancers from the submucosal deep invasive (SM2/3) ones with a sensitivity of 90.1%, a specificity of 95.8%, and an accuracy of 91.0%. The performance of the system was comparable to that of experienced physicians.37 AI also showed comparable results to expert clinicians when classifying IPCLs from video clips.45 A CNN model showed accuracy, sensitivity, and specificity of 91.7%, 93.7%, and 92.4%, respectively, in the recognition of abnormal IPCLs during the analysis of video frames.

Growing proficiency in AI systems allowed the development of real-time operating CAD tools. A CAD video model33 was capable of processing at least 25 frames/s of NBI images in < 100 ms with encouraging performance. The dataset included precancerous lesions, early ESCC, and nonpathologic findings. When analyzing non-ME videos, the per-frame and per-lesion sensitivity of the AI system were 60.8% and 100%, respectively. Notably, the per-frame sensitivity increased to 96.1%, and the per-lesion sensitivity remained stable to 100% with ME videos. Another recent study46 compared the ability of a CAD system to that of 13 expert endoscopists to identify and characterize suspicious lesions from video clips of NBI esophagoscopies. Regarding detection performance, AI sensitivity was significantly higher than that of experts, being 91% versus 79%, whereas AI specificity and accuracy were lower, being 51% versus 72%, and 63% versus 75%, respectively. As for differentiating cancerous from noncancerous lesions, the AUC showed that the AI system had significantly better diagnostic performance than physicians, being the sensitivity 86% versus 74%, the specificity 89% versus 76%, and the accuracy 88% versus 75%. Yang et al47 developed a real-time operating AI system that could detect 100% and 95% of early ESCC from ME and non-ME WLE video clips, respectively. Waki et al48 challenged the AI to diagnose ESCC from video clips simulating a situation in which the CAD tool could assist clinicians during routine EGD. The assistance of AI significantly improved the sensitivity for ESCC diagnosis by 2.7%. Similarly, Li et al49 compared AI-aided detection of ESCC under WLE and NBI, and evaluated the yield of its support to endoscopists. The accuracy of the AI system with NBI and WLE was 94.3% and 89.5%, respectively, whereas the average accuracy of endoscopists was 81.9%. Remarkably, the assistance of AI allowed endoscopists to achieve the highest accuracy of 94.9% with NBI and WLE.

A recently developed CAD system38 achieved a favorable performance at estimating the invasion depth of ESCC in video images with nonmagnifying WLE or ME with NBI/BLI. Accuracy, sensitivity, and specificity of the AI system with non-ME versus those of endoscopists with up to 15 years of experience, were 87% versus 85%, 50% versus 45%, and 99% versus 97%. Good performance was also achieved when comparing the AI system with ME, where accuracy, sensitivity, and specificity were 89% versus 84%, 71% versus 42%, and 95% versus 97%, respectively. AI could potentially help with other endoscopic techniques that require much experience. Accordingly, a clinical trial aiming to evaluate the automatic diagnosis of early ESCC with probe-based CLE is currently recruiting (NCT04136236). The primary outcome will be to test the real-time diagnostic performance of the AI system with probe-based CLE and the secondary to compare AI performance to that of endoscopists.

AI in Benign Esophageal Diseases: Gastroesophageal Reflux, Motility Disorders, Esophagitis, and Varices

The huge computing power of AI facilitates and optimizes the analysis of large amounts of data at once, allowing to recognize complex nonlinear interactions between variables. Accordingly, AI has been applied to esophageal benign disorders (Table 3) including GERD, primary motility disorders, EoE, cytomegalovirus (CMV) and herpes simplex virus (HSV) esophagitis, and esophageal varices.

TABLE 3.

AI in the Diagnosis of Benign Esophageal Diseases

Performance
References AI Model Study Type Aim Diagnostic Tool Accuracy Sensitivity Specificity
Pace et al50 ML Prospective Distinction between GERD and non-GERD based on symptoms Questionnaire 100%
Horowitz et al51 Data mining Prospective Distinction between GERD and non-GERD based on symptoms Questionnaire AUC=0.78 70%-75% 63%-78%
Pace et al52 ML Prospective Distinction between NERD and EE Questionnaire NERD 62.2% EE 70.9%
Rogers et al53 Decision tree analysis Prospective Automate extraction of pH-impedance metrics pH-impedance tracings 88.5%
Rogers et al53 Decision tree analysis Prospective Predict response to GERD management pH-impedance tracings AUC=0.77
Gulati et al54 DL Prospective Endoscopic diagnosis of GERD NF-NBI AUC=0.83 67% 92%
Sallis et al55 ML Prospective Diagnosis of EoE based on mRNA transcripts from esophageal biopsies Esophageal biopsies AUC=0.98 91% 93%
Santos et al56 Multilayered back-propagation ANN Prospective Diagnosis of esophageal motility pattern Stationary esophageal manometry tracings 82%
Lee et al57 DL Retrospective AI vs. endoscopist in differential diagnosis HSV vs. CMV esophagitis WLE AI Endoscopists 100% 52.7% 100% — 100% —

AI indicates artificial intelligence; ANN, artificial neural network; AUC, area under the curve; CMV, cytomegalovirus; DL, deep learning; EE, erosive esophagitis; EoE, eosinophilic esophagitis; GERD, gastroesophageal reflux disease; HSV, herpes simplex virus; ML, machine learning; NERD, nonerosive reflux disease; NF-NBI, near-focus narrow-band imaging; WLE, white-light endoscopy.

GERD is defined as the presence of troublesome symptoms caused by gastroesophageal reflux and/or esophageal mucosal lesions.58 It has a spectrum of symptoms (eg, heartburn, regurgitation, noncardiac chest pain) that overlap with reflux hypersensitivity and functional heartburn,5962 which can make clinical distinction difficult.

In 2005, an AI system based on ML was developed to discriminate, based on symptoms solely, between patients with a pathologic esophageal acid pH exposure with or without esophagitis, and normal individuals.50 Patients were asked to fill in the Gastro-Esophageal Reflux Questionnaire (GERQ) proposed by the Mayo Clinic,63 and underwent an EGD and/or a 24-hour esophageal pH-metry. Among patients, 103 had an objectively confirmed GERD, and 56 were normal. The AI system automatically selected the 45 most relevant variables from the GERQ (collectively referred to as QUID, “QUestionario Italiano Diagnostico”) (Table 4), and this allowed the CAD tool to reach a predictive accuracy up to 100%, as it correctly predicted GERD in all 103 patients. Another study with ML combined AI and the QUID questionnaire to differentiate between GERD patients with and without erosive esophagitis (ie, nonerosive reflux disease).52 The CAD tool successfully distinguished between GERD and normal patients but failed to discriminate erosive esophagitis from nonerosive reflux disease based solely on symptoms evaluation. Horowitz et al51 developed and validated a shorter questionnaire of 15 variables (Table 5) aiming to discriminate, through an ANN-based algorithm, GERD from non-GERD patients. The sensitivity of the model was 70% to 75%, the specificity 63% to 78%, and the AUC 0.787.

TABLE 4.

Variables Included in the QUestionario Italiano Diagnostico Questionnaire52

Heartburn Persistent Gastric/Intestinal Pain Prescribed Examinations for GERD Relative With Gastroduodenal Disorder Medical Visits for GERD
Hiatus hernia Cough intensity Aspirin, frequency of use Marital status Episodes of breathlessness
Frequency of chest pain Periodic frequency of swallowing problems Antirheumatic drugs, frequency of use Cough frequency/year Asthma
Waking at night–retrosternal pain Intensity of swallowing problems esophageal dilation Slow walk—chest pain Belching
Pneumonia Frequency of medical check-ups Variation in weight in past year Heart therapy Sibilant rhonchi or wheezing
Cough Heart disorders Coffee drinker Intensity of chest pain Cough at night
Retrosternal pain Vomiting (frequency) Regular smoker Chest pain in past year Ingestion of beverages—chest pain
Interference with daily activities Acid reflux in mouth Lump in throat Alcohol units/week in past year Total acid reflux
Sleeping semisupine Esophageal surgery Health state during past year Hiccups Intensity of acid reflux

GERD indicates gastroesophageal reflux disease.

TABLE 5.

Symptoms/Signs Evaluated in the Artificial Neural Network-based Questionnaire Proposed by Horowitz et al51 to Primary Care Physicians to Diagnose Gastroesophageal Reflux Disease

Abdominal Pain Belching Halitosis
Chest pain Heartburn Relief with antacid medications
Bloating Regurgitation Stress
Nausea Dysphagia Bend/lie aggravation
Vomiting Sour taste Heavy meal aggravation

Moving from categorical to numerical data, AI has also been applied to pH-impedance studies. Twenty-four-hour pH-impedance monitoring is used to quantify reflux episodes and acid exposure time (AET) to rule in or out the diagnosis of GERD.6468 The mean nocturnal baseline impedance is a novel metric that demonstrated to increase GERD diagnostic yield and predict treatment outcome.65,6974 A recently published proof-of-concept study showed that AI has the potential to automate the extraction and to elaborate useful novel AI pH-impedance metrics. CNN was considered inadequate to be applied to impedance tracings, and a python-based decision tree analysis algorithm was developed to analyze raw pH-impedance data.53 The AI system autonomously evaluated 2049 pH-impedance events with an accuracy of 88.5%, and calculated recumbent and upright values of baseline impedance (AIBI). The upright AI divided by recumbent AI ratio (U:R AIBI ratio) segregated responders to treatment from controls and nonresponders regardless of treatment status upon pH-impedance recording. Moreover, the U:R AIBI ratio at 5 cm above the lower esophageal sphincter outperformed total AET in predicting response to medical therapy in those with AET >6% (AUC: 0.766 vs. 0.606, respectively).

In a recent clinical trial (NCT04268719), an image-driven AI model for the diagnosis of GERD was developed. A published abstract demonstrated the potential of AI-driven near-focus NBI endoscopy for the real-time diagnosis of GERD through the recognition of regions of interests and IPCLs.54 The CNN-based model could diagnose GERD in real-time during endoscopic procedures with a sensitivity, specificity, and AUC up to 67%, 92%, and 0.83, respectively.

The feasibility of applying ANNs in the recognition and classification of primary esophageal motor disorders has also been investigated.56 Two different ANN models were trained to recognize normal and abnormal swallow sequences of pressure wave patterns of conventional stationary manometry recordings. The model correctly classified >80% of swallow sequences, diagnosing 100% of cases of achalasia, 100%, of nutcracker esophagus, 80% of ineffective esophageal motility, 60% of diffuse esophageal spasm, and 80% of normal motility. However, the study took place in 2006, when high-resolution manometry recordings and current highly sophisticated AI algorithms were not available.

Early encouraging applications of AI to EoE have also been reported. EoE is a chronic, local, progressive, T-helper type 2 immune-mediated esophageal disorder.7578 Clinical manifestations vary according to the age of diagnosis, and a timely diagnosis may be difficult.77,79 Accordingly, an AI-based automated algorithm was developed to assist in the diagnosis of EoE.55 The AI system elaborated a diagnostic probability score for eosinophilic esophagitis (pEoE) based on esophageal mRNA transcripts from biopsies of EoE patients, including genes encoded by the EoE transcriptome.76 During the process, individual transcripts were automatically assigned weights by the system. Interestingly, established EoE markers (eg, eotaxin and periostin)76 were weighed higher. For validation, the pEoE score was applied to a set of external patients in a blinded fashion. A pEoE score ≥25 detected EoE patients with a sensitivity of 91%, a specificity of 93%, and AUC 0.985. Importantly, the pEoE score improved the diagnosis of equivocal EoE cases with 84.6% accuracy, distinguishing EoE from GERD. In treatment-responsive patients (ie, < 5 eosinophils/HPF), the pEoE score decreased below the diagnostic cutoff of 25 and remained ≥25 in the 1 patient whose eosinophilia did not resolve completely (5 <eosinophils/HPF<15). To date, no studies or clinical trials combining AI and endoscopic evaluation of EoE have been published.

AI has shown to be useful also in the optical endoscopic differential diagnosis of CMV versus HSV esophagitis. CMV and HSV esophagitis have overlapping endoscopic findings and are infrequent in clinical practice, which makes the diagnosis challenging and hampers the start of a targeted treatment before the histopathologic diagnosis. Accordingly, Lee et al57 trained a DL system to differentiate CMV from HSV esophageal ulcers with impressive results. The accuracy, sensitivity, and specificity of the CAD tool were 100%, largely outperforming endoscopists, whose accuracy was only 52.7%.

Finally, Guo et al80 trained a DL system to classify multiple GI lesions. Among these, esophageal varices could be automatically diagnosed by the system with 90.5% sensitivity.

AI IN GASTRIC DISEASES

AI in Gastric Cancer (GC) and Chronic Atrophic Gastritis (CAG)

GC is one of the 5 most common cancer-related diagnoses globally and the third leading cause of death in cancer patients. Early recognition of premalignant lesions is the main goal to reduce the burden of the disease. Endoscopic evaluation is the standard of care in assessing the gastric mucosa. A recent systematic review and meta-analysis evaluated the data regarding AI and GC demonstrating the expanding role and the favorable impact of this technology for current and upcoming years.81 Hirasawa et al82 developed the first AI system for the detection of GC using DL. Although demonstrating a high sensitivity (92.2% for lesions of 5 mm or less; 98.6% in lesions of 6 mm or more), its positive predictive value (PPV) was 30.6%, due to misdiagnosis of mild-moderate CAG and intestinal metaplasia lesions. In this regard, another study showed that a DL approach had a significantly higher accuracy than experts in the assessment of CAG.83 Other AI systems have been developed since (Table 6), such as the one proposed by Horiuchi et al89 which differentiated GC from gastritis with 85.3% accuracy, 95.4% sensitivity, and 71.0% specificity. Wu et al,85 tested a DL system in the detection of GC. The sensitivity, specificity, accuracy, PPV, and NPV were 94.0%, 91.0%, 92.5%, 91.3%, and 93.8%, respectively. Li et al84 developed an AI model based on CNN to differentiate between noncancerous and early gastric mucosal lesions under ME. The proposed system showed a sensitivity, specificity, and accuracy of 91.1%, 90.6%, and 90.9%, respectively, in the diagnosis of early GC, which were significantly higher than those of nonexpert endoscopists. Lee et al88 combined a residual network with transfer learning to distinguish between GC, ulcers, and normal gastric mucosa with almost 90% accuracy. Zhang et al87 used the CNN DenseNet to identify CAG lesions, with diagnostic accuracy, sensitivity, and specificity of 0.94, 0.95, and 0.94. The classification accuracy between mild, moderate, and severe CAG was 0.93, 0.95, and 0.99 demonstrating higher detection rates for moderate and severe cases. Ueyama et al86 tested a DL system on 2300 ME-NBI images (1430 GC images), achieving 98.7% accuracy, 98% sensitivity, and 100% specificity for the diagnosis of GC. In a multicenter study,92 the performance of AI with ME-NBI was similar to that of senior endoscopists and better than that of junior endoscopists. Interestingly, the diagnostic ability of endoscopists improved significantly after referring to the results provided by the AI system, providing insights into a useful application of AI in routine practice.

TABLE 6.

AI in the Diagnosis of Gastric Cancer and CAG

Performance
References AI Model Study type Aim Diagnostic Tool Accuracy Sensitivity Specificity
Li et al84 DL Retrospective Diagnosis of GC ME-NBI 90.9% 91.1% 90.6%
Hirasawa et al82 DL Retrospective Diagnosis of GC WLE, NBI, IC 92.2%-98.6%
Wu et al85 DL Retrospective Diagnosis of GC WLE, NBI, BLI 92.5% 94.0% 91.0%
Ueyama et al86 DL Retrospective Diagnosis of GC ME-NBI 98.7% 98.0% 100%
Zhang et al87 DL Retrospective Detection of CAG WLE 94.0% 95.0% 94.0%
Lee et al88 DL Retrospective Differential diagnosis GC vs. gastric ulcer WLE 77.1%-90%
Horiuchi et al89 DL Retrospective Differential diagnosis GC vs. gastritis ME-NBI 85.3% 95.4% 71.0%
Zhu et al90 DL Retrospective Characterization of GC invasion depth WLE 89.1% 76.5% 95.5%
Nagao et al91 DL Retrospective Characterization of GC invasion depth WLE, NBI, IC 94.5% 84.4 99.4%

AI indicates artificial intelligence; BLI, blue-laser imaging; CAG, chronic atrophic gastritis; DL, deep learning; GC, gastric cancer; IC, indigo-carmine dye contrast; ME, magnified endoscopy; NBI, narrow-bad imaging; WLE, white-light endoscopy.

AI has also been applied to the recognition of GC invasion depth. Importantly, patients with cancers extending within the mucosa or submucosal layer could benefit from curative endoscopic resection regardless of lymph node involvement. However, the prediction of invasion depth is challenging in clinical practice. Accordingly, a CNN detection system was developed to determine GC invasion depth based on WLE images.90 The model reached an overall accuracy of 89.1%, sensitivity of 76.5%, and specificity of 95.5%. Of note, the system had a significantly higher accuracy of 17.2% and a higher specificity of 32.2% than expert endoscopists. Similarly, Nagao et al91 compared AI with WLE, NBI, and indigo-carmine dye contrast imaging in characterizing GC invasion depth. The authors calculated accuracies of 94.5%, 94.2%, and 95.5%, respectively, without significant differences among imaging techniques.

AI in the Detection of Helicobacter pylori (Hp) Infection

Hp is a gram-negative bacterial pathogen that selectively colonizes the gastric epithelium.93 The infection is a leading cause of gastroduodenal pathology.94 In 1994, the International Agency for Research on Cancer labeled Hp as a definite (group I) carcinogen for GC.95 Indeed, the infection is the major cause of chronic gastritis that sequentially causes precancerous modifications, namely atrophic gastritis, intestinal metaplasia, dysplasia and ultimately, GC. Accordingly, the eradication of Hp has the potential to prevent the development of preneoplastic lesions of gastric mucosa.96 More than 50% of the world’s population is infected with Hp,97 which makes the widest possible diagnosis and eradication of the infection a matter of global health interest. No single endoscopic approach provides validated accuracy for optical diagnosis of Hp infection at present, and endoscopic biopsies are required.7 Recently, several AI models have shown potential to overcome biopsy sampling, or provide “heat maps” for targeted biopsies, and allow accurate optical detection of Hp-infected gastric mucosa to achieve a prompter diagnosis (Table 7).

TABLE 7.

AI in the Diagnosis of Helicobacter pylori Infection

Performance
References AI Model Study Type Aim Diagnostic Tool Accuracy Sensitivity Specificity
Shichijo et al98 DL Retrospective Detection Hp infection WLE 87.7% 88.9% 87.4%
Yasuda et al99 DL Retrospective Detection Hp infection LCI 87.6% 90.5% 85.7%
Zheng et al100 DL Retrospective Detection Hp infection WLE Per-image analysis Per-patient analysis 84.5%  93.8% 81.4%  91.6% 90.1%  98.6%
Shichijo et al101 DL Prospective Detection of Hp-eradicated gastric mucosa WLE 84%
Nakashima et al102 DL Prospective Detection Hp infection WLE  BLI-bright  LCI AUC=0.66  AUC=0.96  AUC=0.95 66.7%  96.7%  96.7% 60%  86.7%  83.3%
Itoh et al103 DL Prospective Detection Hp infection WLE 0.956 86.7% 86.7%
Huang et al104 RFSNN Prospective Detection Hp infection WLE 85.4% 90.9%
Huang et al105 SVM Retrospective Detection Hp infection WLE 86%

AI indicates artificial intelligence; AUC, area under the curve; BLI, blue-laser imaging; DL; deep learning; Hp, Helicobacter pylori; LCI, linked color imaging; RFSNN, refined feature selection with neural network; SVM, support vector machine; WLE, white-light endoscopy.

An early study by Huang et al104 investigated a refined feature selection with neural network model for prediction of Hp-related gastric histologic features. The authors trained the AI system with endoscopic images of 30 prospectively enrolled dyspeptic patients with and without histologically confirmed Hp infection, and then tested the performance of the model on endoscopic images on 74 dyspeptic patients previously unknown to the system. The refined feature selection with neural network model showed a sensitivity of 85.4% and a specificity of 90.9% for the detection of Hp infection, and an accuracy >80% in predicting the presence of gastric atrophy, intestinal metaplasia, and the severity of gastric inflammation. Another study105 aimed to automate the recognition of histologic parameters proposed by the updated Sydney system.106 The authors applied a SVM-based AI model to WLE images, which achieved an accuracy rate of 86% for the detection of Hp by image analysis for any topographic locations over the gastric antrum, body, or cardia. More recently, several CNN models have been engineered to support the optical diagnosis of Hp. In 2017, Shichijo et al98 developed 2 CNN-based CAD tools. The first was trained with WLE images classified according to the presence or absence of Hp infection. The system achieved an accuracy of 83.1%, a sensitivity of 81.9%, and a specificity of 83.4% in 198 seconds for the detection of Hp infection. The second CNN system was trained with images classified according to anatomic locations, achieving an accuracy of 87.7%, a sensitivity of 88.9%, and a specificity of 87.4% in 194 seconds for the diagnosis of Hp infection. Notably, the diagnostic performance of the second CNN model was significantly higher than that of 23 endoscopists, who achieved an accuracy of 82.4% in the considerably longer time of 230 minutes. A subsequent study further challenged a CNN model to recognize the gastric mucosa of Hp-eradicated patients.101 The AI system showed 84% accuracy for the detection of Hp-eradicated gastric mucosa on WLE images. Itoh et al103 trained a CNN-based CAD system with WLE images of Hp-positive and negative patients. The model performed well for detection of Hp infection on prospectively collected endoscopic images, with an AUC of 0.956, a sensitivity and a specificity of 86.7%. Similarly, another study trained and validated a DL model with 2 different retrospectively collected WLE sets of images from a total of 1959 patients.100 The system showed high accuracy, with a per-image AUC of 0.93, sensitivity of 81.4%, specificity of 90.1%, and accuracy of 84.5%. Per-patient AUC was 0.97, with sensitivity 91.6%, specificity 98.6%, and accuracy 93.8%. Recently, AI has been applied to advanced endoscopic imaging for the diagnosis of Hp infection, including linked color imaging (LCI) and BLI-bright. LCI is a new image-enhanced endoscopy system developed by Fujifilm Co. (Tokyo, Japan) that enhances even slight mucosal color differences. The technique has already proven accuracy in the optical diagnosis of active Hp infection.99,107 Yasuda et al99 developed an SVM-based classification algorithm to detect Hp infection on LCI endoscopic images. The authors trained an AI system by SVM, using LCI images of infected and uninfected patients, and then compared the performance of the CAD system to that of an expert in LCI, a gastroenterologist specialist, and a senior resident. For the diagnosis of Hp infection, the AI system proved accuracy of 87.6%, sensitivity of 90.5%, and specificity of 85.7%. Of note, the system significantly outperformed the senior resident. The performance of the LCI expert and the gastroenterology specialist were comparable to that of AI system, with nonsignificant differences. Nakashima et al102 assessed the performance of a CNN-based AI system for the diagnosis of Hp infection comparing WLE and image-enhanced endoscopy. Three still images of the lesser curvature of the stomach were taken with WLE, BLI-bright, and LCI for each of the 222 included patients (105 infected). The images of 162 patients were used to train the AI system, those of 60 patients were used for validation. BLI-bright and LCI significantly outperformed WLE in a comparable amount of time. In detail, the AUC for BLI-bright and LCI were 0.96 and 0.95, whereas the AUC for WLE was 0.66. BLI-bright had a sensitivity of 96.7% and a specificity of 86.7%, LCI had a sensitivity of 96.7% and a specificity of 83.3%, and WLE had a sensitivity of 66.7% and a specificity of 60.0%.

DISCUSSION AND CONCLUSION

AI is attracting increasing attention in diagnostic imaging and complex medical data analysis. It has been observed that CAD tools provide an interpretable universal method for clinical instrumental and endoscopic diagnosis of GI diseases (Table 8), virtually eliminating interobserver variability and reducing the rate of blind spots during EGDS.108 Despite the in vivo application of AI is relatively recent, exciting results have already been achieved.

TABLE 8.

Summary Table: Current Application of AI in the Upper Gastrointestinal Tract

Target Organ Type of Disease Disease Current Application of AI Future Perspectives of AI
Esophagus Malignant/premalignant Barrett’s esophagus Detection Anticipate the diagnosis of cancer and improve the prognosis of the disease
Adenocarcinoma Detection Characterization (invasion depth)
Squamous dysplasia Detection
Interpapillary capillary loops Detection Characterization (morphology)
Squamous cell carcinoma Detection Characterization (invasion depth)
Benign GERD Endoscopic automated diagnosis of abnormal acid exposure Noninvasive diagnosis based on symptoms Differential diagnosis NERD vs. ERD Automated extraction of metrics from pH-impedance tracings Prediction of response to treatment Automate the diagnosis of NERD during endoscopy, provide a noninvasive conclusive diagnosis of GERD, and make quicker the interpretation of pH-impedance tracings, which currently require a time-consuming manual revision
Esophagitis Differential diagnosis between CMV and HSV esophagitis Anticipate diagnosis and start treatment without waiting for the histologic diagnosis
Dysmotility Detection Characterization of motility pattern (manometric diagnosis) Automate and speed-up the diagnosis of motility disorders
Varices Detection Diagnosis Automate the detection and classification of varices during routine endoscopy
Stomach Malignant/premalignant Chronic atrophic gastritis Detection Anticipate the diagnosis of cancer and improve the prognosis of the disease
Gastric cancer Detection Characterization (invasion depth)
Benign Helicobacter pylori infection Detection Automated diagnosis regardless of confounding factors (eg, PPI therapy), start treatment without waiting for the histologic diagnosis
Gastric ulcer Differential diagnosis with cancer Anticipate the diagnosis and start treatment without waiting for the histologic diagnosis

AI indicates artificial intelligence; CMV, cytomegalovirus; ERD, erosive reflux disease; GERD, gastroesophageal reflux disease; HSV, herpes simplex virus; NERD, nonerosive reflux disease; PPI, proton-pump inhibitors.

A recent meta-analysis confirmed the potential of AI to increase the diagnostic yield and reduce underdiagnosis of neoplastic lesions.109 AI systems had a sensitivity of 90%, a specificity of 89%, a PPV of 87%, and an NPV of 91% without significant performance differences in the diagnosis of ESCC, BE-related EAC, or GC. In addition, CAD tools showed good performance in the characterization of submucosal invasion of esophageal and gastric cancerous lesions.3640,43,44,90 This has relevant therapeutic and prognostic implications as early lesions are amenable of endoscopic treatment.110

AI also proved utility to support both clinical and endoscopic diagnosis of benign upper GI pathology. AI systems helped in the development of questionnaires that accurately predict GERD and differentiate disease from normality.5052 AI models autonomously extracted and analyzed pH-impedance tracings and also individuated a novel pH-impedance metric, that is, mean nocturnal baseline impedance, which enables to segregate responders from nonresponders to GERD treatment, thus reducing reporting times and virtually improving reflux management.53 The possibility of a real-time endoscopic GERD diagnosis was also shown.54

A CAD tool could recognize stationary manometry motor patterns with accuracy,56 but the application of novel CAD tools to high-resolution manometry recordings is yet to be evaluated.

The ability of AI to analyze complex variables interactions allowed the development of a score that could predict EoE with good accuracy.55 Although NBI can accurately differentiate EoE from control patients,111 there are no published studies that combine NBI and AI for the diagnosis of EoE. It is tempting to speculate that AI could support the endoscopic diagnosis of EoE when typical endoscopic findings are absent.77

In addition, AI demonstrated impressive utility in the real-time endoscopic diagnosis of infrequent forms of esophagitis (ie, CMV and HSV), which are often misdiagnosed even by experts.57

As regards Hp, a recent meta-analysis confirmed that AI algorithms are reliable predictors of infection with AUC of 0.92 (95% CI, 0.90-0.94), pooled sensitivity of 0.87 (95% CI, 0.72-0.94), and specificity of 0.86 (95% CI, 0.77-0.92).112

The world of AI sets the groundwork for the medicine of the future, and we should brace ourselves for the integration of CAD tools into clinical practice. However, reliance on AI tools should not replace clinical judgment. This is because AI has a black-box nature and is currently burdened with high costs. In addition, the high computational power of AI algorithms carries the risk of overfitting, in which the model is too tightly fitted to the training data and does not generalize towards new data.113 More real-time high-quality studies are needed to expand these early results that, if confirmed, will represent a revolution of routine clinical gastroenterological practice.

Footnotes

E.S. has received lecture or consultancy fees from Abbvie, Alfasigma, Amgen, Aurora Pharma, Bristol-Myers Squibb, EG Stada Group, Fresenius Kabi, Grifols, Janssen, Johnson&Johnson, Innovamedica, Malesci, Medtronic, Merck & Co, Novartis, Reckitt Benckiser, Sandoz, Shire, SILA, Sofar, Takeda, Unifarco. The remaining authors declare that they have nothing to disclose.

Contributor Information

Pierfrancesco Visaggi, Email: pierfrancesco.visaggi@gmail.com.

Nicola de Bortoli, Email: nicola.debortoli@unipi.it.

Brigida Barberio, Email: brigida.barberio@gmail.com.

Vincenzo Savarino, Email: vsavarin@unige.it.

Roberto Oleas, Email: robertoleas@gmail.com.

Emma M. Rosi, Email: emma.maria.rosi@gmail.com.

Santino Marchi, Email: santino.marchi@unipi.it.

Mentore Ribolsi, Email: m.ribolsi@unicampus.it.

Edoardo Savarino, Email: edoardo.savarino@unipd.it.

REFERENCES

  • 1.Le Berre C, Sandborn WJ, Aridhi S, et al. Application of artificial intelligence to gastroenterology and hepatology. Gastroenterology. 2020;158:76.e2–94.e2. [DOI] [PubMed] [Google Scholar]
  • 2.Ebigbo A, Palm C, Probst A, et al. A technical review of artificial intelligence as applied to gastrointestinal endoscopy: clarifying the terminology. Endosc Int Open. 2019;7:E1616–E1623. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Huang L-M, Yang W-J, Huang Z-Y, et al. Artificial intelligence technique in detection of early esophageal cancer. World J Gastroenterol. 2020;26:5959–5969. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Sana MK, Hussain ZM, Shah PA, et al. Artificial intelligence in celiac disease. Comput Biol Med. 2020;125:103996. [DOI] [PubMed] [Google Scholar]
  • 5.Mori Y, Kudo S-e, Mohmed HEN, et al. Artificial intelligence and upper gastrointestinal endoscopy: current status and future perspective. Dig Endosc. 2019;31:378–388. [DOI] [PubMed] [Google Scholar]
  • 6.Holm EA. In defense of the black box. Science. 2019;364:26. [DOI] [PubMed] [Google Scholar]
  • 7.Glover B, Teare J, Patel N. A systematic review of the role of non-magnified endoscopy for the assessment of H. pylori infection. Endosc Int Open. 2020;8:E105–E114. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Tan WK, Sharma AN, Chak A, et al. Progress in screening for Barrett’s esophagus: beyond standard upper endoscopy. Gastrointest Endosc Clin N Am. 2021;31:43–58. [DOI] [PubMed] [Google Scholar]
  • 9.Ferlay J, Colombet M, Soerjomataram I, et al. Estimating the global cancer incidence and mortality in 2018: GLOBOCAN sources and methods. Int J Cancer. 2019;144:1941–1953. [DOI] [PubMed] [Google Scholar]
  • 10.Yamashina T, Ishihara R, Nagai K, et al. Long-term outcome and metastatic risk after endoscopic resection of superficial esophageal squamous cell carcinoma. Am J Gastroenterol. 2013;108:544–551. [DOI] [PubMed] [Google Scholar]
  • 11.Sharma P, Savides TJ, Canto MI, et al. The American Society for Gastrointestinal Endoscopy PIVI (Preservation and Incorporation of Valuable Endoscopic Innovations) on imaging in Barrett’s esophagus. Gastrointest Endosc. 2012;76:252–254. [DOI] [PubMed] [Google Scholar]
  • 12.Thosani N, Abu Dayyeh BK, Sharma P, et al. ASGE Technology Committee systematic review and meta-analysis assessing the ASGE Preservation and Incorporation of Valuable Endoscopic Innovations thresholds for adopting real-time imaging-assisted endoscopic targeted biopsy during endoscopic surveillance of Barrett’s esophagus. Gastrointest Endosc. 2016;83:684.e7–698.e7. [DOI] [PubMed] [Google Scholar]
  • 13.Visaggi P, Barberio B, Ghisa M, et al. Modern diagnosis of early esophageal cancer: from blood biomarkers to advanced endoscopy and artificial intelligence. Cancers. 2021;13. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.de Groof J, van der Sommen F, van der Putten J, et al. The Argos Project: The development of a computer-aided detection system to improve detection of Barrett’s neoplasia on white light endoscopy. United European Gastroenterol J. 2019;7:538–547. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.de Groof AJ, Struyvenberg MR, Fockens KN, et al. Deep learning algorithm detection of Barrett’s neoplasia with high accuracy during live endoscopic procedures: a pilot study (with video). Gastrointest Endosc. 2020;91:1242–1250. [DOI] [PubMed] [Google Scholar]
  • 16.Ebigbo A, Mendel R, Probst A, et al. Real-time use of artificial intelligence in the evaluation of cancer in Barrett’s oesophagus. Gut. 2020;69:615–616. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.de Groof AJ, Struyvenberg MR, van der Putten J, et al. Deep-learning system detects neoplasia in patients with Barrett’s esophagus with higher accuracy than endoscopists in a multistep training and validation study with benchmarking. Gastroenterology. 2020;158:915.e4–929.e4. [DOI] [PubMed] [Google Scholar]
  • 18.Swager A-F, van der Sommen F, Klomp SR, et al. Computer-aided detection of early Barrett’s neoplasia using volumetric laser endomicroscopy. Gastrointest Endosc. 2017;86:839–846. [DOI] [PubMed] [Google Scholar]
  • 19.van der Sommen F, Zinger S, Curvers WL, et al. Computer-aided detection of early neoplastic lesions in Barrett’s esophagus. Endoscopy. 2016;48:617–624. [DOI] [PubMed] [Google Scholar]
  • 20.Ebigbo A, Mendel R, Rückert T, et al. Endoscopic prediction of submucosal invasion in Barrett’s cancer with the use of artificial intelligence: a pilot Study. Endoscopy. 2021;53:878–883. [DOI] [PubMed] [Google Scholar]
  • 21.Lui TKL, Tsui VWM, Leung WK. Accuracy of artificial intelligence-assisted detection of upper GI lesions: a systematic review and meta-analysis. Gastrointest Endosc. 2020;92:821.e9–830.e9. [DOI] [PubMed] [Google Scholar]
  • 22.Van Riel SV, Sommen FVD, Zinger S, et al. Automatic detection of early esophageal cancer with CNNS using transfer learning. In 2018 25th IEEE International Conference on Image Processing (ICIP); 2018.
  • 23.de Souza LA, Jr, Passos LA, Mendel R, et al. Assisting Barrett’s esophagus identification using endoscopic data augmentation based on Generative Adversarial Networks. Comput Biol Med. 2020;126:104029. [DOI] [PubMed] [Google Scholar]
  • 24.Liu G, Hua J, Wu Z, et al. Automatic classification of esophageal lesions in endoscopic images using a convolutional neural network. Ann Transl Med. 2020;8:486. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Iwagami H, Ishihara R, Aoyama K, et al. Artificial intelligence for the detection of esophageal and esophagogastric junctional adenocarcinoma. J Gastroenterol Hepatol. 2021;36:131–136. [DOI] [PubMed] [Google Scholar]
  • 26.Wu Z, Ge R, Wen M, et al. ELNet: Automatic classification and segmentation for esophageal lesions using convolutional neural network. Med Image Anal. 2021;67:101838. [DOI] [PubMed] [Google Scholar]
  • 27.Struyvenberg MR, de Groof AJ, van der Putten J, et al. A computer-assisted algorithm for narrow-band imaging-based tissue characterization in Barrett’s esophagus. Gastrointest Endosc. 2021;93:89–98. [DOI] [PubMed] [Google Scholar]
  • 28.Jamel S, Tukanova K, Markar S. Detection and management of oligometastatic disease in oesophageal cancer and identification of prognostic factors: a systematic review. World J Gastrointest Oncol. 2019;11:741–749. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Hashimoto CL, Iriya K, Baba ER, et al. Lugol’s dye spray chromoendoscopy establishes early diagnosis of esophageal cancer in patients with primary head and neck cancer. Am J Gastroenterol. 2005;100:275–282. [DOI] [PubMed] [Google Scholar]
  • 30.Tomie A, Dohi O, Yagi N, et al. Blue laser imaging-bright improves endoscopic recognition of superficial esophageal squamous cell carcinoma. Gastroenterol Res Pract. 2016;2016:6140854. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Ishihara R, Takeuchi Y, Chatani R, et al. Prospective evaluation of narrow-band imaging endoscopy for screening of esophageal squamous mucosal high-grade neoplasia in experienced and less experienced endoscopists. DisEsophagus. 2010;23:480–486. [DOI] [PubMed] [Google Scholar]
  • 32.Cai SL, Li B, Tan WM, et al. Using a deep learning system in endoscopy for screening of early esophageal squamous cell carcinoma (with video). Gastrointest Endosc. 2019;90:745.e2–753.e2. [DOI] [PubMed] [Google Scholar]
  • 33.Guo L, Xiao X, Wu C, et al. Real-time automated diagnosis of precancerous lesions and early esophageal squamous cell carcinoma using a deep learning model (with videos). Gastrointest Endosc. 2020;91:41–51. [DOI] [PubMed] [Google Scholar]
  • 34.Kumagai Y, Takubo K, Kawada K, et al. Diagnosis using deep-learning artificial intelligence based on the endocytoscopic observation of the esophagus. Esophagus. 2019;16:180–187. [DOI] [PubMed] [Google Scholar]
  • 35.Ohmori M, Ishihara R, Aoyama K, et al. Endoscopic detection and differentiation of esophageal lesions using a deep neural network. Gastrointest Endosc. 2020;91:301.e1–309.e1. [DOI] [PubMed] [Google Scholar]
  • 36.Tokai Y, Yoshio T, Aoyama K, et al. Application of artificial intelligence using convolutional neural networks in determining the invasion depth of esophageal squamous cell carcinoma. Esophagus. 2020;17:250–256. [DOI] [PubMed] [Google Scholar]
  • 37.Nakagawa K, Ishihara R, Aoyama K, et al. Classification for invasion depth of esophageal squamous cell carcinoma using a deep neural network compared with experienced endoscopists. Gastrointest Endosc. 2019;90:407–414. [DOI] [PubMed] [Google Scholar]
  • 38.Shimamoto Y, Ishihara R, Kato Y, et al. Real-time assessment of video images for esophageal squamous cell carcinoma invasion depth using artificial intelligence. J Gastroenterol. 2020;55:1037–1045. [DOI] [PubMed] [Google Scholar]
  • 39.Everson M, Herrera L, Li W, et al. Artificial intelligence for the real-time classification of intrapapillary capillary loop patterns in the endoscopic diagnosis of early oesophageal squamous cell carcinoma: a proof-of-concept study. United European Gastroenterol J. 2019;7:297–306. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Zhao YY, Xue DX, Wang YL, et al. Computer-assisted diagnosis of early esophageal squamous cell carcinoma using narrow-band imaging magnifying endoscopy. Endoscopy. 2019;51:333–341. [DOI] [PubMed] [Google Scholar]
  • 41.Wang YK, Syu HY, Chen YH, et al. Endoscopic images by a single-shot multibox detector for the identification of early cancerous lesions in the esophagus: a pilot study. Cancers (Basel). 2021;13:321. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Horie Y, Yoshio T, Aoyama K, et al. Diagnostic outcomes of esophageal cancer by artificial intelligence using convolutional neural networks. Gastrointest Endosc. 2019;89:25–32. [DOI] [PubMed] [Google Scholar]
  • 43.Inoue H, Kaga M, Ikeda H, et al. Magnification endoscopy in esophageal squamous cell carcinoma: a review of the intrapapillary capillary loop classification. Ann Gastroenterol. 2015;28:41–48. [PMC free article] [PubMed] [Google Scholar]
  • 44.Sato H, Inoue H, Ikeda H, et al. Utility of intrapapillary capillary loops seen on magnifying narrow-band imaging in estimating invasive depth of esophageal squamous cell carcinoma. Endoscopy. 2015;47:122–128. [DOI] [PubMed] [Google Scholar]
  • 45.García-Peraza-Herrera LC, Everson M, Lovat L, et al. Intrapapillary capillary loop classification in magnification endoscopy: open dataset and baseline methodology. Int J Comput Assist Radiol Surg. 2020;15:651–659. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Fukuda H, Ishihara R, Kato Y, et al. Comparison of performances of artificial intelligence versus expert endoscopists for real-time assisted diagnosis of esophageal squamous cell carcinoma (with video). Gastrointest Endosc. 2020;92:848–855. [DOI] [PubMed] [Google Scholar]
  • 47.Yang XX, Li Z, Shao XJ, et al. Real-time artificial intelligence for endoscopic diagnosis of early esophageal squamous cell cancer (with video). Dig Endosc. 2020. [Epub ahead of print]. [DOI] [PubMed] [Google Scholar]
  • 48.Waki K, Ishihara R, Kato Y, et al. Usefulness of an artificial intelligence system for the detection of esophageal squamous cell carcinoma evaluated with videos simulating overlooking situation. Dig Endosc. 2021. [Epub ahead of print]. [DOI] [PubMed] [Google Scholar]
  • 49.Li B, Cai SL, Tan WM, et al. Comparative study on artificial intelligence systems for detecting early esophageal squamous cell carcinoma between narrow-band and white-light imaging. World J Gastroenterol. 2021;27:281–293. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Pace F, Buscema M, Dominici P, et al. Artificial neural networks are able to recognize gastro-oesophageal reflux disease patients solely on the basis of clinical data. Eur J Gastroenterol Hepatol. 2005;17:605–610. [DOI] [PubMed] [Google Scholar]
  • 51.Horowitz N, Moshkowitz M, Halpern Z, et al. Applying data mining techniques in the development of a diagnostics questionnaire for GERD. Dig Dis Sci. 2007;52:1871–1878. [DOI] [PubMed] [Google Scholar]
  • 52.Pace F, Riegler G, de Leone A, et al. Is it possible to clinically differentiate erosive from nonerosive reflux disease patients? A study using an artificial neural networks-assisted algorithm. Eur J Gastroenterol Hepatol. 2010;22:1163–1168. [DOI] [PubMed] [Google Scholar]
  • 53.Rogers B, Samanta S, Ghobadi K, et al. Artificial intelligence automates and augments baseline impedance measurements from pH-impedance studies in gastroesophageal reflux disease. J Gastroenterol. 2021;56:34–41. [DOI] [PubMed] [Google Scholar]
  • 54.Gulati S, Bernth J, Liao J, et al. OTU-07 Near focus narrow and imaging driven artificial intelligence for the diagnosis of gastro-oesophageal reflux disease. Gut. 2019;68(suppl 2):A4. [Google Scholar]
  • 55.Sallis BF, Erkert L, Moñino-Romero S, et al. An algorithm for the classification of mRNA patterns in eosinophilic esophagitis: integration of machine learning. JAllergy Clin Immunol. 2018;141:1354.e9–1364.e9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Santos R, Haack HG, Maddalena D, et al. Evaluation of artificial neural networks in the classification of primary oesophageal dysmotility. Scand J Gastroenterol. 2006;41:257–263. [DOI] [PubMed] [Google Scholar]
  • 57.Lee JS, Yun J, Ham S, et al. Machine learning approach for differentiating cytomegalovirus esophagitis from herpes simplex virus esophagitis. Sci Rep. 2021;11:3672. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Vakil N, van Zanten SV, Kahrilas P, et al. The Montreal definition and classification of gastroesophageal reflux disease: a global evidence-based consensus. Am J Gastroenterol. 2006;101:1900–1920; quiz 1943. [DOI] [PubMed] [Google Scholar]
  • 59.Savarino V, Savarino E, Parodi A, et al. Functional heartburn and non-erosive reflux disease. Dig Dis. 2007;25:172–174. [DOI] [PubMed] [Google Scholar]
  • 60.Savarino E, Pohl D, Zentilin P, et al. Functional heartburn has more in common with functional dyspepsia than with non-erosive reflux disease. Gut. 2009;58:1185–1191. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Savarino V, Marabotto E, Zentilin P, et al. Esophageal reflux hypersensitivity: non-GERD or still GERD? Dig Liver Dis. 2020;52:1413–1420. [DOI] [PubMed] [Google Scholar]
  • 62.Savarino E, Frazzoni M, Marabotto E, et al. A SIGE-SINGEM-AIGO technical review on the clinical use of esophageal reflux monitoring. Dig Liver Dis. 2020;52:966–980. [DOI] [PubMed] [Google Scholar]
  • 63.Locke GR, Talley NJ, Weaver AL, et al. A new questionnaire for gastroesophageal reflux disease. Mayo Clin Proc. 1994;69:539–547. [DOI] [PubMed] [Google Scholar]
  • 64.Frazzoni M, Frazzoni L, Tolone S, et al. Lack of improvement of impaired chemical clearance characterizes PPI-refractory reflux-related heartburn. Am J Gastroenterol. 2018;113:670–676. [DOI] [PubMed] [Google Scholar]
  • 65.Gyawali CP, Kahrilas PJ, Savarino E, et al. Modern diagnosis of GERD: the Lyon Consensus. Gut. 2018;67:1351–1362. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Frazzoni L, Frazzoni M, de Bortoli N, et al. Postreflux swallow-induced peristaltic wave index and nocturnal baseline impedance can link PPI-responsive heartburn to reflux better than acid exposure time. Neurogastroenterol Motil. 2017. doi:10.1111/nmo.13116. [DOI] [PubMed] [Google Scholar]
  • 67.Roman S, Gyawali CP, Savarino E, et al. Ambulatory reflux monitoring for diagnosis of gastro-esophageal reflux disease: Update of the Porto consensus and recommendations from an international consensus group. Neurogastroenterol Motil. 2017;29:1–15. [DOI] [PubMed] [Google Scholar]
  • 68.Savarino E, Bredenoord AJ, Fox M, et al. Expert consensus document: advances in the physiological assessment and diagnosis of GERD. Nat Rev Gastroenterol Hepatol. 2017;14:665–676. [DOI] [PubMed] [Google Scholar]
  • 69.Frazzoni M, Savarino E, de Bortoli N, et al. Analyses of the post-reflux swallow-induced peristaltic wave index and nocturnal baseline impedance parameters increase the diagnostic yield of impedance-pH monitoring of patients with reflux disease. Clin Gastroenterol Hepatol. 2016;14:40–46. [DOI] [PubMed] [Google Scholar]
  • 70.Patel A, Wang D, Sainani N, et al. Distal mean nocturnal baseline impedance on pH-impedance monitoring predicts reflux burden and symptomatic outcome in gastro-oesophageal reflux disease. Aliment Pharmacol Ther. 2016;44:890–898. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71.Frazzoni M, Penagini R, Frazzoni L, et al. Role of reflux in the pathogenesis of eosinophilic esophagitis: comprehensive appraisal with off- and on PPI impedance-pH monitoring. Am J Gastroenterol. 2019;114:1606–1613. [DOI] [PubMed] [Google Scholar]
  • 72.Rengarajan A, Savarino E, Della Coletta M, et al. Mean nocturnal baseline impedance correlates with symptom outcome when acid exposure time is inconclusive on esophageal reflux monitoring. Clin Gastroenterol Hepatol. 2020;18:589–595. [DOI] [PubMed] [Google Scholar]
  • 73.Frazzoni M, de Bortoli N, Frazzoni L, et al. The added diagnostic value of postreflux swallow-induced peristaltic wave index and nocturnal baseline impedance in refractory reflux disease studied with on-therapy impedance-pH monitoring. Neurogastroenterol Motil. 2017. doi:10.1111/nmo.12947. [DOI] [PubMed] [Google Scholar]
  • 74.Frazzoni M, de Bortoli N, Frazzoni L, et al. Impedance-pH monitoring for diagnosis of reflux disease: new perspectives. Dig Dis Sci. 2017;62:1881–1889. [DOI] [PubMed] [Google Scholar]
  • 75.Liacouras CA, Furuta GT, Hirano I, et al. Eosinophilic esophagitis: updated consensus recommendations for children and adults. J Allergy Clin Immunol. 2011;128:3.e6–20.e6; quiz 21–22. [DOI] [PubMed] [Google Scholar]
  • 76.Sciumé GD, Visaggi P, Sostilio A, et al. Eosinophilic esophagitis: novel concepts regarding pathogenesis and clinical manifestations. Minerva Gastroenterol Dietol. 2021. [Epub ahead of print]. [DOI] [PubMed] [Google Scholar]
  • 77.Visaggi P, Savarino E, Sciume G, et al. Eosinophilic esophagitis: clinical, endoscopic, histologic and therapeutic differences and similarities between children and adults. Therap Adv Gastroenterol. 2021;14:1756284820980860. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.Visaggi P, Mariani L, Pardi V, et al. Dietary management of eosinophilic esophagitis: tailoring the approach. Nutrients. 2021;13:1630. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Savarino EV, Tolone S, Bartolo O, et al. The GerdQ questionnaire and high resolution manometry support the hypothesis that proton pump inhibitor-responsive oesophageal eosinophilia is a GERD-related phenomenon. Aliment Pharmacol Ther. 2016;44:522–530. [DOI] [PubMed] [Google Scholar]
  • 80.Guo L, Gong H, Wang Q, et al. Detection of multiple lesions of gastrointestinal tract for endoscopy using artificial intelligence model: a pilot study. Surg Endosc. 2020. [Epub ahead of print]. [DOI] [PubMed] [Google Scholar]
  • 81.Jin P, Ji X, Kang W, et al. Artificial intelligence in gastric cancer: a systematic review. J Cancer Res Clin Oncol. 2020;146:2339–2350. [DOI] [PubMed] [Google Scholar]
  • 82.Hirasawa T, Aoyama K, Tanimoto T, et al. Application of artificial intelligence using a convolutional neural network for detecting gastric cancer in endoscopic images. Gastric Cancer. 2018;21:653–660. [DOI] [PubMed] [Google Scholar]
  • 83.Guimarães P, Keller A, Fehlmann T, et al. Deep-learning based detection of gastric precancerous conditions. Gut. 2020;69:4. [DOI] [PubMed] [Google Scholar]
  • 84.Li L, Chen Y, Shen Z, et al. Convolutional neural network for the diagnosis of early gastric cancer based on magnifying narrow band imaging. Gastric Cancer. 2020;23:126–132. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Wu L, Zhou W, Wan X, et al. A deep neural network improves endoscopic detection of early gastric cancer without blind spots. Endoscopy. 2019;51:522–531. [DOI] [PubMed] [Google Scholar]
  • 86.Ueyama H, Kato Y, Akazawa Y, et al. Application of artificial intelligence using a convolutional neural network for diagnosis of early gastric cancer based on magnifying endoscopy with narrow-band imaging. J Gastroenterol Hepatol. 2021;36:482–489. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Zhang Y, Li F, Yuan F, et al. Diagnosing chronic atrophic gastritis by gastroscopy using artificial intelligence. Dig Liver Dis. 2020;52:566–572. [DOI] [PubMed] [Google Scholar]
  • 88.Lee JH, Kim YJ, Kim YW, et al. Spotting malignancies from gastric endoscopic images using deep learning. Surg Endosc. 2019;33:3790–3797. [DOI] [PubMed] [Google Scholar]
  • 89.Horiuchi Y, Aoyama K, Tokai Y, et al. Convolutional neural network for differentiating gastric cancer from gastritis using magnified endoscopy with narrow band imaging. Dig Dis Sci. 2020;65:1355–1363. [DOI] [PubMed] [Google Scholar]
  • 90.Zhu Y, Wang QC, Xu MD, et al. Application of convolutional neural network in the diagnosis of the invasion depth of gastric cancer based on conventional endoscopy. Gastrointest Endosc. 2019;89:806.e1–815.e1. [DOI] [PubMed] [Google Scholar]
  • 91.Nagao S, Tsuji Y, Sakaguchi Y, et al. Highly accurate artificial intelligence systems to predict the invasion depth of gastric cancer: efficacy of conventional white-light imaging, nonmagnifying narrow-band imaging, and indigo-carmine dye contrast imaging. Gastrointest Endosc. 2020;92:866.e1–873.e1. [DOI] [PubMed] [Google Scholar]
  • 92.Hu H, Gong L, Dong D, et al. Identifying early gastric cancer under magnifying narrow-band images with deep learning: a multicenter study. Gastrointest Endosc. 2021;93:1333.e3–1341.e3. [DOI] [PubMed] [Google Scholar]
  • 93.Wroblewski LE, Peek RM, Jr, Wilson KT. Helicobacter pylori and gastric cancer: factors that modulate disease risk. Clin Microbiol Rev. 2010;23:713–739. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 94.Chey WD, Leontiadis GI, Howden CW, et al. ACG Clinical Guideline: treatment of Helicobacter pylori infection. Am J Gastroenterol. 2017;112:212–239. [DOI] [PubMed] [Google Scholar]
  • 95.Kakinoki R, Kushima R, Matsubara A, et al. Re-evaluation of histogenesis of gastric carcinomas: a comparative histopathological study between Helicobacter pylori-negative and H. pylori-positive cases. Dig Dis Sci. 2008;54:614. [DOI] [PubMed] [Google Scholar]
  • 96.Malfertheiner P, Megraud F, O'Morain C, et al. Current concepts in the management of Helicobacter pylori infection: the Maastricht III Consensus Report. Gut. 2007;56:772–781. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 97.Hooi JKY, Lai WY, Ng WK, et al. Global prevalence of Helicobacter pyloriinfection: systematic review and meta-analysis. Gastroenterology. 2017;153:420–429. [DOI] [PubMed] [Google Scholar]
  • 98.Shichijo S, Nomura S, Aoyama K, et al. Application of convolutional neural networks in the diagnosis of Helicobacter pyloriinfection based on endoscopic images. EBioMedicine. 2017;25:106–111. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 99.Yasuda T, Hiroyasu T, Hiwa S, et al. Potential of automatic diagnosis system with linked color imaging for diagnosis of Helicobacter pylori infection. Dig Endosc. 2020;32:373–381. [DOI] [PubMed] [Google Scholar]
  • 100.Zheng W, Zhang X, Kim JJ, et al. High accuracy of convolutional neural network for evaluation of Helicobacter pylori infection based on endoscopic images: preliminary experience. Clin Transl Gastroenterol. 2019;10:e00109. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 101.Shichijo S, Endo Y, Aoyama K, et al. Application of convolutional neural networks for evaluating Helicobacter pylori infection status on the basis of endoscopic images. Scand J Gastroenterol. 2019;54:158–163. [DOI] [PubMed] [Google Scholar]
  • 102.Nakashima H, Kawahira H, Kawachi H, et al. Artificial intelligence diagnosis of Helicobacter pylori infection using blue laser imaging-bright and linked color imaging: a single-center prospective study. Ann Gastroenterol. 2018;31:462–468. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 103.Itoh T, Kawahira H, Nakashima H, et al. Deep learning analyzes Helicobacter pylori infection by upper gastrointestinal endoscopy images. Endosc Int Open. 2018;6:E139–E144. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 104.Huang CR, Sheu BS, Chung PC, et al. Computerized diagnosis of Helicobacter pylori infection and associated gastric inflammation from endoscopic images by refined feature selection using a neural network. Endoscopy. 2004;36:601–608. [DOI] [PubMed] [Google Scholar]
  • 105.Huang CR, Chung PC, Sheu BS, et al. Helicobacter pylori-related gastric histology classification using support-vector-machine-based feature selection. IEEE Trans Inf Technol Biomed. 2008;12:523–531. [DOI] [PubMed] [Google Scholar]
  • 106.Dixon MF, Genta RM, Yardley JH, et al. Classification and grading of gastritis. The updated Sydney System. International Workshop on the Histopathology of Gastritis, Houston 1994. Am J Surg Pathol. 1996;20:1161–1181. [DOI] [PubMed] [Google Scholar]
  • 107.Dohi O, Yagi N, Onozawa Y, et al. Linked color imaging improves endoscopic diagnosis of active Helicobacter pylori infection. Endosc Int Open. 2016;4:E800–E805. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 108.Wu L, Zhang J, Zhou W, et al. Randomised controlled trial of WISENSE, a real-time quality improving system for monitoring blind spots during esophagogastroduodenoscopy. Gut. 2019;68:2161–2169. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109.Arribas J, Antonelli G, Frazzoni L, et al. Standalone performance of artificial intelligence for upper GI neoplasia: a meta-analysis. Gut. 2020. [Epub ahead of print]. [DOI] [PubMed] [Google Scholar]
  • 110.Kuwano H, Nishimura Y, Oyama T, et al. Guidelines for diagnosis and treatment of carcinoma of the esophagus April 2012 edited by the Japan Esophageal Society. Esophagus. 2015;12:1–30. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 111.Ichiya T, Tanaka K, Rubio CA, et al. Evaluation of narrow-band imaging signs in eosinophilic and lymphocytic esophagitis. Endoscopy. 2017;49:429–437. [DOI] [PubMed] [Google Scholar]
  • 112.Bang CS, Lee JJ, Baik GH. Artificial intelligence for the prediction of Helicobacter pylori infection in endoscopic images: systematic review and meta-analysis of diagnostic test accuracy. J Med Internet Res. 2020;22:e21983. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 113.van der Sommen F, de Groof J, Struyvenberg M, et al. Machine learning in GI endoscopy: practical guidance in how to interpret a novel field. Gut. 2020;69:2035–2045. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Journal of Clinical Gastroenterology are provided here courtesy of Wolters Kluwer Health

RESOURCES