Skip to main content
. 2024 Jun 5;76(5):1655–1667. doi: 10.1007/s13304-024-01892-6

Table 3.

A summary of the articles describing the role of AI applied to LC

Authors Year of publication Dataset and characteristics of the methods Algorithm and AI
Mascagni et al. [37] 2020 78 endoscopic videos of consecutive LC procedures performed at the Digestive Surgery Department of the Nouvel Hospital Civil The inter-observer agreement was calculated with Cohen’s kappa; doublet view method and the binary scheme
Mascagni et al. [38] 2022 2854 images from 201 LC videos were annotated and 402 images were segmented The inter-observer agreement was calculated with Cohen’s kappa
Madani et al. [39] 2022

308 anonymized LC videos from 37 countries (including all continents), 153 surgeons and 136 different institutions

AI predictions were evaluated using tenfold cross-validation against annotations by expert surgeons. A tenfold cross-validation technique evaluated the performance of GoNoGoNet and CholeNet

ffmpeg 4.1 software extracted frames from videos;

Deep Convolutional Neural Network (CNN; ResNet50);

Pyramid Scene Parsing Network (PSPNet) was used for pixel-wise semantic segmentation

GoNoGoNet AI and CholeNet

Laplante et al. [40] 2023

308 anonymized videos from 37 countries, 153 surgeons and 136 different institutions

High-volume expert surgeons delineated the boundaries of the Go and No-Go zones (semantic segmentation)

ffmpeg 4.1 software extracted frames from videos;

Think Like A Surgeon software was used by expert surgeons to delineate the boundaries of the Go and No-Go zones

Visual Concordance Test (VCT) calculated the expert consensus of annotated pixels for Go and No-Go zones

GoNoGoNet AI

Khalid et al. [6] 2023 The database was composed of 308 anonymized LC videos originating from 37 countries across 153 surgeons and 136 different institutions. Two groups: BDI group and Control group. 11 LC videos of BDI group was annotated by GoNoGoNet and compared to another 11 LC videos with cholecystitis of control group

Convolutional Deep Neural Network (DNN);

U-Net architecture (the optimized version of the GoNoGoNet algorithm);

Endo et al. [41] 2023

230 videos of LC conducted from 2019 to 2021 in a single center; 95 cases that remained on video with mild inflammation to be noted; cases with severe inflammation and abnormal biliary anatomy were excluded

1754 images from Calot triangle dissection scenes, 1610 images of LM-EHBD, 1503 images of LM-EHBD, 1623 images of LM-S4, and 1505 images of LM-RS

Test datasets: 19 cases, comprising 190 images of LM-EHBD, 186 images of LM-EHBD, 192 images of LM-S4, and 190 images of LM-RS

Questionnaire survey was given to beginners and experts surgeons

YOLOv3
Nakanuma et al. [42] 2023

10 cases of LC performed in a single center to evaluate the feasibility of LC using an intraoperative AI-assisted landmark recognition

The landmark detection system was connected to the endoscopic integrated operating room (EndoALPHA; Olympus Corp.)

The method for quantitatively evaluating the accuracy of landmark detection by AI was the DICE coefficient. The DICE coefficient is the reference value for detecting the accuracy of AI and shows the degree of agreement between expert surgeons for the annotations of the anatomic regions of interest

The Kruskal–Wallis test was used to evaluate the DICE coefficient

YOLOv3 connected with the endoscopic integrated operating room (EndoALPHA; Olympus Corp.)
Fujinaga et al. [43] 2023

20 cases of LC performed in a single center

External evaluation committee (EEC) evaluated the adequacy of the detection times of the landmarks detected using AI

4-point rubric questionnaire was used to evaluate the accuracy of landmark detection and the contribution of cross-AI in preventing BDI

Cross-AI system for landmark detection and surgical phase recognition: an endoscope (OLYMPUS LTF-S190-10; Olympus Corp., Tokyo, Japan), a video processor (VISERA ELITE II; Olympus Corp.), and a desktop computer that had two graphics processing units (Quadro RTX 6000 and Quadro RTX 4000; NVIDIA Corp., Santa Clara, CA, USA)
Kawamura et al. [36] 2023 72 LC videos, and 23,793 images were used for training data, performed in a single center

EfficientNet-B5: a CNN to classify images and perform prediction;

Sharpness-Aware Minimization: a SAM an optimizer that perfects the learning parameters of EfficientNet-B5 to smooth the information derived from the diversified images for a label

Golany et al.[44] 2022

371 LC videos from 4 hospitals and data set Cholec80: 294 videos were used to train the AI model, 77 videos were used to test the AI model. Each video was divided into 10 steps by 2 expert surgeons

Experienced surgeons noted the phases, adverse events and CSV of LCs

The inter-rater agreement score was used to confirm the quality of the annotations

MS-TCN—Multi-Stage Temporal Convolution Network. is a set of temporal convolution layers that capture temporal connections. The final layer of the MS-TCN gives the prediction of the surgical stage for each frame of the LC video

Resnet50 independently classifies and predicts the surgical phase. For each frame (input), Resnet50 produces a numerical vector of visual features. The set of input vectors are combined to form a sequence of feature vectors representing the entire LC video and fed into the MS-TCN model

Cheng et al. [48] 2022

Dataset: 163 LC videos collected from four medical centers. 90 videos were labeled by expert surgeons. 63 LC videos were used to test the AI model

The mean concordance correlation coefficient was used to compare the reliability between the two surgeons

Super Video Converter software (version 1.0);

The Anvil Video Annotation Research Tool software was used to annotate the videos;

FFmpeg software was used to extract frames from videos;

Convolutional neural network (CNN) captured the spatial characteristics;

Long Short-Term Memory (LSTM) analyzed the temporal information

Anaconda (Anaconda, Inc, Austin, TX) in Python 3.6.5. NVIDIA (NVIDIA, Santa Clara, CA) Tesla V100 graphics processing unit (GPU) trained the visual and temporal models