Skip to main content
JAMA Network logoLink to JAMA Network
. 2022 Aug 19;5(8):e2226265. doi: 10.1001/jamanetworkopen.2022.26265

Development and Validation of a Model for Laparoscopic Colorectal Surgical Instrument Recognition Using Convolutional Neural Network–Based Instance Segmentation and Videos of Laparoscopic Procedures

Daichi Kitaguchi 1,2, Younae Lee 1, Kazuyuki Hayashi 1, Kei Nakajima 1,2, Shigehiro Kojima 1,2, Hiro Hasegawa 1,2, Nobuyoshi Takeshita 1,2, Kensaku Mori 3, Masaaki Ito 1,2,
PMCID: PMC9391983  PMID: 35984660

Key Points

Question

Is an automatic surgical instrument recognition model with wide applicability to multiple types of instruments and pixel-level high recognition accuracy feasible?

Findings

This quality improvement study for development and validation used a multi-institutional data set that consisted of 337 laparoscopic colorectal surgical videos. The mean average precisions of the instance segmentation of surgical instruments were 90.9% for 3 instruments, 90.3% for 4 instruments, 91.6% for 6 instruments, and 91.8% for 8 instruments.

Meaning

This study suggests that deep learning can be used to simultaneously recognize multiple types of instruments with high accuracy.


This quality improvement study evaluates the recognition performance of a deep learning model that can simultaneously recognize 8 types of surgical instruments frequently used in laparoscopic colorectal operations.

Abstract

Importance

Deep learning–based automatic surgical instrument recognition is an indispensable technology for surgical research and development. However, pixel-level recognition with high accuracy is required to make it suitable for surgical automation.

Objective

To develop a deep learning model that can simultaneously recognize 8 types of surgical instruments frequently used in laparoscopic colorectal operations and evaluate its recognition performance.

Design, Setting, and Participants

This quality improvement study was conducted at a single institution with a multi-institutional data set. Laparoscopic colorectal surgical videos recorded between April 1, 2009, and December 31, 2021, were included in the video data set. Deep learning–based instance segmentation, an image recognition approach that recognizes each object individually and pixel by pixel instead of roughly enclosing with a bounding box, was performed for 8 types of surgical instruments.

Main Outcomes and Measures

Average precision, calculated from the area under the precision-recall curve, was used as an evaluation metric. The average precision represents the number of instances of true-positive, false-positive, and false-negative results, and the mean average precision value for 8 types of surgical instruments was calculated. Five-fold cross-validation was used as the validation method. The annotation data set was split into 5 segments, of which 4 were used for training and the remainder for validation. The data set was split at the per-case level instead of the per-frame level; thus, the images extracted from an intraoperative video in the training set never appeared in the validation set. Validation was performed for all 5 validation sets, and the average mean average precision was calculated.

Results

In total, 337 laparoscopic colorectal surgical videos were used. Pixel-by-pixel annotation was manually performed for 81 760 labels on 38 628 static images, constituting the annotation data set. The mean average precisions of the instance segmentation for surgical instruments were 90.9% for 3 instruments, 90.3% for 4 instruments, 91.6% for 6 instruments, and 91.8% for 8 instruments.

Conclusions and Relevance

A deep learning–based instance segmentation model that simultaneously recognizes 8 types of surgical instruments with high accuracy was successfully developed. The accuracy was maintained even when the number of types of surgical instruments increased. This model can be applied to surgical innovations, such as intraoperative navigation and surgical automation.

Introduction

Minimally invasive surgery (MIS), such as laparoscopic and robotic types, continues to play an important role in general surgery as an alternative to traditional open surgery. Since the 1980s, technological advancement and innovation have led to a rapid increase in the number of MIS surgical techniques, because they are viewed as more desirable. Minimally invasive surgery has better cosmetic results than traditional open surgery and has become increasingly common during the past years.1 On the other hand, the reliance on a surgeon’s sense of touch when performing surgical procedures has waned, and surgeons have become highly dependent on the visual information provided by laparoscopic monitors. To be more dependable, laparoscopic cameras must be able to rapidly capture and track the surgical instrument being used so that unsafe blind surgical procedures can be avoided.2 Therefore, a prerequisite in MIS is the ability to accurately recognize the location of surgical instruments and track them; in other words, we can extract much of the information associated with the surgery from such kinematic data.

In recent years, computer-based automation, which self-captures and self-analyzes large amounts and a wide variety of data—exemplified by self-driving cars—has been advancing and rapidly pervading all aspects of daily life. Artificial intelligence, especially its subdomain of deep learning, is used to make big data interpretable and usable.3 One area of artificial intelligence that has particularly bourgeoned during the last decade is computer vision (CV), which is an interdisciplinary scientific field that deals with how computers gain a high-level understanding of digital images or videos and use the knowledge to perform functions, such as object identification, tracking, and scene recognition.4

Deep learning–based CV technology has been applied to various tasks in laparoscopic surgery, such as anatomical structures, surgical action, and step or phase recognition.4,5,6 Applying this technology to surgical instrument recognition can enable the development of robotic camera holders with automated real-time instrument tracking during surgery,7,8 and it can be used for surgical skill assessment,9 augmented reality,10 and depth enhancement11 during MIS. Although the goal of computer-assisted surgery is full automation, it must be realized in stages (eg, starting with intestinal anastomosis).12 In any stage, automatic surgical instrument recognition using deep learning–based CV is an indispensable fundamental technology. Therefore, an automatic surgical instrument recognition model with wide applicability to multiple types of instruments and pixel-level high recognition accuracy is required for research and development in the field of computer-assisted surgery.

The objectives of this study were to develop a deep learning model that can simultaneously recognize 8 types of surgical instruments frequently used in laparoscopic colorectal operations and to evaluate its recognition performance. In addition, the applicability of the recognition results of the developed model to surgical progress monitoring was examined.

Methods

Study Design

This quality improvement study for development and validation was conducted at a single institution using a multi-institutional intraoperative video data set. This study followed the reporting guidelines of the Standards for Quality Improvement Reporting Excellence (SQUIRE) and Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) reporting guidelines. The study protocol was reviewed and approved by the Ethics Committee of the National Cancer Center Hospital East, Chiba, Japan. Informed consent was obtained from patients in the form of an opt-out on the study website, and data from those who rejected participation were excluded. The study conformed to the provisions of the Declaration of Helsinki13 in 1964 (and revised in Brazil in 2013).

Annotation Data Set

Laparoscopic colorectal surgical videos recorded between April 1, 2009, and December 31, 2021, were included in the video data set in this study. Every intraoperative video was converted to the MP4 video format with a display resolution of 1280 × 720 pixels and frame rate of 30 frames per second. A total of 38 628 static images capturing the surgical instruments to be recognized were randomly extracted from the video data set and included in the annotation data set for this study.

The following 8 types of surgical instruments were used for recognition: surgical shears (Harmonic HD 1000i and Harmonic ACE, Ethicon Inc) (SI-1), spatula-type electrode (Olympus Co Ltd) (SI-2), atraumatic universal forceps with grooves (Aesculap AdTec, B Braun AG) (SI-3), dissection forceps (HiQ+ Maryland Dissection Forceps, Olympus Co Ltd) (SI-4), endoscopic clip appliers (Ligamax 5 Endoscopic Clip Applier, Ethicon Inc, and Endo Clip III 5-mm Clip Applier, Medtronic Plc) (SI-5), staplers (Echelon Stapler, Ethicon Inc, and Endo GIA Reload with Tri-Staple Technology, Medtronic Plc) (SI-6), grasping forceps (Croce-Olmi grasping forceps, Karl Storz SE & Co KG) (SI-7), and a suction/irrigation system (HiQ+ suction/irrigation system, Olympus Co Ltd) (SI-8). Reference images of these 8 types of surgical instruments are shown in Figure 1A.

Figure 1. Reference Images and Annotations of the 8 Types of Surgical Instruments.

Figure 1.

Panel A shows reference images of the 8 types of surgical instruments. Panel B shows reference images of the surgical instrument annotation.

Annotation was performed by 29 individuals who were not physicians under the supervision of 2 board-certified surgeons (D.K. and H.H.), and 8 types of annotation labels were manually assigned pixel by pixel by drawing directly on the area of each surgical instrument on the static images using digital drawing pens (Wacom Cintiq Pro and Wacom Pro Pen 2, Wacom Co Ltd, or Microsoft Surface Pro 7 and Microsoft Surface Pen, Microsoft Corp) equipment. Reference images of the annotations are shown in Figure 1B.

CNN-Based Instance Segmentation

Convolutional neural network (CNN)–based instance segmentation was performed with Mask R-CNN14 as the architecture network and ResNet-5015 as the backbone network. Instance segmentation is an image recognition approach that recognizes each object individually and pixel by pixel, instead of enclosing all objects within a bounding box. The network weight was initialized to a pretrained one on the Microsoft Common Objects in Context data set,16 which is a large-scale data set for object detection, segmentation, and captioning that contains 2 500 000 labeled instances in 328 000 images for 91 common object categories. Every annotated image in the training set was input into the deep learning model, and fine-tuning was then performed for the training set. The best epoch model, based on the model performance on the validation set, was selected. For data augmentation, the resize, randomhorizontalflip, randomverticalflip, and normalize functions were used.

Code and Computer Specifications

The code was written in Python, version 3.6 (Python Software Foundation) and is available via GitHub on reasonable request. The implementation of the model was based on MMDetection,17 which is an open-source Python library for object detection and instance segmentation. A computer equipped with a Tesla T4 GPU with 16 GB of VRAM (NVIDIA) and a Xeon CPU Gold 6230 (2.10 GHz, 20 Core) with 48 GB of RAM (Intel Corp) was used for network training.

Validation and Evaluation

Average precision (AP) was used as an evaluation metric for the CNN-based surgical instrument recognition task in this study. Average precision is calculated from the area under the precision-recall curve, which is described based on the number of true-positive (TP), false-positive (FP), and false-negative (FN) results. Intersection over union (IoU) is a metric that signifies the degree of overlap between the area annotated as the ground truth (A) and the predicted area output by the CNN (B); an IoU of 1 signifies that there is perfect overlap for the corresponding pair, A and B, whereas an IoU of 0 signifies that there is no overlap. An IoU reference image is shown in Figure 2A.

Figure 2. Illustration of Intersection Over Union and 5-Fold Cross-validation.

Figure 2.

A, IoU is a metric that signifies the degree of overlap between the area annotated as the ground truth (blue) and the predicted area output by the convolutional neural network (yellow). IoU is calculated by dividing the area of overlap (green) by the area of union. B, In 5-fold cross-validation, the annotation data set is split into 5 segments, of which 4 are used as training sets and the remainder as the validation set. The validation is performed for all 5 validation sets, and the mean value is calculated as the performance metric.

In this study, the threshold value of IoU was set to 0.75; that is, when the IoU of the corresponding pair of A and B was more than 0.75, it was defined as TP; when the IoU was less than 0.75, it was defined as FN; and when there was no A corresponding to B, it was defined as FP. The formulas for precision, recall, and IoU are as follows:

graphic file with name jamanetwopen-e2226265-iea.jpg

The mean AP (mAP) metric is widely used in object detection and instance segmentation tasks.18,19,20 In this study, mAP denotes the mean AP value for 8 types of surgical instruments.

Five-fold cross-validation and hold-out validation were used as the validation methods in this study as appropriate. In 5-fold cross-validation, the annotation data set was split into 5 segments, of which 4 were used as training sets and the remainder as the validation set. The data split was performed at the per-case level instead of the per-frame level; thus, the images extracted from an intraoperative video in the training set never appeared in the validation set. The validation was performed for all 5 validation sets, and the average mAP was calculated. A reference image of the 5-fold cross-validation is shown in Figure 2B. In addition, a misrecognition pattern analysis was performed.

Automatic Surgical Progress Monitoring

In colorectal surgery, transection of the inferior mesenteric artery (IMA) and rectum is an important landmark for monitoring surgical progress. In the standard IMA root transection procedure, the dorsal side of the IMA root is dissected with SI-4 and clipped with SI-5; SI-6 is only used for rectal transections.

To examine the applicability of the proposed model to an automatic surgical progress monitoring system, we used the results of the surgical instrument recognition on a new intraoperative video of laparoscopic colorectal surgery. Specifically, we evaluated whether continuous recognition of SI-4 followed by continuous recognition of SI-5 and the timing of continuous recognition of SI-6 were included in the previously defined surgical step of IMA and rectum transections,21 respectively.

Results

Annotation Data Set

In total, 337 laparoscopic colorectal surgical videos were used in this study, and 81 760 labels were annotated for 38 628 static images extracted from the video data set. The distribution of the number of annotation labels for each surgical instrument is shown in Figure 3A: 5694 for SI-1, 4358 for SI-2, 41 778 for SI-3, 5248 for SI-4, 5254 for SI-5, 5810 for SI-6, 7022 for SI-7, and 6596 for SI-8. SI-3 had numerous annotation labels because up to 4 could appear in a single static image, and they frequently appeared with other surgical instruments.

Figure 3. Distribution of Surgical Instrument (SI) Annotation Labels, Mean Average Precision (mAP) Changes, and Instance Segmentation mAP .

Figure 3.

Instance Segmentation Accuracy

The average (SD) mAP of instance segmentation for the 8 types of surgical instruments obtained from 5-fold cross-validation was 91.9% (0.34%). In this study, the threshold value of IoU was set to 0.75, but as supplemental data, the mAP values at multiple thresholds ranging from 0.5 to 0.95 are plotted and shown in the eFigure in the Supplement.

The change in the mAP when the number of types of surgical instruments to be recognized was increased is shown in Figure 3B. The mAPs of instance segmentation were 90.9% for 3 surgical instruments, 90.3% for 4 surgical instruments, 91.6% for 6 surgical instruments, and 91.8% for 8 surgical instruments.

The AP of instance segmentation for each surgical instrument is shown in Figure 3C: 91.4% for SI-1, 95.6% for SI-2, 85.4% for SI-3, 90.7% for SI-4, 96.4% for SI-5, 91.2% for SI-6, 85.6% for SI-7, and 95.7% for SI-8. The recognition results of instance segmentation for the 8 types of surgical instruments in a complete laparoscopic sigmoid colon resection procedure are shown in Figure 4 and the Video.

Figure 4. Recognition Results of Instance Segmentation for the 8 Types of Surgical Instruments in a Complete Laparoscopic Sigmoid Colon Resection Procedure.

Figure 4.

The vertical axis indicates each surgical instrument, the horizontal axis is the operating time, and the timing of the appearance of each surgical instrument is plotted with black lines to show the time series. The first and second outlined areas represent the surgical steps of inferior mesenteric artery (IMA) transection and rectal transection, respectively. The third and fourth outlined areas indicate extracorporeal procedures and represent the surgical steps of specimen extraction and wound closure, respectively.

Video. Video of Laparoscopic Colorectal Surgical Instrument Recognition Using Convolutional Neural Network–Based Instance Segmentation.

Download video file (329.1MB, mp4)

Recognition results of instance segmentation for 8 types of surgical instruments in laparoscopic sigmoid colon resection.

Misrecognition Pattern Analysis

The misrecognition pattern obtained when instance segmentation was performed on 1 validation set is shown in the confusion matrix of Figure 5A; 213 misrecognitions were observed in the validation set. Intracorporeal artifacts other than surgical instruments (eg, anvil shaft of a circular stapler, drainage tube, and trocar) tended to be misrecognized as SI-3, which was the main cause of misrecognition (125 of 213 misrecognitions [58.7%]). Representative images of the misrecognition patterns are shown in Figure 5B.

Figure 5. Confusion Matrix and Representative Image of Misrecognition Patterns.

Figure 5.

A, Confusion matrix of misrecognition pattern. Besides a total of 11 458 correct surgical instrument recognitions (eg, SI-1 was correctly recognized as SI-1 666 times), a total of 213 misrecognitions were observed in the validation set (eg, SI-4 was misrecognized as SI-1 14 times). B, Representative images of misrecognition patterns. The blue circle indicates the anvil shaft of the circular stapler in the left image and the drainage tube in the right image.

Automatic Surgical Progress Monitoring

As shown in Figure 4, the continuous recognition of SI-4 followed by the continuous recognition of SI-5 was included in the surgical step of the IMA transection. In addition, the timing of the continuous recognition of SI-6 was included in the surgical step of rectal transection.

Discussion

In this quality improvement study, we successfully developed a CNN-based instance segmentation model that can simultaneously recognize 8 types of surgical instruments with high recognition accuracy. Almost all surgical instruments used in laparoscopic colorectal operations are covered by these 8 types; therefore, the developed model can be applied to intraoperative navigation for a series of laparoscopic colorectal surgical procedures.

There are several approaches to CNN-based CV for image recognition tasks. The bounding box–based object detection approach has often been used22,23; however, in future developments in the surgical research field, such as the automation of surgery, more detailed pixel-by-pixel recognition information will be essential, instead of bounding box–based object detection.

Semantic segmentation is a CNN-based CV approach that enables pixel-by-pixel image recognition. Recently, it has also been actively used in the research field of CV in surgery.24,25 Semantic segmentation attempts to specifically understand the role of each pixel in an image through the process of dividing whole images into pixel groupings that can then be labeled and classified. The boundaries of each object can be delineated; therefore, dense pixel-based predictions can be achieved. However, semantic segmentation cannot distinguish them individually when objects of the same type overlap. Instance segmentation overcomes this limitation because it can distinguish overlapping objects, even when they are of the same type. It enables the visualization of complicated images with multiple overlapping objects and identifies their relationships. In this study, we selected instance segmentation as the image recognition approach because the crossing of multiple surgical instruments during surgery is a common occurrence.

In general, multiclass recognition is difficult—the more types of objects there are to be recognized, the more difficult the task becomes. However, the accuracy of the recognition task for the 3, 4, 6, and 8 types of surgical instruments was almost the same in this study (Figure 3B). One possible reason for this is that we prepared a sufficient amount of annotated data. These data comprised more than 80 000 annotation labels for the training of the CNN-based instant segmentation model, enabling us to overcome the difficulties besetting the multiclass recognition task.

Although the recognition accuracy was quite high, several misrecognition patterns needed to be measured, which were mainly caused by intracorporeal artifacts other than surgical instruments in this study (Figure 5). However, we believe that this problem can be solved in the future by intentionally adding annotation images that capture such artifacts and training the model such that they are part of the background. Otherwise, such artifacts can be added to the recognition target with annotations. In addition, although the situation that the surgical instrument tip was out of view was also expected to cause misrecognition, this pattern was minor in this study. This pattern was thought to be because such situations were rare during surgery to begin with, and much of the information on shafts and tips was used in the predictions. Differential identification among surgical instruments with the same shaft and different tip options is a challenge that will be tackled in future work.

The proposed instance segmentation model for surgical instrument recognition has various applications in surgical research and developing fields, such as automated robotic camera holders and automatic surgical skill assessments based on surgical instrument tracking. This study demonstrated the feasibility of applying the model to automatic surgical progress monitoring.

Accurately monitoring surgical progress in real time for optimal operating room (OR) logistics is necessary.26 This monitoring enables OR resource optimization, and the spare time and human resources lead to patient comfort and safety. In addition, if a certain surgical step takes longer than planned and causes delays, this may indicate that some kind of intraoperative adverse event has occurred. Automatic self-extracting, analyzing, and monitoring systems for surgical progress enable the entire OR staff to easily share crisis awareness, which leads to improved OR team performance in terms of quick decisions and actions.

In this study, we focused on the appearance information of SI-4/5 and SI-6 in the surgical steps of IMA transection and rectal transection in laparoscopic colorectal surgery, respectively; however, other surgical instruments appear in specific surgical steps in every type of surgery. Therefore, the proposed approach for automatic surgical progress monitoring based on surgical instrument recognition is promising because of its high applicability. In the future, it may be possible to monitor not only the progress of the surgery but also the event of intraoperative bleeding based on the appearance information of surgical instruments used for hemostasis, such as SI-8.

Limitations

This study has several limitations. First, although the proposed instance segmentation model for surgical instrument recognition has various applications in surgical research and development, only a proof-of-concept evaluation was performed; the model was not applied to actual surgical practice. Second, only validation was conducted; tests using data from an external cohort were not performed. However, the annotation data set was sufficiently large, and stable recognition accuracy was obtained from 5-fold cross-validation; therefore, the generalization performance was high. Third, for automatic surgical progress monitoring, only preliminary results were obtained from a single intraoperative video. For future validation, we must determine the minimum frame rate for continuous recognition. Fourth, because SI-3 tended to appear with complex intersections with other surgical instruments, the corresponding mAP was lower than for the other instruments, despite having more annotation labels. Improving the recognition accuracy of SI-3 is also a future task.

Conclusions

We developed an instance segmentation model that can simultaneously recognize 8 types of surgical instruments that are frequently used in laparoscopic colorectal operations. In this quality improvement study, the recognition accuracy of the model was high and was maintained even when the number of types of surgical instruments to be recognized increased. Surgical instrument recognition is an important fundamental technology for various surgical research and development areas. The proposed model can be applied not only to the proposed automatic surgical progress monitoring in this study but also for future computer-assisted surgery realization or surgical automation.

Supplement.

eFigure. Mean Average Precision Values at Multiple Thresholds Ranging From 0.5 to 0.95

References

  • 1.Siddaiah-Subramanya M, Tiang KW, Nyandowe M. A new era of minimally invasive surgery: progress and development of major technical innovations in general surgery over the last decade. Surg J (N Y). 2017;3(4):e163-e166. doi: 10.1055/s-0037-1608651 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Huber T, Paschold M, Schneble F, et al. Structured assessment of laparoscopic camera navigation skills: the SALAS score. Surg Endosc. 2018;32(12):4980-4984. doi: 10.1007/s00464-018-6260-7 [DOI] [PubMed] [Google Scholar]
  • 3.Garrow CR, Kowalewski KF, Li L, et al. Machine learning for surgical phase recognition: a systematic review. Ann Surg. 2021;273(4):684-693. doi: 10.1097/SLA.0000000000004425 [DOI] [PubMed] [Google Scholar]
  • 4.Hashimoto DA, Rosman G, Rus D, Meireles OR. Artificial intelligence in surgery: promises and perils. Ann Surg. 2018;268(1):70-76. doi: 10.1097/SLA.0000000000002693 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Anteby R, Horesh N, Soffer S, et al. Deep learning visual analysis in laparoscopic surgery: a systematic review and diagnostic test accuracy meta-analysis. Surg Endosc. 2021;35(4):1521-1533. doi: 10.1007/s00464-020-08168-1 [DOI] [PubMed] [Google Scholar]
  • 6.Ward TM, Mascagni P, Ban Y, et al. Computer vision in surgery. Surgery. 2021;169(5):1253-1256. doi: 10.1016/j.surg.2020.10.039 [DOI] [PubMed] [Google Scholar]
  • 7.Zhang J, Gao X. Object extraction via deep learning-based marker-free tracking framework of surgical instruments for laparoscope-holder robots. Int J Comput Assist Radiol Surg. 2020;15(8):1335-1345. doi: 10.1007/s11548-020-02214-y [DOI] [PubMed] [Google Scholar]
  • 8.Wijsman PJM, Broeders IAMJ, Brenkman HJ, et al. First experience with THE AUTOLAP™ SYSTEM: an image-based robotic camera steering device. Surg Endosc. 2018;32(5):2560-2566. doi: 10.1007/s00464-017-5957-3 [DOI] [PubMed] [Google Scholar]
  • 9.Levin M, McKechnie T, Khalid S, Grantcharov TP, Goldenberg M. Automated methods of technical skill assessment in surgery: a systematic review. J Surg Educ. 2019;76(6):1629-1639. doi: 10.1016/j.jsurg.2019.06.011 [DOI] [PubMed] [Google Scholar]
  • 10.Burström G, Nachabe R, Persson O, Edström E, Elmi Terander A. Augmented and virtual reality instrument tracking for minimally invasive spine surgery: a feasibility and accuracy study. Spine (Phila Pa 1976). 2019;44(15):1097-1104. doi: 10.1097/BRS.0000000000003006 [DOI] [PubMed] [Google Scholar]
  • 11.De Paolis LT, De Luca V. Augmented visualization with depth perception cues to improve the surgeon’s performance in minimally invasive surgery. Med Biol Eng Comput. 2019;57(5):995-1013. doi: 10.1007/s11517-018-1929-6 [DOI] [PubMed] [Google Scholar]
  • 12.Saeidi H, Opfermann JD, Kam M, Raghunathan S, Leonard S, Krieger A. A confidence-based shared control strategy for the smart tissue autonomous robot (STAR). Rep U S. 2018;2018:1268-1275. doi: 10.1109/IROS.2018.8594290 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.World Medical Association. World Medical Association Declaration of Helsinki: ethical principles for medical research involving human subjects. JAMA. 2013;310(20):2191-2194. doi: 10.1001/jama.2013.28105 [DOI] [PubMed] [Google Scholar]
  • 14.He K, Gkioxari G, Dollar P, Girshick R. Mask R-CNN. IEEE Trans Pattern Anal Mach Intell. 2020;42(2):386-397. doi: 10.1109/TPAMI.2018.2844175 [DOI] [PubMed] [Google Scholar]
  • 15.Fulton LV, Dolezel D, Harrop J, Yan Y, Fulton CP. Classification of Alzheimer’s disease with and without imagery using gradient boosted machines and ResNet-50. Brain Sci. 2019;9(9):E212. doi: 10.3390/brainsci9090212 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Ouyang N, Wang W, Ma L, et al. Diagnosing acute promyelocytic leukemia by using convolutional neural network. Clin Chim Acta. 2021;512:1-6. doi: 10.1016/j.cca.2020.10.039 [DOI] [PubMed] [Google Scholar]
  • 17.Chen K, Wang J, Pan J, et al. MMDetection: open MMLab detection toolbox and benchmark. arXiv. Preprint posted online June 17, 2019. doi: 10.48550/arXiv.1906.07155 [DOI]
  • 18.Korfhage N, Mühling M, Ringshandl S, Becker A, Schmeck B, Freisleben B. Detection and segmentation of morphologically complex eukaryotic cells in fluorescence microscopy images via feature pyramid fusion. PLoS Comput Biol. 2020;16(9):e1008179. doi: 10.1371/journal.pcbi.1008179 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Zhang L, Zhang Y, Zhang Z, Shen J, Wang H. Real-time water surface object detection based on improved faster R-CNN. Sensors (Basel). 2019;19(16):3523. doi: 10.3390/s19163523 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Wang Q, Bi S, Sun M, Wang Y, Wang D, Yang S. Deep learning approach to peripheral leukocyte recognition. PLoS One. 2019;14(6):e0218808. doi: 10.1371/journal.pone.0218808 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Kitaguchi D, Takeshita N, Matsuzaki H, et al. Automated laparoscopic colorectal surgery workflow recognition using artificial intelligence: experimental research. Int J Surg. 2020;79:88-94. doi: 10.1016/j.ijsu.2020.05.015 [DOI] [PubMed] [Google Scholar]
  • 22.Yamazaki Y, Kanaji S, Matsuda T, et al. Automated surgical instrument detection from laparoscopic gastrectomy video images using an open source convolutional neural network platform. J Am Coll Surg. 2020;230(5):725-732.e1. doi: 10.1016/j.jamcollsurg.2020.01.037 [DOI] [PubMed] [Google Scholar]
  • 23.Zhao Z, Cai T, Chang F, Cheng X. Real-time surgical instrument detection in robot-assisted surgery using a convolutional neural network cascade. Healthc Technol Lett. 2019;6(6):275-279. doi: 10.1049/htl.2019.0064 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Madani A, Namazi B, Altieri MS, et al. Artificial intelligence for intraoperative guidance: using semantic segmentation to identify surgical anatomy during laparoscopic cholecystectomy. Ann Surg. Published online November 13, 2020. doi: 10.1097/SLA.0000000000004594 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Mascagni P, Vardazaryan A, Alapatt D, et al. Artificial intelligence for surgical safety: automatic assessment of the critical view of safety in laparoscopic cholecystectomy using deep learning. Ann Surg. 2022;275(5):955-961. doi: 10.1097/SLA.0000000000004351 [DOI] [PubMed] [Google Scholar]
  • 26.Twinanda AP, Yengera G, Mutter D, Marescaux J, Padoy N. RSDNet: learning to predict remaining surgery duration from laparoscopic videos without manual annotations. IEEE Trans Med Imaging. 2019;38(4):1069-1078. doi: 10.1109/TMI.2018.2878055 [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplement.

eFigure. Mean Average Precision Values at Multiple Thresholds Ranging From 0.5 to 0.95


Articles from JAMA Network Open are provided here courtesy of American Medical Association

RESOURCES