Abstract
Objective
To verify the reliability and clinical feasibility of a self-developed navigation system based on an augmented reality technique for endoscopic sinus and skull base surgery.
Materials and Methods
In this study we performed a head phantom and cadaver experiment to determine the display effect and accuracy of our navigational system. We compared cadaver head-based simulated operations, the target registration error, operation time, and National Aeronautics and Space Administration Task Load Index scores of our navigation system to conventional navigation systems.
Results
The navigation system developed in this study has a novel display mode capable of fusing endoscopic images to three-dimensional (3-D) virtual images. In the cadaver head experiment, the target registration error was 1.28 ± 0.45 mm, which met the accepted standards of a navigation system used for nasal endoscopic surgery. Compared with conventional navigation systems, the new system was more effective in terms of operation time and the mental workload of surgeons, which is especially important for less experienced surgeons.
Conclusion
The self-developed augmented reality navigation system for endoscopic sinus and skull base surgery appears to have advantages that outweigh those of conventional navigation systems. We conclude that this navigational system will provide rhinologists with more intuitive and more detailed imaging information, thus reducing the judgment time and mental workload of surgeons when performing complex sinus and skull base surgeries. Ultimately, this new navigational system has potential to increase the quality of surgeries. In addition, the augmented reality navigational system could be of interest to junior doctors being trained in endoscopic techniques because it could speed up their learning. However, it should be noted that the navigation system serves as an adjunct to a surgeon’s skills and knowledge, not as a substitute.
Introduction
Endoscopic sinus surgery techniques have expanded and are now used to treat orbital and skull base diseases. However, an increasing number of complications related to nasal endoscopic surgery have been reported, including severe bleeding, blindness, and damage to the central nervous system. To address these concerns, navigational systems for nasal endoscopic surgery have been developed and used in clinics to prevent and reduce endoscopy-related issues [1]. However, correlative research showed that the benefits conferred by navigation systems are extremely limited [2]. Based on our review of the literature and personal experience, we determined that navigational systems fail for the following reasons: (1) it is difficult for doctors to match the tomographic image on the screen to the actual structures in the operated area, especially for doctors with less experience (Fig 1); (2) in conventional navigation (C-N) systems, images are displayed on another screen, so the surgeon has to look away from the surgical field when the anatomical location is being defined, which can be dangerous during key surgical steps [3]; and (3) a great amount of time is spent repeatedly verifying surgical sites using a probe during the operation.
Due to these limitations, we developed an augmented reality (AR)-based navigation system for endoscopic sinus and skull base surgery. In this system, three-dimensional (3-D) virtual images, based on preoperative computed tomography scans or/and magnetic resonance imaging (CT/MRI) images, are used as the background, and the distortion-corrected real endoscopic images are semi-transparently fused to it. This new display mode allows surgeons to stereoscopically observe the subsurface and surrounding anatomical structures of the surgical field, providing more detailed and intuitive information for safer surgeries. In this study, we used a series of experiments to evaluate the accuracy of our AR navigation (AR-N) system and to demonstrate its advantages, analyze its benefits to surgeons, and determine its clinical feasibility.
Materials and Methods
1. Platform setup
The AR-N system consists of the following components (Fig 2): (1) nasal endoscopy imaging system (RU0101-25, Rudolf Medizintechnik GmbH & Co. KG, Tuttlingen, Germany) and 0° endoscope (Karl Storz, Tuttlingen, Germany); (2) tracking system POLARIS Vicra (Northern Digital Inc., Waterloo, Canada) including an infrared emission device and reflection markers fixed on the endoscope (Fig 3) and the patient’s head; and (3) a workstation (Intel Core i7) equipped with a self-developed open-source software.
The newly developed navigation system has a similar configuration to conventional systems. However, improvements include binding of the reflection markers to the endoscope, a novel navigation software package that can track the movement of the endoscope, and a merging of the operative view with stereoscopic 3-D virtual images of anatomical structures.
During system design, we ensured that real endoscopic images could not be seamlessly superimposed onto the virtual model until radial distortion of endoscopic images was corrected. We utilized a spherical surface stereographic projection technique, as described in our previous study [4]. After estimating the field of view of the endoscope by detecting the corners, the coefficients of the spherical projection model are fitted using the least-squares method and the distorted images can be corrected according to the spherical projection maps. The method is fully automatic and requires no human intervention.
2. Experimental procedures
To evaluate the accuracy of our system, we used a PVC qualitative skull phantom with an openable top with six non-coplanar reference markers, which was previously determined as the most appropriate number of markers [5]. These markers were fixed at the parietal, frontal, and bilateral temples and mastoid processes. A CT scan (thickness 0.75 mm) (GE Brightspeed Elite, General Electric Company) was performed to obtain imaging data. The data were transferred into the workstation and the skull phantom with reflection markers was fixed to the operating platform. Reference markers were individually touched with an accurately calibrated endoscope tip by the participants; corresponding points on the virtual image were then selected to complete registration. Participants were then asked to use the calibrated probe, which was used only for the test, to point to the center of specific anatomical markers (e.g., piriform aperture vertex, bilateral supraorbital foramen, the anterior end of the inferior turbinate attachment, anterior clinoid process and posterior clinoid process: nine points in total) to obtain the 3-D coordinates of those points (Fig 4). The probe was then moved slightly so that its virtual tip on the screen was pointing at the corresponding position on the virtual image to obtain another 3-D coordinate. The spatial distance between these two coordinates was calculated as the target registration error (TRE). After measuring the nine points, the average TRE was calculated and recorded, with this whole process representing a single test. If the distance obtained was smaller than 2 mm, the recognized allowable error in navigation systems for nasal endoscopic surgeries [6], the points were considered as accurately corresponding between the real and virtual images (i.e., the two images can be perfectly superimposed). A total of nine participants took the test, all of whom were asked to complete the test on nine separate occasions; therefore, we obtained 81 average TRE values. Each operation included rebooting the workstation, restarting the programs, and registering the phantom. Experimental results were recorded and analyzed.
The above results are theoretical values based on a rigid model. To evaluate the system’s performance in practice, we utilized 15 fixed cadaver heads to simulate surgery aided by the system. We compared several key indexes about the procedure from our navigational system with those of a C-N system. The experiment was performed after receiving ethical approval from the Committee on Ethics of Biomedicine Research at Chinese PLA General Hospital (Beijing, China). Experimental cadaver heads were provided by the Beijing Society for Anatomical Sciences (http://test.bjmu.edu.cn/). The process of donating remains to the society involves the donating individuals and their families submitting an application and providing written informed consent to participate in medical research. The Ethics Committee confirmed that the cadaver heads were being used legally and that the entire research process satisfied all ethical requirements. Prior to the experiment, the arteries were infused with a mixture of paint, latex, and developing agent to plump them and allow observation of the surgical field and images. Image collection, data processing, and registration before and prior to the operation were performed as for the phantom experiment. A total of 15 otorhinolaryngologists participated in this study, and each performed maxillary sinus expansion, ethmoidectomy, frontal sinus expansion, sphenoidotomy (Messerklinger method was used as the standard), and intracavernous internal carotid artery anatomical dissection on both nasal cavities of one cadaver head, guided on one side by AR-N and on the other by C-N. The order of sides and order of navigation system employment for each specimen was randomly assigned. TRE and operation time (OT) were recorded after each operation. Meanwhile, to determine the surgeon’s perceived workload and mental demands, a dialog-based computer application developed by Cao et al. [7] based on the National Aeronautics and Space Administration Task Load Index (NASA-TLX) (Table 1) was used during performance of the operation on each side. After the operation, each surgeon completed a computer-based survey to rate his or her experience. At the end of the experiment, the quantified scores in each subscale and the overall score obtained through the program were used to assess the level of assistance provided by the navigational systems. Data were statistically analyzed using IBM SPSS Statistics version 22, and P < 0.05 was considered statistically significant.
Table 1. Description of the NASA-TLX rating scale [7].
Title | Description |
---|---|
Mental Demand | How much mental and perceptual activity was required (e.g., thinking, deciding, calculating, remembering, looking, searching)? Was the task easy or demanding, simple or complex, exacting or forgiving? |
Physical Demand | How much physical activity was required (e.g., pushing, pulling, turning, controlling, activating)? Was the task easy or demanding, slow or brisk, slack or strenuous, restful or laborious? |
Temporal Demand | How much time pressure did you feel due to the rate or pace at which the tasks or task elements occurred? Was the pace slow and leisurely or rapid and frantic? |
Performance | How successful do you think you were in accomplishing the goals of the task set by the experimenter (or yourself)? How satisfied were you with your performance in accomplishing these goals? |
Effort | How hard did you have to work (mentally and physically) to accomplish your level of performance? |
Frustration Level | How insecure, discouraged, irritated, stressed, and annoyed or secure, gratified, content, relaxed, and complacent did you feel during the task? |
Results
The newly developed AR-N system display mode was able to fuse nasal endoscopic images into 3-D virtual images. The head phantom experiment demonstrated the satisfactory display effect (Fig 5): when we examined the nasal cavity with the 0° lens, as mentioned in the Materials and Methods section, the contours of the anatomical structure on both images overlapped accurately (S1 Video). Moreover, by measuring nine different points for each of the nine participants (S1 Table), the average TRE of the system was calculated as 1.19 ± 0.42 mm. The frequency distributions of all of the recorded data are shown in Fig 6.
The cadaver head experiment further demonstrated the effectiveness of the display effect of the AR-N system in an anatomical, structural environment. Fig 7 shows the seamless fusion of semitransparent endoscopic images with 3-D virtual images with satisfying anatomical structure contour continuity at the boundary. The two types of images changed synchronously when the endoscope moved and rotated in any direction. By adjusting the transparency of the endoscopic image or the virtual image rendering method, the operator is able to clearly observe the anatomical structures (e.g., the arteries) located in deep positions and around the surgical field (S2 Video). Table 2 compares the performance of the AR-N to the C-N system during the simulated operations; the average TRE rates were 1.28 ± 0.45 and 1.32 ± 0.41 mm, respectively, and the average OT were 88.27 ± 20.45 and 104.93 ± 24.61 min, respectively. Although we did not observe significant differences in TRE between the two systems (P > 0.05), task completion time differed significantly (P < 0.05) (Fig 8).
Table 2. Performance of augmented reality navigation (AR-N) system and conventional navigation (C-N) system during the simulated operations.
Years of Experience | TRE (mm) | OT (min) | |||
---|---|---|---|---|---|
AR-N | C-N | AR-N | C-N | ||
Participant 1 | 22 | 1.19 | 1.14 | 54 | 62 |
Participant 2 | 6 | 1.81 | 1.51 | 98 | 117 |
Participant 3 | 8 | 1.70 | 1.78 | 82 | 108 |
Participant 4 | 19 | 1.47 | 1.64 | 67 | 76 |
Participant 5 | 9 | 1.16 | 0.76 | 88 | 107 |
Participant 6 | 8 | 0.90 | 1.09 | 94 | 112 |
Participant 7 | 11 | 1.24 | 1.71 | 69 | 83 |
Participant 8 | 16 | 0.71 | 0.80 | 63 | 70 |
Participant 9 | 5 | 0.56 | 2.00 | 110 | 132 |
Participant 10 | 10 | 1.60 | 1.53 | 82 | 96 |
Participant 11 | 13 | 2.02 | 1.26 | 73 | 89 |
Participant 12 | 4 | 0.83 | 1.13 | 105 | 125 |
Participant 13 | 4 | 1.36 | 1.65 | 121 | 136 |
Participant 14 | 6 | 0.86 | 0.55 | 102 | 122 |
Participant 15 | 4 | 1.77 | 1.30 | 116 | 139 |
Average | 1.28±0.45 | 1.32±0.41 | 88.27±20.45 | 104.93±24.61 | |
P>0.05 | P<0.05 |
Performance was also analyzed according to surgeons’ years of experience. We divided the participants into two groups: a junior group (surgeons with <10 years of experience [6, 8, 9, 8, 5, 4, 4, 6, and 4 years; n = 9]) and a senior group (surgeons with ≥10 years of experience [22, 19, 11, 16, 10, and 13 y; n = 6]). The difference in time required for surgery between the two systems (OTC-N-OTAR-N) was greater for the junior group than for the senior group (P < 0.05) (Fig 9). Furthermore, the extent of time reduction decreased with increasing experience (Fig 10).
All participants completed the computer-based survey as required after the simulated operation. The results showed that the average score on each item (mental demand, physical demand, temporal demand, performance, effort, frustration level, and overall scores) differed significantly (Wilcoxon rank sum test) between the AR-N and C-N systems (S2 Table), with the AR-N system outperforming the C-N system in all categories measured (Fig 11).
When surgeons were divided by years of experience, the junior group had the greatest reduction in overall score for mental workload (SC-N-SAR-N) when using the developed navigational system compared with the senior group (P < 0.05) (Fig 12). This result was similar to that for operation time: the reduction in mental workload decreased with increasing experience (Fig 13).
Discussion
Commensurate with a development period of more than 20 years, the performance of navigation systems during nasal endoscopic surgery has improved significantly. However, several recent studies have indicated that these systems may not be as effective as their users had expected. An evidence-based review by Ramakrishnan et al. [2] showed that the use of navigation systems was not directly associated with decreased surgical complications or improved surgical outcomes. One possible reason, according to our analysis, is that it is not easy for a doctor to match the tomographic image quickly and accurately to the endoscopic image in triplanar view mode of C-N system. Therefore, we considered the possibility that a navigation system based on the AR technique, which can reconstruct the stereoscopic form of surgical site and fuse the endoscopic image to 3-D reconstruction image, could be better able to meet the doctors’ observation habit. The most common application of AR techniques is in the field of neurosurgery, which requires a high degree of navigational accuracy [8]. More recently, AR techniques have been used during spinal [9], plastic [10], maxillofacial [11], and several other highly technical surgeries. However, few studies have focused on the application of AR to nasal endoscopic navigation. Although Caversaccio et al. [12], Thoranaghatte et al. [13], and Dixon et al. [14] reported dramatic improvements in this respect, the display mode of the navigation systems still fails to restore the original forms and positions of important structures within and surrounding the surgical field.
The navigational system for nasal endoscopic surgeries proposed in this study was developed based on the AR technique. It innovatively matches and semi-transparently fuses distortion-corrected nasal endoscopic images into 3-D virtual images based on preoperative imaging data. Thus, this system provides surgeons with a real-time stereoscopic image in a real sense, allowing surgeons to view deep anatomical structures, such as tumors, blood vessels, the brain, and eyeballs, in their original forms and positions by a superficial surgical field prior to sectioning. By operating in this manner, this system expands in depth and breadth the previously limited visual field of the nasal endoscope. Because some participating surgeons suggested that virtual images of bony structures in the display interface might interfere with observation of deeper structures, we made improvements to allow separate adjustment of the transparencies of real images, virtual images, and even different structures in virtual images.
In this study, we conducted head phantom and cadaver head experiments. The TRE was calculated during testing. Maurer et al. [15] and Citardi and Batra [16] suggest that this index could be the best indicator of the accuracy of a system. The results of the phantom experiment revealed that the theoretical accuracy of this system meets the recognized standard, while the cadaver head experiment revealed that the system was satisfactory in terms of display and accuracy in real operations. Furthermore, comparisons of the AR-N and C-N systems revealed that the AR-N system allowed surgeons to complete the simulated operation in less time. One possible explanation for this result is the improved display, which facilitates the cognitive processes required to connect imaging data to real structures and eliminates the need to look away from the screen or use probes to verify surgical sites.
Participants rated their mental workload, a key factor that affects a surgeon’s performance [7], during the operation using the NASA-TLX scale. The results indicate that using the AR-N system reduced mental workload compared to the C-N system. One possible explanation for this result is that this system allows participants to intuitively learn the anatomical structures within the surgical field, thus relieving mental stress during operation, reducing frustration, and increasing confidence.
Grouped analyses of experimental results indicated that doctors with less experience benefited more significantly from use of the AR-N system than surgeons with more experience, in terms of task completion time and mental workload. We suspect that this difference might be due to the anatomical knowledge already possessed by surgeons with more experience (i.e., surgeons who can easily connect imaging data to real images even under C-N monitoring). Although the intuitive process is not as well developed in less experienced doctors, a more intuitive display could improve this process.
In summary, the new AR-N system was accurate and intuitive for the proposed applications. It enables doctors to more easily learn the shape and positional relationships of anatomical structures in the surgical field, providing the necessary tools to achieve a more successful operation. Given these advantages, this system will likely play an important role in nasal endoscopic surgery, and specifically in endoscopic skull base surgery.
Besides, the system itself and the study has several limitations: (1) soft tissues might be displaced due to gravity or tumor resection during surgery, and navigational systems based on preoperative imaging data may not reflect tissue deformations. However, intraoperative CT/MRI devices to update imaging data may solve this problem; (2) in cadaver head experiment, we did not break down the task analysis by surgical sub-procedure, because we considered that in the initial validation, the overall differences in performance between the two systems may well depend on the differences in some key procedures. Thus we analyzed the task as a whole. In the next step of the research, we plan to compare the system performances in sub-procedures respectively, so as to find out the especial advantages of the AR-N system in complex surgical procedures.
Additionally, this study had a limited sample size. Further preclinical tests will need to be performed, and further improvements will be necessary according to the usage and habits of surgeons. We plan to combine this navigation system with a simulated surgical system based on force feedback technology. These integrated systems will be more appropriate for delivering the training necessary for skull-based surgery through an endoscopic endonasal approach, as well as for planning complex surgeries.
Conclusion
The newly developed AR-N system uses a novel display mode that fuses corrected endoscopic images into a 3-D reconstructed background. Our findings indicate that this display mode can more successfully restore the original forms and positions of important structures within and surrounding the surgical field, and may thus provide surgeons with intuitive, real-time imaging information, which could ultimately reduce their judgment time and mental workload. This system might eventually help doctors perform safer operations based on more rational decisions. In addition, the AR-N system could be of interest to junior doctors being trained in endoscopic techniques: the anatomical information provided by the endonasal endoscope, the greater maneuverability of the instruments, and the clearly defined surgical steps may improve their learning efficiency. However, it should be noted that the navigation system serves as an adjunct to a surgeon’s skills and anatomical knowledge, not as a substitute [17]. Moreover, in cases of nasal and sinus lesions involving the brain or orbit, the assistance that can be provided by navigation systems is limited, such that multidisciplinary collaboration during the operation remains the preferred approach.
Supporting Information
Data Availability
All relevant data are within the paper and its Supporting Information files.
Funding Statement
The study was supported by the National science and technology Support Plan (Funding institution: Ministry of Science and Technology of the People's Republic of China, URLs: http://program.most.gov.cn/, Grant number:2013BAI01B01, Grant author: PL.) The funder had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
References
- 1.Anon JB. Computer-aided endoscopic sinus surgery. The Laryngoscope. 1998;108(7):949–61. Epub 1998/07/17. . [DOI] [PubMed] [Google Scholar]
- 2.Ramakrishnan VR, Orlandi RR, Citardi MJ, Smith TL, Fried MP, Kingdom TT. The use of image-guided surgery in endoscopic sinus surgery: an evidence-based review with recommendations. International forum of allergy & rhinology. 2013;3(3):236–41. Epub 2012/10/10. 10.1002/alr.21094 . [DOI] [PubMed] [Google Scholar]
- 3.Edwards PJ, Hawkes DJ, Hill DL, Jewell D, Spink R, Strong A, et al. Augmentation of reality using an operating microscope for otolaryngology and neurosurgical guidance. Journal of image guided surgery. 1995;1(3):172–8. Epub 1995/01/01. . [DOI] [PubMed] [Google Scholar]
- 4.Weijian C, Jian Y, Weijiang D, Jianzhun Z. Automatic radial distortion correction for endoscope image. 6th International Congress on Image and Signal Processing (CISP 2013). 2013:932–7.
- 5.Snyderman C, Zimmer LA, Kassam A. Sources of registration error with image guidance systems during endoscopic anterior cranial base surgery. Otolaryngology—head and neck surgery: official journal of American Academy of Otolaryngology-Head and Neck Surgery. 2004;131(3):145–9. Epub 2004/09/15. 10.1016/j.otohns.2004.03.002 . [DOI] [PubMed] [Google Scholar]
- 6.Chu ST. Endoscopic sinus surgery under navigation system—analysis report of 79 cases. Journal of the Chinese Medical Association: JCMA. 2006;69(11):529–33. Epub 2006/11/23. 10.1016/s1726-4901(09)70323-5 . [DOI] [PubMed] [Google Scholar]
- 7.Cao A, Chintamani KK, Pandya AK, Ellis RD. NASA TLX: software for assessing subjective mental workload. Behavior research methods. 2009;41(1):113–7. Epub 2009/02/03. 10.3758/brm.41.1.113 . [DOI] [PubMed] [Google Scholar]
- 8.Edwards PJ, King AP, Hawkes DJ, Fleig O, Maurer CR Jr, Hill DL, et al. Stereo augmented reality in the surgical microscope. Studies in health technology and informatics. 1999;62:102–8. Epub 1999/10/28. . [PubMed] [Google Scholar]
- 9.Bastien S, Peuchot B, Tanguy A. Augmented reality in spine surgery: critical appraisal and status of development. Studies in health technology and informatics. 2002;88:153–6. Epub 2004/10/01. . [PubMed] [Google Scholar]
- 10.Blackwell M, Morgan F, DiGioia AM 3rd. Augmented reality and its future in orthopaedics. Clinical orthopaedics and related research. 1998;(354):111–22. Epub 1998/10/02. . [DOI] [PubMed] [Google Scholar]
- 11.Mischkowski RA, Zinser MJ, Kubler AC, Krug B, Seifert U, Zoller JE. Application of an augmented reality tool for maxillary positioning in orthognathic surgery—a feasibility study. Journal of cranio-maxillo-facial surgery: official publication of the European Association for Cranio-Maxillo-Facial Surgery. 2006;34(8):478–83. Epub 2006/12/13. 10.1016/j.jcms.2006.07.862 . [DOI] [PubMed] [Google Scholar]
- 12.Caversaccio M, Garcia Giraldez J, Thoranaghatte R, Zheng G, Eggli P, Nolte LP, et al. Augmented reality endoscopic system (ARES): preliminary results. Rhinology. 2008;46(2):156–8. Epub 2008/06/26. . [PubMed] [Google Scholar]
- 13.Thoranaghatte R, Garcia J, Caversaccio M, Widmer D, Gonzalez Ballester MA, Nolte LP, et al. Landmark-based augmented reality system for paranasal and transnasal endoscopic surgeries. The international journal of medical robotics + computer assisted surgery: MRCAS. 2009;5(4):415–22. Epub 2009/07/23. 10.1002/rcs.273 . [DOI] [PubMed] [Google Scholar]
- 14.Dixon BJ, Daly MJ, Chan H, Vescan A, Witterick IJ, Irish JC. Augmented real-time navigation with critical structure proximity alerts for endoscopic skull base surgery. The Laryngoscope. 2014;124(4):853–9. Epub 2013/10/15. 10.1002/lary.24385 . [DOI] [PubMed] [Google Scholar]
- 15.Maurer CR Jr, Fitzpatrick JM, Wang MY, Galloway RL Jr, Maciunas RJ, Allen GS. Registration of head volume images using implantable fiducial markers. IEEE transactions on medical imaging. 1997;16(4):447–62. Epub 1997/08/01. 10.1109/42.611354 . [DOI] [PubMed] [Google Scholar]
- 16.Citardi MJ, Batra PS. Intraoperative surgical navigation for endoscopic sinus surgery: rationale and indications. Current opinion in otolaryngology & head and neck surgery. 2007;15(1):23–7. Epub 2007/01/11. 10.1097/MOO.0b013e3280123130 . [DOI] [PubMed] [Google Scholar]
- 17.Irugu DV, Stammberger HR. A note on the technical aspects and evaluation of the role of navigation system in endoscopic endonasal surgeries. Indian journal of otolaryngology and head and neck surgery: official publication of the Association of Otolaryngologists of India. 2014;66(Suppl 1):307–13. Epub 2014/02/18. 10.1007/s12070-011-0380-x ; PubMed Central PMCID: PMCPmc3918301. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
All relevant data are within the paper and its Supporting Information files.