Abstract
Surgery in the cranial area includes complex anatomic situations with high-risk structures and high demands for functional and aesthetic results. Conventional surgery requires that the surgeon transfers complex anatomic and surgical planning information, using spatial sense and experience. The surgical procedure depends entirely on the manual skills of the operator. The development of image-guided surgery provides new revolutionary opportunities by integrating presurgical 3D imaging and intraoperative manipulation. Augmented reality, mechatronic surgical tools, and medical robotics may continue to progress in surgical instrumentation, and ultimately, surgical care. The aim of this article is to review and discuss state-of-the-art surgical navigation and medical robotics, image-to-patient registration, aspects of accuracy, and clinical applications for surgery in the cranial area.
Keywords: image-guided surgery, mechatronic surgical tools, medical robotics, image-to-patient registration accuracy
INTRODUCTION
Surgery in the cranial area includes operations of the fronto-zygomatico-maxillary complex, nasal cavity, paranasal sinuses, ear, and the skull base that have close proximity to highly critical structures such as nerves, vessels, the eye, cochlear and labyrinth organ, or the brain. Such operations often require re-establishing functional and aesthetic anatomy by repositioning displaced skeletal elements, or by grafting and contouring abnormal bony contours and transplants [1-5]. The need for accurate preoperative determination of the proposed surgical procedure is essential, and excellent intraoperative orientation and manual skills are required for surgical precision and reliable protection of vital anatomic structures [6-12]. Next generation surgical systems should explore and enhance imaging or manipulation, the two basic components of a surgical procedure [14]. The development of image-guided surgery provides new revolutionary opportunities by integration of presurgical 3D imaging, obtained by computed tomography (CT) or magnetic resonance imaging (MRI), and intraoperative manipulation through three fundamental issues [4,15,16]:
Localisation - determination of a target’s locus (for example, tumour, foreign body, and so on) that defines a task the surgeon performs,
Orientation - information on current location on the patient’s anatomy that defines where the surgeon (with respect to the surgical tool) is operating, and
Navigation - the process of (passive) guidance to reach a desired target from the current location (for example, biopsy, tumour resection, bone segment manipulation, implant positioning, and so on).
As a logical extension of image-guided surgery, the development of mechatronic surgical tools, tele-manipulated robotic arms, and semi- or fully- automated surgical robots are beginning to introduce the next revolution [17,18].
SURGICAL NAVIGATION SYSTEMS
Surgical navigation systems generally consist of a (transportable) work station, a monitor, a graphical user interface with software to plan and guide therapy, and a position measuring system (a three-dimensional coordinate-detection or tracking system, which can be either mechanical, electromagnetic, or optical) [6,12,19-22]. By providing a spatial coordinate system relative to the patient's anatomy (see chapter on image-to-patient transformation), the actual position of a probe or tracked surgical tool is shown with respect to cross-sectional images of the preoperative dataset (see chapter on image guidance).
Mechanical navigation systems
A mechanical navigation system consists of an articulated arm with six degrees of freedom [23-26]. Calculation of position is based on measurement of temperature changes recorded by a semiconductor temperature sensor within the gear of movable angles. As the spatial system is entirely self-referential, rigid fixation of both the patient and the navigation arm is an important prerequisite [19,20,23,24,27].
The advantages of the mechanical systems are acceptable precision, low susceptibility to failure, and sterile covering with a tube [21,23,25,29]. The disadvantages are impractical handling during some surgeries, restricted range (circa 60 cm), and mobility as well as the space requirements in the operating table [6,19,25,27]. Due to their bulkiness, the mechanical systems have been generally replaced by more flexible electromagnetic and optical navigation systems.
Electromagnetic navigation systems
The position of electromagnetic navigation is measured by detecting of magnetic field changes with coils [19,21,29]. The electromagnetic transmitter is located near the operative site and the receiver is inside the surgical instrument. The advantages of electromagnetic navigation systems are the use of very small detector coils, absence of visual contact between instrument and sensor system, rapid computation of the signals, and easy sterilisation [21,25].
However, due to interference by external magnetic fields and metal objects, particularly those associated with drilling and sawing tools [12,21,30,31], incorrect position sensing of up to 4 mm may occur . To reduce the incorrect position sensing special titanium or ceramic instrument set is required [19,20,32]. Electromagnetic navigation systems are relatively contraindicated for use of patients with pacemakers and cochlear implants [29].
Optical navigation systems
Optical based systems are used for intra-operative navigation [4,16,21,25,33,34]. Position calculation is provided by a minimum of three infrared diodes or passive light reflecting reference elements mounted to the registered patient using dynamic reference frame (DRF) and the surgical tool (tracker), and recognition of the obtained patterns with a stereotactic camera. The advantages of optical navigation systems are high technical accuracy in the range of 0.1-0.4 mm [35,36], convenient handling, and easy sterilisation. The disadvantages are the necessity of constant visual contact between camera-array, DRF and instruments, and the potential susceptibility to interference through light reflexes on metallic surfaces in the operating environment [21,25,37-39].
MEDICAL ROBOTICS
Robots are generally defined as computer controlled devices with five to six degrees of freedom that can execute complex movements with high accuracy [14,40]. Medical robots can be classified based on technology, application, or role [14].
Using a technology-based classification, two groups of systems that differ substantially from each other can be distinguished:
telemanipulators robots (not pre-programmed)
pre-programmed surgical robots (automated or semi-automated)
Application-based taxonomy distinguishes robots on the basis of surgical disciplines and operative procedures. Role-based taxonomy distinguishes robots into three discrete categories:
passive (the role of the robot is limited in scope or its involvement is largely low risk)
restricted (the robot is responsible for more invasive tasks with higher risk but still restricted from essential portions of the procedure)
active (the robot is intimately involved in the procedure and carries high responsibility and risk).
Telemanipulated robots
Telemanipulated robots are non-autonomously working robotic arms (manipulator) that are controlled remotely by the surgeon using force-feedback joysticks or more advanced haptic devices (master console) [18,41]. Compared to conventional endoscopic arms with limited mechanical control, telemanipulated robots provide a greater degree of freedom and have a computer controlled men-machine interface that allows for automatically processing of the input for the manipulator system without active interaction by the surgeon for motion scaling, tremor filtering, indexing, and so on. [18,41-44].
Pre-programmable surgical robots
Pre-programmable surgical robots can automatically or semi-automatically execute surgical tasks directly on the patient. These systems include:
floor or operating table mounted robots with six degrees of freedom
roof mounted modified surgical microscopes with generally six to seven active and one passive degree of freedom [45-49]
The surgeon in the operating theatre supervises the execution of the plan by the robot [7,50].
Interactive assistant robots are navigated tool support systems that carry, guide, and move surgical instruments. The robot is primarily moved passively by the surgeon but the robot can limit the degrees of freedom of the movements. Favourable positions can be saved and reached again with high precision. The surgeon has a spatial interval in which free movements are allowed, preventing movement into high-risk areas [21,40].
Mechatronic surgical tools
As a separate development in surgical instrumentation, mechatronic surgical tools are dedicated to special tasks such as drilling or bone shaving [5,51,52]. These tools may include force feedback sensors to prevent bone perforation or navigated controlled systems that only work within a certain surgical accuracy threshold.
IMAGE-TO-PATIENT TRANSFORMATION
Image-to-patient (IP) transformation or registration is the essential determination of a one-to-one mapping between the coordinates in the image data and those in the patient [53,54]. The registration procedure is based on anatomical landmarks (bone or skin), artificial markers (fiducials, bone affixed or skin applied), teeth supported registration templates, external registration frames, and laser surface scanning [12,25,55-60].
Anatomical landmarks
Registration with anatomical landmarks uses clearly defined external (such as nasion, spina nasalis, tragi, medial canthi, mastoid, umbo, and so on) and/or internal landmarks [61,62]. However, precise identification of the landmarks in both the patient and the image dataset is subjective and depends on the experience of the operator [63]. Surface matching, which is done by touching about 40-80 points on the patient’s skin or bone, can refine anatomical registration [62,64]. However, this method is generally inaccurate and time-consuming.
Fiducial markers
The advantage of fiducial markers over anatomical landmarks is the enhanced localisation accuracy on the image data and the patient. Consequently, registration with skin-applied fiducials is more accurate than registration with surface anatomical landmarks [65-67]. However, the use of skin-applied fiducials is associated with high logistics because the markers must be placed prior to data set acquisition and must be kept in their position until the patient enters the operating room. The time lag between imaging and surgery, and the sensitivity to skin shift can lead to unfavourable inaccuracies [25,56,63,68-70]. Bone-implanted fiducials provide invariant spatial registration points with the highest possible accuracy and generally serve as the reference gold standard in registration [21,53,66,68,71-73]. The drawbacks of bone-implanted fiducials are their invasiveness, the need for additional surgery, and possible major patient discomfort for which they should not be left in place for an extended period [55,63,70,71].
Registration templates
Registration templates are non-invasive, denture fixed acrylic splints with integrated fiducial markers [36,39,60,71,74-81]. Proven accuracy similar to bone implanted fiducials is available for the regions of the maxilla, mandible, orbit and face [36,72,81]. Registration templates cannot be applied to edentulous patient, except when the templates are invasively secured to the underlying bone.
Vogele-Bale-Hohner (VBH) mouthpiece / external registration frame
The Vogele-Bale-Hohner (VBH) vacuum mouthpiece is an individualised mouthpiece that can be objectively and rigidly secured against the maxilla with submillimetric repositioning control, that is regulated by the amount of negative pressure on the scale of a vacuum pump [56,82-84]. Alternatively, the VBH mouthpiece can be glued to an acrylic template, similar to registration templates. Compared to registration templates, where the markers are integrated in the template, an external registration frame is connected to the VBH mouthpiece. The VBH mouthpiece can be removed after registration [55,59,82,83,85-89]. The external registration frame allows for broad marker distribution around the entire head volume. Supported with exchangeable markers for CT/MRI/PET/SPECT, the external registration frame can serve as a single reference device for multimodal surgical navigation and fusion imaging [56,84,90-92].
Laser surface registration
Laser surface registration is based on projection of visible laser beams on the patient’s skin [67-70,93]. The skin reflections are detected by a camera array and a virtual three-dimensional matrix of the skin anatomy of the patient is generated. The matrix, which is an advanced surface-matching algorithm, is then matched to the surface matrix of the pre-operative image-data set.
Currently, up to 300,000 skin surface points can be registered. This allows the registration accuracy reach comparable values to bone markers or registration templates [67]. However, the shift of the patient’s skin surface or different tension in muscles of expression when performing CT-data acquisition and during preoperative and intraoperative recording, may lead to an invalid data set correlation [68,69,93]. Though the patient might to be continuously tracked during surgery, the original geometry of the facial soft tissue may be destroyed by intraoperative swelling, surgical cuts, or during repositioning osteotomies [21,33,69,94]. To compensate, a combination with dynamic reference frames must be available for intraoperative tracking after the initial laser registration has been reported [94]. Laser surface registration is unsuitable for surgery in the mandible but is expected to serve as a sufficiently stable and relatively invariable reference base for many applications in cranio-maxillofacial surgery [66,67,70, 93,94].
IMAGE-GUIDANCE
For image-guidance, the correlation between the space coordinates of the image-data in the navigation system and the patient’s coordinates defined during registration are preserved during the surgical procedure. The coordinates are obtained by rigid fixation of the patient on the operating table, for example invasively via the Mayfield head clamp, or non-invasively via the vacuum mouthpiece based VBH head holder [56,64,95]. Alternatively, bone (invasive) or registration template (non-invasive) affixed DRFs are used for continuous patient tracking after initial registration [33,36,60,74,76].
During surgery, the navigation software indicates the actual real-time position of the tracked surgical tool within the patient’s presurgical 3D-data for intraoperative orientation, and shows the calculated accuracy of the tool’s position and angulations related to the predefined surgical plan. Integrated mechatronic surgical tools provide automatic on/off-regulation depending on the current position of the patient in relation to the planned working space or the connection of the drill speed to the operator accuracy. Integrated mechatronic surgical tools are immediate stopped when possible damage to vital structures occurs (= navigated control) [5,51,52]. In addition, the development of adjustable rigid aiming devices enables a steady linear approach to defined targets [56,85,96-98].
Visualisation of the navigation process is generally provided via the computer screen of the navigation system’s transportable workstation. A disadvantage of such a display is that the surgeon has to look up at the screen and therefore, cannot simultaneously view the surgical field [4,13,39,51,71].
In contrast, augmented reality (AR) provides navigational support by direct projection of segmented structures from the preoperative image data (surgical targets, resection lines, and planned implant position) to the patient. Therefore allowing complete interaction with the real world, while simultaneously making the virtual environment accessible [30,58,71,99,100-104,111]. AR can be based on monocular projection in the operating microscope or the binocular optics of a tracked surgical microscope projection for the purpose of building semitranslucent screens placed between the operating screen and the surgeon or the head mounted displays [4,22,30,58,71,99,100,105-110]. Recently, a promising AR concept using laser registration and stereotactic optical projection of tumour margins and osteotomy lines directly on the patient was presented. This concept does not necessitate navigation instruments [104,111,112].
ASPECTS OF ACCURACY
Terminology
Accuracy is of utmost importance for clinical application of image-guided surgery and medical robotics. Use of standardised terminology and measurement types is essential for correct understanding and comparability of accuracy reports [113].
Accuracy is qualitatively determined as the amount of approximation of the mean of the measurements to the true value (which refers to the term trueness) and quantitatively determined through the margin of error and the uncertainty of measurement, which is characterised by the variation of the mean value from several single measurements.
Precision is the inner accuracy of measurements obtained by repeated measurements (under the same circumstances and with the same measurement technique and system) and refers to the quantitative characterisation of the concision of the measuring instrument and its readout. Although often used as a synonym for accuracy, precision must be clearly distinguished from the term accuracy.
For evaluation of image-guided surgery, the suggested measurement types are as follows: [66,72,113-115]:
Fiducial Localising Error (FLE): the error in locating the fiducial points.
Fiducial Registration Error (FRE) : the error between corresponding fiducial points after registration
Target Registration Error (TRE) : the error between corresponding points other than the fiducial points after registration
Target Positioning Error (TPE): the error between the real position of the navigated surgical tool and the calculated position during the actual surgical procedure (TRE plus additional factors).
The best indicator for a navigation system’s or medical robot’s accuracy is represented by the TPE, but the definitive overall accuracy of the surgical procedure has to be ultimately evaluated by directly comparing the achieved surgical result to the initial planning data.
Influential factors of accuracy
The overall accuracy of image-guided / robotic surgery depends on all systematic and non-systematic (random) errors, from the data-set acquisition to the surgical procedure [116]. The accurate linking of the virtual planning to the surgical site depends on the accuracy of the registration procedure, which includes limitations in the image space and the device space (see chapter image-to-patient transformation). Image quality depends on the image resolution as represented by the voxel size and slice thickness. The thinner the slice thickness and the smaller the voxel size, the higher is the accuracy of determining the centre of the fiducial markers (fiducial-based registration) or the accuracy of the calculated 3D surface model (surface based registration) [88,119,120]. In principle, multi detector CT is more accurate than MRI, because MRI is prone to inhomogeneities of the magnetic field and, due to the longer examination time, more susceptible to motion artefacts [64,117-119]. The arrangement of fiducial markers is a critical factor and it is important to use as many points as possible (although the return diminishes rapidly after five or six markers are used), avoid near-collinear configurations, and ensure that the centroid of the fiducial points is as near as possible to the target [12,54]. The typical feedback provided by the registration software is a measure of the degree of alignment of the points used in the registration. Unfortunately these measures show no direct correlation to the TRE and to reliably control the registration accuracy intraoperatively, the real error between the image and the patient’s anatomy has to be checked prior to surgery by a few independent markers not used for initial registration and/or by anatomic landmarks [10,12,36,39,77,93]. This can be performed with the probe of the navigation system by comparing the probe’s real position (device space) to the virtual position displayed on the computer screen (image space). The accuracy of the surgical transfer is dependant on the technical accuracy of the navigation system, mechatronic, semi-active, or active robotic system and the surgical application accuracy. Notably, human error is attributed to imaging, registration, and transfer errors, for which every step has to be carefully managed.
CLINICAL APPLICATIONS
Image-guided surgery
Successful clinical applications of image-guided surgery in the cranial area have been already described for many procedures, such as the following (neurosurgical procedures excluded): oral implant surgery [10,16,37,38,52,73,77,79,103,121], removal of tumours and foreign bodies [16,33,58,76,81,122], bone segment navigation [60,122,123], temporo-mandibular-joint surgery [74,124], biopsy [16], frameless stereotactic interstitial brachytherapy [28,87], percutaneous radio frequency ablation of the Gasserion ganglion in medically untreatable trigeminal neuralgia [88,95,125], functional endoscopic sinus surgery and skull base surgery [5,9,12,22,107,126-128]. Use of mechatronic surgical tools has been tested for navigate-controlled drilling in oral implant surgery [52] and shaving in functional endoscopic sinus surgery [13,51].
Medical robotics
In the cranial area, robotic systems were considered to help the surgeon interactively with the following tasks [1,7,21,40,45,129]: (1) the drilling of holes with an automatic stop after penetrating the bone to protect the tissue lying deep to the bone, (2) the defined drilling of the implant bed for positioning of implants or bone fixtures for anaplastology, (3) the milling of the bone surfaces in plastic surgery according to a 3D-operation plan, (4) performing deep saw-cuts for osteotomies and allowing for the precise three-dimensional transportation of the subsequent bone segments or CAD/CAM (computer aided design / computer aided manufacturing) transplant, (5) the preoperative automatic selection of the necessary osteosynthesis plates, their bending by a special machine and their intraoperative positioning in defined positions, or (6) the automated guidance for non-flexible catheter implantation at brachytherapy.
Pre-clinical and clinical studies have been started around the millennium in Germany, France, USA and Japan for robot-assisted placement of craniofacial implants in ear anaplastology [130], resection of frontotemporal bone segments [131], implant fabrication combined with CAD/CAM technology in reconstructive surgery [21,79,131], model surgery in orthognatic surgery [26], passive guidance for the positioning of oral implants [133-135], and videoendoscopic ENT and skull base surgery [18,47-49,132].
Cost-benefit ratio
Image-guided surgery is considered to be more accurate than standard surgery. Comparative studies in oral implant surgery indicate significantly more accuracy compared to the manual freehand procedure even if performed by experienced surgeons [79,136,137]. In addition, no significant difference between experienced surgeons and trainees was found, which demonstrates that image-guidance is a valuable means for achieving a predictable and reproducible result without heavy reliance on the clinician’s surgical experience [10,79,136,138]. In other procedures, such as percutaneous interventions (which are generally a “blind” surgical procedure), removal of foreign bodies, access to deep seated locations, orientation in complex and changed anatomic regions, etc., clear benefit of image-guidance is evident [4,12,16,33,128,143]. Generally, shorter operation time, safer manipulation around delicate structures and higher intraoperative accuracy have been reported [9,16,20,60,128,139,140]. Further, image-guidance may allow for more thorough surgical resection and potentially decreasing the need for revision procedures [140].
In a large clinical study for image-guided ENT surgery, it was found that image-guidance can provide additional relevant information that was not available to the surgeon solely by virtue of his existing knowledge and that every second application of the navigation system may lead to a change in surgical strategy [5]. Accordingly more benefit is obtained from additional orientation and resulting cognitive relief at the moment of stressed and distracted surgical situations. Another clinical study including 158 surgical procedures in cranio-maxillo-facial surgery showed high to very high medical benefits for image-guided biopsies, punctures of the trigeminal ganglion, removal of foreign bodies, osteotomies of the facial skeleton, arthroscopies of the temporomandibular joint and positioning of dental implants [16].
Image-guided surgery is more expensive than the standard procedure (navigation systems cost about USD 60,000 to USD 200,000) and requires presurgical imaging with registration elements, intraoperative image-to-patient registration and specialised equipment for tool tracking. However, these systems can be used for a wide range surgical procedures in different medical specialities [56,57,59,83-85,89,144] and thus may represent a valuable acquisition for an institution [16,33]. A further beneficial aspect is the associated automatic and complete electronic documentation of the intervention [16,116].
Robots are expected to be more accurate and more reliable than a human being. Robots can work as part of an interactive system, are immune to radiation and can be automatically programmed for documentation, evaluation and training protocols [14,40,45,46,129]. Except for very few cases, surgical robots will not execute operations fully autonomously but will support the physician to achieve optimal results [1,7,21,40,44, 45,129,141].
Considering the advantages mentioned above, image-guided surgery and medical robotics may have a positive cost/effort–benefit ratio, depending on the individual surgical task and the developmental stage of each system. The necessity of special knowledge for this technology is indisputable and the relationship between cost and benefit may additionally be dependent on familiarity and availability [15,113].
CONCLUSION
Due to the complex anatomic situations with high-risk structures and the high demands for functional and aesthetic results, surgery in the cranial area is a prototype for application of image-guided surgery and medical robotics. Successful clinical use has been already described for many different procedures and clear benefit is proved in terms of intraoperative orientation, surgical accuracy, safety and reduced operation time. The development of mechatronic surgical tools may additionally improve safety and surgical accuracy. For appropriate clinical application of image-guided surgery, it is important that the surgeon is aware of all influential factors of accuracy and the maximum error of each system / technique regarding the required surgical accuracy for the individual operation.
In the future, surgical navigation with integration of intraoperative imaging, improved augmented reality techniques, sophisticated mechatronic surgical tools and new robotic developments which are smaller, less expensive and easier to operate will enable continued progress in surgical instrumentation, and ultimately, surgical care.
REFERENCES
- 1.Hassfeld S, Brief J, Krempien R, et al. [Computer-assisted oral, maxillary and facial surgery] Radiologe. 2000;40(3):218–26. doi: 10.1007/s001170050660. [DOI] [PubMed] [Google Scholar]
- 2.Xia J, Samman N, Yeung RW, et al. Computer-assisted three-dimensional surgical planing and simulation. 3D soft tissue planning and prediction. Int J Oral Maxillofac Surg. 2000;29(4):250–8. [PubMed] [Google Scholar]
- 3.Troulis MJ, Everett P, Seldin EB, et al. Development of a three-dimensional treatment planning system based on computed tomographic data. Int J Oral Maxillofac Surg. 2002;31(4):349–57. doi: 10.1054/ijom.2002.0278. [DOI] [PubMed] [Google Scholar]
- 4.Nijmeh AD, Goodger NM, Hawkes D, et al. Image-guided navigation in oral and maxillofacial surgery. Br J Oral Maxillofac Surg. 2005;43(4):294–302. doi: 10.1016/j.bjoms.2004.11.018. [DOI] [PubMed] [Google Scholar]
- 5.Strauss G, Koulechov K, Rottger S, et al. Evaluation of a navigation system for ENT with surgical efficiency criteria. Laryngoscope. 2006;116(4):564–72. doi: 10.1097/01.MLG.0000202091.34295.05. [DOI] [PubMed] [Google Scholar]
- 6.Hassfeld S, Muhling J, Zoller J. Intraoperative navigation in oral and maxillofacial surgery. Int J Oral Maxillofac Surg. 1995;24(1 Pt 2):111–9. doi: 10.1016/s0901-5027(05)80871-9. [DOI] [PubMed] [Google Scholar]
- 7.Lueth T, Heissler E, Bier J. Schlag PM. Tele- und computergestützte Chirurgie. Springer Verlag; 1998. Evaluierung von navigations- und Robotersystemen für den Einsatz in der Chirurgie. Navigations- und Robotersysteme als Führungshilfen. [Google Scholar]
- 8.Gaggl A, Schultes G, Karcher H. Navigational precision of drilling tools preventing damage to the mandibular canal. J Craniomaxillofac Surg. 2001;29(5):271–5. doi: 10.1054/jcms.2001.0239. [DOI] [PubMed] [Google Scholar]
- 9.Caversaccio M, Bachler R, Ladrach K, et al. Frameless computer-aided surgery system for revision endoscopic sinus surgery. Otolaryngol Head Neck Surg. 2000;122(6):808–13. doi: 10.1016/S0194-59980070006-4. [DOI] [PubMed] [Google Scholar]
- 10.Siessegger M, Schneider BT, Mischkowski RA, et al. Use of an image-guided navigation system in dental implant surgery in anatomically complex operation sites. J Craniomaxillofac Surg. 2001;29(5):276–81. doi: 10.1054/jcms.2001.0242. [DOI] [PubMed] [Google Scholar]
- 11.Xia J, Ip HHS, Samman N, et al. Three-dimensional virtual-reality surgcial planning and soft tissue prediction for orthognatic surgery. IEEE Trans Inform Tech Biomed. 2001;5(2):97–107. doi: 10.1109/4233.924800. [DOI] [PubMed] [Google Scholar]
- 12.Wise SK, DelGaudio JM. Computer-aided surgery of the paranasal sinuses and skull base. Expert Rev Med Devices. 2005;2(4):395–408. doi: 10.1586/17434440.2.4.395. [DOI] [PubMed] [Google Scholar]
- 13.Strauss G, Koulechov K, Stopp S, et al. [Improved accuracy and precision of the automated shaver (navigated control) in functional endoscopic sinus surgery] Laryngorhinootologie. 2006;85(8):559–66. doi: 10.1055/s-2006-925194. [DOI] [PubMed] [Google Scholar]
- 14.Camarillo DB, Krummel TM, Salisbury JKJ. Robotic technology in surgery: past, present, and future. Am J Surg. 2004;188(4A Suppl):2–15. doi: 10.1016/j.amjsurg.2004.08.025. [DOI] [PubMed] [Google Scholar]
- 15.Vannier MW, Haller JW. Navigation in diagnosis and therapy. Eur J Radiol. 1999;31(2):132–40. doi: 10.1016/s0720-048x(99)00091-1. [DOI] [PubMed] [Google Scholar]
- 16.Ewers R, Schicho K, Undt G, et al. Basic research and 12 years of clinical experience in computer-assisted navigation technology: a review. Int J Oral Maxillofac Surg. 2005;34(1):1–8. doi: 10.1016/j.ijom.2004.03.018. [DOI] [PubMed] [Google Scholar]
- 17.Marescaux J, Solerc L. Image-guided robotic surgery. Semin Laparosc Surg. 2004;11(2):113–22. [PubMed] [Google Scholar]
- 18.Strauss G, Winkler D, Jacobs S, et al. Mechatronic in functional endoscopic sinus surgery. First experiences with the daVinci Telemanipulatory System. HNO. 2005;53(7):623–30. doi: 10.1007/s00106-005-1242-1. [DOI] [PubMed] [Google Scholar]
- 19.Marmulla R, Hilbert M, Niedert-Ellmann H. Introperative precision of mechanical, electromechanical, infrared and laser-guided navigation systems in computer-assisted surgery. Mund Kiefer Gesichtschir. 1998;2(Suppl 1):145–8. doi: 10.1007/PL00014462. [DOI] [PubMed] [Google Scholar]
- 20.Gunkel AR, Freysinger W, Thumfart WF. Experience with various 3-dimensional navigation systems in head and neck surgery. Arch Otolaryngol Head Neck Surg. 2000;126(3):390–5. doi: 10.1001/archotol.126.3.390. [DOI] [PubMed] [Google Scholar]
- 21.Hassfeld S, Muhling J. Computer assisted oral and maxillofacial surgery--a review and an assessment of technology. Int J Oral Maxillofac Surg. 2001;30(1):2–13. doi: 10.1054/ijom.2000.0024. [DOI] [PubMed] [Google Scholar]
- 22.Caversaccio, Freysinger Computer assistance for intraoperative navigation in ENT surgery. Minim Invasive Ther Allied Technol. 2003;12(1):36–51. doi: 10.1080/13645700310001577. [DOI] [PubMed] [Google Scholar]
- 23.Freysinger W, Gunkel AR, Martin A, et al. Advancing ear, nose, and throat computer-assisted surgery with the arm-based ISG viewing wand: the stereotactic suction tube. Laryngoscope. 1997;107(5):690–3. doi: 10.1097/00005537-199705000-00025. [DOI] [PubMed] [Google Scholar]
- 24.Freysinger W, Gunkel AR, Bale R, et al. Three-dimensional navigation in otorhinolaryngological surgery with the viewing wand. Ann Otol Rhinol Laryngol. 1998;107(11 Pt 1):953–8. doi: 10.1177/000348949810701109. [DOI] [PubMed] [Google Scholar]
- 25.Hassfeld S, Muhling J. Comparative examination of the accuracy of a mechanical and an optical system in CT and MRT based instrument navigation. Int J Oral Maxillofac Surg. 2000;29(6):400–7. [PubMed] [Google Scholar]
- 26.Theodossy T, Bamber MA. Model surgery with a passive robot arm for orthognatic surgery planning. Am Ass Oral Maxillofac Surg. 2003;61:1210–317. doi: 10.1016/s0278-2391(03)00733-x. [DOI] [PubMed] [Google Scholar]
- 27.Hassfeld S, Muhling J. Navigation in maxillofacial and craniofacial surgery. Comput Aided Surg. 1998;3(4):183–7. doi: 10.1002/(SICI)1097-0150(1998)3:4<183::AID-IGS8>3.0.CO;2-F. [DOI] [PubMed] [Google Scholar]
- 28.Kremer B, Klimek L, Andreopoulos D, et al. A new method for the placement of brachytherapy probes in paranasal sinus and nasopharynx neoplasms. Int J Radiat Oncol Biol Phys. 1999;43(5):995–1000. doi: 10.1016/s0360-3016(98)00521-5. [DOI] [PubMed] [Google Scholar]
- 29.Koele W, Stammberger H, Lackner A, et al. Image guided surgery of paranasal sinuses and anterior skull base--five years experience with the InstaTrak-System. Rhinology. 2002;40(1):1–9. [PubMed] [Google Scholar]
- 30.Wagner A, Ploder O, Enislidis G, et al. Image-guided surgery. Int J Oral Maxillofac Surg. 1996;25(2):147–51. doi: 10.1016/s0901-5027(96)80062-2. [DOI] [PubMed] [Google Scholar]
- 31.Wagner A, Schicho K, Birkfellner W, et al. Quantitative analysis of factors affecting intraoperative precision and stability of optoelectronic and electromagnetic tracking systems. Med Phys. 2002;29(5):905–12. doi: 10.1118/1.1469625. [DOI] [PubMed] [Google Scholar]
- 32.Birkfellner W, Watzinger F, Wanschitz F, et al. Systematic distortions in magnetic position digitizers. Med Phys. 1998;25(11):2242–8. doi: 10.1118/1.598425. [DOI] [PubMed] [Google Scholar]
- 33.Heiland M, Habermann CR, Schmelzle R. Indications and limitations of intraoperative navigation in maxillofacial surgery. J Oral Maxillofac Surg. 2004;62(9):1059–63. doi: 10.1016/j.joms.2004.02.013. [DOI] [PubMed] [Google Scholar]
- 34.Henderson JM, Holloway KL, Gaede SE, et al. The application accuracy of a skull-mounted trajectory guide system for image-guided functional neurosurgery. Comput Aided Surg. 2004;9(4):155–60. doi: 10.3109/10929080500050249. [DOI] [PubMed] [Google Scholar]
- 35.Khadem R, Yeh CC, Sadeghi-Tehrani M, et al. Comparative tracking error analysis of five different optical tracking systems. Comput Aided Surg. 2000;5(2):98–107. doi: 10.1002/1097-0150(2000)5:2<98::AID-IGS4>3.0.CO;2-H. [DOI] [PubMed] [Google Scholar]
- 36.Casap N, Wexler A, Persky N, et al. Navigation surgery for dental implants: assessment of accuracy of the Image Guided Implantology System. J Oralmaxillofac Surg. 2004;62(Suppl 2):116–9. doi: 10.1016/j.joms.2004.06.028. [DOI] [PubMed] [Google Scholar]
- 37.Watzinger F, Birkfellner W, Wanschitz F, et al. Positioning of dental implants using computer-aided navigation and an optical tracking system: case report and presentation of a new method. J Craniomaxillofac Surg. 1999;27(2):77–81. doi: 10.1016/s1010-5182(99)80017-1. [DOI] [PubMed] [Google Scholar]
- 38.Wagner A, Wanschitz F, Birkfellner W, et al. Computer-aided placement of endosseous oral implants in patients after ablative tumour surgery: assessment of accuracy. Clin Oral Implants Res. 2003;14(3):340–8. doi: 10.1034/j.1600-0501.2003.110812.x. [DOI] [PubMed] [Google Scholar]
- 39.Meyer U, Wiesmann HP, Runte C, et al. Evaluation of accuracy of insertion of dental implants and prosthetic treatment by computer-aided navigation in minipigs. Br J Oral Maxillofac Surg. 2003;41(2):102–8. doi: 10.1016/s0266-4356(02)00297-8. [DOI] [PubMed] [Google Scholar]
- 40.Hassfeld S, Raczkowsky J, Bohner P, et al. Robotik in der Mund-, Kiefer- und Gesichtschirurgie. Möglichkeiten - Chancen - Risiken. Mund Kiefer GesichtsChir. 1997;1:316–23. doi: 10.1007/BF03043575. [DOI] [PubMed] [Google Scholar]
- 41.Rockall TA, Darzi AW. Tele-manipulator robots in surgery. Br J Surg. 2003;90(6):641–3. doi: 10.1002/bjs.4166. [DOI] [PubMed] [Google Scholar]
- 42.Darzi A, Mackay S. Recent advances in minimal access surgery. BMJ. 2002;324(7328):31–4. doi: 10.1136/bmj.324.7328.31. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Das H, Ohm T, Boswell C, et al. Dexterity-enhanced telerobotic microsurgery; Proceedings of the 8th International Conference on Advanced Robotics. IEEE; 1997. pp. 5–10. [Google Scholar]
- 44.Cleary K, Nguyen C. State of the art in surgical robotics: clinical applications and technology challenges. Comput Aided Surg. 2001;6(6):312–28. doi: 10.1002/igs.10019. [DOI] [PubMed] [Google Scholar]
- 45.Lueth T, Bier J. Robot assisted intervention in surgery. In: Gilsbach JM, Stiehl HS, editors. Neuronavigation - neurosurgical and computer scientific aspects. Springer Verlag; 1999. [Google Scholar]
- 46.Bier J. Robotik. Mund Kiefer GesichtsChir. 2000;4(Suppl 1):356–68. doi: 10.1007/PL00014559. [DOI] [PubMed] [Google Scholar]
- 47.Steinhart H, Bumm K, Wurm J, et al. Surgical application of a new robotic system for paranasal sinus surgery. Ann Otol Rhinol Laryngol. 2004;113(4):303–9. doi: 10.1177/000348940411300409. [DOI] [PubMed] [Google Scholar]
- 48.Bumm K, Wurm J, Rachinger J, et al. An automated robotic approach with redundant navigation for minimal invasive extended transsphenoidal skull base surgery. Minim Invasive Neurosurg. 2005;48(3):159–64. doi: 10.1055/s-2005-870903. [DOI] [PubMed] [Google Scholar]
- 49.Wurm J, Bumm K, Steinhart H, et al. [Development of an active robot system for multi-modal paranasal sinus surgery] HNO. 2005;53(5):446–54. doi: 10.1007/s00106-004-1155-4. [DOI] [PubMed] [Google Scholar]
- 50.Woern H, Muehling J. Computer- and robot-based operation theater of the future in cranio-facial surgery. In: Lemke HU, Vannier MV, Inamura K, editors. International Congress Series. Vol. 1230. 2001. pp. 753–9. [Google Scholar]
- 51.Strauss G, Koulechov K, Richter R, et al. [Navigated control: a new concept in computer assisted ENT-surgery] Laryngorhinootologie. 2005;84(8):567–76. doi: 10.1055/s-2005-861451. [DOI] [PubMed] [Google Scholar]
- 52.Brief J, Edinger D, Hassfeld S, et al. Accuracy of image-guided implantology. Clin Oral Implants Res. 2005;16(4):495–501. doi: 10.1111/j.1600-0501.2005.01133.x. [DOI] [PubMed] [Google Scholar]
- 53.Maurer CR, Jr, Fitzpatrick JM, Wang MY, et al. Registration of head volume images using implantable fiducial markers. IEEE Trans Med Imaging. 1997;16(4):447–62. doi: 10.1109/42.611354. [DOI] [PubMed] [Google Scholar]
- 54.West JB, Fitzpatrick JM, Toms SA, et al. Fiducial point placement and the accuracy of point-based, rigid body registration. Neurosurgery. 2001;48(4):810–6. doi: 10.1097/00006123-200104000-00023. discussion 816-7. [DOI] [PubMed] [Google Scholar]
- 55.Alp MS, Dujovny M, Misra M, et al. Head registration techniques for image-guided surgery. Neurol Res. 1998;20(1):31–7. doi: 10.1080/01616412.1998.11740481. [DOI] [PubMed] [Google Scholar]
- 56.Bale RJ, Burtscher J, Eisner W, et al. Computer-assisted neurosurgery by using a noninvasive vacuum-affixed dental cast that acts as a reference base: another step toward a unified approach in the treatment of brain tumors. J Neurosurg. 2000;93(2):208–13. doi: 10.3171/jns.2000.93.2.0208. [DOI] [PubMed] [Google Scholar]
- 57.Bale RJ, Hoser C, Rosenberger R, et al. Osteochondral lesions of the talus: computer-assisted retrograde drilling--feasibility and accuracy in initial experiences. Radiology. 2001;218(1):278–82. doi: 10.1148/radiology.218.1.r01ja18278. [DOI] [PubMed] [Google Scholar]
- 58.Schultes G, Gaggl A. [Ct-assisted navigation for insertion of dental implants in maxilla models] Schweiz Monatsschr Zahnmed. 2001;111(7):828–33. [PubMed] [Google Scholar]
- 59.Widmann G, Widmann R, Widmann E, et al. In vitro accuracy of a novel registration and targeting technique for image-guided template production. Clin Oral Implants Res. 2005;16(4):502–8. doi: 10.1111/j.1600-0501.2005.01131.x. [DOI] [PubMed] [Google Scholar]
- 60.Klug C, Schicho K, Ploder O, et al. Point-to-point computer-assisted navigation for precise transfer of planned zygoma osteotomies from the stereolithographic model into reality. J Oral Maxillofac Surg. 2006;64(3):550–9. doi: 10.1016/j.joms.2005.11.024. [DOI] [PubMed] [Google Scholar]
- 61.Anon JB, Lipman SP, Oppenheim D, et al. Computer-assisted endoscopic sinus surgery. Laryngoscope. 1994;104(7):901–5. doi: 10.1288/00005537-199407000-00023. [DOI] [PubMed] [Google Scholar]
- 62.Caversaccio M, Zulliger D, Bachler R, et al. Practical aspects for optimal registration (matching) on the lateral skull base with an optical frameless computer-aided pointer system. Am J Otol. 2000;21(6):863–70. [PubMed] [Google Scholar]
- 63.Wolfsberger S, Rossler K, Regatschnig R, et al. Anatomical landmarks for image registration in frameless stereotactic neuronavigation. Neurosurg Rev. 2002;25(1-2):68–72. doi: 10.1007/s10143-001-0201-x. [DOI] [PubMed] [Google Scholar]
- 64.Germano IM, Villalobos H, Silvers A, et al. Clinical use of the optical digitizer for intracranial neuronavigation. Neurosurgery. 1999;45(2):261–9. doi: 10.1097/00006123-199908000-00013. discussion 269-70. [DOI] [PubMed] [Google Scholar]
- 65.Helm PA, Eckel TS. Accuracy of registration methods in frameless stereotaxis. Comput Aided Surg. 1998;3(2):51–6. doi: 10.1002/(SICI)1097-0150(1998)3:2<51::AID-IGS1>3.0.CO;2-J. [DOI] [PubMed] [Google Scholar]
- 66.Maurer CR, Jr, Maciunas RJ, Fitzpatrick JM. Registration of head CT images to physical space using a weighted combination of points and surfaces. IEEE Trans Med Imaging. 1998;17(5):753–61. doi: 10.1109/42.736031. [DOI] [PubMed] [Google Scholar]
- 67.Marmulla R, Muhling J, Luth T, et al. Advanced surface-recording techniques for computer-assisted oral and maxillofacial surgery. Br J Oral Maxillofac Surg. 2004;42(6):511–9. doi: 10.1016/j.bjoms.2004.06.021. [DOI] [PubMed] [Google Scholar]
- 68.Marmulla R, Hassfeld S, Luth T, et al. Laser-scan-based navigation in cranio-maxillofacial surgery. J Craniomaxillofac Surg. 2003;31(5):267–77. doi: 10.1016/s1010-5182(03)00056-8. [DOI] [PubMed] [Google Scholar]
- 69.Marmulla R, Luth T, Muhling J, et al. Automated laser registration in image-guided surgery: evaluation of the correlation between laser scan resolution and navigation accuracy. Int J Oral Maxillofac Surg. 2004;33(7):642–8. doi: 10.1016/j.ijom.2004.01.005. [DOI] [PubMed] [Google Scholar]
- 70.Marmulla R, Muhling J, Wirtz CR, et al. High-resolution laser surface scanning for patient registration in cranial computer-assisted surgery. Minim Invasive Neurosurg. 2004;47(2):72–8. doi: 10.1055/s-2004-818471. [DOI] [PubMed] [Google Scholar]
- 71.Edwards PJ, King AP, Maurer CR, Jr, et al. Design and evaluation of a system for microscope-assisted guided interventions (MAGI) IEEE Trans Med Imaging. 2000;19(11):1082–93. doi: 10.1109/42.896784. [DOI] [PubMed] [Google Scholar]
- 72.Birkfellner W, Solar P, Gahleitner A, et al. In-vitro assessment of a registration protocol for image guided implant dentistry. Clin Oral Implants Res. 2001;12(1):69–78. doi: 10.1034/j.1600-0501.2001.012001069.x. [DOI] [PubMed] [Google Scholar]
- 73.Ewers R, Schicho K, Truppe M, et al. Computer-aided navigation in dental implantology: 7 years of clinical experience. J Oral Maxillofac Surg. 2004;62(3):329–34. doi: 10.1016/j.joms.2003.08.017. [DOI] [PubMed] [Google Scholar]
- 74.Schmelzeisen R, Gellrich NC, Schramm A, et al. Navigation-guided resection of temporomandibular joint ankylosis promotes safety in skull base surgery. J Oral Maxillofac Surg. 2002;60(11):1275–83. doi: 10.1053/joms.2002.35724. [DOI] [PubMed] [Google Scholar]
- 75.Kniha F, Gahlert M, Lassen T, et al. Anwendung der CT-unterstützten Navigation in der dentalen Implantation unter besonderer Berücksichtigung der Sofortbelastung. Quintessenz Zahntech. 2003;29(7):842–63. [Google Scholar]
- 76.Schultes G, Zimmermann V, Feichtinger M, et al. Removal of osteosynthesis material by minimally invasive surgery based on 3-dimensional computed tomography-guided navigation. J Oral Maxillofac Surg. 2003;61(3):401–5. doi: 10.1053/joms.2003.50067. [DOI] [PubMed] [Google Scholar]
- 77.Neugebauer J, Karapetian VE, Schuler M, et al. Fabrication of surgical template for CT-based implant planning. Int Poster J Dent Oral Med. 2004;6(4) Poster 248. [Google Scholar]
- 78.Blanchet E, Lucchini JP, Jenny R, et al. An image-guided system based on custom templates: case reports. Clin Implant Dent Relat Res. 2004;6(1):40–7. doi: 10.1111/j.1708-8208.2004.tb00026.x. [DOI] [PubMed] [Google Scholar]
- 79.Kramer FJ, Baethge C, Swennen G, et al. Navigated vs. conventional implant insertion for maxillary single tooth replacement. Clin Oral Impl Res. 2005;16:60–8. doi: 10.1111/j.1600-0501.2004.01058.x. [DOI] [PubMed] [Google Scholar]
- 80.Eggers G, Haag C, Hassfeld S. Image-guided removal of foreign bodies. Br J Oral Maxillofac Surg. 2005;43(5):404–9. doi: 10.1016/j.bjoms.2005.01.016. [DOI] [PubMed] [Google Scholar]
- 81.Eggers G, Muhling J, Marmulla R. Template-based registration for image-guided maxillofacial surgery. J Oral Maxillofac Surg. 2005;63(9):1330–6. doi: 10.1016/j.joms.2005.05.312. [DOI] [PubMed] [Google Scholar]
- 82.Martin A, Bale RJ, Vogele M, et al. Vogele-Bale-Hohner mouthpiece: registration device for frameless stereotactic surgery. Radiology. 1998;208(1):261–5. doi: 10.1148/radiology.208.1.9646822. [DOI] [PubMed] [Google Scholar]
- 83.Sweeney RA, Bale R, Auberger T, et al. A simple and non-invasive vacuum mouthpiece-based head fixation system for high precision radiotherapy. Strahlenther Onkol. 2001;177(1):43–7. doi: 10.1007/pl00002357. [DOI] [PubMed] [Google Scholar]
- 84.Sweeney RA, Bale RJ, Moncayo R, et al. Multimodality cranial image fusion using external markers applied via a vacuum mouthpiece and a case report. Strahlenther Onkol. 2003;179(4):254–60. doi: 10.1007/s00066-003-1031-2. [DOI] [PubMed] [Google Scholar]
- 85.Bale RJ, Vogele M, Freysinger W, et al. Minimally invasive head holder to improve the performance of frameless stereotactic surgery. Laryngoscope. 1997;107(3):373–7. doi: 10.1097/00005537-199703000-00018. [DOI] [PubMed] [Google Scholar]
- 86.Bale RJ, Vogele M, Martin A, et al. VBH head holder to improve frameless stereotactic brachytherapy of cranial tumors. Comput Aided Surg. 1997;2(5):286–91. doi: 10.1002/(SICI)1097-0150(1997)2:5<286::AID-IGS4>3.0.CO;2-W. [DOI] [PubMed] [Google Scholar]
- 87.Bale RJ, Freysinger W, Gunkel AR, et al. Head and neck tumors: fractionated frameless stereotactic interstitial brachytherapy-initial experience. Radiology. 2000;214(2):591–5. doi: 10.1148/radiology.214.2.r00fe33591. [DOI] [PubMed] [Google Scholar]
- 88.Bale RJ, Laimer I, Martin A, et al. Frameless stereotactic cannulation of the foramen ovale for ablative treatment of trigeminal neuralgia. Neurosurgery. 2006;59(4) Suppl 2:ONS394–401. doi: 10.1227/01.NEU.0000232770.97616.D0. discussion ONS402. [DOI] [PubMed] [Google Scholar]
- 89.Burtscher J, Sweeney R, Bale RJ, et al. Neuroendoscopy based on computer assisted adjustment of the endoscope holder in the laboratory. Minim Invas Neurosurg. 2003;46:208–14. doi: 10.1055/s-2003-42348. [DOI] [PubMed] [Google Scholar]
- 90.Kainz H, Bale R, Donnemiller E, et al. Image fusion analysis of 99m Tc-HYNIC-octreotide scintigraphy and CT/MRI in patients with thyroid-associated orbitopathy: the importance of the lacrimal gland. Eur J Nucl Med Mol Imaging. 2003;30(8):1155–9. doi: 10.1007/s00259-003-1207-0. [DOI] [PubMed] [Google Scholar]
- 91.Profanter C, Prommegger R, Gabriel M, et al. Computed axial tomography-MIBI image fusion for preoperative localization in primary hyperparathyroidism. Am J Surg. 2004;187(3):383–7. doi: 10.1016/j.amjsurg.2003.12.012. [DOI] [PubMed] [Google Scholar]
- 92.Profanter C, Wetscher GJ, Gabriel M, et al. CT-MIBI image fusion: a new preoperative localization technique for primary, recurrent, and persistent hyperparathyroidism. Surgery. 2004;135(2):157–62. doi: 10.1016/s0039-6060(03)00396-9. [DOI] [PubMed] [Google Scholar]
- 93.Raabe A, Krishnan R, Wolff R, et al. Laser surface scanning for patient registration in intracranial image-guided surgery. Neurosurgery. 2002;50(4):797–801. doi: 10.1097/00006123-200204000-00021. discussion 802-3. [DOI] [PubMed] [Google Scholar]
- 94.Marmulla R, Eggers G, Muhling J. Laser surface registration for lateral skull base surgery. Minim Invasive Neurosurg. 2005;48(3):181–5. doi: 10.1055/s-2005-870906. [DOI] [PubMed] [Google Scholar]
- 95.Hajioff D, Dorward NL, Wadley JP, et al. Precise cannulation of the foramen ovale in trigeminal neuralgia complicating osteogenesis imperfecta with basilar invagination: technical case report. Neurosurgery. 2000;46(4):1005–8. [PubMed] [Google Scholar]
- 96.Dorward NL, Alberti O, Dijkstra A, et al. Clinical introduction of an adjustable rigid instrument holder for frameless stereotactic interventions. Comput Aided Surg. 1997;2(3-4):180–5. [PubMed] [Google Scholar]
- 97.Patel N, Sandeman D. A simple trajectory guidance device that assists freehand and interactive image guided biopsy of small deep intracranial targets. Comput Aided Surg. 1997;2(3-4):186–92. [PubMed] [Google Scholar]
- 98.Germano IM, Queenan JV. Clinical experience with intracranial brain needle biopsy using frameless surgical navigation. Comput Aided Surg. 1998;3(1):33–9. doi: 10.1002/(SICI)1097-0150(1998)3:1<33::AID-IGS5>3.0.CO;2-M. [DOI] [PubMed] [Google Scholar]
- 99.Wagner A, Rasse M, Millesi W, et al. Virtual reality for orthognathic surgery: the augmented reality environment concept. J Oral Maxillofac Surg. 1997;55(5):456–62. doi: 10.1016/s0278-2391(97)90689-3. discussion 462-3. [DOI] [PubMed] [Google Scholar]
- 100.Wagner A, Kremser J, Watzinger F, et al. [Telenavigation and expert consultation using a stereotaxic surgical videoserver] Mund Kiefer Gesichtschir. 2000;(4) Suppl 1:S369–74. doi: 10.1007/PL00014560. [DOI] [PubMed] [Google Scholar]
- 101.Enislidis G, Wagner A, Ploder O, et al. Computed intraoperative navigation guidance--a preliminary report on a new technique. Br J Oral Maxillofac Surg. 1997;35(4):271–4. doi: 10.1016/s0266-4356(97)90046-2. [DOI] [PubMed] [Google Scholar]
- 102.Edwards PJ, King AP, Hawkes DJ, et al. Stereo augmented reality in the surgical microscope. Stud Health Technol Inform. 1999;62:102–8. [PubMed] [Google Scholar]
- 103.Wanschitz F, Birkfellner W, Figl M, et al. Computer-enhanced stereoscopic vision in a head-mounted display for oral implant surgery. Clin Oral Implants Res. 2002;13(6):610–6. doi: 10.1034/j.1600-0501.2002.130606.x. [DOI] [PubMed] [Google Scholar]
- 104.Kahrs LA, Hoppe H, Eggers G, et al. Visualization of surgical 3D information with projector-based augmented reality. Stud Health Technol Inform. 2005;111:243–6. [PubMed] [Google Scholar]
- 105.Roessler K, Ungersboeck K, Dietrich W, et al. Frameless stereotactic guided neurosurgery: clinical experience with an infrared based pointer device navigation system. Acta Neurochir (Wien) 1997;139(6):551–9. doi: 10.1007/BF02750999. [DOI] [PubMed] [Google Scholar]
- 106.Levy ML, Day JD, Albuquerque F, et al. Heads-up intraoperative endoscopic imaging: a prospective evaluation of techniques and limitations. Neurosurgery. 1997;40(3):526–30. doi: 10.1097/00006123-199703000-00020. discussion 530-1. [DOI] [PubMed] [Google Scholar]
- 107.Zheng G, Caversaccio M, Bachler R, et al. Frameless optical computer-aided tracking of a microscope for otorhinology and skull base surgery. Arch Otolaryngol Head Neck Surg. 2001;127(10):1233–8. doi: 10.1001/archotol.127.10.1233. [DOI] [PubMed] [Google Scholar]
- 108.Feiner SK. Augmented reality: a new way of seeing. Sci Am. 2002;286(4):44–55. doi: 10.1038/scientificamerican0402-48. [DOI] [PubMed] [Google Scholar]
- 109.Birkfellner W, Figl M, Huber K, et al. A head-mounted operating binocular for augmented reality visualization in medicine--design and initial evaluation. IEEE Trans Med Imaging. 2002;21(8):991–7. doi: 10.1109/TMI.2002.803099. [DOI] [PubMed] [Google Scholar]
- 110.Birkfellner W, Figl M, Matula C, et al. Computer-enhanced stereoscopic vision in a head-mounted operating binocular. Phys Med Biol. 2003;48(3):49–57. doi: 10.1088/0031-9155/48/3/402. [DOI] [PubMed] [Google Scholar]
- 111.Marmulla R, Hoppe H, Muhling J, et al. An augmented reality system for image-guided surgery. Int J Oral Maxillofac Surg. 2005;34(6):594–6. doi: 10.1016/j.ijom.2005.05.004. [DOI] [PubMed] [Google Scholar]
- 112.Eggers G, Salb T, Hoppe H, et al. Intraoperative augmented reality: the surgeons view. Stud Health Technol Inform. 2005;111:123–5. [PubMed] [Google Scholar]
- 113.Strauss G, Hofer M, Korb W, et al. Accuracy and precision in the evaluation of computer assisted surgical systems. A definition. HNO. 2006;54(2):78–84. doi: 10.1007/s00106-005-1321-3. [DOI] [PubMed] [Google Scholar]
- 114.Fitzpatrick JM, West JB, Maurer CR., Jr Predicting error in rigid-body point-based registration. IEEE Trans Med Imaging. 1998;17(5):694–702. doi: 10.1109/42.736021. [DOI] [PubMed] [Google Scholar]
- 115.Fitzpatrick JM, West JB. The distribution of target registration error in rigid-body point-based registration. IEEE Trans Med Imaging. 2001;20(9):917–27. doi: 10.1109/42.952729. [DOI] [PubMed] [Google Scholar]
- 116.Widmann G, Bale RJ. Accuracy in computer-aided implant surgery--a review. Int J Oral Maxillofac Implants. 2006;21(2):305–13. [PubMed] [Google Scholar]
- 117.Cohen DS, Lustgarten JH, Miller E, et al. Effects of coregistration of MR to CT images on MR stereotactic accuracy. J Neurosurg. 1995;82(5):772–9. doi: 10.3171/jns.1995.82.5.0772. [DOI] [PubMed] [Google Scholar]
- 118.Golfinos JG, Fitzpatrick BC, Smith LR, et al. Clinical use of a frameless stereotactic arm: results of 325 cases. J Neurosurg. 1995;83(2):197–205. doi: 10.3171/jns.1995.83.2.0197. [DOI] [PubMed] [Google Scholar]
- 119.Dorward NL, Alberti O, Palmer JD, et al. Accuracy of true frameless stereotaxy: in vivo measurement and laboratory phantom studies. Technical note. J Neurosurg. 1999;90(1):160–8. doi: 10.3171/jns.1999.90.1.0160. [DOI] [PubMed] [Google Scholar]
- 120.Galloway RL, Jr, Maciunas RJ, Latimer JW. The accuracies of four stereotactic frame systems: an independent assessment. Biomed Instrum Technol. 1991;25(6):457–60. [PubMed] [Google Scholar]
- 121.Casap N, Kreiner B, Wexler A, et al. Flapless approach for removal of bone graft fixing screws and placement of dental implants using computerized navigation: a technique and case report. Int J Oral Maxillofac Implants. 2006;21(2):314–9. [PubMed] [Google Scholar]
- 122.Marmulla R, Niederdellmann H. Computer-assisted bone segment navigation. J Cranio Maxillofac Surg. 1998;26:347–59. doi: 10.1016/s1010-5182(98)80067-x. [DOI] [PubMed] [Google Scholar]
- 123.Marmulla R, Niederdellmann H. Surgical planning of computer-assisted repositioning osteotomies. Plast Reconstr Surg. 1999;104(4):938–44. doi: 10.1097/00006534-199909040-00007. [DOI] [PubMed] [Google Scholar]
- 124.Wagner A, Undt G, Watzinger F, et al. Principles of computer-assisted arthroscopy of the temporomandibular joint with optoelectronic tracking technology. Oral Surg Oral Med Oral Pathol Oral Radiol Endod. 2001;92(1):30–7. doi: 10.1067/moe.2001.114384. [DOI] [PubMed] [Google Scholar]
- 125.Bale RJ, Laimer I, Schlager A, et al. Frameless stereotactic cannulation of the foramen ovale for ablative treatment of trigeminal neuralgia. Neurosurgery. 2006;59(4) Suppl 2:ONS394–ONS402. doi: 10.1227/01.NEU.0000232770.97616.D0. [DOI] [PubMed] [Google Scholar]
- 126.Carrau RL, Snyderman CH, Curtin HD, et al. Computer-assisted intraoperative navigation during skull base surgery. Am J Otolaryngol. 1996;17(2):95–101. doi: 10.1016/s0196-0709(96)90003-4. [DOI] [PubMed] [Google Scholar]
- 127.Caversaccio M, Nolte LP, Hausler R. Present state and future perspectives of computer aided surgery in the field of ENT and skull base. Acta Otorhinolaryngol Belg. 2002;56(1):51–9. [PubMed] [Google Scholar]
- 128.Caversaccio M, Romualdez J, Baechler R, et al. Valuable use of computer-aided surgery in congenital bony aural atresia. J Laryngol Otol. 2003;117(4):241–8. doi: 10.1258/00222150360600814. [DOI] [PubMed] [Google Scholar]
- 129.Korb W, Marmulla R, Raczkowsky J, et al. Robots in the operating theatre--chances and challenges. Int J Oral Maxillofac Surg. 2004;33(8):721–32. doi: 10.1016/j.ijom.2004.03.015. [DOI] [PubMed] [Google Scholar]
- 130.Klein M, Hein A, Lueth T, et al. Robot-assisted placement of craniofacial implants. Int J Oral Maxillofac Implants. 2003;18(5):712–8. [PubMed] [Google Scholar]
- 131.Weihe S, Wehmoller M, Schliephake H, et al. Synthesis of CAD/CAM, robotics and biomaterial implant fabrication: single-step reconstruction in computer-aided frontotemporal bone resection. Int J Oral Maxillofac Surg. 2000;29(5):384–8. [PubMed] [Google Scholar]
- 132.Kavanagh KT. Applications of image-directed robotics in otolaryngologic surgery. Laryngoscope. 1994;104(3 Pt 1):283–93. doi: 10.1288/00005537-199403000-00008. [DOI] [PubMed] [Google Scholar]
- 133.Brief J, Hassfeld S, Redlich T, et al. Robot assisted insertion of dental implants - a clinical evaluation. In: Lemke HU, Vannier MV, Inamura K, editors. CARS 2000. Elsevier Science B.V.; 2000. pp. 932–7. [Google Scholar]
- 134.Brief J, Hassfeld S, Boeseke R, et al. Int Poster J Dent Oral Med. 2002;4(1) Poster 109. [Google Scholar]
- 135.Goulette F, Dutreuil J, Laurgeau, et al. A new method and clinical case for computer assisted dental implantology. In: Lemke HU, Vannier MV, Inamura K, editors. CARS 2002. Elsevier Science B.V.; 2002. pp. 953–8. [Google Scholar]
- 136.Schermeier O, Hildebrand D, Lueth TC, et al. Lemke HU, Vannier MV, Inamura K, editors. Accuracy of an image-guided system for oral implantology. International Congress Series. Vol. 1230. 2001:748–52. [Google Scholar]
- 137.Sarment DP, Sukovic P, Clinthorne N. Accuracy of implant placement with a stereolithographic surgical guide. Int J Oral Maxillofac Implants. 2003;18(4):571–7. [PubMed] [Google Scholar]
- 138.Sarment DP, Al-Shammari K, Kazor CE. Stereolithographic surgical templates for placement of dental implants in complex cases. Int J Periodont Restor Dent. 2003;23(3):287–95. [PubMed] [Google Scholar]
- 139.Paleologos TS, Dorward NL, Wadley JP, et al. Clinical validation of true frameless stereotactic biopsy: analysis of the first 125 consecutive cases. Neurosurgery. 2001;49(4):830–5. doi: 10.1097/00006123-200110000-00009. discussion 835-7. [DOI] [PubMed] [Google Scholar]
- 140.Fried MP, Moharir VM, Shin J, et al. Comparison of endoscopic sinus surgery with and without image guidance. Am J Rhinol. 2002;16(4):193–7. [PubMed] [Google Scholar]
- 141.Tabaee A, Kacker A, Kassenoff TL, et al. Outcome of computer-assisted sinus surgery: a 5-year study. Am J Rhinol. 2003;17(5):291–7. [PubMed] [Google Scholar]
- 142.Howe RD, Matsuoka Y. Robotics for surgery. Annu Rev Biomed Eng. 1999;1:211–40. doi: 10.1146/annurev.bioeng.1.1.211. [DOI] [PubMed] [Google Scholar]
- 143.Wittwer G, Adeyemo WL, Schicho K, et al. Computer-guided flapless transmucosal implant placement in the mandible: a new combination of two innovative techniques. Oral Surg Oral Med Oral Pathol Oral Radiol Endod. 2006;101(6):718–23. doi: 10.1016/j.tripleo.2005.10.047. [DOI] [PubMed] [Google Scholar]
- 144.Widmann G, Widmann R, Widmann E, et al. Use of surgical navigation systems for CT-guided template production. Int J Oral Maxillofac Implants. 2007;22(1):72–78. [PubMed] [Google Scholar]