Abstract
During MRI-guided interventions, navigation support is often separated from the operating field on displays, which impedes the interpretation of positions and orientations of instruments inside the patient's body as well as hand–eye coordination. To overcome these issues projector-based augmented reality can be used to support needle guidance inside the MRI bore directly in the operating field. The authors present two visualisation concepts for needle navigation aids which were compared in an accuracy and usability study with eight participants, four of whom were experienced radiologists. The results show that both concepts are equally accurate ( and ), useful and easy to use, with clear visual feedback about the state and success of the needle puncture. For easier clinical applicability, a dynamic projection on moving surfaces and organ movement tracking are needed. For now, tests with patients with respiratory arrest are feasible.
Keywords: augmented reality, data visualisation, medical image processing, biomedical MRI, needles
Keywords: augmented reality visualisation, needle guidance, MRI-guided interventions, operating field, displays, positions, patient, hand–eye coordination, needle navigation aids, needle puncture, projector-based augmented reality, visual feedback
1. Introduction
During minimally invasive interventions, target lesions are accessed through small entry points with instruments such as laparoscopes, catheters or needle applicators.
Because the human body is not opened for such treatments, radiological images are required to locate target lesions, surgical instruments and risk structures. Many needle-based interventions are thus carried out with the help of ultrasound (US) or computed tomography (CT) scanners [1, 2]. US is widely available, compact and cheap, but does not reach deeper structures, especially under bones [3]. In contrast, CT images represent the whole operating area, but with limited soft tissue contrast [4] and only by using harmful ionising radiation. As an alternative, MRI does not emit ionising radiation and provides excellent soft tissue contrast, so that target lesions invisible for US or CT can be identified. In addition, image planes in oblique orientations can be acquired and morphologic as well as functional information (temperature changes, blood flow, diffusion) can be monitored. When using tracked instruments that are calibrated with the MRI the live imaging planes can even be aligned along the needle to always get the view of the needle's surroundings [5].
To improve interventional MRI, appropriate instrument guidance is essential to simplify and shorten the intervention. Such assistance is often used on a display [6] (as shown in Fig. 1), which separates the useful information from the patient and increases mental load [7]. To overcome this issue, AR can be used to fuse the separate virtual data directly to the operating field.
Different Augmented Reality (AR)-based needle guidance systems have been introduced for the interventional MRI before. These include a head mounted display (HMD) to augment the radiologist's view of the patient with two-dimensional (2D) anatomical images and a virtual needle to navigate to the target [8], and a semi-transparent mirror placed in front of the MRI to visualise an overlay of 2D images and a needle navigation overlay [9]. Both techniques have the drawback in common that they require the patient to be translated out of the MRI because the HMD is not fully MR safe and the mirror occupies too much space to fit inside the bore. This interrupts the workflow and prevents the use of most of the MRI's advantages mentioned above, first and foremost live imaging, which is required to control the precise position of the needle during the puncture.
This work is based on a projector-based AR system for interventional MRI that is described in [10]. Based on this, we propose an AR visualisation approach that provides needle guidance information directly in the operating field inside the MRI bore. We present different concepts for AR needle navigation visualisation, their prototypical implementation and an assessment of their usability and accuracy.
2. State-of-the-art
Supporting needle guidance during image-guided interventions is a widely discussed topic. Separating the needle insertion task into three subtasks (tip positioning, needle alignment, needle insertion) and applying a cross-hair visualisation for needle positioning/alignment as well as a progress bar indicating the needle depth has been proposed [11, 12] and is nowadays implemented in many commercial navigation systems. Another approach is to overlay a video stream with the preview visualisation of the needle into the proper position that the user needs to target [6]. Because it is difficult to interpret navigation data relating to the patient on separate monitors, spatial AR approaches have emerged. One of them is a handheld projector augmenting arbitrary surfaces with explicit navigation aids and a needle path preview [13]. Alternatively, depth hints can also be given as concentric circles, where the inner expands to the outer circle with increasing depth, as described in [14]. Here, the needle orientation was led by an arrow. Besides projector-based AR, optical see-through AR using HMDs [8, 15] or head-up displays [16, 9, 17] are proposed. In both cases, the operating field was augmented with 2D images, needle positioning aids and basic distance cues. The latter plays a crucial role in AR applications. Especially in the medical context, it is important to correctly perceive the distance to risk structures to prevent injuries. Different visualisation approaches have been proposed to improve depth perception in AR, such as adaptive alpha blending on volumetric representations [18] or focus-and-context methods that consider the surrounding structures, i.e. bones and skin, for rendering and to create a window-like effect while angle and distance to the focused target objects are taken into account for transparency calculation [19].
Other approaches try to optimise the visualisation of 3D models themselves, in order to improve the perception of spatial relations and distances. Pseudo-chromadepth, a rendering technique where distance is represented by colours that follow the visible spectrum of light, was shown to improve relative depth perception [20]. Furthermore, illustrative visualisation encodes the shape and distance of vascular structures with different styles of hatching lines and can support depth perception in AR environments as well [21]. Similar approaches were presented by Wang et al. [22], who applied contour enhancing, occlusion and depth colour-coding to improve the perception of vascular structures.
In addition, augmented reality in the form of auditory feedback has been proposed to guide needles [23]. Bork et al. [24] combined both auditory and visual feedbacks to encode distance information between a surgical instrument and regions of interest. Their method includes a propagating shape around the instrument tip, which increases in size over time. Acoustic signals emphasise the growth process and speed of that shape. However, acoustic feedback may only be of limited use because of the noise from the MRI device.
3. Methods
To support needle guidance inside the interventional MRI, we developed two visualisation concepts. In order to accurately display navigation aids for needle insertion tasks directly on a patient, several steps were necessary. These steps are described in the following section. The navigation concepts are presented afterwards.
3.1. AR hardware setup
The basis of our navigation approach is a projector-camera system set up in the MRI scanning room. The long-throw projector with a resolution of 1024 × 768 px is located outside the room and its light is directed into the MRI bore via a waveguide through the wall and three mirrors (see Fig. 2). The detailed setup of the AR system, the extraction of the patient geometry, and the calibration process are described in [10].
For needle guidance applications, it is possible to track instruments with a commercial MPT system that is calibrated with the MRI. Therefore, three moiré phase markers are attached to a needle and calibrated with a tracked calibration body. Due to the fixed focus of the MPT camera, the tracking rate ranges from 115FPS, depending on the distance to the markers.
The projector is used to 3D scan the arbitrarily shaped projection surface via a structured-light approach. The result is a semi-dense point cloud of the projection surface which can then be further processed. The process of 3D scanning and generating the point cloud takes about 8 s in total. Then, a surface mesh is created from the point cloud that is used for later intersection calculations.
3.2. Needle interaction
In each frame, the needle's position and orientation are compared to the initially set needle path. The needle's intersection with the surface is determined by tracing the needle through the surface triangle mesh. The resulting intersection point is then compared with the planned needle insertion point. As a measure for the needle's orientation, the angle between the needle's direction vector and the vector between insertion and target point is calculated. This data will then be visualised depending on which visualisation concept is used.
3.3. Navigation by explicit aids (2D)
The first concept uses explicit aids to guide the user through the single insertion tasks positioning, alignment and insertion, as proposed in [11] and is illustrated in Fig. 3. The positioning of the needle at the planned insertion point is supported by a circle visualisation of adaptive size. The nearer the user positions the needle to the specified position, the smaller the radius of the circle. This is further clarified by a change of colour from red (high distance) to green (small distance).
The alignment process is supported by displaying an arrow originating at the insertion point and pointing towards the direction in which the needle shaft has to be tilted. The length of the arrow results from the angle between needle direction and designated needle path. The smaller the angle gets, the shorter the arrow is drawn. The length is logarithmically interpolated so that alignment changes at smaller differences to the planned orientation have less effect on the arrow length than changes at larger angle differences. That way, the arrow remains visible at small alignment deviations and thus allows fine adjustments. At an angular difference below 1°, the arrow's colour changes from red to yellow indicating an acceptable alignment. Further reducing the deviation below 0.5° causes the arrow to be replaced by a green dot, thus signalling a successful needle alignment.
Finally, the insertion step is supported by the visualisation of a progress bar. The filling of the bar is linearly dependent on the Euclidean distance between the needle tip and target structure. The bar begins filling up after inserting the needle. Reaching a distance below 0.4 mm, the bar's colour turns from red to yellow. After further reducing that distance to 0.2 mm, the progress bar switches its colour to green, thus implying a successful needle insertion. Over inserting will cause the colour to change back to red and continue downwards from the progress bar.
Due to the flat visualisation of explicit aids, this concept will be called concept 2D in the following.
3.4. Navigation by virtual vision through the skin (3D)
The second implemented approach enables the user to virtually see through the projection surface by visualising segmented structures, the aimed target and the needle inside the body. The concept is illustrated in Fig. 4. Those visualisations are rendered during a ray tracing process and show structures perspectively correct from the fixed viewing position. For each pixel, a ray is traced through the scene using the world coordinate correspondences. These rays are then checked for intersection with the surface triangle mesh and the cylinder representing the needle and its path. The resulting depth values at the intersection sites are then compared to initially calculated depth maps of the static objects. The depth values are then sorted and a pixel colour is computed with respect to the intersected objects’ colour and transparency values. The resultant colour of an object is the product of a set base colour, an applied Phong shading coefficient and a depth encoding coefficient. The latter is a linearly interpolated factor that depends on the distance to the viewing position. The higher the distance, the smaller the coefficient and the darker the final colour. Because the final image is displayed on the scene using a projector, dark colours cannot be visualised. Therefore, darkening colours results in applying transparency to them.
In order to support the user while using this concept, the needle is displayed virtually elongated as a cylinder. That way, the user can see the path the needle would follow if it was inserted at that moment. A small dot along the path line signifies where the needle would intersect an object. That dot's colour thereby depends on the kind of intersected object. Red illustrates the intersection of risk structures while green signalises the intersection of the target structure.
Furthermore, the target region is displayed as a two-zone sphere. That way, the target is visualised big enough to be quickly perceived while the inner zone emphasises the target's centre more precisely. The two zones are used for colour coding as well. The inner zone's colour represents the state of the needle's planned alignment and is equivalent to the previously described colours of the alignment arrow. The outer zone's colour represents the Euclidean distance from the needle tip to the target's centre and is coloured similarly to the previously described progress bar. A perfectly aligned needle at the target's centre thereby results in a target sphere with two green zones.
In addition, this concept uses a different approach to visualise the planned insertion point. The point is illustrated by a circle on the projection surface with a constant radius. The border of the circle is coloured red, while its centre is coloured green. The space in between is interpolated between these colours. The nearer the needle is positioned to the insertion point, the greater the green colour of the centre is weighted. A perfectly positioned needle results in a totally green circle. To further emphasise that, the colour lights up when the needle is positioned correctly.
This concept makes use of the visualisation of 3D structures inside the body. Therefore, this concept will be called concept 3D in this work. A real-world view of the needle navigation visualisation on a phantom is shown in Fig. 5. For demonstration purposes, the image shows both visualisation concepts.
4. Evaluation
In order to find out if the navigation concepts are useful and accurate, we conducted a user study with eight participants, four of whom were radiologists with 4–15 years of experience with minimally invasive needle-based image-guided interventions and four of whom were inexperienced users with a technical background in the medical domain. The users had the task to insert a tracked needle into a phantom body inside the MRI bore without image guidance, relying only on the AR interface described in this paper. The users were also told to keep their heads in the approximate predetermined viewing position because no eye tracking could be used to keep track of the correct viewpoint. However, the space inside the bore is very limited so that the users were not able to move the head around considerably. The needle needed to be positioned on the pre-planned entry point, aligned as desired and then inserted until the target was reached. Each user performed the task three times for each concept individually with different entry points and targets.
The phantom consisted of an opaque box filled with candle gel, which provided haptics similar to skin and other soft tissue and can be seen in a T1 MRI sequence dataset. In addition, rings were embedded in the candle gel as symbolic targets, which are visible in the MRI images. However, they were covered with a paper tissue so that the users were not influenced. The order of the used visualisation concepts was varied between the subjects via the biased coin method: The concepts were assigned to the two sides of the coin. It was flipped for the first user to determine the order of concepts, e.g. 3D first, then 2D. The second user then performed the tasks in the opposite order. The coin was flipped for every second user.
We measured the puncture duration as well as errors between the planned and reached entry and target positions in a post-operatively acquired dataset with the MRI's built-in distortion correction to best preserve the validity of the 3D positions. After each concept, each user answered the meCUE [25] and Nasa RawTLX [26] questionnaires so that insight into usefulness, usability, intention of use and subjective workload were given.
5. Results
The study did not reveal substantial differences between the concepts nor between experienced and inexperienced users (see Table 1) regarding positioning errors or puncture duration. Besides the duration of the insertion, there were no notable differences between user groups with regard to accuracy, subjective workload or user experience. Therefore, the results of the two concepts for all users are presented in the following. The subjective workload score from RawTLX for concept 2D (33.5) is slightly lower than for concept 3D (38.1), but still very similar. Frustration is higher with 3D (65.6) than with 2D (35), but mental demand was higher with 2D (37.5) than with 3D (25.6). The meCUE scores for both concepts are shown in Table 2.
Table 1.
Experience | Concept | Puncture duration, s | Entry point error, mm | Target distance error, mm |
---|---|---|---|---|
med | 2D | |||
3D | ||||
tech | 2D | |||
3D | ||||
all | 2D | |||
3D |
Table 2.
scale | 2D | 3D |
---|---|---|
usefulness | ||
usability | ||
intention of use | ||
overall |
6. Discussion
We successfully demonstrated and tested two projector-based AR visualisation concepts for navigating a tracked needle from a planned entry point to a target position inside an MRI bore during minimally-invasive interventions. The accuracy measurements serve as an indicator of the quality of the registration as well as the operability. The values for the target point errors (2D: , 3D: ) are comparable to those in the related literature ( [27], [28], [29], [30], [8]). The achieved accuracy can differ in a clinical setup due to needle deflection, which highly depends on the needle length and stiffness. During our experiments, the needle stiffness was very high and did not cause considerable deviations from the calibrated tip position with regard to the needle marker. Because of the virtual malformation of MRI scanned objects when processing the received signals, i.e. during the pre- and post-operative scans used for measurements, and the low resolution, these error values are inherently faulty and can only serve as an indication for accuracy.
The visualisation of the navigation aids and the virtual needle was perceived as perspective-correct even though it was fixed at an assumed viewing position. Nevertheless, the user's viewing position should be tracked in the future to guarantee a correct perspective and to make use of motion parallaxing and to thus create a more immersive visual representation of 3D objects.
The low MPT rate caused frustration with the users. Sometimes, when the MPT markers were too close to the MPT camera, the tracking stopped and the users did not have any feedback on the position or orientation of the needle. This led to more correcting movements, i.e. pulling the needle back and inserting it again, in both concepts. In addition, according to the RawTLX results, concept 3D shows a higher potential for frustration, because the users could not aim for the target as precisely as with concept 2D. Thus, many users had to correct more punctures as compared with concept 2D. However, the accuracy of both concepts and the usability and usefulness scores are similar, thus no superior concept can be determined.
Because some users commented they were more comfortable with one concept than with the other, it should remain the user's choice which one to use. Furthermore, no notable difference between the two user groups could be found in the study regarding the accuracy or subjective perception of workload or usability. This could be due to the easy-to-understand and unambiguous visualisations, which do not presume medical experience, as some users commented. The clear feedback during the whole process with colour and shape changes of the visual aids compensates for the professional differences.
By tracking the user's head position further development steps could enable kinetic depth cues by interactively adapting the visualisation to the viewing position. In addition, further illustrative visualisation approaches could be used to better encode depth information [21]. These additional depth cues could improve user performance and reduce frustration by better clarifying the 3D virtual scene. This navigation information should be visualised alongside explicit navigation aids and could thereby serve for visual inspection of the inserted needle position.
In total, the users confirmed with high meCUE usability and usefulness scores, as well as in comments, that the proposed needle guidance visualisation concepts serve as a suitable support for MRI-guided interventions. As suggested by the intention of use, the users would use this AR system beyond the study scope. To achieve clinical applicability the radiologists commented that organ movement tracking with ultrasound could be integrated and visualised and a dynamic surface registration to maintain position correctness of the projection is required. In the current state of the system, needle punctures could only be carried out during the artificially induced respiratory arrest of the intubated patient based on a planning dataset that is acquired breath triggered. The strength of the AR projection system shows during the first stage of a puncture when the needle can already be oriented accurately outside the body so that only small corrections are required when approaching the target. Thus, less healthy tissue will be damaged. This makes it also valuable for teaching radiologists.
7. Conclusion
In this work, we introduced two concepts for visualising needle guidance aids via a projector-based AR system inside the MRI bore directly in the operating field. In the 3D concept, a virtual depth-encoded needle, the planned entry point, and the target are projected. The virtual needle is elongated to facilitate the correct orientation in relation to the target. Changing colours reflect the current state as well as the correctness of the puncture. The 2D concept provides no 3D objects. Instead, the planned entry point, an orientation indicating arrow, and a progress bar showing the current distance to the target are visualised. A user study with four experienced radiologists and four participants with a technical background in the medical domain did not reveal a superior concept. Furthermore, no differences between the user groups could be found in terms of accuracy or usability assessment. This confirms that the visualisations are easy to use with clear feedback that can be understood without prior experience with the AR needle guidance system.
In summary, the proposed projector-based AR visualisation concepts represent a new approach to facilitate needle guidance for interventional MRI procedures, and the performed evaluation provides valuable insights. Further studies in experimental operating rooms and in the clinical environment are necessary to further improve the visualisation concepts and to evaluate their clinical applicability.
8. Funding and declaration of interests
This work is funded by the Federal Ministry of Education and Research (BMBF) within the STIMULATE research campus (grant no. 13GW0095A) and the German Research Foundation (grant nos. HA 7819/1-1 and LA 3855/1-1). Frank Wacker declares grants from Siemens Healthineers outside of this work.
9. Conflict of interest
None declared.
10 References
- 1.Schullian P., Widmann G., Lang T.B., et al. : ‘Accuracy and diagnostic yield of CT-guided stereotactic liver biopsy of primary and secondary liver tumors’, Comput. Aided Surg., 2011, 16, (4), pp. 181–187 (doi: 10.3109/10929088.2011.578367) [DOI] [PubMed] [Google Scholar]
- 2.Fischbach F., Bunke J., Thormann M., et al. : ‘MR-guided freehand biopsy of liver lesions with fast continuous imaging using a 1.0-t open MRI scanner: experience in 50 patients’, Cardiovasc. Intervent. Radiol., 2011, 34, (1), pp. 188–192 (doi: 10.1007/s00270-010-9836-8) [DOI] [PubMed] [Google Scholar]
- 3.Rempp H., Loh H., Hoffmann R., et al. : ‘Liver lesion conspicuity during real-time MR-guided radiofrequency applicator placement using spoiled gradient echo and balanced steady-state free precession imaging’, J. Magn. Reson. Imag., 2014, 40, (2), pp. 432–439 (doi: 10.1002/jmri.24371) [DOI] [PubMed] [Google Scholar]
- 4.Hoffmann R., Thomas C., Rempp H., et al. : ‘Performing MR-guided biopsies in clinical routine: factors that influence accuracy and procedure time’, Eur. Radiol., 2012, 22, (3), pp. 663–671 (doi: 10.1007/s00330-011-2297-x) [DOI] [PubMed] [Google Scholar]
- 5.Kägebein U., Godenschweger F., Boese A., et al. : ‘Automatic image plane alignment for percutaneous MR guided interventions using an optical moiré’ phase tracking system’, Magn. Reson. Mater. Phys. Biol. Med., 2015, 28, pp. S159–S160 [Google Scholar]
- 6.Seitel A., Bellemann N., Hafezi M., et al. : ‘Towards markerless navigation for percutaneous needle insertions’, Int. J. Comput. Assist. Radiol. Surg., 2016, 11, (1), pp. 107–117 (doi: 10.1007/s11548-015-1156-7) [DOI] [PubMed] [Google Scholar]
- 7.Manzey D., Röttger S., Bahner-Heyne J.E., et al. : ‘Image-guided navigation: the surgeon's perspective on performance consequences and human factors issues’, Int. J. Med. Robot. Comput. Assisted Surg., 2009, 5, (3), pp. 297–308 (doi: 10.1002/rcs.261) [DOI] [PubMed] [Google Scholar]
- 8.Wacker F.K., Vogt S., Khamene A., et al. : ‘An augmented reality system for mr image-guided needle biopsy: initial results in a swine model 1’, Radiology, 2006, 238, (2), pp. 497–504 (doi: 10.1148/radiol.2382041441) [DOI] [PubMed] [Google Scholar]
- 9.Weiss C.R., Nour S.G., Lewin J.S.: ‘MR-guided biopsy: a review of current techniques and applications’, J. Magn. Reson. Imag., 2008, 27, (2), pp. 311–325 (doi: 10.1002/jmri.21270) [DOI] [PubMed] [Google Scholar]
- 10.Mewes A., Heinrich F., Kägebein U., et al. : ‘Projector-based augmented reality system for interventional visualization inside MRI scanners’, J. Med. Robot. Comput. Assisted Surg., 2018, p. e1950, doi: 10.1002/rcs.1950 [DOI] [PubMed] [Google Scholar]
- 11.Seitel A., Maier-Hein L., Schawo S., et al. : ‘In vitro evaluation of different visualization approaches for computer assisted targeting in soft tissue’, Int. J. Comput. Assisted Radiol. Surg., 2007, 2, pp. S188–S190 [Google Scholar]
- 12.Oliveira-Santos T., Klaeser B., Weitzel T., et al. : ‘A navigation system for percutaneous needle interventions based on PET/CT images: design, workflow and error analysis of soft tissue and bone punctures’, Comput. Aided Surg., 2011, 16, (5), pp. 203–219 (doi: 10.3109/10929088.2011.597566) [DOI] [PubMed] [Google Scholar]
- 13.Gavaghan K., Oliveira-Santos T., Peterhans M., et al. : ‘Evaluation of a portable image overlay projector for the visualisation of surgical navigation data: phantom studies’, Int. J. Comput. Assisted Radiol. Surg., 2012, 7, (4), pp. 547–556 (doi: 10.1007/s11548-011-0660-7) [DOI] [PubMed] [Google Scholar]
- 14.Krempien R., Hoppe H., Kahrs L., et al. : ‘Projector-based augmented reality for intuitive intraoperative guidance in image-guided 3d interstitial brachytherapy’, Int. J. Radiat. Oncol. Biol. Phys., 2008, 70, (3), pp. 944–952 (doi: 10.1016/j.ijrobp.2007.10.048) [DOI] [PubMed] [Google Scholar]
- 15.Das M., Sauer F., Schoepf U.J., et al. : ‘Augmented reality visualization for CT-guided interventions: system description, feasibility, and initial evaluation in an abdominal phantom 1’, Radiology, 2006, 240, (1), pp. 230–235 (doi: 10.1148/radiol.2401040018) [DOI] [PubMed] [Google Scholar]
- 16.Fichtinger G., Deguet A., Masamune K., et al. : ‘Image overlay guidance for needle insertion in CT scanner’, IEEE Trans. Biomed. Eng., 2005, 52, (8), pp. 1415–1424 (doi: 10.1109/TBME.2005.851493) [DOI] [PubMed] [Google Scholar]
- 17.Fritz J., Thainual P.U., Ungi T., et al. : ‘Augmented reality visualization with image overlay for MRI-guided intervention: accuracy for lumbar spinal procedures with a 1.5-t MRI system’, Am. J. Roentgenolog., 2012y, 198, (3), pp. W266–W273 (doi: 10.2214/AJR.11.6918) [DOI] [PubMed] [Google Scholar]
- 18.Marques B., Haouchine N., Plantefeve R., et al. : ‘Improving depth perception during surgical augmented reality’. ACM SIGGRAPH 2015 Posters, Los Angeles, USA, 2015, p. 24 [Google Scholar]
- 19.Bichlmeier C., Wimmer F., Heining S.M., et al. : ‘Contextual anatomic mimesis hybrid in-situ visualization method for improving multi-sensory depth perception in medical augmented reality’. IEEE Int. Symp. on Mixed and Augmented Reality (ISMAR), Nara, Japan, 2007, pp. 129–138 [Google Scholar]
- 20.Kersten-Oertel M., Chen S.J., Collins D.L.: ‘An evaluation of depth enhancing perceptual cues for vascular volume visualization in neurosurgery’, IEEE Trans. Vis. Comput. Graphics, 2014, 20, (3), pp. 391–403 (doi: 10.1109/TVCG.2013.240) [DOI] [PubMed] [Google Scholar]
- 21.Lawonn K., Luz M., Hansen C.: ‘Improving spatial perception of vascular models using supporting anchors and illustrative visualization’, Comput. Graph., 2017, 63, pp. 37–49 (doi: 10.1016/j.cag.2017.02.002) [Google Scholar]
- 22.Wang X., Zu Berge C.S., Demirci S., et al. : ‘Improved interventional x-ray appearance’. IEEE and ACM Int. Symp. on Mixed and Augmented Reality, Munich, Germany, 2014, pp. 237–242 [Google Scholar]
- 23.Black D., Hettig J., Luz M., et al. : ‘Auditory feedback to support image-guided medical needle placement’, Int. J. Comput. Assist Radiol. Surg., 2017, 12, (9), pp. 1655–1663 (doi: 10.1007/s11548-017-1537-1) [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Bork F., Fuerst B., Schneider A.-K., et al. : ‘Auditory and visio-temporal distance coding for 3-dimensional perception in medical augmented reality’. IEEE Int. Symp. on Mixed and Augmented Reality (ISMAR), Fukuoka, Japan, 2015, pp. 7–12 [Google Scholar]
- 25.Minge M., Thüring M., Wagner I., et al. : ‘The meCUE questionnaire. A modular evaluation tool for measuring user experience’, in Soares M., Falcão C., Ahram T.Z. (Eds.): ‘Advances in Ergonomics Modeling, Usability & Special Populations. Proceedings of the 7th Applied Human Factors and Ergonomics Society Conference 2016’ (Springer International Press, Cham, 2016), pp. 115–128 [Google Scholar]
- 26.Hart S.G., Staveland L.E.: ‘Development of nasa-tlx (task load index): results of empirical and theoretical research’, Adv. Psychol., 1988, 52, pp. 139–183 (doi: 10.1016/S0166-4115(08)62386-9) [Google Scholar]
- 27.Rothgang E., Gilson W.D., Wacker F., et al. : ‘Rapid freehand mr guided percutaneous needle interventions: an image-based approach to improve workflow and feasibility’, J. Magn. Reson. Imag., 2013, 37, (5), pp. 1202–1212 (doi: 10.1002/jmri.23894) [DOI] [PubMed] [Google Scholar]
- 28.Meyer B.C., Brost A., Kraitchman D.L., et al. : ‘Percutaneous punctures with MR imaging guidance: comparison between mr imaging-enhanced fluoroscopic guidance and real-time MR imaging guidance’, Radiology, 2013, 266, (3), pp. 912–919 (doi: 10.1148/radiol.12120117) [DOI] [PubMed] [Google Scholar]
- 29.Busse H., Garnov N., Thörmer G., et al. : ‘Flexible add-on solution for MR image guided interventions in a closed-bore scanner environment’, Magn. Reson. Med., 2010, 64, (3), pp. 922–928 (doi: 10.1002/mrm.22464) [DOI] [PubMed] [Google Scholar]
- 30.Moche M., Zajonz D., Kahn T., et al. : ‘MRI-guided procedures in various regions of the body using a robotic assistance system in a closed-bore scanner: preliminary clinical experience and limitations’, J. Magn. Reson. Imag., 2010, 31, (4), pp. 964–974 (doi: 10.1002/jmri.21990) [DOI] [PubMed] [Google Scholar]