Abstract
Mixed reality opens interesting possibilities as it allows physicians to interact with both, the real physical and the virtual computer-generated environment and objects, in a powerful way. A mixed reality system, based in the HoloLens 2 glasses, has been developed to assist cardiologists in a quite complex interventional procedure: the ultrasound-guided femoral arterial cannulations, during real-time practice in interventional cardiology. The system is divided into two modules, the transmitter module, responsible for sending medical images to HoloLens 2 glasses, and the receiver module, hosted in the HoloLens 2, which renders those medical images, allowing the practitioner to watch and manage them in a 3D environment. The system has been successfully used, between November 2021 and August 2022, in up to 9 interventions by 2 different practitioners, in a large public hospital in central Spain. The practitioners using the system confirmed it as easy to use, reliable, real-time, reachable, and cost-effective, allowing a reduction of operating times, a better control of typical errors associated to the interventional procedure, and opening the possibility to use the medical imagery produced in ubiquitous e-learning. These strengths and opportunities were only nuanced by the risk of potential medical complications emerging from system malfunction or operator errors when using the system (e.g., unexpected momentary lag). In summary, the proposed system can be taken as a realistic proof of concept of how mixed reality technologies can support practitioners when performing interventional and surgical procedures during real-time daily practice.
Keywords: Femoral arterial, Cannulation, Puncture, Interventional cardiology, Mixed reality, HoloLens glasses
Background and Significance
Virtual reality technologies (and its variations augmented reality and mixed reality) are playing an increasing prominent role in everyday life, from leisure to education and healthcare.
Virtual reality (VR) offers an immersive experience in an entirely artificial scenario. The new environment is entirely synthetic (computer-generated). The user must wear a VR headset which completely blocks perception of the surroundings. Virtual reality systems offer various degrees of immersive experiences, which can be classified in non-immersive, semi-immersive, and immersive [1].
Augmented reality (AR) offers an enhanced version of the real world by layering computer-generated virtual information (such as text, images, 3D models, music, video) over the real world [2] using a device combining a camera and a screen or a transparent headset. Mixed reality (MR) is an advanced form of the AR paradigm in which synthetic data is not only layered: instead, virtual objects are completely integrated within the real environment and can interact with it. The physical real world and the virtual computer-generated objects are difficult to distinguish [3].
Table 1 briefly describes virtual reality, augmented reality and mixed reality.
Table 1.
Virtual reality |
User interacts with an entirely computer-generated scenario Perception of real surrounding × Users need to always wear some type of device ✓ Example: Virtual traveling or virtual visits to museums. During the entire experience, the utilization of a wearable device, such as Meta Quest glasses, is required for scene rendering |
Augmented reality |
User interacts with an enhanced version of the real world containing computer-generated objects Perception of the real surrounding ✓ Users need to always wear some type of device × Example: Games in which virtual objects or creatures appear as if they are part of the player’s real environment. In such cases, individuals can actively participate in the game by donning augmented reality glasses like Magic Leap. Alternatively, they can integrate the game’s scenarios and objects into their environment without requiring additional wearable hardware. For instance, they can use the camera of a mobile phone to capture the real scene, overlay virtual objects on it and display the augmented view on the mobile phone screen |
Mixed reality |
User interacts with an enhanced version of the real world where physical and virtual objects are difficult to distinguish Perception of the real surrounding ✓ Users need to always wear some type of device × Example: Furniture fitting. Mixed reality allows users to interact with virtual objects, fostering a genuine sense of manipulation. In this particular case, advanced devices such as HoloLens 2 can be used to render furniture in their environment and place it according to their preferences. Another possibility is to use a mobile phone to view virtual objects on its screen while moving freely around the physical space |
These technologies are opening new promising opportunities in the healthcare area as revealed by the literature [4–9].
More specifically, virtual reality–related technologies have started to be successfully used for the training of different techniques and interventions. Liang et al. [10] developed a simulation application for a mixed reality device that projected a human face displaying facial drooping (a symptom of stroke) onto a computerized training mannequin. Nursing students wearing a mixed reality device participated in the simulation to perform an assessment of their mannequin patient to identify the symptom of stroke and act accordingly. Students’ feedback was quite positive. Medellin-Castillo et al. [11] conducted three case studies to study haptics and virtual reality for the training of the orthognathic surgery (OGS), which is a very complex procedure that allows to correct a wide range of skeletal and dental irregularities, requiring surgical skills traditionally acquired through extended hands-on training using cadavers or models. Lin et al. [12] evaluated the use of virtual reality technologies to assist in the training of the lateral ventricle puncture necessary for external ventricular drain. Two groups of neurosurgical interns were trained with and without the system to validate its effectiveness. In general, these experiments provide initial evidence of the usefulness of virtual reality technologies in procedure training, showing an improvement of surgical skills, performance, and confidence of trainees, accompanied by a reduction of execution times and typical errors.
The integration of virtual reality technologies as real-time assistance in surgical procedures is more complex as a system malfunction can have undesirable catastrophic consequences and extremely accurate translation of virtual coordinates into the real world are required. Initial experiences are also found in the literature, although in a far more initial state than in training. Zhou et al. [13] presented a system to facilitate brachytherapy, a useful treatment for lung cancer, where the accuracy of needle positioning is critical. Using mixed reality for medical image fusion, the system allowed doctors to have an intuitive understanding of the patient’s tumor. The system was validated through a phantom experiment with good results for the needle insertion accuracy.
The use of mixed reality in surgeries and interventions is however very promising, as mixed reality allows physicians to easily interact with and manipulate physical and virtual environments and objects. An example is a pelvic ultrasound which is used for monitoring pregnancies, check for causes of abnormal bleeding or pain, or controlling intrauterine devices. When the doctor performs a pelvic ultrasound placing a transducer on the stomach or in the vagina, the medical images and related relevant information are presented on a screen. The physician must constantly turn his face to alternatively look at the patient where he is manipulating the instrument and the screen. The use of mixed reality in this procedure would allow to project ultrasound imaging data on the patient’s body.
Another good example are image-guided punctures performed during a variety of clinical procedures and surgeries, where accuracy is critical. Again, in these procedures, the practitioner needs to divide attention between the hand holding the puncture needle and a screen showing relevant medical images: another suitable scenario for the application of mixed reality. Table 2 summarizes different experiences where virtual reality technologies, and especially mixed reality, have been used to assist at punctures performed during clinical and surgical procedures in real-time practice in surgery or in phantom experiments.
Table 2.
Reference | Type of virtual reality technologies | Objective |
---|---|---|
Lin et al. [12] | Virtual reality | Improving the lateral ventricle puncture skill which is an important step of external ventricular drain |
Zhou et al. [13] | Mixed reality | Facilitate brachytherapy, which is an effective method for curing lung cancer, where the accuracy of needle positioning is vital and can influence the treatment effect |
Uhl et al. [14] | Mixed reality | Facilitating the puncture of the common femoral artery in a phantom model |
Li et al. [15] | Mixed reality, assisted by HoloLens glasses | Improving accuracy in respiratory liver tumor punctures in radiofrequency ablation (RFA) |
Porpiglia et al. [16] | Mixed reality | Evaluating the feasibility of three-dimensional mixed reality holograms in establishing the access point and guiding the needle during percutaneous kidney puncture when performing percutaneous nephrolithotomy |
Demerath et al. [17] | Augmented/mixed reality | Comparing the accuracy of augmented/mixed reality–guided versus frame-based stereotaxy-guided freehand drain placement in an intracerebral hemorrhage phantom model |
Park et al. [18] | Mixed reality, assisted by HoloLens glasses | Improving procedural efficiency and reducing radiation dose for CT-guided lesion targeting in an abdominal phantom study |
Davrieux et al. [19] | Mixed reality | Comparing ultrasound-guided needle punctures with the use of a mixed reality navigation system for percutaneous punctures |
Jing et al. [20] | Virtual reality | Improving pediatric lumbar puncture clinical skill training of medical students through a virtual reality lumbar puncture model vs. the traditional rubber lumbar puncture model |
Vrillon et al. [21] | Virtual reality | Improving the training of the practice of lumbar punctures through a three-dimensional video |
Schneideret al [22] | Augmented reality, assisted by HoloLens glasses | Improving ventricular drain (a common neurosurgical procedure), providing guidance for ventriculostomy in a custom-made 3D-printed head model |
Su et al. [23] | Virtual reality | Improving the implementation of lateral ventricle puncture operation through a haptic-based virtual reality simulator |
Ulh et al. [24] | Mixed reality | Facilitating the puncture of the common femoral artery in a phantom model |
Pratt et a.l [25] | Augmented reality, assisted by HoloLens glasses | Precise and efficient localization of perforating vessels during extremity reconstruction surgery using 3D vascular models |
Tai et al. [26] | Virtual reality | Implementing a renal puncture surgical simulator to improve percutaneous nephrolithotomy surgery for renal calculi |
Objective
Virtual reality technologies open new and exciting possibilities in healthcare. Among virtual reality technologies, mixed reality outstands going one step beyond blurring boundaries between the real world and the virtual environment and allowing the user to interact with both the physical and the virtual objects, which become practically indistinguishable.
The objective of this work is to present a use case where mixed reality is used to assist physicians in quite complex interventional procedures, and specifically, in real-time ultrasound-guided femoral arterial cannulations. This use case should be taken as a good example of how virtual reality technologies can successfully support practitioners during clinical and surgical procedures.
Interventional cardiology is probably one of the most advanced fields in terms of minimally invasive approaches to complex cardiac diseases and the professionals are used to work with angiographic, echocardiographic, and computed tomography–based images simultaneously. Therefore, the additive value of mixed reality in this field is particularly relevant to increase situational awareness, integrate multimodal data, and provide hands-free system interfaces.
Materials and Methods
Method Overview
In this work, a mixed reality system, based in the HoloLens 2 glasses, has been developed to assist practitioners during real-time ultrasound-guided femoral arterial cannulations. In this surgical procedure, the doctor must look at a screen placed over the patient while holding an ultrasound scanner and a puncture needle. Although it is a common procedure, it involves a specific technique that requires challenging skills that must be mastered by practitioners in the field, and for this reason, this procedure may have difficulties for less experienced surgeons or operators. The main difficulty lies in the visualization of medical images and relevant related information and relative positioning of the needle used for the puncture in relation to the blood vessel. To fulfill these purposes, and to provide a helpful system, a mixed reality solution has been developed to assist the doctor during this surgical procedure, allowing to improve accuracy, and consequently reducing the time and cost of interventions, as well as the potential associated risks [27].
The physicians using the system participated in its design to produce a system adapted to their needs and requirements, taking into consideration working conditions at hospital, and the practitioners’ skills and attitude.
The main requirements posed by the operator and medical management staff can be summarized as follows:
Cost-effective enough to be integrated in daily practice in surgery and procedures in public hospitals.
Simple enough to be used by practitioners with extended experience in the procedure, i.e., femoral arterial cannulations, but with average technological skills.
Reliable enough to be used in real-time surgeries and interventions, allowing surgeons and operators to increase accuracy when performing punctures in femoral arterial, as well as to reduce time consumed in the procedure.
Taking into account the above premises, a system divided into two modules was proposed, each of them with specific objectives: (1) transmitter module, which is responsible for sending medical images to HoloLens 2 and (2) receiver module, which is hosted in the HoloLens 2 and is responsible for rendering those medical images, placing them on a specific space location, and allowing the doctor to see and manage them in a 3D environment. These two main modules and their associated tasks are explained below.
Transmitter Module
This module is in the machine in charge of hosting the video transmission application. The main tasks of this module are the following:
Collecting a video stream of medical signals of interest.
Processing the video stream collected to make it suitable for the application.
Encapsulating and transmitting the video frames.
Receiver Module
This module, located in the mixed reality HoloLens 2 glasses, hosts the video reception application. Its tasks are divided into:
Decoding the flow of incoming frames.
Detecting and tracking an object of interest. In this case, a QR code is utilized as a straightforward and dependable target for tracking purposes across various objects, eliminating the necessity for training a model to track specific objects of interest. In this particular scenario, when employing an ultrasound scanner, the QR code will be affixed onto it for the tracking.
Projecting the video stream onto the object of interest.
Allowing the doctor to interact with the video projection (zoom in, zoom out, move it, close it).
Materials: Hardware and Software Overview
Different hardware and software elements have been used for the transmitter and receiver modules:
Hardware
- Desktop computer for processing and programming 3D model virtualization.
- Model: HP OMEN 875-1029NS.
- CPU: Intel Core i9-9900K.
- RAM: HyperX 64 GB DDR4-2666 SDRAM (4 x 16 GB).
- GPU: NVIDIA GeForce RTX 2080Ti (11GB dedicated GDDR6).
HoloLens 2 Mixed Reality device: As a visualization system for 3D modelling.
Router 4G LTE Wi-Fi TP-Link TL-MR6400: As a common access point to the network for the communication between the computer and the HoloLens 2.
Video capture card AVerMedia Live Gamer Mini 1080p 60fps: For the extraction and adaptation of images from different devices used in medical practice, that must be processed on the computer.
Ultrasound scanner.
Software
Visual Studio Code: IDE (Integrated Development Environment) for development and debugging code.
Unity 2019.4.xf1: To visualize and configure the applications developed in the IDE.
Unity packages: FMTEP Stream and AVPro Live. The first for the video transmission and reception service, and the second for projecting video onto a 3D object.
Vuforia Engine: As Mixed Reality tool for object tracking (QR code in this application).
System Explanation
For a better understanding, Fig. 1 illustrates how the different elements in the hardware system are connected.
First, the ultrasound scanner extracts images from the patient, which can be sent through HDMI (high-definition multimedia interface), if necessary, an adapter can be used for the output. In this research, a tablet with a commercial application has been used for image conditioning for the capture machine. Then, the video capture machine collects these images through the HDMI IN port and transmits them to the computer through the HDMI OUT port; the computer processes these images, encapsulates them in a UDP (User Datagram Protocol) datagram, and transmits them via broadcast on the network shared with the HoloLens 2. Finally, the HoloLens 2 receives the stream of UDP datagrams, decapsulates them, and renders them on a 3D object.
Once the hardware system has been explained, it is time to describe the two main modules that host the software applications: transmitter module and receiver module.
The transmitter module is hosted in the computer. This module is responsible for the following tasks: receiving the flow of images (video stream) from the video capture device and the ultrasound scanner, processing those images to encapsulate them in UDP datagrams, and sending them broadcast to its local network so that they can be consumed by each web client connected to that network. To achieve this, the computer runs a Unity application which includes two main packages or assets called AVPro Live and FMTEP. AVPro Live allows the application to detect HDMI input video stream and render it onto a plane surface inside Unity 3D scene. Thus, FMTEP uses that plane surface to capture the video stream and encapsulate it in UDP datagrams to send it through the local network. To sum up, AVPro Live asset allows the application to detect and insert video from video capture card to FMTEP encapsulate and transmission system as shown in Fig. 2.
Once the computer sends the transmission, Router 4G must read the destination of the UDP datagrams, which is broadcast. That means the destination is each device connected to the network provided by Router 4G. That allows HoloLens 2 (receiver) to easily detect and extract the transmission without manual IP configuration, because every time a device leaves the local network, its IP is available for other devices (that also happens to HoloLens 2, which is the main destination of this system). For this reason, a configuration of the transmitter destination IP is needed every time the HoloLens 2 leaves the local network. So, to avoid manual configurations and to build an easy plug and play system, the destination will be broadcast (anyway, manual IP is available if the user wants to change the default configuration).
Finally, the receiver module has several functions to perform in parallel. On the one hand, it must render the content of the UDP datagrams in a plane, and on the other hand, it must detect a QR code and position the plane on it as shown in Fig. 3. The doctor can then place the QR code in an area of interest on which the image will be projected. An area of interest could be, for example, the ultrasound scanner. Thus, the ultrasound images could be presented right above the patient and follow the ultrasound scanner using QR tracking. Furthermore, the doctors can lock the position of the plane (thereby terminating the tracking) by utilizing a button situated at the top of the window. This enables them to conveniently reposition the plane to a more suitable location when needed, simply by grabbing and moving it manually. Additionally, they can reactivate the tracking functionality by pressing the same button.
Another consideration is that, although real-time streaming is necessary for this medical practice, it is important to find a balance between resolution and delay depending on the procedure’s significance. In this context, real-time capability is of great relevance, as any latency, such as screen freezing or delays, could potentially hinder the doctor’s ability to effectively perform the puncture and compromise patient safety.
To achieve this goal, UDP is selected. UDP is a lightweight and fast protocol, and base64 encoding is used to simplify image data transfer. However, higher resolution requires more data to be transmitted, which compromises the stability of the connection for applications where resolution is crucial. To meet this challenge and achieve a near real-time streaming experience, it is crucial to keep the latency near to 200 ms for any given image resolution. In this case, for a 640 × 480 image resolution (which turned out to be sufficient for the medical practice of arterial puncture), an average latency of 200 ms is achieved, with occasional outliers reaching around 250 ms, primarily due to wireless connectivity issues.
It is crucial to incorporate efficient logic for encoding, transmitting, decoding, and rendering data in order to minimize latency, considering the potential inconsistency of connection integrity. To achieve a more efficient system, it is highly recommended to leverage the computational power of a computer or laptop instead of burdening the HoloLens 2 with processing requirements.
Validation: Pilot Study
The developed mixed reality system, based on HoloLens 2 glasses, was validated as a proof of concept in a pilot study conducted in a tertiary public hospital in central Spain. A single use case was selected to focus validation on a single task: the system was employed to facilitate ultrasound-guided femoral arterial cannulation in interventional cardiology procedures. In the period between November 2021 and August 2022, 9 patients with severe aortic stenosis selected for trascatheter aortic valve replacement (TAVR), with planned femoral access in the pre-intervention CT-Scan, were included in the study. Two different practitioners carried out the interventions (6 and 3 interventions each).
This pilot study was reviewed and approved by the local Ethics Committee (Comité de Ética de la Investigación con Medicamentos del Área de Salud Valladolid Este) with code PI-21–2183. Ethical and regulatory concerns have been carefully considered: although the system is only a visualization component (it does not process any medical information, it only presents a video stream in a more convenient way than a conventional 2D screen), there are critical aspects in the interaction with the system (such as lag or image aspect ratio) that could theoretically influence the medical practice in specific applications. For this reason, the system cannot currently be employed outside a regulated research project.
Results
Figure 4 shows a picture of the transmitter module setup and different moments of the use of the system in real-time ultrasound-guided femoral arterial cannulations in interventional cardiology procedures. This picture shows how helpful the system can be in a real-life environment.
As shown in Fig. 4 top-right quadrant, the doctor may want to manually handle the plane. In this scene, the doctor fixes the plane over the patient. According to the requirements of this system established by the practitioners, some situations need reallocating the plane on different places. This feature may be helpful to clean the view during an intervention or to allocate the plane in a more comfortable place for the doctor depending on different factors such as the scenario illumination and textures of the background.
Furthermore, Fig. 4 bottom-right and -left quadrants show two different situations in which the interventional cardiologist follows the puncture and the guide used to proceed in the safest possible manner while using the system. As shown, the cardiologist can operate looking at the point where the puncture is performed, and at the ultrasound scan, simultaneously, reducing the risks and time of the intervention. Thus, according to the practitioners’ report, the possibility of operating looking at the point of interest is more intuitive and comfortable than the old method, which was based on a screen located vertically above the patient, that forced the doctor to look away from the operating table and the patient, as shown in the top-right quadrant of Fig. 4.
As a proof of concept, the main objective of the study was to demonstrate the feasibility of the approach. Nine out of 9 punctures were considered successful: no need to re-access, no recorded adverse events associated with puncture (bleeding). After the experience, the two participating interventional cardiologists rated the experience according to 5 subjective parameters, presented in Table 3.
Table 3.
Parameter | Average score |
---|---|
Ease of use | 4.5 |
Responsiveness | 4.5 |
Image quality | 4 |
Lack of technological burden | 4.5 |
Headset comfort | 3.5 |
In the post-experience interview, operators mentioned two main limitations:
Low image contrast in a brightly illuminated environment.
Headset could be uncomfortable to wear for long periods of time (such as entire shifts in the cath lab).
Discussion
The practitioners using the mixed reality system, based in HoloLens 2 glasses, also participated in a focus group-based SWOT (strengths, weaknesses, opportunities, and threats) qualitative analysis. In this analysis, related medical management staff were also involved. Main conclusions are summarized in Table 4.
Table 4.
Strengths |
Easy to use Reliable Real-time Reachable Cost-effective |
Opportunities |
Reduction of operating time Better control of typical errors associated to the procedure Medical information and images produced can be broadcasted and practitioners can be assisted by more expert colleagues Medical information and images produced can be broadcasted and used for training purposes with interns or novice practitioners integrated in a ubiquitous e-learning system |
Weaknesses |
Digital divide for practitioners used to more classical surgical procedures and techniques not involving the use of virtual reality technologies |
Threats |
Some practitioners are afraid of potential system errors or of the learning curve of this new approach |
The thoughts of the users revealed by the SWOT analysis confirmed their perception of that their requirements have been met.
As main strengths, practitioners highlighted the possibility to be used during real-time practice in interventions, as well as its reliability, as the system allowed them to perform punctures in femoral arterial cannulations with high accuracy. Moreover, practitioners confirmed that the system has been easy to use and helpful. Medical management staff added that the system was reachable for the public health system as it can be deployed and managed at a reasonable cost.
Regarding the opportunities, practitioners believe that its use in interventional cardiology can reduce operating times and allow a better control of the procedure, especially of the typical mistakes made by interns and novice doctors. Operators involved in the training of medical students also highlighted that the system produce real-time valuable medical imagery that could be used for training purposes in the context of a ubiquitous e-learning system, even with students abroad. The possibility to broadcast real-time medical imagery would also open the possibility to ask for advice of more expert colleagues, even from other hospitals, during interventions, which can be especially useful for less experienced practitioners, or in the case of complex cases.
Regarding the weaknesses and considering the possibility of using the system in other hospitals, the practitioners suggested that interventional cardiologists used to more classical procedures, not involving the use of virtual reality technologies, may have difficulties adopting the new system. They highlighted that there was a digital divide affecting mainly to older practitioners, and remarked that, in their case, they have both volunteered to use the system. Related to this fear, practitioners suggested as a threat, which some colleagues may be afraid of potential system errors or of the learning curve of this new approach.
Conclusion
In conclusion, the mixed reality system developed, based in HoloLens 2 glasses, is reliable and simple enough to be used by practitioners during real-time practice in interventional cardiology. This was confirmed by practitioners’ feedback during a qualitative analysis of the system. The system has been used to assist physicians in femoral arterial cannulations in up to 9 interventions in a large public hospital in central Spain since November 2021. The system is also cost-effective enough to be incorporated into the public health system. As a summary, the proposed system is a realistic proof of concept of how virtual reality technologies, and more specifically mixed reality, can successfully support practitioners during clinical and surgical procedures.
QR code has proven to be an effective method for tracking various objects in well illuminated scenarios like surgery rooms. One potential direction for future investigation would be to explore object tracking utilizing other HoloLens 2 sensors, such as LiDAR (light detection and ranging), and the potential opportunities it presents for low light conditions. In addition, the main line of research open in order to transfer the prototype into clinical practice is to conduct a randomized clinical trial with a twofold purpose:
• First, to quantitatively characterize the potential clinical advantages for the patient and the practitioner (shorter intervention times and radiation).
• To obtain solid scientific evidence on which to base future potential certification claims to regulatory authorities.
Author Contribution
Conceptualization and methodology, J.M.A.P. and C.B.Z.; formal analysis and investigation, M.A.F.; writing, review, and editing, M.A.F and M.A.P.J.; supervision, J.P.O. and I.J.A.S. All authors read and approved the final manuscript.
Funding
Open Access funding provided thanks to the CRUE-CSIC agreement with Springer Nature. This work has received funding from Gerencia Regional de Salud de Castilla y León (grant number GRS 2275/A/2020) and Instituto de Salud Carlos III (grant number DTS21/00158).
Data Availability
There are no new data associated with this article.
Declarations
Competing interests
The authors declare no competing interests.
Ethics Approval
Approval for this study was obtained from the local ethics committee “Comité de Ética de la investigación con Medicamentos Área de Salud Valladolid Este” in Valladolid (Spain), at its meeting on February 18, 2021.
Conflict of Interest
The authors declare no competing interests.
Footnotes
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Cipresso P, Chicchi-Giglioli IA, Alcañiz-Raya M, et al. The past, present, and future of virtual and augmented reality research: a network and cluster analysis of the literature. Front Psychol. 2018;9:2086. doi: 10.3389/fpsyg.2018.02086. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Chen Y, Wang Q, Chen H, et al. An overview of augmented reality technology. J Phys: Conf Ser. 2019;1237:0220822. [Google Scholar]
- 3.Egliston B, Carter M. The interface of the future’: mixed reality, intimate data and imagined temporalities. Big Data Soc. 2022;9:1. doi: 10.1177/20539517211063689. [DOI] [Google Scholar]
- 4.Gao Y, Chang C, Yu X, et al. A VR-based volumetric medical image segmentation and visualization system with natural human interaction. Virtual Real. 2022;26:415–424. doi: 10.1007/s10055-021-00577-4. [DOI] [Google Scholar]
- 5.Fong KNK, Tang YM, Sie K, et al. Task-specific virtual reality training on hemiparetic upper extremity in patients with stroke. Virtual Real. 2022;26:453–464. doi: 10.1007/s10055-021-00583-6. [DOI] [Google Scholar]
- 6.VanHorn K, Çobanoğlu MC. Democratizing AI in biomedical image classification using virtual reality. Virtual Real. 2022;26:159–171. doi: 10.1007/s10055-021-00550-1. [DOI] [Google Scholar]
- 7.Garcia-Hernandez N, Guzman-Alvarado M, Parra-Vega V. Virtual body representation for rehabilitation influences on motor performance of cerebral palsy children. Virtual Real. 2021;25:669–680. doi: 10.1007/s10055-020-00481-3. [DOI] [Google Scholar]
- 8.Heinrich C, Cook M, Langlotz T, et al. My hands? Importance of personalised virtual hands in a neurorehabilitation scenario. Virtual Real. 2021;25:313–330. doi: 10.1007/s10055-020-00456-4. [DOI] [Google Scholar]
- 9.Houzangbe S, Christmann O, Gorisse G, et al. Effects of voluntary heart rate control on user engagement and agency in a virtual reality game. Virtual Real. 2020;24:665–681. doi: 10.1007/s10055-020-00429-7. [DOI] [Google Scholar]
- 10.Liang C-J, Start C, Boley H, et al. Enhancing stroke assessment simulation experience in clinical training using augmented reality. Virtual Real. 2021;25:575–584. doi: 10.1007/s10055-020-00475-1. [DOI] [Google Scholar]
- 11.Medellin-Castillo HI, Zaragoza-Siqueiros J, Govea-Valladares EH, et al. Haptic-enabled virtual training in orthognathic surgery. Virtual Real. 2021;25:53–67. doi: 10.1007/s10055-020-00438-6. [DOI] [Google Scholar]
- 12.Lin W, Zhu Z, He B, et al. A novel virtual reality simulation training system with haptic feedback for improving lateral ventricle puncture skill. Virtual Real. 2022;26:399–411. doi: 10.1007/s10055-021-00578-3. [DOI] [Google Scholar]
- 13.Zhou Z, Jiang S, Yang Z, et al. Surgical navigation system for brachytherapy based on mixed reality using a novel stereo registration method. Virtual Real. 2021;25:975–984. doi: 10.1007/s10055-021-00503-8. [DOI] [Google Scholar]
- 14.Uhl C, Hatzl J, Meisenbacher K, et al. Mixed-reality-assisted puncture of the common femoral artery in a phantom model. J Imaging. 2022;8(2):47. doi: 10.3390/jimaging8020047. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Li R, Si W, Liao X, et al. Mixed reality based respiratory liver tumor puncture navigation. Comput Vis Media. 2019;5:363–374. doi: 10.1007/s41095-019-0156-x. [DOI] [Google Scholar]
- 16.Porpiglia F, Checcuccia E, Amparore D, et al. Percutaneous kidney puncture with three-dimensional mixed-reality hologram guidance: from preoperative planning to intraoperative navigation. Eur Urol. 2022;81(6):588–597. doi: 10.1016/j.eururo.2021.10.023. [DOI] [PubMed] [Google Scholar]
- 17.Demerath T, Stanicki A, Roelz R, et al. Accuracy of augmented reality-guided drainage versus stereotactic and conventional puncture in an intracerebral hemorrhage phantom model. J NeuroInterv Surg. 2022 doi: 10.1136/neurintsurg-2022-018678. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Park BJ, Hunt SJ, Nadolski GJ, et al. Augmented reality improves procedural efficiency and reduces radiation dose for CT-guided lesion targeting: a phantom study using HoloLens 2. Sci Rep. 2020;10:18620. doi: 10.1038/s41598-020-75676-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Davrieux CF, Giménez ME, González CA, et al. Mixed reality navigation system for ultrasound-guided percutaneous punctures: a pre-clinical evaluation. Surg Endosc. 2019;34(1):226–230. doi: 10.1007/s00464-019-06755-5.pmid:30911919. [DOI] [PubMed] [Google Scholar]
- 20.Jing D, Jing L, Yihan X, et al. The application of the virtual reality in training skill of pediatric lumbar puncture by eight-year program students in medical college. Chin J Med Educ. 2020;40(11):913–916. doi: 10.3760/cma.j.cn115259-20200209-00117. [DOI] [Google Scholar]
- 21.Vrillon A, Gonzales-Marabal L, Ceccaldi PF, et al. Using virtual reality in lumbar puncture training improves students learning experience. BMC Medical Educ. 2022;22:244. doi: 10.1186/s12909-022-03317-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Schneider M, Kunz C, Pal’a A, et al. Augmented reality–assisted ventriculostomy. J Neurosurg. 2020;50(1):E16. doi: 10.3171/2020.10.FOCUS20779. [DOI] [PubMed] [Google Scholar]
- 23.Su XH, Deng Z, He BW, et al. Haptic-based virtual reality simulator for lateral ventricle puncture operation. Int J Med Robot Comput Asist Surg. 2020;16(6):1–10. doi: 10.1002/rcs.2176. [DOI] [PubMed] [Google Scholar]
- 24.Uhl C, Hatzl J, Meisenbacher K et al (2022) Mixed-reality-assisted puncture of the common femoral artery in a phantom model. J Imag 8(2):47. 10.3390/jimaging8020047 [DOI] [PMC free article] [PubMed]
- 25.Pratt P, Ives M, Lawton G, et al. Through the HoloLens™ looking glass: augmented reality for extremity reconstruction surgery using 3D vascular models with perforating vessels. Eur Radiol Exp. 2018;2:2. doi: 10.1186/s41747-017-0033-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Tai Y, Wei L, Zhou H et al (2016) Integrating virtual reality and haptics for renal puncture surgical simulator. 2nd IEEE International Conference on Computer and Communications (ICCC), Chengdu, pp. 480–484. 10.1109/CompComm.2016.7924747
- 27.Seto AH, Abu-Fadel MS, Sparling JM, et al. Real-time ultrasound guidance facilitates femoral arterial access and reduces vascular complications: FAUST (Femoral Arterial Access With Ultrasound Trial) JACC Cardiovasc Interv. 2010;3(7):751–758. doi: 10.1016/j.jcin.2010.04.015. [DOI] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
There are no new data associated with this article.