Skip to main content
Springer logoLink to Springer
. 2022 Nov 15;47(2):495–501. doi: 10.1007/s00264-022-05619-3

Next step trauma and orthopaedic surgery: integration of augmented reality for reduction and nail implantation of tibial fractures

Tim Klopfer 1, Thomas Notheisen 2, Heiko Baumgartner 2,, Dorien Schneidmueller 3, Ryan Giordmaina 4, Tina Histing 2, Christoph Emanuel Gonser 2
PMCID: PMC9877081  PMID: 36378324

Abstract

Introduction

There is a tremendous scope of hardware and software development going on in augmented reality (AR), also in trauma and orthopaedic surgery. However, there are only a few systems available for intra-operative 3D imaging and guidance, most of them rely on peri- and intra-operative X-ray imaging. Especially in complex situations such as pelvic surgery or multifragmentary multilevel fractures, intra-operative 3D imaging and implant tracking systems have proven to be of great advantage for the outcome of the surgery and can help reduce X-ray exposure, at least for the surgical team (Ochs et al. in Injury 41:1297 1305, 2010). Yet, the current systems do not provide the ability to have a dynamic live view from the perspective of the surgeon. Our study describes a prototype AR-based system for live tracking which does not rely on X-rays.

Materials and methods

A protype live-view intra-operative guidance system using an AR head-mounted device (HMD) was developed and tested on the implantation of a medullary nail in a tibia fracture model. Software algorithms that allow live view and tracking of the implant, fracture fragments and soft tissue without the intra-operative use of X-rays were derived.

Results

The implantation of a medullar tibia nail is possible while only relying on AR-guidance and live view without the intra-operative use of X-rays.

Conclusions

The current paper describes a feasibility study with a prototype of an intra-operative dynamic live tracking and imaging system that does not require intra-operative use of X-rays and dynamically adjust to the perspective of the surgeons due to an AR HMD. To our knowledge, the current literature does not describe any similar systems. This could be the next step in surgical imaging and education and a promising way to improve patient care.

Keywords: Intra-operative imaging, Intra-operative 3D imaging, Augmented reality in surgery, Navigation, Head-mounted display

Introduction

There is a tremendous scope of hardware and software development going on in augmented reality (AR). This development also spreads into medicine; however, there are only a few established applications in use so far, especially as for trauma surgery. This also applies to intra-operative imaging and guidance. There are some systems available for intra-operative 3D imaging and guidance, most of them rely on peri- or intra-operative imaging deploying X-rays. Trauma surgeons are often faced with complex situations which arise intra-operatively, might it be in emergency procedures or complex elective cases. Especially in these complex situations such as pelvic surgery or multifragmentary multilevel fractures, intra-operative 3D imaging and implant tracking systems have proven to be of great advantage for the outcome of the surgery and can help reduce X-ray exposure, at least for the surgical team [1]. Moreover, intra-operative dynamic life tracking and imaging can help improve and facilitate surgical education. However, there is still space for improvements and further development. Although currently available imaging devices deliver excellent static 3D images. Yet, these methods do not provide the ability to have a dynamic live view from the perspective of the surgeon. In complex situations such as multifragmentary fractures the fit of the implant and the fragments must be determined intra-operatively; this is mostly done by X-ray imaging, regardless of 2D or 3D. Moreover, these methods do not show the perspective as the surgeon sees the fracture site.

To our mind, intra-operative AR-enhanced imaging using overlays of pre-operatively gathered CT or MRI scans on to the site greatly improves the knowledge of the orientation of the different fragments and the implant. This can lead to reduced X-ray exposure and faster procedures with better results.

The aim of this study was to develop a system that allows an intra-operative overlay of different layers of the site starting with the skin up to the bone using a head mounted display at the example of closed reduction and nailing of a tibial fracture. Moreover, we examined if AR technologies can help to make surgical procedures safer and deepen the understanding of different intra-operative situations. We wanted to investigate if an AR device can be used under real-world conditions in the OR and be controlled with a virtual user interface while wearing sterile surgical gloves. Moreover, we wanted to examine if closed reduction and implant placement is possible without the use of X-rays just relying on the AR guidance und real-world conditions in the OR using a realistic sawbone model.

Materials and methods

For the prototype, we used the latest generation AR-HMD of Microsoft’s HoloLens (HoloLens 2, Microsoft Cooperation, Redmond, WA, USA). This device fulfills all the stated prerequisites. The conceptual idea and as well as the model were developed by the authors, additional coding support was provided by AmbiGate Motion Sensing GmbH, Tuebingen, Germany. First of all, a model consisting of an artificial tibia bone with a multifragmentary fracture and artificial soft tissue (Sawbones, A Pacific Research Company, Vashon Island, Washington, USA) forming a model of a lower leg was fitted with an external fixator (DepuySynthes, Umkirch, Germany).

The tibia fracture was chosen for the following reasons:

  • Tibia fractures are rather common and closed reduction and nail osteosyntheses is an established treatment [2].

  • The successful treatment depends on the exact entry point and correct reposition as well as axis reconstruction [3].

  • Closed reduction and insertion of the nail into the core canal of the distal fragment can be quite demanding and require several tries and multiple X-ray images resulting in a high X-ray exposure for the surgical team and patient [4].

In this situation, the AR augmented live tracking can provide significant advantages. To be able to use this technique intra-operatively, we defined the following prerequisites:

  1. The AR glasses must not disturb or restrict the surgeon.

  2. Sterility must absolutely be guaranteed.

  3. The system must be easy to use with a clean user interface (UI).

  4. Direct visualization of the positional relation of the different fragments with high precision and low latency of the overlay CT/MRI data.

  5. The system must allow customizable 3D visualization with adjustable views as for zoom, rotation, position, and transparency.

  6. The surgeon should be able to select different layers of the segmented CT/MRI.

  7. Simultaneous view of the AR Phantom with bone and implant overlayed on the site.

  8. Live view of the reduction of the bone fragments.

  9. Implementation of recording and casting functions (audio, video).

This model (see Fig. 1) was then scanned with a 64-line CT scanner (Siemens Health Care, Erlangen, Germany) using a protocol with 120 kV, 90 mAs, 0.75 mm slice thickness. In a second step, a regular tibia nail including the aiming arm (DepuySynthes, Umkirch, Germany) which is also used in clinical routine, was scanned with the same protocol.

Fig. 1.

Fig. 1

The model consisting of a sawbone lower leg with soft tissue and an external fixator with attached QR codes

These data were than segmented using Mimics Innovation suite (Materialise GmbH, Munich, Germany). This provided separate high-quality three-dimensional image of all relevant structures:

  1. Tibia nail

  2. Aiming arm

  3. External fixator

  4. Soft tissues and skin

  5. Bone

In order to project the data on the site, a user interface (UI) was developed which allow to live-visualize the implantation of the tibia nail as a hologram of the segmented data in all planes in space. Figure 2 shows part of the virtual user interface. We used QR codes which were applied to the external fixator and the implant (Fig. 3) to track these parts. We developed an algorithm that allowed the HoloLens2 AR-HMD to precisely locate the marked elements and visualize them in physical space and project the overlay to the site without the use of additional X-ray imaging. The algorithm was refined to optimized precision and latency and to offer additional virtualization options:

Fig. 2.

Fig. 2

The virtual user interface with sliders to adjust the opacity of each layer seen in the lower right corner

Fig. 3.

Fig. 3

QR codes used to identify the different parts of the model

With the prototype finished to the abovementioned specifications, the following scenarios were tested in the OR using the phantom:

  • Use of the system under real life conditions in the operating room: The phantom was placed on an operating table, sterile draping was applied after disinfection, sterile surgical gowns were used to assess the impact of the AR-HMD on the surgeon and to test operationality of the system and the UI.

  • The quality of the reduction and the implant position were visually assessed using a medial incision on the model (Fig. 6).

  • Closed reduction and nail insertion relying only on the AR-HMD were tested.

  • The accuracy of the overlay was evaluated.

  • The usability and operationality of the system regarding OR lighting conditions.

Fig. 6.

Fig. 6

Medial incision to manually check the reduction and implant placement

The tests were performed by four surgeons in a set of 5 runs a day using a rotation system so that each surgeon had a break between the individual implantations to reduce a habituation effect. Altogether, 40 runs were performed. Figures 4 and 5 show the intra-operative use of the system.

Fig. 4.

Fig. 4

Intra-operative use of the system with the surgeon performing different gestures to zoom/rotate the virtual model in the physical space

Fig. 5.

Fig. 5

Intra-operative use of the system with the surgeon performing different gestures to zoom/rotate the virtual model in the physical space

Results

While using the prototype under OR conditions, the user interface (UI) was easy and comfortable to use with an intuitive control set.

The correct insertion of the nail into the tibia using only AR guidance was achieved reproducibly. In all 40 test implantations, a correct implant position with an anatomical reduction was achieved relying only on the AR navigation (Fig. 6).

The system showed some deviation of the overlay of the phantom to the site under severe head movements of the surgeons, these could be corrected by recalibrating the system intra-operatively by scanning the QR codes.

The three-dimensional visualization and overlays were deemed excellent by all four users. All four users stated that the AR-augmented visualization delivered valuable information and facilitated the process. Figure 7 shows a sample image of an overlay on the site with the user interface of the AR-HMD.

Fig. 7.

Fig. 7

Image of the overlay of the implant and bones of the virtual model on the real site

None of the test users found the system to negatively affect the surgeon’s comfort, sterility, or liberty of movement.

The visualization options were assessed excellent by the test users. However, during test implantation, several cases of loss of the tracking of the nail occurred. This happened when the QR codes of the aiming arm disappeared from the field of view and is attributed to the movement of the head of the surgeon.

The changing lighting conditions in the OR with high contrast scenarios did not cause any problems; however, two cases of reduced picture quality of the HMD were reported when the HMD was in the direct ray of light of a surgical light.

Last, the added functionality of showing the virtual OP Site of the phantom in any random position in space with any zoom level as a hologram was tested. This functionality was assessed as very useful by three surgeons, one surgeon assessed it as useful.

Due to the wireless construction of the HMD, sterility was not compromised in the usage scenario. The tracking of the surgeon’s hand and the operation of the virtual user interface were not negatively affected using sterile surgical gloves.

To check how easy and intuitively the system is operable, we defined the core prerequisites:

  1. The AR glasses must not disturb or restrict the surgeon.

  2. Sterility must absolutely be guaranteed.

  3. The system must be easy to use with a clean user interface (UI).

  4. Direct visualization of the positional relation of the different fragments with high precision and low latency of the overlay CT/MRI data.

  5. The system must allow customizable 3D visualization with adjustable views as for zoom, rotation, position, and transparency.

  6. The surgeon should be able to select different layers of the segmented CT/MRI.

  7. Simultaneous view of the AR Phantom with bone and implant overlayed on the site.

  8. Live view of the reduction of the bone fragments.

  9. Implementation of recording and casting functions (audio, video).

All of the abovementioned prerequisites were met with the exemption of prerequisite 4: several loss of tracking incidences occurred under severe head movements of the surgeon and precision of the overlay was reduced during very fast head movements.

Discussion

Up to date, to our knowledge, no other comparable feasibility study for the AR guided implantation of a tibia nail in a real-world scenario [567] although fractures of the lower leg are commonly found in all age clusters [2]. AR integration in surgical procedures still seems to be rare, although we think there is a tremendous potential to be unleashed. Some workgroups investigated the implantation of K-wires into the pelvis or pedicles of the vertebra. In a cadaver model, Wang et al. could show that placing IS-Joint screws with the help of AR is possible with minimal deviation and without lesions to vessels or nerval structures [8]. Ochs and Gonser demonstrated that with the help of three-dimensional navigation, radiation exposure could be reduced [9]. A reduction in radiation dose is very important, since especially in these procedures, radiation doses experienced by the surgical team are high [10]. Studies combining AR and a conventional C-arm X-ray device showed promising results as for the reduction X-ray exposure and procedure duration [11, 12]. This was also found by Fischer when placing K-wires into different models [13]. Gibby et al. used the HoloLens1 for pedicle screw placement and found only minimal misalignment compared to the placement with X-ray support. Liu et al. found only minimal deviation of an AR-guided placement of a K-wire for hip-resurfacing [14].

Hybrid models such as the integration of C-Arm images and sonographically acquired images have been tried as well: Hajek et al. have used a HoloLens and a C-Arm for percutaneous interventions [15]. Heide et al. could show that X-ray exposure could be reduced by 46% by using such a hybrid method [16].

All these findings concur with our findings showing that a high level of precision can be reached with AR-guided implant placement, even without the use of X-rays.

So far, we can conclude that there are several approaches to the use of AR in surgery. These approaches are on different evolutional levels. The described prototype opens a new perspective by using AR to guide the implantation of a tibia nail without the intra-operative use of X-ray.

To our opinion, a key factor for clinical use of AR technology in the OR is the ease of use. The described prototype was easy to use for all testers. One future development should be the implementation of smaller QR markers or even a markerless system—further development is necessary. Further use-scenarios can be found in endoprosthetics, e.g., for planning and live-tracking of component placement. Reliable results for placement of the acetabular component have already been found by Ogawa [17]. Further enhancements in precision are to be expected with the evolution of technology and corrections algorithms. Oliveria et al. could already show that reliability and precision were enhanced with added algorithms to tackle latency and orientation in physical space [18].

The benefit of being able to intra-operatively adjust the views as for angle and zoom as well as other parameters was already shown by Gregory et al. when overlaying a CT scan and a three-dimensional planning sketch during implantation of an inverse shoulder prothesis [19]. Ochs and Gonser could show that visualization of different levels and reference points facilitates complex procedures [20]. We could show that the developed user interface was easy to use in the OR und real-life conditions without compromising on sterility or patient safety. We could show that a safe integration into the surgical workflow is possible. The UI was found easy to use and all the testers confirmed that the ability to intra-operatively navigate through the projected hologram of the fracture freely in physical space helped to gain better insight. Some reviews show an increasing interest in the use of AR-technology in surgery, however there are still problems to be solved, e.g., motion sickness or precision of the overlay. But apart from all these problems, even the currently available systems were found to be able to reduce the degree of difficulty and the X-ray exposure [6, 7]. A further advantage can be found in the ability to actively interact with other experts on the field which are not physically present through teleconsultation with the possibility to interact [21]. This is also applied to neurosurgery: Incekara et al. showed that focusing on a tumor areal was eased with the help of AR with high precision overlays [1]. For such highly complex procedures, we think, at least as phase of transition, the combination of several imaging sources and intra-operative imaging would help to enhance safety. Vessel and important neural structure could precisely be shown in the overlay and through the tracking of the instruments, a feedback would be possible if the surgeons approaches too closely to one of these structures. Solely AR-guided complex surgery is a future perspective.

We found a high level of precision of the AR device which enabled us to precisely implant a tibia nail. However, there were some limitations: There were some cases of loss of tracking. We attribute this to a to some extent limited angle of view of the built-in cameras of the HMD. The limited field of view of the built in-cameras also limited the ability of the system to recognize hand gestures. Future technological development will probably also address this. Another limitation of our system is the need to mark all the relevant surfaces with QR markers. So far, all other systems described used some way of markers as well, another way is using a pattern of reflective cue points. Our focus right now lies on the development of a markerless system with enhanced precision. For clinical use in humans, the use of a backup X-ray system to verify implant positions seems inevitable at the current point. The future aim is a system that can be used as standalone device without intra-operative X-ray; this would not only reduce X-ray exposure for the patient and the surgical team but also the number of devices required in the OR.

A markerless system would also allow to overcome another limitation of the present system: not all boney fragments were marked and thus not live tracked during surgery. In our model, all major fragments were tracked with the attached external fixator. A system which will be able to track at least all fragments bigger than 1 cm2 is in development.

In our opinion, the presented system provides possibilities of intra-operative planning and step by step verification that is unpreceded; also, there are some limitations which have to be tackled in future development.

Conclusions

The current paper describes a feasibility study with a prototype of an intra-operative dynamic live tracking and imaging system that does not require intra-operative use of X-rays and dynamically adjust to the perspective of the surgeons due to an AR head-mounted display (HMD). To our knowledge, this is the first description of such a system. Our study showed that that the implantation of a tibia nail in a real-life model is possible relying only on AR guidance. Excellent visualization of the pre-operatively gained image data was shown, the ability to free rotate the hologram in physical space and to zoom as well as the precise overlay onto the surgical site were highly rated.

We could prove that AR guidance works under real-life conditions in the OR without compromising sterility and that the system was easy to use even when sterilely dressed. Further development should improve precision and the ability to track small boney fragments. The development and use of hybrid models should be examined in further studies.

Author contribution

T.K. and C.G. compiled the data and authored the manuscript and took part in the field trials in the OR. H.B. and T.N. planned and carried out the trials in the OR. R.G. and D.S. contributed to the preparation and contributed to the interpretation of the results. T.H. proofread the manuscript. All authors provided critical feedback and helped shape the research, analysis, and manuscript.

Funding

Open Access funding enabled and organized by Projekt DEAL, we acknowledge support by Open Access Publishing Fund of University of Tübingen. The study was financed by a grant of the federal state of Baden-Wuerttemberg of Germany (Grant: Innovationsgutschein Hightech 43–4310.026/2791*01). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Data availability

All data is available.

Declarations

Ethics approval

Not applicable.

Consent to participate

Not applicable.

Consent to publish

Not applicable.

Conflict of interest

The authors declare no competing interests.

Footnotes

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Ochs BG, Gonser C, Shiozawa T, Badke A, Weise K, Rolauffs B, Stuby FM. Computer-assisted periacetabular screw placement: comparison of different fluoroscopy-based navigation procedures with conventional technique. Injury. 2010;41:1297–1305. doi: 10.1016/j.injury.2010.07.502. [DOI] [PubMed] [Google Scholar]
  • 2.Larsen P, Elsoe R, Hansen SH, Graven-Nielsen T, Laessoe U, Rasmussen S. Incidence and epidemiology of tibial shaft fractures. Injury. 2015;46:746–750. doi: 10.1016/j.injury.2014.12.027. [DOI] [PubMed] [Google Scholar]
  • 3.Hu L, Xiong Y, Mi B, Panayi AC, Zhou W, Liu Y, Liu J, Xue H, Yan C, Abududilibaier A, Chen L, Liu G. Comparison of intramedullary nailing and plate fixation in distal tibial fractures with metaphyseal damage: a meta-analysis of randomized controlled trials. J Orthop Surg Res. 2019;14:30. doi: 10.1186/s13018-018-1037-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Williamson M, Iliopoulos E, Williams R, Trompeter A. Intra-operative fluoroscopy time and radiation dose during suprapatellar tibial nailing versus infrapatellar tibial nailing. Injury. 2018;49:1891–1894. doi: 10.1016/j.injury.2018.07.004. [DOI] [PubMed] [Google Scholar]
  • 5.Casari FA, Navab N, Hruby LA, Kriechling P, Nakamura R, Tori R, de Lourdes Dos Santos Nunes F, Queiroz MC, Fürnstahl P, Farshad M. Augmented reality in orthopedic surgery is emerging from proof of concept towards clinical studies: a literature review explaining the technology and current state of the art. Curr Rev Musculoskelet Med. 2021;14:192–203. doi: 10.1007/s12178-021-09699-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Laverdière C, Corban J, Khoury J, Ge SM, Schupbach J, Harvey EJ, Reindl R, Martineau PA. Augmented reality in orthopaedics: a systematic review and a window on future possibilities. Bone J J. 2019;101-B:1479–1488. doi: 10.1302/0301-620X.101B12.BJJ-2019-0315.R1. [DOI] [PubMed] [Google Scholar]
  • 7.Jud L, Fotouhi J, Andronic O, Aichmair A, Osgood G, Navab N, Farshad M. Applicability of augmented reality in orthopedic surgery - a systematic review. BMC Musculoskelet Disord. 2020;21:103. doi: 10.1186/s12891-020-3110-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Wang H, Wang F, Leong APY, Xu L, Chen X, Wang Q. Precision insertion of percutaneous sacroiliac screws using a novel augmented reality-based navigation system: a pilot study. Int Orthop. 2016;40:1941–1947. doi: 10.1007/s00264-015-3028-8. [DOI] [PubMed] [Google Scholar]
  • 9.Ochs BG, Schreiner AJ, de Zwart PM, Stöckle U, Gonser CE. Computer-assisted navigation is beneficial both in primary and revision surgery with modular rotating-hinge knee arthroplasty. Knee surgery, sports traumatology, arthroscopy : official journal of the ESSKA. 2016;24:64–73. doi: 10.1007/s00167-014-3316-7. [DOI] [PubMed] [Google Scholar]
  • 10.Kim KP, Miller DL, Berrington de Gonzalez A, Balter S, Kleinerman RA, Ostroumova E, Simon SL, Linet MS. Occupational radiation doses to operators performing fluoroscopically-guided procedures. Health Phys. 2012;103:80–99. doi: 10.1097/HP.0b013e31824dae76. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Lee SC, Fuerst B, Tateno K, Johnson A, Fotouhi J, Osgood G, Tombari F, Navab N. Multi-modal imaging, model-based tracking, and mixed reality visualisation for orthopaedic surgery. Healthcare technology letters. 2017;4:168–173. doi: 10.1049/htl.2017.0066. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Befrui N, Fischer M, Fuerst B, Lee S-C, Fotouhi J, Weidert S, Johnson A, Euler E, Osgood G, Navab N, Böcker W. „3D-augmented-reality“-Visualisierung für die navigierte Osteosynthese von Beckenfrakturen (3D augmented reality visualization for navigated osteosynthesis of pelvic fractures) Unfallchirurg. 2018;121:264–270. doi: 10.1007/s00113-018-0466-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Fischer M, Fuerst B, Lee SC, Fotouhi J, Habert S, Weidert S, Euler E, Osgood G, Navab N. Preclinical usability study of multiple augmented reality concepts for K-wire placement. Int J Comput Assist Radiol Surg. 2016;11:1007–1014. doi: 10.1007/s11548-016-1363-x. [DOI] [PubMed] [Google Scholar]
  • 14.Liu H, Auvinet E, Giles J, Rodriguez Y, Baena F. Augmented reality based navigation for computer assisted hip resurfacing: a proof of concept study. Ann Biomed Eng. 2018;46:1595–1605. doi: 10.1007/s10439-018-2055-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Hajek J, Unberath M, Fotouhi J, Bier B, Lee SC, Osgood G, Maier A, Armand M, Navab N, et al. Closing the calibration loop: an inside-out-tracking paradigm for augmented reality in orthopedic surgery. In: Frangi AF, Schnabel JA, Davatzikos C, et al., editors. Medical Image Computing and Computer Assisted Intervention – MICCAI 2018. Cham: Springer International Publishing; 2018. pp. 299–306. [Google Scholar]
  • 16.Heide AM von der, Fallavollita P, Wang L, Sandner P, Navab N, Weidert S, Euler E (2018) Camera-augmented mobile C-arm (CamC): a feasibility study of augmented reality imaging in the operating room. The international journal of medical robotics + computer assisted surgery : MRCAS 14. 10.1002/rcs.1885 [DOI] [PubMed]
  • 17.Ogawa H, Hasegawa S, Tsukada S, Matsubara M. A pilot study of augmented reality technology applied to the acetabular cup placement during total hip arthroplasty. J Arthroplasty. 2018;33:1833–1837. doi: 10.1016/j.arth.2018.01.067. [DOI] [PubMed] [Google Scholar]
  • 18.de Oliveira ME, Debarba HG, Lädermann A, Chagué S, Charbonnier C. A hand-eye calibration method for augmented reality applied to computer-assisted orthopedic surgery. Int J Med Robotics + Comput Assisted Surg : MRCAS. 2019;15:e1969. doi: 10.1002/rcs.1969. [DOI] [PubMed] [Google Scholar]
  • 19.Gregory TM, Gregory J, Sledge J, Allard R, Mir O. Surgery guided by mixed reality: presentation of a proof of concept. Acta Orthop. 2018;89:480–483. doi: 10.1080/17453674.2018.1506974. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Ochs BG, Stuby FM, Stoeckle U, Gonser CE. Virtual mapping of 260 three-dimensional hemipelvises to analyse gender-specific differences in minimally invasive retrograde lag screw placement in the posterior acetabular column using the anterior pelvic and midsagittal plane as reference. BMC Musculoskelet Disord. 2015;16:240. doi: 10.1186/s12891-015-0697-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Borgmann H, Rodríguez Socarrás M, Salem J, Tsaur I, Gomez Rivas J, Barret E, Tortolero L. Feasibility and safety of augmented reality-assisted urological surgery using smartglass. World J Urol. 2017;35:967–972. doi: 10.1007/s00345-016-1956-6. [DOI] [PubMed] [Google Scholar]
  • 22.Incekara F, Smits M, Dirven C, Vincent A. Clinical feasibility of a wearable mixed-reality device in neurosurgery. World neurosurgery. 2018;118:e422–e427. doi: 10.1016/j.wneu.2018.06.208. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

All data is available.


Articles from International Orthopaedics are provided here courtesy of Springer

RESOURCES