Skip to main content
Journal of Medical Imaging logoLink to Journal of Medical Imaging
. 2021 Feb 12;8(1):015002. doi: 10.1117/1.JMI.8.1.015002

Drill-mounted video guidance for orthopaedic trauma surgery

Prasad Vagdargi a,, Niral Sheth b,, Alejandro Sisniega b, Ali Uneri b, Tharindu De Silva b, Greg M Osgood c, Jeffrey H Siewerdsen a,b,*
PMCID: PMC7880243  PMID: 33604409

Abstract.

Purpose: Percutaneous fracture fixation is a challenging procedure that requires accurate interpretation of fluoroscopic images to insert guidewires through narrow bone corridors. We present a guidance system with a video camera mounted onboard the surgical drill to achieve real-time augmentation of the drill trajectory in fluoroscopy and/or CT.

Approach: The camera was mounted on the drill and calibrated with respect to the drill axis. Markers identifiable in both video and fluoroscopy are placed about the surgical field and co-registered by feature correspondences. If available, a preoperative CT can also be co-registered by 3D–2D image registration. Real-time guidance is achieved by virtual overlay of the registered drill axis on fluoroscopy or in CT. Performance was evaluated in terms of target registration error (TRE), conformance within clinically relevant pelvic bone corridors, and runtime.

Results: Registration of the drill axis to fluoroscopy demonstrated median TRE of 0.9 mm and 2.0 deg when solved with two views (e.g., anteroposterior and lateral) and five markers visible in both video and fluoroscopy—more than sufficient to provide Kirschner wire (K-wire) conformance within common pelvic bone corridors. Registration accuracy was reduced when solved with a single fluoroscopic view (TRE=3.4  mm and 2.7 deg) but was also sufficient for K-wire conformance within pelvic bone corridors. Registration was robust with as few as four markers visible within the field of view. Runtime of the initial implementation allowed fluoroscopy overlay and/or 3D CT navigation with freehand manipulation of the drill up to 10  frames/s.

Conclusions: A drill-mounted video guidance system was developed to assist with K-wire placement. Overall workflow is compatible with fluoroscopically guided orthopaedic trauma surgery and does not require markers to be placed in preoperative CT. The initial prototype demonstrates accuracy and runtime that could improve the accuracy of K-wire placement, motivating future work for translation to clinical studies.

Keywords: surgical navigation, drill guidance, computer vision, 3D–2D registration, pelvic surgery

1. Introduction

Orthopaedic trauma is a prominent healthcare burden in terms of cost and quality of life.1 Fractures of the pelvic ring present a particular challenge, affecting 37 out of 100,000 individuals/year (3% to 7% of skeletal fractures).24 Pelvic fracture surgery involves (1) reduction of bone fragments to their natural position and (2) fixation of the reduced fracture by insertion of a guidewire [e.g., a Kirschner wire (K-wire)] along bone corridors—e.g., the pubis, ischium, and/or ilium. A cannulated screw is then delivered over the wire for fixation of the fracture, and the wire is removed.5

Surgical treatment increasingly involves percutaneous approach under fluoroscopy,68 with intermittent exposures acquired during placement of the guidewire to assess position relative to surrounding anatomy—namely conformance within bone corridors.9 Surgeons qualitatively estimate the 3D position of the K-wire from multiple projection views (e.g., inlet, outlet, PA, and lateral views). However, 3D reckoning within the complex morphology of the pelvis is challenging even to experienced surgeons, and accurate K-wire placement often requires “fluoro hunting” and trial and error.5,10 It is not uncommon for the guidewire to be completely withdrawn if the K-wire appears in danger of breaching the bone corridor and reinserted along a new trajectory, leading to extended procedure time and fluoroscopic exposure, often exceeding 120 s of fluoroscopy time and hundreds of radiographic views.5

Surgical navigation based on optical or electromagnetic (EM) trackers is fairly common in brain and spine surgery.11,12 However, orthopaedic trauma surgery has not seen widespread adoption of these systems, primarily due to factors of cost, additional equipment, line of sight, and workflow bottlenecks in setup and registration. Given the steep requirements in cost and workflow of orthopaedic trauma surgery, fluoroscopic guidance has remained the mainstay for decades.

Navigation solutions that are potentially better suited to the demanding workflow of orthopaedic trauma surgery are an active area of research. For example, mounting a camera on the C-arm provides a basis for augmented fluoroscopic guidance and tool tracking.1315 Mounting a stereoscopic camera on the C-arm demonstrated improved accuracy and better line-of-sight compared to a conventional infrared tracking arrangement and provided stereoscopic video for augmented views of the surgical field and/or fluoroscopy.16 EM trackers (despite sensitivity to metal objects and workflow limitations similar to infrared trackers) have been similarly investigated—e.g., incorporating an open frame EM field generator into the operating table 17 for real-time tracking while compatibility with fluoroscopy through the open frame. Others have mounted tracking equipment on the instrument itself.1820 For example, Magaraggia et al.18 described a video guidance system involving an implant-specific drill sleeve for distal radial fracture surgery and a video camera mounted on the surgical drill. While the approach relies on the use of a custom drill sleeve, the study reported improvements in screw positioning accuracy compared to conventional fluoroscopic guidance.

In this work, we present a video-based surgical drill guidance system that is potentially suitable to the workflow of fluoroscopy-guided orthopaedic trauma procedures. The system uses a miniature camera mounted on a surgical drill for real-time visualization of the drill trajectory in fluoroscopy and/or CT. The relationship between camera images and intraoperative fluoroscopy is established via multimodal (optical and radio-opaque) fiducial markers placed about the surgical field at the time of surgery. Notably (and likely essential to realistic workflow), the markers are not required in preoperative CT. The solution couples 3D–2D registration with vision-based tracking to register the drill axis without additional equipment in the operating room (OR). The proposed solution also has the potential to reduce “fluoro hunting,” and registration can be performed with commonly acquired (e.g., inlet and outlet) views. In principle, the system could reduce the number of views required for K-wire placement from hundreds5 to as few as two—one for initial registration and one for visual confirmation of device placement.

The proposed system is distinct from previously reported systems in a variety of ways. Compared to the drill-mounted video system described by Magaraggia et al.,18 the system uses multimodal markers (rather than custom drill sleeves) to register video images to fluoroscopy. Furthermore, the system can register the drill axis to a preoperative CT [or intraoperative cone-beam CT (CBCT)] for 3D navigation. Other video-based systems require that markers be placed on the patient in preoperative CT.2125 The need to place markers in preoperative imaging presents a workflow barrier that is not compatible with emergent trauma, where the preoperative CT is acquired quickly for many diagnostic purposes (e.g., to rule-out hemorrhage, detect and characterize the fracture, etc.). Such systems also require the markers to remain unperturbed throughout the case. The proposed system allows the markers to be placed during the procedure (e.g., after the patient is draped, immediately prior to K-wire insertion), and perturbations of the markers are accommodated by updating the registration with as little as one fluoroscopic view.

The following sections detail the video-guided drill system, the video-to-fluoroscopy and fluoroscopy-to-CT registration methods, and integration with two modes of guidance—“video-fluoro” and “video-CT.” Experiments quantify the geometric accuracy and runtime in placing K-wires within bone corridors in a phantom study. We conclude with a discussion of improvements in future design iterations that will support translation to clinical studies.

2. Materials and Methods

2.1. System for Surgical Drill Guidance Using Computer Vision and 3D–2D Registration

Figure 1 summarizes the high-level workflow of the drill guidance system. As detailed below: if a preoperative CT is not available, the system enables video-fluoro navigation; and if a preoperative CT is available, the system enables both video-fluoro and/or video-CT navigation. The system assumes two calibrations [Fig. 1(b)]: camera calibration (a one-time step to correct lens distortion) and drill axis calibration (a relatively quick step that could be performed offline or in the OR if the camera mount is swapped between drills).

Fig. 1.

Fig. 1

Workflow and algorithms for surgical guidance with a drill-mounted video camera. (a) If available, preoperative CT is registered to the patient (dashed lines) via 3D–2D registration to intraoperative fluoroscopy, allowing video-CT guidance. (b) Offline calibration of the camera and drill axis. (c), (d) Intraoperative registration: (c) 3D–2D registration to localize marker poses in C-arm fluoroscopy and register preoperative CT to the patient; and (d) a continuous, real-time loop that registers the drill axis to fluoroscopy and/or CT via the video scene.

Intraoperative workflow includes two main steps. The first [Fig. 1(c)] is to place multimodal markers about the surgical field such that at least three are in the fluoroscopy field of view (FOV). It is worthy to note that the markers need not be rigidly affixed with respect to each other or the patient, since the registration is updated with each fluoroscopic view. The drill itself does not need to be visible in the fluoroscopic view; rather, the drill-mounted camera only need have clear view of the markers as present in the fluoroscopic image. The second step [Fig. 1(d)] involves registration of the drill trajectory to the fluoroscopy image (and/or optionally to CT) via the video scene. This registration can be computed continuously in real time, noting that the overlaid trajectory shows the drill axis (not the K-wire tip). Such guidance may therefore be most useful in the challenging initial step of finding an acceptable entry point and trajectory orientation.

Table 1 summarizes relevant notation, with 3D vectors denoted in uppercase and 2D vectors denoted in lowercase. Throughout this work, the hat symbol denotes an estimate (subject to measurement error) of the given quantity. Figure 2(a) illustrates the coordinate systems and transforms.

Table 1.

Summary of notation for video-based surgical drill guidance. Coordinate frames: D (for drill camera), C (for the C-arm), and V (for preoperative CT volume). Multimodal markers (m=1,,M) are co-registered between the video frame, one or more fluoroscopic images at view angles (θ), and the CT volume via 3D–2D registration.

System parameters
θ Projection view angle for a C-arm fluoroscopy frame
m Index for a particular multimodal marker (m: 1M)
κ 3D mesh representation of radio-opaque marker features
μ Preoperative CT volume
Drill camera coordinate frame (D)
KD Drill camera calibration matrix
X^D(m) 3D feature point estimate for marker m in drill camera 3D coordinate frame
xD(m) 2D feature point for marker m in drill camera 2D image plane
P^D(m) Pose estimate of marker m in the drill camera 3D coordinate frame
LD Drill axis in drill camera 3D coordinate frame
C-arm coordinate frame (C)
PC(θ) C-arm projection matrix for fluoroscopy frame at view angle
X^C(m) 3D feature point estimate for marker m in the C-arm 3D coordinate frame
xC(m,θ) 2D feature point for marker m in the C-arm 2D image plane for a fluoroscopy frame at view angle θ
P^C(m) Pose estimate of marker m in the C-arm 3D coordinate frame
T^CD Estimated transformation from the drill camera to C-arm 3D coordinate frames
LC Drill axis in the C-arm 3D coordinate frame
lC Drill axis projected to the C-arm 2D image plane (i.e., 2D fluoroscopy frame)
Preoperative CT coordinate frame (V)
T^CV Estimated transformation from the preoperative CT to C-arm 3D coordinate frames
LV Drill axis in the preoperative CT coordinate frame

Fig. 2.

Fig. 2

Coordinate frames and multimodal marker design. (a) Coordinate frames for the video-guided drill setup, including the drill camera (D) and C-arm (C). The zoomed-in view shows the drill camera (and frame D) in relation to the drill axis (LD). (b) Schematic of multimodal marker features visible in both video (center of the ArUco tag) and fluoroscopy (central BB). Additional BBs and wires encode individual markers and provide a basis for 3D–2D registration between the marker design (κ) and fluoroscopy.

2.2. Multimodal Markers

2.2.1. Marker design

Previous work used “multimodal” markers visible optically and radiographically to relate surgical trackers to fluoroscopic images—for example, Hamming et al.26 and Dang et al.27 Figure 2(b) shows an exploded view of multimodal markers to relate the drill camera and C-arm frames via optical and radio-opaque features. Optical features were presented by ArUco marker tags28 consisting of a 6×6 inner matrix of bits to uniquely identify each marker. Each marker contained a central ball bearing (BB) (tungsten, Ø2.0  mm) and a unique arrangement of outer BBs differing in size (Ø2.0 or 3.5 mm) and location. To aid feature extraction in fluoroscopy, BBs were constrained to be collinear with at least two other BBs. Each marker was encircled by a tungsten wire (Ø0.8  mm) to create a 48-mm periphery that assisted in marker detection (by Hough transform) and pose estimation (circle-to-ellipse perspective relationship).

The marker base was 3D-printed (Vero PureWhite, Connex-3 Objet 260, Stratasys, Eden Prairie Minnesota) with pockets to hold the BBs, a peripheral groove to hold the wire, and a 30×30  mm2 square recess (2-mm deep) in which the ArUco tag was placed such that its center coincides with the central BB. The depth of the recess introduces a small offset between the ArUco tag center and central BB along the axis normal to the marker. Since this offset is known, it can be corrected prior to registration (discussed in Sec. 2.4). The initial design has Ø50  mm and can generate up to 48 unique markers, with future work to investigate more compact designs.

2.2.2. Marker detection

The detection of ArUco tags in video was based on open-source tools available in OpenCV.29,30 The algorithm first performs adaptive thresholding and contour detection of grayscale images to isolate candidate regions for multiple tags. The inner area of each candidate is analyzed by correcting the tag perspective to a square region and then binarizing the resulting region to a regularly spaced grid upon which marker identification can be performed.

Marker detection in the fluoroscopic scene was performed first by ellipse detection (using the peripheral wire) to coarsely identify individual marker positions. A Canny edge filter was followed by morphological closing to yield binarized elliptical contours and filter out smaller objects (e.g., BBs). A Hough-based ellipse detector31 returned elliptical fits ordered by accumulator score, which was cutoff according to the known number of markers (M). Within each ellipse, a morphological top-hat filter isolated marker features from surrounding anatomy, and the ellipse was removed by morphological opening to isolate the BB features. Hough-based circle detection was used to identify the position and radius of BBs in each marker, and candidate BBs were filtered based on the known range of BB radii. Detections within a certain proximity to each other were also filtered based on the known marker designs, and collinearity was enforced to eliminate any remaining false positives. The resulting BB detections were then hierarchically clustered in two groups according to size, and markers were uniquely identified according to a lookup table.

2.3. Camera and Drill Axis Calibration

Camera calibration determined the intrinsic parameters of the camera (pinhole model) and distortion coefficients of the lens. Calibration was performed as in the work by Zhang32 using multiple images of a planar checkboard to obtain a closed-form solution for the camera matrix (KD). Lens distortion was corrected using the Brown–Conrady even-order polynomial model,32 describing both radial and tangential distortion.

The orientation of the drill axis in the camera coordinate frame (LD) was solved using a calibration jig described in previous work.33 The jig consisted of a drill sleeve centered on an ArUco board. As the jig (drill sleeve) rotates about the drill axis, the pose of the ArUco board in the 3D frame of the camera was estimated in multiple images yielding a 3D cylindrical point cloud. A RANSAC-based cylindrical fit (MLESAC34) was computed to obtain the central axis of the point cloud—i.e., the drill axis (LD).

2.4. Video-to-Fluoroscopy Registration

Video images were registered to the fluoroscopic scene through localization and registration of the point-based feature correspondences of markers discussed in Sec. 2.2. For the ArUco tags, the 3D pose of each marker in the drill camera coordinate frame [P^D(m)] was estimated with the perspective-N-points algorithm for m=1,,M markers. The translational component of the resulting pose estimate was extracted and represents the center of the ArUco tag. To correct for the offset between the tag center and central BB mentioned in Sec. 2.2.1, the translational component of the pose estimate was adjusted along the axis normal to the surface of the marker [third column of the rotational component of P^D(m)] by the known offset (2 mm). The resulting estimate represents the central marker feature point in the drill camera frame [X^D(m)] corresponding to the location of the central BB.

A mobile C-arm (Cios Spin, Siemens Healthineers, Forchheim, Germany) was used in all studies described below. The projective relationship (PC) relating 3D points in the C-arm frame (XC) to 2D fluoroscopic image points on the C-arm detector plane (xC) (in homogenous coordinates) was determined by standard C-arm geometry calibration,35 with the projection matrix defined by

PC=KC[RθTθ]. (1)

The intrinsic matrix KC describes the geometric relationship between the C-arm source and detector. The extrinsic parameters ([RθTθ]) describe the pose of the C-arm source–detector assembly for a fluoroscopy frame at view angle θ in a common coordinate frame, referred to as the C-arm coordinate frame C.

To extract marker pose from fluoroscopic images, 3D–2D “known-component” registration (KC-Reg) was performed36 to register markers according to the known 3D design (κ) of each marker. The general framework for KC-Reg is shown in Fig. 3(a). A generic mesh model (κ) representing the radio-opaque features of the marker is iteratively transformed to optimize the image similarity between digitally reconstructed radiographs (DRRκ) and one or more intraoperative fluoroscopic images (Iθ). Fluoroscopic views were selected such that markers were not overlapping in projection data. 3D–2D registration of the fluoroscopy image (Iθ) and the corresponding DRR was computed by optimization of gradient information (GI), which has been shown to be robust in a variety of realistic fluoroscopy scenarios.37,38 To solve for the pose of marker m, the cumulative GI across θ=1,,Nview fluoroscopic views was maximized:

P^C(m)=argmaxPC(m)θGI(Iθ,rθPC(m)(κ)drθ), (2)

where the integral represents the DRR at view θ. Optimization was performed using the covariance matrix adaptive evolution strategy (CMA-ES) algorithm39 initialized using features extracted during marker detection (Sec. 2.2.2). For each multimodal marker m, an initial estimate of the central 3D feature point in the C-arm frame [X^C(m)] was obtained by first backprojecting a ray from the corresponding 2D image feature point in homogenous coordinates [denoted xC(m,θ)] toward the x-ray source. The backprojected ray was estimated for each fluoroscopic view θ, from which an initial estimate of the 3D feature point could be reconstructed by

X^C(m)=1Nviewθ=1Nviewλ(m,  θ)(KCRθ)1xC(m,θ)(KCRθ)1xC(m,θ)+(Rθ1Tθ), (3a)

where

λ(m,  θ)=SDDM(m,θ) (3b)

describes the estimated distance along the backprojected source–detector ray. The magnification M was estimated for each marker m in each fluoroscopic view using the perspective relationship between the diameter of a circle and the major axis length of its elliptical projection. Once the 3D position of each marker was initialized, a rotational initialization was performed with a planar fit of the global marker arrangement. The computed plane normal was used as an initial estimate of the out-of-plane axis for each marker.

Fig. 3.

Fig. 3

3D–2D registration of intraoperative fluoroscopy to (a) multimodal markers and (b) preoperative CT.

With corresponding point estimates obtained in both the drill camera (D) and C-arm (C) coordinate frames, a transformation between the two was estimated using point-based registration.40 The resulting video-to-fluoroscopy transform (T^CD) was used to represent the surgical drill axis in the C-arm coordinate frame:

LC=T^CDLD (4a)

and its projection on the C-arm detector plane:

lC=PCLC. (4b)

Augmentation of fluoroscopic views with lC realizes the video-fluoro navigation mode. For drill guidance, marker localization in the camera image is continuously updated to allow freehand motion of the drill. In the event that the markers are perturbed relative to anatomy (or if the surgeon desires overlay in different fluoroscopic view), then the video-to-fluoroscopy registration is updated with the acquisition of one or more fluoroscopy views.

2.5. Fluoroscopy-to-CT Registration

Intraoperative fluoroscopy was used to register the patient to a preoperative CT (or intraoperative CBCT) volume using 3D–2D registration.38,41 The workflow for patient registration is shown in Fig. 3(b). The CT volume (μ) is iteratively transformed to optimize similarity between fluoroscopic images (Iθ) and forward projections of the rigidly transformed CT (DRRμ). An objective function similar to Eq. (2) was solved using CMA-ES with gradient correlation (GC) as the similarity metric. GC was used for CT registration because it is independent of absolute gradient magnitudes and is robust for images in which corresponding anatomical gradients may differ due to differences in imaging technique or mismatch in image content (e.g., tools in the fluoroscopic scene that are not in the CT).41

The resulting CT-to-fluoroscopy transform (T^CV) was used to transform the surgical drill axis into the preoperative CT coordinate frame as

LV=(T^CV)1T^CDLD. (5)

Augmentation of the CT image (e.g., orthogonal CT slices, MIPs, or volume renderings) realizes the video-CT navigation mode.

2.6. Experimental Evaluation

2.6.1. Experimental setup

The performance of the video-guided drill system was evaluated in terms of the accuracy of drill axis registration and guidance along K-wire trajectories common in pelvic trauma surgery. The imaging setup is illustrated in Fig. 4(a), consisting of a mobile C-arm, the video-guided drill, and an anthropomorphic pelvis phantom. The mobile C-arm was used to acquire fluoroscopic images and CBCT (for truth definition). CBCT images were acquired with 400 projections over a 195-deg semi-circular orbit and reconstructed on a 0.512×0.512×0.512  mm3 voxel grid with a standard bone kernel. Figure 4(b) shows the initial prototype for the drill guidance system, consisting of a Stryker System 6 (Kalamazoo, Michigan) surgical drill with a pin collet for 3–mm-diameter K-wires and a Logitech C900 USB webcam. The camera was rigidly attached to the drill body using a custom 3D-printed chassis that positioned the camera with clear view of the drill axis and surgical scene. Experiments were performed in an OR laboratory with lighting conditions that approximated realistic ambient lighting with or without an overhead OR light. To assist in experimental workflow and provide a stable platform for experiments, a UR3e robotic arm (Universal Robotics, Odense, Denmark) was used as a drill holder throughout the experiments. The robot was for convenience in experiments, is not required for the envisioned freehand system, and does not affect the accuracy of the registration and guidance methods.

Fig. 4.

Fig. 4

Experimental setup for video-based drill guidance. (a) Imaging setup showing the C-arm, phantom A, and video-drill held by a robotic positioner. A similar setup was created for phantom B (not shown). (b) Initial prototype of the drill camera and markers. The markers are placed about the entry for the AIIS-to-PSIS trajectory (shown in green).

The system was evaluated in two phantom experiments. The first assessed the accuracy of individual registration steps and the end-to-end accuracy in guiding the initial entry point and orientation of the drill. The effect of relevant experimental parameters (viz., the number of markers, M, and the number of fluoroscopic views, Nview) on drill registration accuracy was also investigated. The second experiment assessed accuracy during K-wire insertion along multiple pelvic trajectories, including potential mechanical sources of error (e.g., bending of the K-wire during drilling).

For the first set of experiments, an anthropomorphic pelvis phantom composed of a natural adult skeleton in tissue-equivalent plastic (Rando®, The Phantom Lab, Greenwich, New York) was used (referred to as “phantom A”). The drill camera was aligned with respect to the anterior inferior iliac spine (AIIS) to posterior superior iliac spine (PSIS) trajectory in the left hip. Five markers were placed about the planned entry site. The robotic arm was used to position the drill camera at a distance of 20  cm from the surface of phantom A to emulate realistic surgical drill positioning. The drill camera was also placed at nine different positions about the planned trajectory to measure the sensitivity of registration accuracy to various camera perspectives. At each camera position, a single camera image was acquired from which the position of each marker was determined. All other experimental parameters (and specifically, the number of markers) were held constant, varying only the position of the camera. An initial CBCT scan (110 kV, 350 mAs) was acquired with only the phantom A and markers in the FOV. Projections in which markers were not overlapping (drawn from the scan arc from θ=20  deg to 70 deg) were selected as views for solving 3D–2D registration. The CBCT defined the true position of the BBs. Fluoroscopic views (100 kV, 0.9  mAs) were collected at common clinical orientations for augmentation in video-fluoro mode, including anteroposterior view (θ=0  deg, ϕ=0  deg), lateral view (θ=90  deg, ϕ=0  deg), inlet view (θ=0  deg, ϕ=25  deg), and outlet view (θ=0  deg, ϕ=30  deg). The drill camera was then positioned with a 3-mm K-wire extending from the drill tip to the surface of phantom A, and video images of the marker arrangement were collected. A final CBCT scan (110 kV, 380 mAs) was acquired with the K-wire in the FOV for truth definition of the drill axis (determined by threshold-based segmentation of the K-wire followed by principal component analysis of the segmented voxels to extract the drill axis). This process was robust to metal artifacts about the K-wire. In total, two CBCT scans and four radiographs were acquired for experiments with phantom A.

For evaluation of the video-CT navigation mode, a preoperative CT (0.82×0.82×0.5  mm3voxel grid) of phantom A was acquired (SOMATOM Definition, Siemens, Erlangen Germany). Pelvic K-wire trajectories were planned in preoperative CT using the atlas-based planning method in Goerres et al.42 and Han et al.43 Acceptance volumes interior to bone corridors were created for three common K-wire trajectories: AIIS-to-PSIS, superior ramus (SR), and iliac crest to posterior column (PC). Acceptance volumes were used to visualize and evaluate drill axis conformance within pelvic bone corridors.

For the second set of experiments, an anthropomorphic (left) hemi-pelvis phantom composed of radiopaque material (Sawbones®, Pacific Research Laboratories Inc., Vashon, Washington) was used (referred to as “phantom B”) since phantom A did not permit drilling. Due to the simple nature of phantom B (lack of surrounding bulk anatomy), an additional 3.2-mm Cu filtration (15-cm water equivalent) was added at the x-ray source to provide realistic exposure levels and quantum noise in the fluoroscopic images. System performance during K-wire insertion was evaluated for the same three pelvic trajectories (AIIS-to-PIIS, PC, and SR). For each trajectory, five markers were placed about the planned entry site. The video-drill system was then used to insert a K-wire (3-mm diameter) along each trajectory. Data were collected at three positions along each path, referred to as: (1) “entry,” acquired at 0 to 5 mm along the trajectory; (2) “midpoint,” acquired at 40 to 60 mm along the trajectory; and (3) “endpoint,” acquired at 80 to 120 mm along the trajectory. At each position, video images of the marker arrangement were collected, and a CBCT scan (110 kV, 55 mAs) was acquired. Projections from the scan were selected (differing in view angle by up to 30 deg) for solving 3D–2D registration, and CBCT images defined ground truth of the K-wire drill axis as described above. In total, nine CBCT scans (three scans for each trajectory) were acquired for experiments with phantom B.

2.6.2. Evaluation of system accuracy

Table 2 summarizes the figures of merit for assessing the accuracy of individual registration steps and the overall system. The performance of video-to-fluoroscopy registration was quantified in terms of errors related to 3D marker localization in the drill camera (D) and C-arm (C) coordinate frames. The positional estimate for each marker in the camera frame [X^D(m)] was evaluated with respect to the truth definition (points defined in CBCT) and quantified in terms of fiducial registration error (FRE):

FRE=T^CDX^D(m)XC(m,true), (6a)

where XC(m,true) represents the true location of marker m in the C-arm frame and T^CD is an estimate of the true camera-to-C-arm transform derived from point-based registration with all markers. The term T^CDX^D(m) is therefore the estimated location of marker m in the C-arm frame. Fiducial errors were further decomposed into in-plane and depth errors with respect to the drill camera coordinate frame (ΔD) as

ΔD(m)=RDC(T^CDX^D(m)XC(m,true)). (6b)

Estimation of RDC (the rotation matrix from the C-arm to the drill camera coordinate frame) was performed independently of the fiducials by solving the rotation between the calibrated drill axis (LD) and the true drill axis segmented from CBCT [denoted as LC(true)].

Table 2.

Figures of merit for evaluating registration accuracy for individual registration methods and end-to-end system performance.

Video-to-fluoroscopy registration
FRE(m) Norm of error in 3D localization of marker m from video images
ΔD(m) In-plane and out-of-plane translational error components in 3D localization of marker m from video images
δC(m) Norm of error in 3D localization of marker m from fluoroscopic images
ΔC(m) In-plane and out-of-plane translational error components in 3D localization of marker m from fluoroscopic images
Fluoroscopy-to-CT registration
ΓV Fluoroscopy-to-CT registration difference transform
ΓVΔ In-plane and out-of-plane translational error components in fluoroscopy-to-CT registration
ΓVΦ In-plane and out-of-plane rotational error components in fluoroscopy-to-CT registration
δV Norm of translational error in fluoroscopy-to-CT registration
Drill axis registration
TREx Norm of translational error between computed and true drill axes
TREφ Angular skew between computed and true drill axes

Marker localization errors in the C-arm frame (C) were estimated relative to the true marker locations [XC(m,true)]. The 3D–2D registration of each marker was solved using Nview=13 fluoroscopic views selected to span a total arc of 30 deg with equiangular spacing. Accuracy was assessed in terms of the norm of the translational error (δC):

δC(m)=X^C(m)XC(m,true). (6c)

Translational errors were further examined with respect to in-plane (parallel to the detector plane) and out-of-plane (depth) components (referred to as ΔC).

Fluoroscopy-to-CT registration was evaluated over the same set of fluoroscopic views used during marker localization for Nview=1 and 2 (over a 30-deg arc). To evaluate accuracy, truth was defined from a large number (Nview=15) of fluoroscopic views to solve the true 3D–2D patient registration (referred to as TCV), selecting distinct views for registration and truth definition to mitigate bias. Performance was calculated in terms of the difference transform (ΓV) between the registration result and the truth definition by

ΓV=T^CV(TCV)1. (7)

The difference transform was further decomposed into translational (ΓVΔ) and rotational error components (ΓVΦ), from which the norm of the translational error (δV) was also computed.

End-to-end system performance was evaluated by comparison of the computed drill axis (LC) [Eq. (4a)] with the true drill axis from CBCT [LC(true)]. Taking the true axis as reference, possible mismatch from the automatically planned trajectory was corrected by applying a transformation to the computed and true drill axes. This transformation places the drill within the context of the bone corridor, as would be done in clinical use, while preserving the relative errors between the computed and true drill trajectories. The error between the computed and true drill axes was reported in terms of target registration error (TRE), separated into positional (TREx) and angular (TREφ) errors by decomposing axes into translational (τC) and rotational (ρC) components:

TREx=τ^CτC(true), (8a)
TREφ=cos1ρ^C·ρC(true)ρ^C·ρC(true), (8b)

where the hat symbol here denotes components of the computed drill axis. The rotational components (ρC) are given by a vector describing the direction of the drill axis, and the translational components (τC) describe a point along the drill axis. The translational components (τC) were selected to evaluate TREx with respect to the bone corridor. In experiments with phantom A, the target location (τC) corresponded to the bone corridor entry point. Since the drill was placed at the surface of the phantom (rather than the surface of the bone corridor), the target locations (τC) were calculated as the intersection of the computed and true drill axes with the surface of the planned acceptance volume. In experiments with phantom B, the target location (τC) corresponded to the K-wire tip position as the K-wire was physically inserted into the bone corridor. Registration errors reported were computed in the CT coordinate frame (V). The truth trajectory in the CT coordinate frame was estimated using the fluoroscopy-to-CT truth definition [LV(true)=(TCV)1LC(true)].

To evaluate the conformance of the drill axis trajectories within pelvic bone corridors, a mesh of the planned acceptance volume was created for each corridor to represent the cortical bone surface. The acceptability of a resulting trajectory was measured by first computing the entry and exit point at which the given trajectory intersects the cortical bone surface. The resulting path from entry to exit point was equidistantly sampled along the trajectory, and the distance from each sample to the nearest cortical bone surface point was calculated.

3. Results

3.1. Accuracy of Video-to-Fluoroscopy Registration

The accuracy of 3D marker localization from video images [FRE in Eq. (6a)] was evaluated over all (M=5) markers to examine how localization errors in the video scene translate to errors in fiducial registration. Among the nine different camera positions measured, the registration performance exhibited considerable variation between markers (interquartilerange(IQR)3.5  mm) with FRE up to 5 mm in some cases due to poor marker corner detection and pose estimation. The median FRE, however, was fairly consistent across all camera poses, with median error of 1.5 mm (IQR = 0.42 mm). Localization errors with respect to the drill camera coordinate frame [ΔD in Eq. (6b)] were decomposed in terms of in-plane (parallel to the camera image plane) and out-of-plane (perpendicular to the camera plane) translational components for the planned trajectory. In-plane translations exhibited median error of 0.46 mm (IQR=0.52  mm), and median out-of-plane error was 1.6 mm (IQR=1.2  mm). Overall, the results demonstrate consistent FRE across the sampled trajectories, suggesting reasonable robustness in localizing markers from video images from a broad variety of camera poses. Such robustness is important as the surgeon maneuvers the drill about the scene to align with the planned trajectory. The out-of-plane errors reflect challenges in resolving depth with a monocular camera, addressed in future work incorporating a stereoscopic camera system.

Table 3 summarizes the errors associated with 3D marker localization from fluoroscopic images [δC in Eq. (6c)] pooled over all (M=5) markers. The localization error was notably higher for registrations solved with a single fluoroscopic view, with median error of 1.8 mm (IQR=1.7  mm). The main directional component of such errors was out-of-plane (i.e., along the x-ray source to detector direction), again illustrating the challenge to depth resolution from a single perspective.

Table 3.

Accuracy of individual registration methods. Values shown are the median (IQR) in registration error across 60 views.

Video-to-fluoroscopy registration
  Nview=1 Nview=2 Nview=3
δC 1.80 mm (IQR=1.7  mm) 0.67 mm (IQR=0.25  mm) 0.66 mm (IQR=0.20  mm)
ΔC(xC) 0.27 mm (0.30 mm) 0.36 mm (0.33 mm) 0.36 mm (0.33 mm)
ΔC(yC) 0.44 mm (0.51 mm) 0.25 mm (0.28 mm) 0.24 mm (0.25 mm)
ΔC(zC) 1.60 mm (1.8 mm) 0.50 mm (0.53 mm) 0.65 mm (0.20 mm)
Fluoroscopy-to-CT registration
  Nview=1 Nview=2
δV 0.41 mm (IQR=0.30  mm) 0.42 mm (IQR=0.30  mm)
ΓVΔ(xV) 0.11 mm (0.16 mm) 0.13 mm (0.12 mm)
ΓVΔ(yV) 0.12 mm (0.15 mm) 0.10 mm (0.09 mm)
ΓVΔ(zV) 0.39 mm (0.40 mm) 0.32 mm (0.31 mm)

3.2. Accuracy of Fluoroscopy-to-CT Registration

Fluoroscopy-to-CT registration accuracy was computed in terms of the difference in true and estimated 3D transformations [ΓV in Eq. (7)]. As shown in Table 3, the translational error norm (δV) for registrations solved with Nview=1 and Nview=2 was similar, with median error of 0.4 mm (0.3-mm IQR) in both cases, noting a few outliers (up to 2-mm error) when Nview=1. The errors were further separated into translational errors and rotational errors (not shown for brevity) relative to the detector plane. For the Nview=1 case, median in-plane translational error (xV,yV) was 0.12 mm (IQR=0.15  mm), and out-of-plane error (zV) was slightly higher (median of 0.39 mm, IQR=0.40  mm) and accounted for the outliers observed. In-plane and out-of-plane rotational errors were consistently low (median of 0.07 deg and 0.03 deg, respectively).

3.3. Accuracy of Drill Axis Registration

Figure 5 shows positional and rotational TRE [TREx and TREφ in Eq. (8)] for the AIIS-to-PSIS trajectory in phantom A evaluated using Nview=1 or 2 fluoroscopic images with a variable number of markers, M. For M=3 or 4, each possible subset among the five markers was evaluated and pooled as a single distribution. As expected, overall performance diminished with fewer markers, with end-to-end performance for M=3 markers yielding median TREx=4.7 to 6.9 mm and TREφ=3.7  deg to 6.9 deg for Nview=1 or 2 with an unacceptably high rate of outliers. For M=4 markers, registration from Nview=1 was still unacceptably high (median TREx=4.2  mm and TREφ=3.8  deg) and improved considerably with Nview=2 (median TREx=2.2  mm and TREφ=2.8  deg with no outliers). For M=5, the end-to-end system performance with a single fluoroscopic view (Nview=1) yielded median TREx=3.4  mm (1.9-mm IQR) and TREφ=2.7  deg (0.79-deg IQR), attributed primarily to out-of-plane error in marker localization in fluoroscopic images. For M=5 markers and Nview=2, the end-to-end accuracy improved to median TREx=0.88  mm (0.16-mm IQR) and TREφ=2.0  deg (0.16-deg IQR).

Fig. 5.

Fig. 5

Registration performance for end-to-end drill axis localization in the preoperative CT coordinate frame (V). Registration error was assessed as a function of the number of markers in terms of (a) translational error (TREx) at the entry point of the planned bone corridor and (b) rotational error (TREφ) between the computed and true drill axes. The outlier rate (fraction of measurements greater than 10 mm or 10 deg) is shown above each violin plot.

The accuracy of drill axis registration was further assessed (for M=5 markers) to analyze and visualize how the measured registration errors relate to conformance of the drill trajectory within bone corridors of the pelvis. A set of simulated drill axis registrations were computed by perturbing the planned trajectory according to a uniform distribution given by the median TREx and TREϕ (Fig. 5), generating a “cone” of perturbed registrations about the plan. Figure 6 shows the conformance analysis for three planned trajectories. In each case, the dispersion of simulated trajectories is shown relative to a volumetric surface rendering of phantom A, with the acceptance corridor (region interior to bone cortex) shown in green and the range of simulated trajectories shown in blue and pink for Nview=1 and 2, respectively.

Fig. 6.

Fig. 6

Analysis and visualization of drill axis trajectories within the bone corridors of the (a) AIIS-to-PSIS, (b) PC, and (c) SR trajectories for Nview=1 and Nview=2 fluoroscopic views. The distributions reflect the distance between the computed drill axis and the outer bone cortex as a function of distance along the drill axis for an ensemble of simulated trajectories (with the solid line representing median distance from cortex and the shaded region representing the interquartile range). The horizontal dashed line represents the thickness of the K-wire (1.5-mm radius, 300-mm length), with values below this threshold indicating a breach of the bone cortex. The distance between the planned trajectory and the outer bone cortex is also shown for reference (solid black line labeled as “plan”).

Also shown in Fig. 6 are plots of the distance between the registered drill axis and the bone cortex as a function of position along the trajectory. Distances below the radius of the K-wire (dashed line in Fig. 6) indicate a breach of the bone cortex. In clinical practice, the K-wire is not necessarily inserted along the entire length of the planned trajectory (enough to guide a cannulated screw). For Nview=1, the average distance to cortex was 4.4 mm for AIIS-to-PSIS, 5.6 mm for PC, and 2.5 mm for SR—each without breach of the cortex, apart from a minor impingement (0.1  mm) for the lower quartile of the data in the narrowest region of the central SR, which would likely correspond to a glancing incidence of the K-wire. For Nview=2, the average distance to cortex was 5.1 mm for AIIS-to-PSIS, 6.4 mm for PC, and 2.9 mm for SR—all well within the acceptance corridor.

The accuracy of drill axis registration was further evaluated (for M=5 markers and Nview=2) to measure registration errors during K-wire insertion along multiple pelvic bone corridors in phantom B. Positional and rotational TRE [TREx and TREφ in Eq. (8)] were estimated at three K-wire positions (pooled across three trajectories) and are illustrated in Fig. 7. Overall, system performance was found to be relatively consistent, noting a slight increase in error at increased depth along the trajectory. At the entry position, median TREx=1.8  mm (1.1-mm IQR) and TREφ=0.4  deg (0.6-deg IQR). At the midpoint position, the error slightly increased with TREx=2.1  mm (1.8-mm IQR) and TREφ=0.6  deg (0.8-deg IQR), primarily attributed to propagation of positional and rotational errors from the entry point. At the endpoint, errors increased further to median TREx=2.5  mm (3.2-mm IQR) and TREφ=0.7  deg (0.6-deg IQR), primarily attributed to mechanical bending of the K-wire as it was advanced freehand into the bone model. Figure 7(c) illustrates an example visualization of augmented guidance from this experiment (video-fluoro navigation mode) for a K-wire inserted along the AIIS-PSIS trajectory. The figure illustrates the error propagation along the length of the trajectory, and an error boundary is shown, describing an “error cone” (cyan) describing the upper quartile in TRE propagated from the entry position.

Fig. 7.

Fig. 7

End-to-end performance in drill axis localization during actual K-wire insertion. Registration error was assessed at the entry, midpoint, and endpoint of each trajectory in terms of (a) translational error (TREx) at the K-wire tip and (b) rotational error (TREφ) between the computed and true drill axes. (c) Visualization of augmented guidance within the AIIS-to-PSIS trajectory, illustrating a con” of trajectories (cyan) about the K-wire bounding the IQR of trajectories based on the TRE at the entry position.

3.4. Video-Fluoro and Video-CT Guidance

Figure 8 illustrates two modes of surgical guidance enabled by the video drill system, taking the AIIS-to-PSIS trajectory from phantom A as an example. Figure 8(a) shows video-fluoro augmentation in which fluoroscopic views are overlaid with the registered drill axis trajectory in real time according to the current pose of the drill—e.g., trajectories overlaid in Fig. 6(a) corresponding to Nview=1 or 2 (cyan or magenta, respectively). This mode of navigation allows the surgeon to adjust freehand drill trajectory in real time to align with the planned trajectory, with the background anatomical scene providing visual context. Since only the axis of the drill (cf., the “tip” of the K-wire) is conveyed, the trajectory overlay may be of primary benefit to the surgeon by helping to guide selection of the entry point and initial K-wire angulation, which are essential (and challenging) in establishing a safe approach into the bone. Visualization in multiple augmented fluoroscopic views gives reliable 3D reckoning of the scene (normally done mentally, requiring years of training in understanding the complex morphology of the pelvis in projection views). To the extent that the presence of markers in the fluoroscopy image is visually distracting [e.g., in Fig. 8(a)], they can be digitally masked (e.g., by median filter), since the precise location of the wire and BB features was accurately determined in the marker detection step. Once the K-wire is advanced into the bone, of course, the K-wire tip is directly visible in the fluoroscopy image, and the overlays are confirmatory. This video-fluoro mode could be well suited to procedures for which no preoperative 3D image is available, as it establishes registration using only the fluoroscopic images collected in standard clinical workflow for fluoroscopically guided procedures.

Fig. 8.

Fig. 8

Illustration of video-fluoro and video-CT modes of guidance with the drill-mounted video system. (a) Augmented fluoroscopy guidance (video-fluoro mode) with commonly acquired fluoroscopic views. If a preoperative CT is available, the planned trajectory and acceptance corridors may also be overlaid (shown in green). (b) Augmented CT guidance (video-CT mode) illustrated with multiplanar slices of the preoperative CT image. The acceptance corridor and planned trajectory (green) are shown along with real-time rendering of the drill axis solved with Nview=1 (cyan) or Nview=2 (magenta). The trajectory shown in this example corresponds to K-wire delivery along the AIIS-to-PIIS bone corridor.

For cases in which a preoperative CT (or intraoperative CBCT) is available, the system permits video-CT guidance as shown in Fig. 8(b), providing real-time 3D navigation analogous to common surgical tracking systems. The CT image and fluoroscopic views are overlaid with acceptance corridors and planned trajectories (green), and the drill axis trajectory is rendered in real time according to the pose of the drill to aid the surgeon in determining whether the current trajectory conforms within the bone corridor. Such a 3D-navigated view is common in neurosurgery and spine surgery via surgical trackers. The video-CT guidance provided by the video drill system could help to bring the precision of 3D navigation to orthopaedic trauma surgery without the additional cost and workflow associated with tracking systems.

The computational runtime of the system is summarized in Table 4, divided in two sections: (top) fluoroscopy/CT registration steps that are performed intermittently during the case (e.g., with each new fluoroscopic view and/or in the event that the markers arrangement is perturbed); and (bottom) video navigation computed continuously according to freehand motion of the drill. In the current research implementation, the runtime for fluoroscopy/CT registration was 3.1 min total for Nview=1 and 4 min for Nview=2. In each case, the most time-consuming step (72% to 74% of total runtime) was fluoroscopy marker localization. It is worthy to note that the fluoroscopy-to-CT registration is performed in parallel. In the current implementation, the 3D–2D localization of each marker was performed sequentially, suggesting a potential 5× speedup (from 175 to 30  s via parallelization). Methods for feature extraction described in Sec. 2.2.2 were also fairly rudimentary for initial feasibility testing, and more efficient implementations could reduce the feature extraction step to a few seconds.44

Table 4.

Computational runtime for the video-guided drill system. Values are mean ± standard deviation in 60 trials from the AIIS-to-PSIS trajectory experiment.

Fluoroscopy/CT Registration Nview=1 Nview=2
Fluoroscopy marker feature extraction 13.1±0.4  s 26.3±0.9  s
Fluoroscopy marker localization 139.0±15.0  s 175.0±20.0  s
Fluoroscopy-to-CT registration 35.5±7.0  s 40.3±3.3  s
Video navigation
Video feature extraction and marker localization 56.9±9.4  ms
Video-to-fluoroscopy registration 1.5±0.6  ms
Fluoroscopy/CT augmentation 40.1±6.4  ms

The lower part of Table 4 summarizes runtime for the continuous video navigation loop, consisting of video feature extraction, marker localization, video-to-fluoroscopy registration, and augmentation of the fluoroscopy image or CT volume with the trajectory. Total runtime for the video guidance loop was 96.2±11.4  ms, permitting an update rate of 10 frames per second. Runtime in the current implementation is limited primarily by the video feature extraction step (59% of runtime in the video navigation loop), with potential speedup through alternative image segmentation schemes for detection of candidate markers and multiresolution, image pyramid strategies30 for marker extraction. Taken together—and recognizing the intermittent nature of the slower (fluoroscopy/CT registration) step—the results are fairly promising with respect to further development and potentially practical clinical implementation.

4. Discussion and Conclusions

A video-based system for surgical drill guidance in pelvic trauma surgery was reported. The system uses a drill-mounted miniature video camera and multimodal markers visible in both video and fluoroscopic images for real-time registration of the drill axis with respect to the patient. The registered drill trajectory is augmented onto fluoroscopic images (or a preoperative CT via 3D–2D registration) in real time to provide navigation and guidance to surgeons, particularly in the important task of initial K-wire entry point placement and orientation.

Feasibility assessment in phantom evaluated system performance in the context of common pelvic bone corridors. Drill axis registration solved with multiple (Nview>1) fluoroscopic views demonstrated conformance within multiple pelvic bone corridors (viz., the AIIS-PSIS, PC, and SR), with margins sufficient for common K-wire sizes (median TREx=0.9 mm and TREφ=2.0  deg for M=5 markers). Registration solved with a single (Nview=1) fluoroscopic view was notably less accurate (median TREx=3.4 mm and TREφ=2.7  deg for M=5 markers), but still yielded acceptable conformance within bone corridors. While registration based on multiple fluoroscopic views is more accurate, single-view registration may be adequate in scenarios of complex setups (e.g., external frames of fixators that constrain C-arm rotation), non-isocentric C-arms (which may require gross repositioning of the C-arm to maintain structures within the FOV), and/or rapid workflow requirements. Experiments showed that registration was robust with as few as four markers (M4) within the FOV. Experiments also quantified errors during K-wire insertion, demonstrating relatively consistent registration performance along the trajectory, subject to additional mechanical sources of error that may arise during the drilling process (e.g., K-wire bending).

Some previously reported video-based guidance solutions2125 require the placement of fiducial markers during preoperative setup to establish CT-to-patient registration. This requirement presents a major bottleneck in workflow and is likely infeasible in emergent trauma scenarios in which the CT is acquired for a multitude of purposes in diagnosis, rule-out, and planning. Moreover, such methods require the markers to be unperturbed during the procedure.45 The proposed system eliminates this bottleneck by establishing image-to-patient registration intraoperatively via 3D–2D registration (i.e., fluoroscopy-to-CT registration) with each fluoroscopic view. The resulting registration accuracy (δV<2  mm) is comparable to that of established surgical tracking systems.46 Furthermore, the system may operate independent of (i.e., without) a preoperative CT via video-fluoro mode to allow navigation in cases for which CT is unavailable or obviated due to clear emergent need for surgery—which is often the case in trauma surgery.47

The system uses routinely acquired fluoroscopic images for registration and a minimum of additional hardware components—viz., a drill-mounted camera that can be swapped between drills (and drill axis calibration established with a simple jig) and multimodal markers placed about the surgical field (e.g., via sticky-back tabs). The system has a strong degree of robustness against variations in marker positions that may commonly occur during the procedure. Unlike conventional navigation systems which often require time consuming, manual re-registration steps, the proposed system automatically updates the registration with each new fluoroscopic view, thereby restoring navigation with minimal impact on workflow. The system therefore appears to be potentially compatible with standard workflows in mainstream orthopaedic trauma surgery, where the adoption of surgical tracking systems has been limited. The system has the potential to reduce trial-and-error in K-wire placement and thereby improve precision, reduce radiation dose, and reduce procedure time. Furthermore, augmentation of the drill trajectory in fluoroscopy and/or CT is anticipated to improve the accuracy of device placement in a manner analogous to the improvements gained by incorporation of trackers in navigated brain or spine surgery—potentially bringing the benefits of precision guidance to mainstream orthopaedic trauma surgery.

Limitations in the current work include practical considerations in the implementation of hardware components (viz., the multimodal markers and camera). The marker size (Ø50  mm) in the initial prototype was somewhat large and limited the number of markers and arrangements that could be conveniently arranged about the surgical field (M5). Reducing the form factor of the markers is an important next step for clinical translation, as it will allow: (1) a larger number of markers to be used for registration; (2) an increase in the operating space about the surgical site, and (3) greater flexibility in placing the markers about the patient. Alternative realizations of the multimodal markers include a flexible mask or frame that bears the marker arrangement all at once and eliminates the need to individually place markers about the surgical field. Similarly, the initial camera configuration used a somewhat bulky drill mount, and ongoing work aims to streamline the mount and refine the camera selection for improved resolution and FOV. Future work will identify methods for sterilizing hardware components prior to surgery, including: (1) use of a sterile, disposable camera drape with a clear camera lens cover; and (2) the use of sterile, disposable multimodal markers.

In the current work, fluoroscopy-to-CT registration operated under the assumption of rigidity between the patient and preoperative CT. This assumption is reasonable in scenarios with simple, non-displaced fractures (which account for the majority of fragility fractures48) but is challenged with complex, comminuted fractures that exhibit strong anatomical mismatch as they are manipulated during the procedure. Methods to overcome such challenges have been proposed, including image masking (to focus registration on a particular region of interest and masking the effect of comminuted/displaced fracture components on the registration) and recently developed multibody 3D–2D registration methods4951 that account for displacement of multiple, fractured bone fragments by computing the pose of multiple rigid components through statistical pose modeling. Integration of such methods with the drill-mounted video system could improve the robustness of fluoroscopy-to-CT registration and extend its applicability beyond non-displaced fractures.

The current work is very specifically focused on the first implementation of the system, development of requisite registration algorithms, and proof of concept with quantitative assessment of registration accuracy and runtime—recognizing that such studies were conducted within the limited context of anthropomorphic phantoms. The current work investigated the feasibility of the video-drill system for fluoroscopically guided procedures, focusing on the particularly challenging step of identifying the initial entry point and orientation. The experiments also provided initial insight on considerations such as K-wire bending during drilling, and future work will investigate additional aspects of practical use, including camera jitter. The current studies were sufficient to provide quantitative insight on registration performance and tradeoffs in accuracy with a variable number of fluoroscopic views (Nview) and markers (M) while quantifying end-to-end-performance and conformance within pelvic bone corridors. Future work will further assess system performance in pre-clinical cadaver studies in which the system will be operated in real time by a fellowship-trained orthopaedic surgeon. Studies will assess the accuracy of K-wire delivery by both novice and experienced surgeons with both the proposed system for video-drill guidance and a conventional approach with fluoroscopy alone. In addition to accuracy (e.g., conformance of the guidewire in bone corridors), the workflow implications of the system will also be evaluated in terms of fluoroscopy time and procedure time compared to conventional fluoroscopic guidance.

In conclusion, a video-based surgical drill guidance system was described and tested in initial feasibility studies, addressing several issues that may be suitable for the fast-paced workflow of orthopaedic trauma procedures. The accuracy and feasibility of the system were quantified in terms of guidewire placement in phantom, serving as an important precursor for future pre-clinical studies. The current studies focused on the context of pelvic trauma surgery, and the proposed solution may be generalizable to a broader range of image-guided procedures (e.g., hip, femur, knee, and wrist surgery) that conventionally use fluoroscopy to guide instrument placement.

Acknowledgments

This work was supported by funding from National Institutes of Health, R21-EB-028330. Special thanks to Philipp Stolka, Pezhman Foroughi, Ehsan Basafa, and Dorothee Heisenberg (Clear Guide Medical, Baltimore, Maryland) for productive discussion on system design.

Biography

Biographies of the authors are not available.

Disclosures

No conflicts of interest, financial or otherwise, are declared by the authors.

Contributor Information

Prasad Vagdargi, Email: prasad@jhu.edu.

Niral Sheth, Email: nsheth8@jhu.edu.

Alejandro Sisniega, Email: asisniega@gmail.com.

Ali Uneri, Email: ali.uneri@jhu.edu.

Tharindu De Silva, Email: tsameera1@gmail.com.

Greg M. Osgood, Email: gosgood2@jhmi.edu.

Jeffrey H. Siewerdsen, Email: jsiewerd@gmail.com.

References

  • 1.Burge R., et al. , “Incidence and economic burden of osteoporosis-related fractures in the United States, 2005–2025,” J. Bone Miner. Res. 22(3), 465–475 (2007). 10.1359/jbmr.061113 [DOI] [PubMed] [Google Scholar]
  • 2.Buller L. T., Best M. J., Quinnan S. M., “A nationwide analysis of pelvic ring fractures,” Geriatr. Orthop. Surg. Rehab. 7(1), 9–17 (2015). 10.1177/2151458515616250 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Balogh Z., et al. , “The epidemiology of pelvic ring fractures: a population-based study,” J. Trauma 63(5), 1066–1073 (2007). 10.1097/TA.0b013e3181589fa4 [DOI] [PubMed] [Google Scholar]
  • 4.Davis D. D., et al. , “Pelvic fracture,” StatPearls Publishing, 2020, http://www.ncbi.nlm.nih.gov/pubmed/28613485 (accessed 22 June 2020).
  • 5.Gras F., et al. , “2D-fluoroscopic navigated percutaneous screw fixation of pelvic ring injuries—a case series,” BMC Musculoskelet. Disorder 11, 153 (2010). 10.1186/1471-2474-11-153 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Van Den Bosch E. W., et al. , “Functional outcome of internal fixation for pelvic ring fractures,” J. Trauma 47(2), 365–371 (1999). 10.1097/00005373-199908000-00026 [DOI] [PubMed] [Google Scholar]
  • 7.Burgess A. R., et al. , “Pelvic ring disruptions: effective classification system and treatment protocols,” J. Trauma 30(7), 848–856 (1990). 10.1097/00005373-199007000-00015 [DOI] [PubMed] [Google Scholar]
  • 8.Gao H., et al. , “Minimally invasive fluoro-navigation screw fixation for the treatment of pelvic ring injuries,” Surg. Innov. 18(3), 279–284 (2011). 10.1177/1553350611399587 [DOI] [PubMed] [Google Scholar]
  • 9.Hilgert R. E., Finn J., Egbers H. J., “Technik der perkutanen SI-verschraubung mit unterstützung durch konventionellen C-bogen,” Unfallchirurg 108(11), 954–960 (2005). 10.1007/s00113-005-0967-3 [DOI] [PubMed] [Google Scholar]
  • 10.Mahajan A., “Occupational radiation exposure from C arm fluoroscopy during common orthopaedic surgical procedures and its prevention,” J. Clin. Diagn. Res. 9(3), RC01 (2015). 10.7860/JCDR/2015/10520.5672 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Helm P. A., et al. , “Spinal navigation and imaging: history, trends, and future,” IEEE Trans. Med. Imaging 34(8), 1738–1746 (2015). 10.1109/TMI.2015.2391200 [DOI] [PubMed] [Google Scholar]
  • 12.Härtl R., et al. , “Worldwide survey on the use of navigation in spine surgery,” World Neurosurg. 79(1), 162–172 (2013). 10.1016/j.wneu.2012.03.011 [DOI] [PubMed] [Google Scholar]
  • 13.Navab N., Bani-Kashemi A., Mitschke M., “Merging visible and invisible: two camera-augmented mobile C-arm (CAMC) applications,” in Proc. 2nd IEEE and ACM Int. Workshop Augmented Reality, Institute of Electrical and Electronics Engineers Inc., pp. 134–141 (1999). 10.1109/IWAR.1999.803814 [DOI] [Google Scholar]
  • 14.Traub J., et al. , “A multi-view opto-xray imaging system,” Lect. Notes Comput. Sci. 4792, 18–25 (2007). 10.1007/978-3-540-75759-7_3 [DOI] [PubMed] [Google Scholar]
  • 15.Navab N., Heining S. M., Traub J., “Camera augmented mobile C-arm (CAMC): calibration, accuracy study, and clinical applications,” IEEE Trans. Med. Imaging 29(7), 1412–1423 (2010). 10.1109/TMI.2009.2021947 [DOI] [PubMed] [Google Scholar]
  • 16.Reaungamornrat S., et al. , “An on-board surgical tracking and video augmentation system for C-arm image guidance,” Int. J. Comput. Assist. Radiol. Surg. 7, 647–665 (2012). 10.1007/s11548-012-0682-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Yoo J., et al. , “An electromagnetic ‘tracker-in-table’ configuration for x-ray fluoroscopy and cone-beam CT-guided surgery,” Int. J. Comput. Assist. Radiol. Surg. 8(1), 1–13 (2013). 10.1007/s11548-012-0744-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Magaraggia J., et al. , “Design and evaluation of a portable intra-operative unified-planning-and-guidance framework applied to distal radius fracture surgery,” Int. J. Comput. Assist. Radiol. Surg. 12, 77–90 (2017). 10.1007/s11548-016-1432-1 [DOI] [PubMed] [Google Scholar]
  • 19.Von Jako C. R., et al. , “A novel accurate minioptical tracking system for percutaneous needle placement,” IEEE Trans. Biomed. Eng. 60(8), 2222–2225 (2013). 10.1109/TBME.2013.2251883 [DOI] [PubMed] [Google Scholar]
  • 20.Vetter S. Y., et al. , “Virtual guidance versus virtual implant planning system in the treatment of distal radius fractures,” Int. J. Med. Rob. Comput. Assist. Surg. 14(5), e1945 (2018). 10.1002/rcs.1945 [DOI] [PubMed] [Google Scholar]
  • 21.Basafa E., Hoßbach M., Stolka P. J., “Fast, intuitive, vision-based: performance metrics for visual registration, instrument guidance, and image fusion,” Lect. Notes Comput. Sci. 9958, 9–17 (2016). 10.1007/978-3-319-46472-5_2 [DOI] [Google Scholar]
  • 22.Shahidi R., et al. , “Implementation, calibration and accuracy testing of an image-enhanced endoscopy system,” IEEE Trans. Med. Imaging 21(12), 1524–1535 (2002). 10.1109/TMI.2002.806597 [DOI] [PubMed] [Google Scholar]
  • 23.Maurer C. R., et al. , “Registration of head volume images using implantable fiducial markers,” IEEE Trans. Med. Imaging 16(4), 447–462 (1997). 10.1109/42.611354 [DOI] [PubMed] [Google Scholar]
  • 24.Roessler K., et al. , “Frameless stereotactic guided neurosurgery: clinical experience with an infrared based pointer device navigation system,” Acta Neurochir. 139(6), 551–559 (1997). 10.1007/BF02750999 [DOI] [PubMed] [Google Scholar]
  • 25.Grunert P., et al. , “Stereotactic biopsies guided by an optical navigation system: technique and clinical experience,” Minimally Invasive Neurosurg. 45(1), 11–15 (2002). 10.1055/s-2002-23576 [DOI] [PubMed] [Google Scholar]
  • 26.Hamming N. M., et al. , “Automatic image-to-world registration based on x-ray projections in cone-beam CT-guided interventions,” Med. Phys. 36(5), 1800–1812 (2009). 10.1118/1.3117609 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Dang H., et al. , “Robust methods for automatic image-to-world registration in cone-beam CT interventional guidance,” Med. Phys. 39(10), 6484–6498 (2012). 10.1118/1.4754589 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Garrido-Jurado S., et al. , “Automatic generation and detection of highly reliable fiducial markers under occlusion,” Pattern Recognit. 47(6), 2280–2292 (2014). 10.1016/j.patcog.2014.01.005 [DOI] [Google Scholar]
  • 29.Garrido-Jurado S., et al. , “Generation of fiducial marker dictionaries using mixed integer linear programming,” Pattern Recognit. 51, 481–491 (2016). 10.1016/j.patcog.2015.09.023 [DOI] [Google Scholar]
  • 30.Romero-Ramirez F. J., Muñoz-Salinas R., Medina-Carnicer R., “Speeded up detection of squared fiducial markers,” Image Vision Comput. 76, 38–47 (2018). 10.1016/j.imavis.2018.05.004 [DOI] [Google Scholar]
  • 31.Yonghong X., Qiang J., “A new efficient ellipse detection method,” in Proc. Int. Conf. Pattern Recognit., Vol. 16, pp. 957–960 (2002). 10.1109/icpr.2002.1048464 [DOI] [Google Scholar]
  • 32.Zhang Z., “A flexible new technique for camera calibration,” IEEE Trans. Pattern Anal. Mach. Intell. 22(11), 1330–1334 (2000). 10.1109/34.888718 [DOI] [Google Scholar]
  • 33.Vagdargi P., et al. , “Calibration and registration of a freehand video-guided surgical drill for orthopaedic trauma,” Proc. SPIE 11315, 113150G (2020). 10.1117/12.2550001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Torr P. H. S., Zisserman A., “MLESAC: a new robust estimator with application to estimating image geometry,” Comput. Vision Image Understanding 78(1), 138–156 (2000). 10.1006/cviu.1999.0832 [DOI] [Google Scholar]
  • 35.Cho Y., et al. , “Accurate technique for complete geometric calibration of cone-beam computed tomography systems,” Med. Phys. 32(4), 968–983 (2005). 10.1118/1.1869652 [DOI] [PubMed] [Google Scholar]
  • 36.Uneri A., et al. , “Known-component 3D–2D registration for quality assurance of spine surgery pedicle screw placement,” Phys. Med. Biol. 60(20), 8007–8024 (2015). 10.1088/0031-9155/60/20/8007 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Otake Y., et al. , “Automatic localization of vertebral levels in x-ray fluoroscopy using 3D–2D registration: a tool to reduce wrong-site surgery,” Phys. Med. Biol. 57(17), 5485–5508 (2012). 10.1088/0031-9155/57/17/5485 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Otake Y., et al. , “Robust 3D-2D image registration: application to spine interventions and vertebral labeling in the presence of anatomical deformation,” Phys. Med. Biol. 58(23), 8535–8553 (2013). 10.1088/0031-9155/58/23/8535 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Hansen N., Ostermeier A., “Completely derandomized self-adaptation in evolution strategies,” Evol. Comput. 9(2), 159–195 (2001). 10.1162/106365601750190398 [DOI] [PubMed] [Google Scholar]
  • 40.Horn B. K. P., “Closed-form solution of absolute orientation using unit quaternions,” J. Opt. Soc. Am. A 4(4), 629 (1987). 10.1364/JOSAA.4.000629 [DOI] [Google Scholar]
  • 41.De Silva T., et al. , “3D–2D image registration for target localization in spine surgery: investigation of similarity metrics providing robustness to content mismatch,” Phys. Med. Biol. 61(8), 3009–3025 (2016). 10.1088/0031-9155/61/8/3009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Goerres J., et al. , “Planning, guidance, and quality assurance of pelvic screw placement using deformable image registration,” Phys. Med. Biol. 62(23), 9018–9038 (2017). 10.1088/1361-6560/aa954f [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Han R., et al. , “Atlas-based automatic planning and 3D-2D fluoroscopic guidance in pelvic trauma surgery,” Phys. Med. Biol. 64(9), 095022 (2019). 10.1088/1361-6560/ab1456 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Fornaciari M., Prati A., Cucchiara R., “A fast and effective ellipse detector for embedded vision applications,” Pattern Recognit. 47(11), 3693–3708 (2014). 10.1016/j.patcog.2014.05.012 [DOI] [Google Scholar]
  • 45.Zheng G., Nolte L. P., “Computer-assisted orthopedic surgery: current state and future perspective,” Front Surg. 2, 66 (2015). 10.3389/fsurg.2015.00066 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Chassat F., Lavallée S., “Experimental protocol of accuracy evaluation of 6-D localizers for computer-integrated surgery: application to four optical localizers,” Lect. Notes Comput. Sci. 1496, 277–284 (1998). 10.1007/bfb0056211 [DOI] [Google Scholar]
  • 47.Stöckle U., Schaser K., König B., “Image guidance in pelvic and acetabular surgery—expectations, success and limitations,” Injury 38(4), 450–462 (2007). 10.1016/j.injury.2007.01.024 [DOI] [PubMed] [Google Scholar]
  • 48.Rommens P. M., Wagner D., Hofmann A., “Fragility fractures of the pelvis,” JBJS Rev. 5(3), e3 (2017). 10.2106/JBJS.RVW.16.00057 [DOI] [PubMed] [Google Scholar]
  • 49.Han R., et al. , “Multi-body 3D–2D registration for image-guided reduction of pelvic dislocation in orthopaedic trauma surgery,” Phys. Med. Biol. 65(13), 135009 (2020). 10.1088/1361-6560/ab843c [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Han R., et al. , “Multi-body registration for fracture reduction in orthopaedic trauma surgery,” Proc. SPIE 11315, 113150F (2020). 10.1117/12.2549708 [DOI] [Google Scholar]
  • 51.Han R., et al. , “Fracture reduction planning and guidance in orthopaedic trauma surgery via multi-body image registration,” Med. Image Anal. 68, 101917 (2021). 10.1016/j.media.2020.101917 [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Journal of Medical Imaging are provided here courtesy of Society of Photo-Optical Instrumentation Engineers

RESOURCES