Abstract
The Challenge on Liver Ultrasound Tracking (CLUST) was held in conjunction with the MICCAI 2014 conference to enable direct comparison of tracking methods for this application. This paper reports the outcome of this challenge, including setup, methods, results and experiences. The database included 54 2D and 3D sequences of the liver of healthy volunteers and tumor patients under free breathing. Participants had to provide the tracking results of 90% of the data (test set) for pre-defined point-landmarks (healthy volunteers) or for tumor segmentations (patient data). In this paper we compare the best six methods which participated in the challenge. Quantitative evaluation was performed by the organizers with respect to manual annotations. Results of all methods showed a mean tracking error ranging between 1.4 mm and 2.1 mm for 2D points, and between 2.6 mm and 4.6 mm for 3D points. Fusing all automatic results by considering the median tracking results, improved the mean error to 1.2 mm (2D) and 2.5 mm (3D). For all methods, the performance is still not comparable to human inter-rater variability, with a mean tracking error of 0.5–0.6 mm (2D) and 1.2–1.8 mm (3D). The segmentation task was fulfilled only by one participant, resulting in a Dice coefficient ranging from 76.7% to 92.3%. The CLUST database continues to be available and the online leader-board will be updated as an ongoing challenge.
Keywords: respiratory motion, ultrasound, tracking, image registration, image guidance, motion estimation, challenge
1. Introduction
Ultrasound (US) imaging is a widely used medical imaging technique. As US has high temporal resolution and is non-ionizing, it is an appealing choice for applications which require tracking and tissue motion analysis, such as motion compensation in image-guided intervention and therapy. Conformal and minimally invasive tumor treatments, such as high intensity focused ultrasound and intensity-modulated radiation and proton therapy, deliver highly localized dose into the target tissues. Yet, the motion of the organs in the treatment region is a critical limitation. Specifically, we want to address the issue of respiratory motion in the liver (Keall et al 2006, Shirato et al 2007). Note that liver tumors are not necessarily visible in US images. Instead, other visible structures (e.g. vessels) are tracked and these tracking results are used as input surrogate data to 4D motion models to predict the tumor position (Tanner et al 2012, McClelland et al 2013).
Despite the rapid development of image-guided therapy, intervention systems and medical imaging tools, the translation into clinical practice of automated motion estimation is limited. One of the main reasons for algorithms not being integrated in clinical practice is the lack of adequate validation. Open datasets for designing and testing tracking algorithms are missing, and private datasets differ in size, image dimension and sequence length. The variations in tracking objective (full organ, anatomical landmarks, tumor) and validation strategies are additional impediments to strategy comparisons. For image-guided therapies, tracking methods should have high accuracy, robustness over the duration of the therapy and real-time capability.
Several methods have been proposed for tracking human liver structures on US sequences. Yet quantitative evaluation of tracking the human liver under free breathing was reported only by Banerjee et al (2015), Harris et al (2010), Lediju Bell et al (2012) and Vijayan et al (2014) for 3D sequences and by Cifor et al (2012), (2013), De Luca et al (2012) and De Luca et al (2013) for 2D sequences. Intensity-based and hybrid approaches achieved good accuracy (∼1.4 mm mean tracking error (Harris et al 2010) and ∼90% mean overlap ratio (Cifor et al 2013)). The non-rigid registration method of Vijayan et al (2014) estimated liver motion with an error of 1 mm (75% percentile of a root-mean squared metric over all datasets), which was lower than the inter-observer variability of 1.4 mm. More recently, a fast 3D affine block-matching algorithm with an outlier rejection strategy achieved a mean tracking error of 1.8 mm (Banerjee et al 2015). However, these methods were only tested off-line on short sequences (<1 min). For longer sequences (5–10 min long), a learning-based block-matching algorithm (De Luca et al 2013) achieved a mean tracking accuracy of mm.
In this paper we present the outcome of the open challenge for the validation of liver motion tracking algorithms. The reported methods are selected from the ones presented at the open challenge on liver US tracking (CLUST, http://clust14.ethz.ch), held in conjunction with the 2014 international conference on medical image computing and computer assisted intervention (MICCAI 2014). The aim of CLUST was to present the current state-of-the-art in automated tracking of anatomical landmarks in the liver (vessel centers (2D), vessel bifurcations (3D) and tumor contours (2D)) and enable comparison between different methods. For the test set, the annotations of the first images were provided, which needed to be tracked over time. This paper reports the results for the full test set, while the challenge proceedings exclude 20% of the test data, which were distributed shortly before the MICCAI conference. Furthermore this publication reports results after imposing restrictions against adjusting parameters per sequence by visual inspection (which is not realistic). Thus, method parameters were either automatically determined or generally fixed for at most each US scanner and task. In addition, we investigate the inter-observer variability of the annotations, analyze various aspects influencing the tracking performance, and explore the tracking performance when fusing results.
The paper is organized as follows. In section 2 we describe the challenge data and tracking objectives. In section 3 the methods proposed by the 6 participant groups are presented. The fusion of all methods, by considering the median of their results, is also considered. The evaluation criteria are described in section 4 and the tracking results are reported and compared in section 5. Discussion and conclusions of the challenge outcome are provided in sections 6 and 7.
2. Materials
2.1. Ultrasound data
The collected database included a total of 54 US sequences of the liver of patients and volunteers under free breathing. The data were provided by 6 groups, namely the Computer Vision Laboratory, ETH Zurich, Switzerland (ETH) (De Luca et al 2013, Preiswerk et al 2014); mediri GmbH, Heidelberg, Germany (MED); Institute of Biomedical Engineering, University of Oxford, UK (OX) (Cifor et al 2013); Biomedical Imaging Group, Departments of Radiology and Medical Informatics, Erasmus MC, Rotterdam, The Netherlands (EMC) (Banerjee et al 2014); Joint Department of Physics, Institute of Cancer Research & Royal Marsden NHS Foundation Trust, London and Sutton, UK (ICR) (Lediju et al 2010, Lediju Bell et al 2012); and SINTEF Medical Technology, Image Guided Therapy, Trondheim, Norway (SMT) (Vijayan et al 2013).
The sequences were acquired with 6 US scanners, 7 types of transducer and different acquisition settings. An overview of the data is given in table 1. The length of the sequences ranges from 4 s to 10 min, with a temporal resolution in the range of 6–25 Hz. The dataset is divided into three subsets, according to the image dimension and annotation type. The first subset is composed of 28 2D sequences from healthy volunteers with point-landmark annotations. The second subset contains 10 2D sequences from 5 patients with segmentation annotations. The third subset consists of 16 3D sequences with point-landmark annotations from healthy volunteers. The data were anonymized and divided into a training set (10% of the sequences) and a test set (90%). For the training set annotations were released, to allow for some tuning of the tracking algorithm. For the test set, the annotations of the first images were provided. These needed to be tracked over time. Examples of the first frames and annotations are shown in figure 1.
Table 1.
Summary of the challenge data with annotation of point-landmarks (2D sequences: ETH, MED1 and MED2; and 3D sequences: EMC, ICR and SMT) and segmentations of tumor areas (2D sequences: OX).
| Sequence | Sequence info | J: No. annota. | No.ann. frames | Acquisition info | |||||
|---|---|---|---|---|---|---|---|---|---|
| Im.size [pix] | Im.res. [mm] | T: No. frames | Im.rate [Hz] | Scanner | Probe | Center Freq. [MHz] | |||
| ETH-01 | 0.71 | 14516 | 25 | 1 | 1453 | Siemens Antares | CH4-1 | 2.22 | |
| ETH-02 | 0.40 | 5244 | 16 | 1 | 525 | Siemens Antares | CH4-1 | 2.00 | |
| ETH-03 | 0.36 | 5578 | 17 | 3 | 559 | Siemens Antares | CH4-1 | 1.82 | |
| ETH-04 | 0.42 | 2620 | 15 | 1 | 263 | Siemens Antares | CH4-1 | 2.22 | |
| ETH-05 | 0.40 | 3652 | 15 | 2 | 366 | Siemens Antares | CH4-1 | 2.22 | |
| ETH-06 | 0.37 | 5586 | 17 | 2 | 560 | Siemens Antares | CH4-1 | 1.82 | |
| ETH-07 | 0.28 | 4588 | 14 | 1 | 460 | Siemens Antares | CH4-1 | 2.22 | |
| ETH-08 | 0.36 | 5574 | 17 | 2 | 558 | Siemens Antares | CH4-1 | 1.82 | |
| ETH-09 | 0.40 | 5247 | 16 | 2 | 525 | Siemens Antares | CH4-1 | 1.82 | |
| ETH-10 | 0.40 | 4587 | 15 | 4 | 460 | Siemens Antares | CH4-1 | 1.82 | |
| ETH-11 | 0.42 | 4615 | 15 | 2 | 463 | Siemens Antares | CH4-1 | 1.82 | |
| ETH-12 | 0.45 | 4284 | 14 | 2 | 429 | Siemens Antares | CH4-1 | 2.22 | |
| MED-01 | 0.41 | 2470 | 20 | 3 | 248 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-02 | 0.41 | 2478 | 20 | 3 | 248 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-03 | 0.41 | 2456 | 20 | 4 | 246 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-04 | 0.41 | 2455 | 20 | 3 | 246 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-05 | 0.41 | 2458 | 20 | 3 | 246 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-06 | 0.41 | 2443 | 20 | 3 | 245 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-07 | 0.41 | 2450 | 20 | 3 | 246 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-08 | 0.41 | 2442 | 20 | 2 | 245 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-09 | 0.41 | 2436 | 20 | 5 | 244 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-10 | 0.41 | 2427 | 20 | 4 | 243 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-11 | 0.41 | 2424 | 20 | 3 | 243 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-12 | 0.41 | 2450 | 20 | 3 | 246 | DiPhAs Fraunhofer | VermonCLA | 5.5 | |
| MED-13 | 0.35 | 3304 | 11 | 3 | 331 | Zonare z.one | C4-1 | 4.0 | |
| MED-14 | 0.35 | 3304 | 11 | 3 | 331 | Zonare z.one | C4-1 | 4.0 | |
| MED-15 | 0.35 | 3304 | 11 | 1 | 331 | Zonare z.one | C4-1 | 4.0 | |
| MED-16 | 0.35 | 3304 | 11 | 2 | 331 | Zonare z.one | C4-1 | 4.0 | |
| OX-01 | 0.30 | 71 | 12 | 1 | 71 | Zonare z.one | P4-1 | 3.6 | |
| OX-02 | 0.40 | 82 | 12 | 1 | 82 | Zonare z.one | P4-1 | 3.6 | |
| OX-03 | 0.38 | 82 | 12 | 1 | 82 | Zonare z.one | P4-1 | 3.4 | |
| OX-04 | 0.36 | 51 | 14.5 | 1 | 51 | Zonare z.one | C6-2 | 4.4 | |
| OX-05 | 0.46 | 101 | 11.7 | 1 | 101 | Zonare z.one | C6-2 | 3.8 | |
| OX-06 | 0.55 | 76 | 11 | 1 | 76 | Zonare z.one | C6-2 | 3.8 | |
| OX-07 | 0.50 | 63 | 10 | 2 | 63 | Zonare z.one | C6-2 | 3.8 | |
| OX-08 | 0.46 | 105 | 11 | 1 | 105 | Zonare z.one | C6-2 | 3.8 | |
| OX-09 | 0.50 | 98 | 10 | 2 | 98 | Zonare z.one | C6-2 | 3.8 | |
| OX-10 | 0.55 | 92 | 11 | 1 | 92 | Zonare z.one | C6-2 | 3.8 | |
| EMC-01 | 79 | 6 | 1 | 8 | iU22 | X6-1 | 3.2 | ||
| EMC-02 | 54 | 6 | 4 | 6 | iU22 | X6-1 | 3.2 | ||
| EMC-03 | 159 | 6 | 1 | 16 | iU22 | X6-1 | 3.2 | ||
| EMC-04 | 140 | 6 | 1 | 15 | iU22 | X6-1 | 3.2 | ||
| EMC-05 | 147 | 6 | 1 | 15 | iU22 | X6-1 | 3.2 | ||
| ICR-01 | 141 | 24 | 1 | 15 | Siemens SC2000 | 4Z1c | 2.8 | ||
| ICR-02 | 141 | 24 | 1 | 20 | Siemens SC2000 | 4Z1c | 2.8 | ||
| SMT-01 | 0.70 | 97 | 8 | 3 | 96 | GE E9 | 4V-D | 2.5 | |
| SMT-02 | 0.70 | 96 | 8 | 3 | 92-93 | GE E9 | 4V-D | 2.5 | |
| SMT-03 | 0.70 | 96 | 8 | 2 | 45-96 | GE E9 | 4V-D | 2.5 | |
| SMT-04 | 0.70 | 97 | 8 | 1 | 96 | GE E9 | 4V-D | 2.5 | |
| SMT-05 | 0.70 | 96 | 8 | 2 | 64-96 | GE E9 | 4V-D | 2.5 | |
| SMT-06 | 0.70 | 97 | 8 | 3 | 49-96 | GE E9 | 4V-D | 2.5 | |
| SMT-07 | 0.70 | 97 | 8 | 2 | 95 | GE E9 | 4V-D | 2.5 | |
| SMT-08 | 0.70 | 97 | 8 | 3 | 96 | GE E9 | 4V-D | 2.5 | |
| SMT-09 | 0.70 | 97 | 8 | 3 | 96 | GE E9 | 4V-D | 2.5 | |
Note: The test set is listed in black font. The training sequences, for which all available annotations were provided, are highlighted in red.
Figure 1.
Examples of first frame I(0) of the training data: (top row) 2D sequences (ETH, MED, OX) and (bottom row) 3D sequences (EMC, SMT). Point-landmarks Pj(0) and the contour of the tumor segmentation Sj(0) are depicted in yellow.
Given a sequence of T images I(t), denotes the intensity (or brightness) of image I(t) at position at frame t, with and (D = 2, 3). Depending on the subset, the tracking objective was to compute either the position of J point-landmarks or the segmentation of J tumor areas in each image, with and in this challenge. The displacement of landmark j at time t is denoted as . For all sequences, annotations of the first frame (Pj(0) or Sj(0)) were provided.
3. Methods
The challenge raised interest from 55 individuals worldwide. All of them successfully downloaded the data. Only 8 from the downloaders submitted their results to CLUST14. After the withdrawal of one of the groups, a total of 7 papers were accepted to the MICCAI workshop. In this paper we included the 6 contributions with the highest mean accuracy, namely (in alphabetical order of the abbreviations) from Konica Minolta Inc., Osaka, Japan (KM); Fraunhofer MEVIS, Lübeck, Germany (MEVIS); Fraunhofer MEVIS, Bremen and Fraunhofer FOKUS, Berlin, Germany (MEVIS + FOKUS); Fraunhofer MEVIS, Bremen and Mediri GmbH, Heidelberg, Germany (MEVIS + MED); Philips Research, Suresnes, France (PhR); and Technische Universität München, Germany (TUM).
In the following we briefly describe each algorithm. An overview of the methods’ main features is presented in table 2. All 6 presented methods were tested on 2D landmark tracking. Of these, 3 were extended to 3D tracking, while only one submission covered all the challenge tasks. A more detailed description of each method can be found in the workshop proceedings15.
Table 2.
Summary of the main features of the evaluated tracking methods: the tracking Task namely 2D and 3D point-landmark (2D p. and 3D p. respectively) and 2D tumor segmentation (2D s.); the key components of the tracking algorithm (Keywords).
| Participant | Task | Keywords | Real-time | ||
|---|---|---|---|---|---|
| 2D p. | 2D s. | 3D p. | |||
| KM | ✓ | ✗ | ✗ | block matching, NCC, local translation, exhaustive search | ✗ |
| MEVIS | ✓ | ✗ | ✗ | variational, large moving ROI, SSD & NGF, curvature regularizer | ✗ |
| MEVIS + FOKUS | ✓ | ✗ | ✓ | optical flow, histogram equalization, 30% downsampling, polynomial expansion, bilateral filtering, outlier detection, 2 orthogonal slices | ✓ |
| MEVIS + MED | ✓ | ✗ | ✓ | Bayesian approach, particle filter, intensity difference, local translation | ✓ |
| PhR | ✓ | ✓ | ✓ | sparse Demons, ROI, SSD, fluid regularizer, gradient descent, drift prevention strategy | ✓ |
| TUM | ✓ | ✗ | ✗ | kernel-based, intensity distribution similarity, adaptive ellipsoidal target descriptor, local affine, failure recovery strategy | ✓ |
Note: Real-time capability was assessed w.r.t. the average frame rate of 20 Hz (50 ms) for 2D sequences and 8 Hz (125 ms) for 3D sequences.
3.1. KM: template matching
Kondo, Konica Minolta Inc. (KM), developed a multiple-template matching method, based assuming that all regions of interest (ROIs) from the same image move along almost the same direction, and the motion has high periodicity due to breathing. Similar to De Luca et al (2013) and Matthews et al (2004), templates are selected from a plurality of recent frames to exploit the highly periodicity. The method consists of five steps:
Step 1: selection of global and long-term templates. From the first image I(0) two templates (i.e. unchanging subimages) are selected: a global template , which is used to determine the motion on the entire frame and is defined by the largest rectangle which can be inscribed in the US image; and a long-term template , which is a squared ROI around the annotated point Pj(0), whose size is based on minimizing the variance of the pixel values inside the ROI. The set of pixel coordinates included in and are denoted as BL,j and , respectively.
Step 2: global motion estimation. For each image I(t), t > 0, the global displacement , of point Pj(0), is estimated by maximizing the normalized cross-correlation (NCC) over using exhaustive search:
| 1 |
where
| 2 |
Step 3: long-term motion estimation. Similarly the long-term motion component over is computed as
| 3 |
where SL is the set of tested pixel displacements. It depends on the global motion estimation (Step2) as follows: if then pixels, otherwise SL=±15 pixels. The motion vectors and are stored in the short-term buffer.
Step 4: short-term motion estimation. Firstly, the cycle length c is estimated from the past tracking results. Then, two short-term templates for the t-th frame are selected: the ROI with the maximum , with ; and the ROI with the minimum in . Principal component analysis (PCA) is applied to the 2D trajectory of the tracking positions, and the motion estimation is performed only in the first principal direction. The resulting is determined by providing the maximum NCC for these templates ().
Step 5: final tracking. The final motion estimation is given by:
| 4 |
where is a weight that prioritizes the long-term motion estimation.
Run-time. This method was implemented in C++ using OpenCV and OpenMP on a computer with Intel Core i7 3.3 GHz GPU, 6 cores and 64 GB memory. The average processing time was approximately 84 ms per frame, which can be improved by optimizing the motion estimation routine, currently based on an OpenCV function.
3.2. MEVIS: variational real-time registration
König et al Fraunhofer MEVIS Project Group Image Registration, Lübeck, Germany, (MEVIS) proposed a novel scheme for point tracking in long US sequences, based on an efficient, state-of-the-art variational non-linear registration method (Modersitzki 2009, König and Rühaak 2014). It is extended by a moving window scheme with additional fallback strategy to minimize tracking failures.
Registration algorithm. Transformation between images A and B was determined by a variational approach with the objective function , where is the sum of squared differences (SSD), is the edge-based normalized gradient fields (NGF) (Haber and Modersitzki 2006), and is the second order curvature regularizer (Fischer and Modersitzki 2003). See König et al (2014), König and Rühaak (2014) and Modersitzki (2009) for details on the numerical optimization.
Tracking scheme. The non-linear registration algorithm is embedded in a tracking framework. By calculating registrations of moving windows on each frame, is tracked over time, see figure 2.
Figure 2.
Tracking scheme. First W(t − 1, I (t)) is registered to W(0, I(0)) providing . If this registration fails, W(t − 1, I(t)) is registered to corresponding window of previous frame ().
is defined as a window of with extent and center position . First, window W(0,I(0)) with center Cj(0)=Pj(0) is chosen. Then W(0,I(1)) is registered to W(0,I(0)) and Pj(0) is transformed by to get Pj(1). Using Cj(t) = Pj(t) as the center of a new window, this process is then repeated for all frames.
As a safeguard, the original intensity at position Pj(0) is compared to the current intensity at Pj(t). If , the registration paradigm is switched and window W(t − 1,I(t)) is registered to W(t − 1,I(t − 1)), yielding .
Parametrization. Parameters for optimization, deformation resolution () and multi-level scheme (2 levels) were kept constant. The window size was 50 mm in each dimension. Weight is automatically determined before each registration as . The parameters and (edge parameter of NGF) were manually chosen per scanner (ETH datasets: , MED1: , MED2: ). The threshold was for ETH and MED2, and for MED1.
Run-time. The algorithm achieved close to real-time performance in all cases (23–90 ms per frame), exceeding acquisition rate in 48% of all cases, computed on a three year old Intel i7-2600 PC with 3.40 GHz. Thus real-time performance is easily within reach when using recent hardware.
3.3. MEVIS + FOKUS: high performance optical flow
The method of Lübke, Fraunhofer MEVIS, Bremen, and Grozea, Fraunhofer FOKUS, Berlin, Germany, (MEVIS + FOKUS) is based on a two-frame motion estimation by an optical flow approach using polynomial expansion (Farnebäck 2003) and builds on the OpenCV function calcOpticalFlowFarneback. It yields a motion vector field for the entire frame which allows to track multiple points in parallel without any computational overhead. The method consists of approximating images locally with quadratic polynomials and then obtaining the dense displacement field by inferring the local displacement analytically. This is based on the coefficients of the fitted polynomial surfaces, with smoothing coming from the assumption of a global parametrized displacement model. In contrast to the original method, the OpenCV implementation did not provide a certainty for each pixel. This reduced to some extent the accuracy when the tracked point approached repeatedly the border of the acquisition region.
Real-time capability is achieved by downscaling the images before tracking and filtering, and subsequent upscaling of the result to the original resolution. A fixed downscaling factor to 30% has been chosen to match the computation time with the frame-rate. Fixed-parameter bilateral filtering and histogram equalization are used to obtain more stable results. Outliers, which are defined as motion vector component changes greater than 12 pixels, are discarded and the previous positions are used instead. Linear fitting (sliding window) is used to smooth the trajectory for each dimension to eliminate high-frequency motion.
This method was applied to 2D and 3D datasets. In the latter case it is in fact a 2.5D method, as we are evaluating the motion in two orthogonal slices intersecting at the manual annotation. This yields two separate tracking results with redundant information on the intersection. The method is sensitive to out-of-plane-motion as any 2D tracking method, since no adjustment based on the combined information has yet been done.
Another approach was proposed by this group and applied on 2D datasets, see the CLUST14 proceedings for details (Lübke and Grozea 2014). It was based on patch-matching, maximizing NCC, random sampling, explicit masking to improve accuracy at the border, and GPU implementation. This approach was not included in this paper, as it was on average slower (84%) and had a similar mean accuracy (3% better) as the group’s included method.
Run-time. The method was implemented using the Mevislab software (MeVis Medical Solutions, Bremen, Germany), Python, OpenCV and Numpy. The run-time was measured on a Intel Core i7-4770k with 32 GB RAM. The average run-time for all 2D points is 40 ms. As the 3D tracking is performed on two orthogonal slices, the average processing time per volume of 61 ms is slightly higher than in the 2D datasets but still below the 3D images frame-rate (real-time), except for the ICR data subset with temporal resolution of 24 Hz.
3.4. MEVIS + MED: Bayesian tracking
Rothlübbers et al Fraunhofer MEVIS, Bremen, Germany, (MEVIS + MED) proposed a Bayesian approach for tracking point landmarks in 2D and 3D US sequences.
Particle Filter. A particle filter based algorithm (Isard and Blake 1996, Arulampalam et al 2002, Zhang et al 2010) is used to track Pj(0) in a 2D or 3D sequence. A small region around target Pj(0) is considered to translate through the sequence, yielding a d = 2,3 dimensional state space for 2D and 3D translations respectively. The state of the target is represented by a probability density function in state space, approximated by a fixed size set of N particles. Each particle n holds the state of a position hypothesis and the associated weight wn(t), normalized by , for . With each tracking step and wn are updated by estimations and observations, respectively.
Target Description. The target region is described by an under-sampled grid, characterized by radius R0 around Pj(0) and grid constant R1 (see figure 3), resulting in grid points , . Each has associated information for belonging to the bright and dark part of the region (,) derived from the intensities in the initial frame (, ) where bmax and are the maximal and minimal intensity among all .
Figure 3.

Initialization: (Left) Within radius R0 of a given position, points on a local triangular grid with grid constant R1 are chosen. (Right) Example of point weights (, see text) in a first frame: area indicates value and color encodes sign (red: negative, green: positive).
Estimation. States are updated by re-sampling (Isard and Blake 1996) and application of stochastic drift and diffusion (equation (5)), taking into account the direct predecessor state (Markov property):
| 5 |
where determines the amount of drift towards the mean state and scales the diffusion by normally-distributed Gaussian noise ().
Observation. Weights are updated from and the incoming image I(t+1) using the translation matrix to transform points into I(t+1) by:
| 6 |
When coinciding with bright image areas, bright grid points increase the weight while dark grid points decrease it. The final weighting function is
| 7 |
with Heaviside function to set negative weights to zero, and with squaring of weights as this improved performance. The tracking result is computed to be the centroid of the target model, transformed into the image by the weighted mean sample. The tracking step was repeated M times to allow the distribution’s mean to settle, leading to reduced lag.
Execution. Per target, R0 and R1 were adapted to fit the given target scenario and resolution of I(0), ranging from small speckle-like features to large dark vessels. Typical values for 2D are , pixels resulting in . The other parameters were set to N = 400, , , and M = 2. For the 3D sets with generally lower resolution, , voxels and was increased due to the additional third dimension. For 3D, parameters were set to N = 200, , , and M = 4.
Run-time. With the semi-manual initialization, the algorithm performs very efficiently, with an average run-time of ms per 2D frame and ms per 3D volume. The core source code is written in C++ and integrated into MeVisLab (MeVis Medical Solutions, Bremen, Germany) for high level evaluation routines using Python scripts. The code was executed single threaded on a Windows 7 machine with an Intel Core i7-2600 CPU @ 3.4 GHz and 32 GB RAM.
3.5. PhR—sparse demons
Somphone et al Medisys Lab, Philips Research, Suresnes, France, (PhR) based their method on the Sparse Demons framework, where a dense displacement field is found by minimizing an energy E defined only on a small number of points of interest :
| 8 |
where R and T are the reference and template images respectively, is the image domain and is the Dirac function. The dissimilarity between the images is measured by (Somphone et al 2013). Regularization is based on filtering the fluid-like field with a Gaussian kernel of scale mm (Mory et al 2012):
| 9 |
Minimizing E w.r.t. is done by gradient descent. Calculus of variations results in the following evolution equation:
| 10 |
where and is the dense gradient of E w.r.t. :
| 11 |
The subsequent algorithm has similarities with the Demons algorithm (Thirion 1998, Mansi et al 2011). Its computational complexity is however lower since image forces are only computed at points .
Algorithm 1.
Sparse Demons—Gradient Descent
| 1: ; |
| 2: ; |
| 3: repeat |
| 4: ; |
| 5: for all xi do |
| 6: ; |
| 7: ; ⊳ according to (11) |
| 8: end for |
| 9: ; |
| 10: ; |
| 11: k = k+1; |
| 12: until steady state |
2D point tracking. is chosen as dissimilarity measure. The reference points of interest are chosen in the neighborhood of the landmarks, based on having a high amplitude of the image gradient. The tracking strategy consists of two phases. In the initial (t − 1)-to-t strategy for , R = I(t − 1) and T = I(t), and a mean reference patch (of size ixels2) is built around Pj(0) from the patches centered on the tracked landmark positions. To prevent error accumulation, the 0-to-t patch tracking is used from t > 100.
3D point tracking. Reference points are chosen in the neighborhood of the landmark in a mm3 region, regularly spaced by 10 mm. The baseline tracking follows the (t − 1)-to-t scheme. To prevent drifting, a 0-to-t tracking is enabled if the difference between the histograms of I(t) and I(0) is lower than 20% of the histogram difference between I(1) and I(0), or landmarks positions are closer than 1.8 mm to Pj(0).
2D segmentation tracking. The entropy of the difference between the reference and the transformed template is used as dissimilarity measure, since the OX sequences display large intensity changes between frames. The points of interest are selected in the neighborhood of the segmentation boundary, again based on the image gradient amplitude. As the OX sequences are short, only the (t − 1)-to-t scheme is used.
Run-time The average processing time of 25 ms for 2D and 100 ms for 3D was obtained on a multithreaded PC platform. Yet, the method was not specifically optimized for run-time.
3.6. TUM: kernel-based tracking
Benz et al Technische Universität München, Germany, (TUM) proposed a kernel-based US tracking method. The target model centered at I(0,Pj(0)) and the target candidate model centered at are represented by normalized weighted intensity histograms, with m = 32 number of histogram bins. Each pixel contribution to a histogram bin u is weighted based on the radially symmetric Epanechnikov kernel (Epanechnikov 1969), which assigns smaller weights to pixel locations farther away from the center.
In each frame I(t) the goal is to find that best matches q. The discrete Bhattacharyya coefficient (Comaniciu et al 2003) is used as similarity measure between q and . In each frame t the procedure to find the location that maximizes is started at location , which initially is set to the previous solution . After linearization through a Taylor series expansion around , can be maximized using the mean shift procedure (Fukunaga and Hostetler 1975). When using the Epanechnikov kernel, the mean shift iteration step to move the kernel center position from to its new position is
| 12 |
The weight wi for each pixel depends on a comparison of the histogram bins of qu and that intensity falls into:
| 13 |
where is the Kronecker delta function and a binning function that maps intensity to a histogram bin number. After each mean shift iteration, convergence is checked (number of iterations >20 or mean shift vector length <0.05 pixels) and if met, is set to , otherwise is set to and the mean shift procedure is repeated.
This method incorporates modifications proposed by Ning et al (2012) to make it adaptive to scale and orientation. First the target’s scale is estimated based on the sum of the weights of all pixels in the search region and adjusted by the current Bhattacharyya coefficient. The estimated area is then used to adjust an ellipsoidal target descriptor to match the width, height and orientation of the ellipsoidal target descriptor, which is manually initialized in I(0). Between two adjacent frames the kernel size is enlarged by parameter , which is set to 1 and 3 for all ETH and MED sequences respectively.
Assuming motion periodicity, two failure recovery strategies are employed. First, if the Bhattacharyya coefficient drops below 0.8, the found target position is discarded and tracking is repeated using the target search area defined in I(0). Second, if the target search area in the current frame is three times larger than the initial target’s size, the search area and its position is reset to the one in the first frame.
Run-time. The algorithm was implemented in MATLAB 2013b, and the experiments were conducted on a machine with an Intel i5-3320M processor at 2.6 GHz clock speed and 8 GB RAM. Tracking speed using this hardware set up was approximately 33 ms.
3.7. Tracking by decision fusion
Fusing the results from different methods or annotations has shown improvements for various applications (Sinha et al 2008) including classification (Kittler et al 1998), segmentation (Rohlfing et al 2003, Heckemann et al 2006), and tracking. Therefore we included an investigation of such a fusion approach. The tracking results of all six previously described methods were combined by computing for each frame t the median position of the tracked points Pj(t) from the automatic methods. Using the median helps to reduce the influence of outlier results. Furthermore we determined the performance when fusing only the two methods which had the lowest mean tracking errors on the training data.
4. Evaluation
We compared the performance of the methods described in section 3 on the test data, consisting of 66 point-landmarks (vessel centers) in 26 2D sequences and 28 point-landmarks (vessel bifurcations) in 13 3D sequences, which the observers were confident to be able to reliably annotate. We evaluated the performance of the segmentation method in section 3.5 (PhR) on 11 manually segmented tumors in 9 2D sequences. In the following we describe the evaluation scheme used to validate and quantify the tracking accuracy.
4.1. Point-landmark tracking error
We randomly selected at least 10% of the images from each sequence and manually annotated the corresponding position of the initial point Pj(0) in each of these selected images , denoted as . The number of annotated frames per sequence is listed in table 1. For the annotated frame and landmark j, we calculated the tracking error (TE) as the Euclidean distance between the estimated landmark position and its annotation (ground truth) :
| 14 |
We summarized the results by the mean (MTE), standard deviation (STD) and 95th percentile of the single distribution including all belonging to a particular subgroup. These subgroups were the individual landmarks j (MTEj), the image groups (MTEETH, MTEMED1 and MTEMED2 for 2D sequences, and MTEEMC, MTEICR and MTESMT for 3D sequences) and the landmark dimensionality (MTE2D and MTE3D).
We also included the motion magnitude of the landmarks, defined as:
| 15 |
We estimated the inter-observer variability of the results. For this two additional experts annotated a randomly selected subset of the images marked by the first observer amounting to a total of 3% of all images. We then defined as ground truth the mean position over the three annotations and computed the tracking error as described above.
Per tracking task, the median results from all employed methods were pair-wise tested for statistically significantly differences at the probability level p = 0.001 using the sign test. The sign test was used as it is a non-parametric test, which neither assumes a normal distribution nor a symmetric distribution.
4.2. Segmentation accuracy
A clinical expert segmented the visible boundaries of the tumors corresponding to the reference segmentation in each frame t of the 2D OX sequences. The results of the segmentation tracking were quantified by computing the Dice coefficient:
| 16 |
where and Sj denote the manually segmented and the estimated tumor region, respectively, in each frame t. % measures the overlap ratio of the two regions and is at best 100%. We summarized the results by the mean, standard deviation and 5th percentile of the Dice coefficient distribution for each segmented tumor in the entire sequence. For comparison, we calculated the initial Dice coefficient, before tracking, as follows:
| 17 |
5. Results
5.1. 2D landmarks
The results for the 2D point-landmark tracking are summarized in table 3. The MTE2D ranges from 1.4 mm to 2.1 mm for the automatic methods, with best results achieved by MEVIS. Fusing the results of all tracking methods improved accuracy by 15–41% in comparison to the individual results, achieving a MTE2D of 1.2 mm.
Table 3.
Results of 2D point-landmark tracking.
| Method | Tracking error | Mean error range of sequences | ||||
|---|---|---|---|---|---|---|
| MTE2D | STD | 95thTE | MTEj∈ ETH | MTEj∈ MED1 | MTEj∈ MED2 | |
| Fusion | 1.23 | 1.52 | 3.26 | [0.36, 2.04] | [0.81, 7.83] | [0.89, 2.97] |
| MEVIS | 1.44 | 2.04 | 3.86 | [0.31, 2.61] | [0.90, 8.75] | [1.02, 3.47] |
| MEVIS + MED | 1.53 | 2.45 | 3.95 | [0.32, 3.02] | [0.94, 5.12] | [1.20, 12.71] |
| TUM | 1.64 | 1.84 | 4.68 | [0.43, 7.48] | [0.56, 4.24] | [0.95, 2.83] |
| KM | 1.83 | 3.16 | 4.82 | [0.37, 1.73] | [0.93, 13.22] | [1.62, 3.63] |
| PhR | 2.00 | 2.87 | 5.59 | [0.51, 3.47] | [0.79, 12.72] | [0.88, 3.54] |
| MEVIS + FOKUS | 2.09 | 2.87 | 6.22 | [0.52, 10.05] | [0.59, 11.27] | [0.88, 3.36] |
| Motion M | 6.64 | 4.81 | 15.53 | [2.90, 13.56] | [3.78, 12.48] | [4.33, 12.31] |
Note: The results are in millimeters and ranked (top to bottom) according to increasing MTE2D.
Three of the proposed methods had for some sequences a higher mean error (MTEj) than the landmark motion Mj.
To assess the robustness of all methods, we quantified the percentage of failures, i.e. the percentage of annotated frames per landmark for which mm or mm. These results, summarized in figure 6, show that the percentage of failures ranges between 6.3% and 15.8% (1.6% and 7.2%) for mm ( mm), with the fusion method being the best one.
Figure 6.
Percentage of failure cases: ratio of annotated 2D landmarks whose mm (orange) or mm (red) shown for all methods. Results are shown (left to right) according to decreasing MTE2D (see table 3). TE is evaluated with respect to one observer.
We illustrate the difference in tracking performance between all methods for the point which had the highest variance of MTE (24.85 mm) across all methods, namely P1 from MED-07 (sequence details are listed in table 1). From figure 4 it can be observed that high tracking errors occur for some of the proposed methods when the landmark position drifted after a deep inhalation (at frame tc) of the subject under investigation. Comparing the associated ROIs, it seems that methods KM and PhR maintain a similar distance to the diaphragm, rather than stay with the vessel.
Figure 4.
Illustration of tracking performance for landmark P1 from sequence MED-07. Tracking errors MTE1∈MED−07 were 13.22 (KM), 3.84 (MEVIS), 7.46 (MEVIS + FOKUS), 2.88 (MEVIS + MED), 12.72 (PhR) and 1.93 mm (TUM). The mean motion for the landmark was 11.23 mm. Frames at ta, tc and te correspond to end-inhalations (with a deep inhale happening at tc), while tb, td and t f correspond to end-exhalations. In ROI(ta) the manual annotation is shown as a yellow circle.
Lower errors were obtained with respect to the mean annotation of the 3 observers with MTE2D reduced by 0.06 to 0.23 mm for all methods, while the mean motion M was similar, see table 5. Note that now the order changed between MEVIS + MED and TUM and between PhR and MEVIS + FOKUS. The MTE2D of the 3 observers was more than 50% lower than any individual method. An overview of these results is given in the box-plots in figure 5.
Table 5.
Summary of the results of point-landmark tracking w.r.t. mean manual annotation of three observers.
| Method | 2D tracking error | 3D tracking error | |||||||
|---|---|---|---|---|---|---|---|---|---|
| MTE2D | STD | MedianE | 95thTE | MTE | MTE3D | STD | MedianE | 95thTE | |
| Fusion | 1.08 | 1.42 | 0.75 | 2.85 | 1.32 | 2.43 | 2.76 | 1.49 | 7.61 |
| KM | 1.75 | 3.05 | 1.03 | 4.76 | ✗ | ✗ | ✗ | ✗ | ✗ |
| MEVIS | 1.33 | 1.94 | 0.88 | 3.56 | ✗ | ✗ | ✗ | ✗ | ✗ |
| MEVIS + FOKUS | 1.90 | 2.75 | 1.11 | 6.02 | 2.33 | 4.79 | 4.72 | 2.99 | 13.48 |
| MEVIS + MED | 1.45 | 2.48 | 0.95 | 3.49 | 1.78 | 2.76 | 4.10 | 1.52 | 8.80 |
| PhR | 1.94 | 2.93 | 1.12 | 5.53 | 2.38 | 2.83 | 2.97 | 1.46 | 9.67 |
| TUM | 1.41 | 1.89 | 0.78 | 4.70 | ✗ | ✗ | ✗ | ✗ | ✗ |
| Motion M | 6.69 | 4.78 | 5.65 | 15.52 | 8.19 | 5.83 | 4.21 | 4.52 | 14.80 |
| Obs1 | 0.58 | 0.42 | 0.48 | 1.40 | 0.71 | 1.21 | 1.21 | 0.79 | 4.43 |
| Obs2 | 0.48 | 0.34 | 0.41 | 1.06 | 0.59 | 1.73 | 2.66 | 0.78 | 8.74 |
| Obs3 | 0.50 | 0.41 | 0.40 | 1.23 | 0.61 | 1.81 | 2.61 | 0.95 | 5.61 |
Note: The results are in millimeters and ranked according to alphabetical order of the methods.
Figure 5.
Box-plot summarizing the 2D tracking error (in mm) w.r.t. mean manual annotation of three observers (Obs). Results are ranked (left to right) according to decreasing MTE2D (in green). On each box, the central red line is the median and the edges of the box are given by q1 = 25th and q3 = 75th percentiles of the error. Outliers are drawn as red crosses if larger than q3 + w(q3 − q1), where w = 1.5 corresponds to approximately ±2.7 STD of the data.
There was low correlation between the motion magnitude of the landmarks and the tracking errors. In details, the sample Pearson correlation coefficients () between landmark motion and tracking errors for the individual methods ranged from 0.11 to 0.37, while for the observers there was no correlation ().
We also studied if tracking errors were influenced by the change in image quality due to the range of center frequencies used during the US acquisitions (see table 1). Only low correlation () was found between MTEj∈2D of the individual methods and center frequency of landmark j.
Tracking landmarks close to the acquisition border can be difficult. We analyzed the correlation between the landmark distance to the acquisition border and the tracking error for the landmark which comes closest to the acquisition border (P2 from MED-08), and found a high correlation () for all methods except TUM ().
The median results with respect to the mean annotation of the 3 observers from all methods were significantly different to each other, except for and . The median tracking errors of the observers were similar, apart from Obs1’s being significantly higher.
5.2. 3D landmarks
Results for all methods on the 3D sequences are shown in table 4. On average, the most accurate results were achieved by PhR, with MTE3D of 2.55 mm. The fusion method slightly improved the average results by 3%. The percentages of failures are shown in figure 8 for all methods. These ranged between 18.7% and 51.3% (6.8% and 29.6%) for mm (5 mm).
Table 4.
Results of 3D point-landmark tracking.
| Method | Tracking error | Mean error range of sequences | ||||
|---|---|---|---|---|---|---|
| MTE3D | STD | 95thTE | MTEj∈ EMC | MTEj∈ ICR | MTEj∈ SMT | |
| Fusion | 2.48 | 2.46 | 6.91 | [1.19, 9.84] | [2.53, 2.59] | [0.94, 8.12] |
| PhR | 2.55 | 2.46 | 7.98 | [1.03, 9.63] | [2.54, 3.89] | [0.99, 11.57] |
| MEVIS + MED | 2.71 | 3.01 | 7.58 | [1.36, 10.40] | [1.59, 2.76] | [1.00, 6.59] |
| MEVIS + FOKUS | 4.63 | 4.03 | 12.44 | [2.41, 11.26] | [4.28, 5.88] | [1.23, 10.10] |
| Motion M | 6.19 | 4.64 | 14.83 | [3.59, 13.16] | [4.47, 5.72] | [2.46, 12.89] |
Note: The results are in millimeters and ranked according to increasing MTE3D.
Figure 8.
Percentage of failure cases: ratio of annotated 3D landmarks whose mm (orange) or mm (red) shown for all methods. Results are shown (left to right) according to decreasing MTE3D (see table 4). TE is evaluated with respect to one observer.
The tracking errors with respect to the mean manual annotation of 3 observers are increased by 0.05–0.28 mm, while the mean motion is reduced by 0.36 mm. The error obtained by fusing all results is slightly reduced, see table 5. An overview of these results is given in the box-plots in figure 7.
Figure 7.
Box-plot summarizing the 3D tracking error (in mm) w.r.t. mean manual annotation of three observers (Obs). Results are ranked (left to right) according to decreasing MTE3D (in green). On each box, the central red line is the median and the edges of the box are given by q1 = 25th and q3 = 75th percentiles of the error. Outliers are drawn as red crosses if larger than q3 + w(q3 − q1), where w = 1.5 corresponds to approximately ±2.7 STD of the data.
Figure 9 shows the result with the highest MTEj for MEVIS + MED and PhR. Tracking methods and observers disagreed mainly along the vessel (x-coordinate), showing the challenge of tracking landmarks within elongated structures without a stable bifurcation.
Figure 9.
Illustration of tracking performance for landmark P1 from sequence EMC-03. Tracking errors MTE1∈EMC−03 with respect to the mean of 3 observers were 7.77 mm (MEVIS + FOKUS), 5.63 mm (MEVIS + MED) and 9.93 mm (PhR). The mean motion for the landmark was 11.71 mm. Inter-observer errors were 4.06 mm (Obs1), 11.59 mm (Obs2) and 9.76 mm (Obs3). The tracking results and annotations are shown at time for the same ROI() with planes cut at the corresponding from each method.
All MTE3D were larger than the estimated equivalent 3D error from MTE2D16 (e.g. +70% Obs1, 84% Fusion, 55% MEVIS + MED), while the estimated 3D motion was on average lower (−29%). Yet this clear relationship was not sustained for error measures calculated in pixels (see table 6). In particular Fusion resulted in very similar results.
Table 6.
Summary of the results of point-landmark tracking w.r.t. mean manual annotation of three observers.
| Method | 2D tracking error | 3D tracking error | |||||||
|---|---|---|---|---|---|---|---|---|---|
| MTE2D | STD | MedianE | 95thTE | MTE | MTE3D | STD | MedianE | 95thTE | |
| Fusion | 2.68 | 3.52 | 1.81 | 7.02 | 3.28 | 3.35 | 3.77 | 2.14 | 10.02 |
| KM | 4.33 | 7.54 | 2.41 | 11.86 | ✗ | ✗ | ✗ | ✗ | ✗ |
| MEVIS | 3.29 | 4.73 | 2.10 | 8.89 | ✗ | ✗ | ✗ | ✗ | ✗ |
| MEVIS + FOKUS | 4.72 | 7.00 | 2.75 | 15.85 | 5.78 | 6.34 | 6.13 | 3.94 | 16.78 |
| MEVIS + MED | 3.68 | 6.89 | 2.32 | 8.86 | 4.51 | 3.66 | 5.12 | 2.11 | 11.36 |
| PhR | 4.71 | 7.12 | 2.71 | 13.56 | 5.77 | 3.80 | 3.88 | 2.09 | 13.22 |
| TUM | 3.37 | 4.39 | 1.98 | 11.48 | ✗ | ✗ | ✗ | ✗ | ✗ |
| Motion M | 16.91 | 12.49 | 14.13 | 40.39 | 20.71 | 8.00 | 5.82 | 6.03 | 21.00 |
| Obs1 | 1.45 | 1.09 | 1.12 | 3.66 | 1.78 | 1.71 | 1.56 | 1.14 | 4.56 |
| Obs2 | 1.19 | 0.87 | 0.99 | 2.70 | 1.46 | 1.60 | 1.55 | 1.11 | 5.11 |
| Obs3 | 1.25 | 1.05 | 0.98 | 3.11 | 1.53 | 1.82 | 1.56 | 1.36 | 4.69 |
Note: The results are in pixels/voxels and ranked according to alphabetical order of the methods.
The median error (see table 5) of MEVIS + FOKUS was statistically different (p < 0.001) with respect to the one of Fusion and PhR. No other significant difference between the methods existed. The median tracking errors of the 3 observers were statistically significantly lower than the ones from the 4 tracking methods, but not significantly different amongst each other.
We investigated if the landmark error is correlated with the motion and found little evidence for such a correlation, as the sample Pearson correlation coefficients was in the range of 0.25–0.45 for all methods and observers, apart from MEVIS + FOKUS ().
For 3D, a low to moderate correlation () was found between tracking errors (MTEj∈3D) and center frequency of the US acquisition protocol (see table 1) for each landmark j.
We analyzed the correlation between the mean tracking error and the distance to the acquisition border for a 3D vessel bifurcation landmark, which was moving close to this border and had good local image contrast throughout the sequence. We found that the distance was moderately correlated with the TE of landmark P1 from SMT-02 for all methods ().
5.3. 2D segmentations
The results of the tumor segmentation task, performed by PhR are summarized in table 7. The segmentation accuracy, expressed in mean Dice coefficient, ranges between 76.7% and 92.3%. The mean accuracy is 8% higher than the initial overlap ratio, but lower in 3 sequences.
Table 7.
Results of 2D tumor segmentation tracking (PhR) w.r.t. manual annotation of a clinical expert.
| Segmentation | Tracking overlap [%] | Initial Dice [%] | ||||
|---|---|---|---|---|---|---|
| Mean | STD | 5th TO | Mean | STD | 5th TO | |
| OX-1, S1 | 86.6 | 5.4 | 78.4 | 49.9 | 25.3 | 20.6 |
| OX-2, S1 | 85.5 | 4.8 | 76.6 | 73.7 | 10.7 | 57.1 |
| OX-4, S1 | 92.3 | 1.9 | 90.0 | 74.9 | 17.6 | 46.3 |
| OX-5, S1 | 79.8 | 6.6 | 68.4 | 87.3 | 5.5 | 77.3 |
| OX-6, S1 | 76.7 | 9.2 | 58.6 | 72.6 | 12.7 | 51.9 |
| OX-7, S1 | 89.6 | 4.3 | 78.8 | 80.5 | 6.2 | 71.5 |
| OX-7, S2 | 77.2 | 4.8 | 70.0 | 58.3 | 15.5 | 33.2 |
| OX-8, S1 | 88.7 | 2.7 | 84.9 | 92.1 | 3.5 | 84.9 |
| OX-9, S1 | 92.2 | 2.2 | 88.8 | 91.8 | 5.0 | 81.9 |
| OX-9, S2 | 77.6 | 5.7 | 66.7 | 80.2 | 8.2 | 65.7 |
| OX-10, S1 | 81.8 | 3.4 | 76.7 | 75.5 | 6.8 | 63.2 |
| OX | 84.1 | 7.5 | 71.1 | 77.8 | 16.7 | 41.5 |
Note: The results are in % of the Dice coefficient.
Figure 10 shows the results of the segmentation on the sequence with the highest variance of the Dice coefficient, i.e. S1 from OX-6. The overlap gradually decreases, likely due to lower image contrast compared to the other sequences.
Figure 10.
Illustration of tracking performance for S1 from OX-6. The Dice coefficient ranged from 92.5 % (at ta) to 53.6% (at tb). ROI(ta) and ROI(tb) show the overlap of the manual (in yellow) and PhR (in light blue) segmentations.
6. Discussion
When considering the individual methods (described in sections 3.1–3.6), determining the local translations of the vessels by a Bayesian approach (MEVIS + MED) was computationally the most efficient algorithm and was runner-up (by 6%) in both landmark tasks. The Sparse Demons method on patches (PhR) performed best for the 3D landmarks, but had a relative poor performance for the 2D landmarks. For the case illustrated in figure 5, the method tracks the diaphragm as this is the main bright feature in the reference patch, resulting in high errors when the distance to the vessel changes. The variational registration approach within relative large regions (MEVIS) provided on average the most accurate results for the 2D landmarks, but was unfortunately not tested for the 3D landmarks. Registration to the initial frame and if needed to the previous frame, allowing smooth deformations, as well as including NGF in the image similarity likely contributed to this good performance.
Fusion of the tracking results from all the independent methods by taking their median value provided on average the highest landmark tracking accuracy, with improvements by at least 15% (3%) for 2D (3D) landmarks. This reduced mean motion by 82% (60%) resulting in a mean accuracy of 1.2 mm (2.5 mm). Yet it requires the run-time of the slowest method and n times the computing power, where n is the number of fused methods (in this paper n = 6 for 2D and n = 3 for 3D). To save resources, fusing only the 2 methods, which performed on average best on the training set, led to worse results for 2D (MTE2D=1.6 mm for median of MEVIS + MED and MEVIS + FOKUS) as the training performance of MEVIS + FOKUS was not representative. More training data will hopefully enable selecting instead one of the better methods (e.g. MEVIS), which would have led to a slightly better fusion result (1.4 mm). Alternatively fusing more methods might circumvent this problem (e.g. incorporating MEVIS as third best from training data would result in an MTE of 1.3 mm). For 3D, similar results were achieved, with the median of MEVIS + MED and PhR lead to MTE3D=2.5 mm.
The tracking performance was generally not dependent on the motion magnitude. Comparing the errors in millimeters, tracking the 3D landmarks appears to be a harder task than tracking the 2D landmarks. This can mostly be explained by the lower image resolution, since 2D and 3D results from Fusion are very similar when considering the pixel resolution and assuming equal errors in each dimension (2D: 2.7 pixels, estimated 3D: 3.3 pixels, 3D: 3.4 pixels). Hence it seems very likely that advances in 3D image resolution will improve tracking results. Other factors include some worse image quality (e.g. EMC), a more cumbersome annotation task and greater difficulties of visually inspecting the 3D images for parameter optimization.
Methods that adjusted their parameters per sequence for the MICCAI 2014 challenge were TUM, MEVIS, MEVIS + MED and MEVIS + FOKUS (only 2D case). The mean performance of these ‘tuned’ methods for all 2D test data was worse (TUM 1.84 mm, MEVIS 1.51 mm), similar (MEVIS + MED 1.52 mm), or better (MEVIS + FOKUS 1.91 mm) than with fixed parameters (table 3). For the 3D case, resubmission was needed only from MEVIS + MED, for which the original submission resulted in worse accuracy (MTE3D=2.80 mm) than fixed parameters (table 4).
We compared the methods to previously published results (Cifor et al 2013, De Luca et al 2013, Vijayan et al 2014) by re-computing the tracking error for the same subset (excluding any training data), the same error measure and annotations. For 2D landmark tracking, two of the individual methods performed on average better than previously published results, namely MEVIS + MED achieved an MTE (95th TE) of 0.6 (1.4) mm and MEVIS of 0.7 (1.6) mm, versus 0.8 (1.7) mm from De Luca et al (2013). This comparison was based on ETH-01 to ETH-09 without ETH-05. For 3D landmark tracking, comparison was done on the SMT subset. The method proposed in Vijayan et al (2014) achieved an MTE (95th TE) of 3.6 (14.9) mm. PhR and MEVIS + MED obtained on average better results, with an MTE (95th TE) of 2.4 (6.8) mm and 2.6 (7.3) mm respectively. The single method attempting the tracking of tumor regions performed similar to the previous baseline method (Cifor et al 2013). The comparison was possible only for 3 sequences, where PhR was on average 5% (OX-01) and 1% (OX-04) worse, but 4% better (OX-02) than the method from Cifor et al (2013).
On average, the tracking performance was worse than the observer annotation accuracy, which hints at potentials for further method improvements. Error reduction beyond the observer accuracy will require improvements in image resolution and quality. The clinically acceptable tracking error depends on the application. The American Association of Physicists in Medicine recommends for external-beam radiation therapy that respiratory management techniques are warranted when the target motion is greater than 5 mm (Keall et al 2006). This implies that techniques for predicting the target motion should have a maximum error of at most 5 mm. Tracking will need to be more accurate as errors are also introduced by temporal prediction to compensate system latencies and spatial prediction if the target cannot be directly tracked. Overall this was clearly not yet achieved by any automatic tracking methods (see figures 4, 6, 7 and 8) and also observers had some difficulties. Only for one method (MEVIS + MED) and the ICR sequences the tracking errors stayed always within the 5 mm limit (maximum TE: 4.7 mm), likely due to good image quality, smaller field of view, no disappearing structures, and high temporal and spatial resolution.
The high 95th percentile errors and number of tracking failures indicate that the methods are not very robust. Visual inspection of failure cases hint at problems due to accumulation of errors, disappearing structures at the image acquisition border, out-of-plane motion, substantial changes in image appearance or insufficient motion capture ranges. These might be the reasons of the high variance in the individual tracking performances across the sequences.
Tracking run-times per frame range from 1.8 ms to 84 ms for 2D and from 51 ms to about 100 ms for 3D. Processing times generally varied depending on the number of landmarks and the image appearance. Real-time performance was achieved by 4 methods for 2D and 3 for 3D landmarks (MEVIS + MED, MEVIS + FOKUS, PhR and TUM), assuming average acquisition times of 50 ms (20 Hz) and 125 ms (8 Hz). The remaining methods are not far off, indicating that further code optimization and hardware improvements are likely to provide real-time performance for the current dataset. Furthermore, off-line applications, such as dose accumulation during radiation therapy, do not require real-time performance. Hence, the fusion of several tracking results can be a promising approach, despite having the cumulative computational burden of all fused methods (which is however easily parallelizable), as it achieved the highest accuracy.
For the purpose of evaluating different tracking approaches, specific anatomical landmarks in the first frame were provided in CLUST. These landmarks were manually selected after inspecting the US images, to ensure that they would not disappear during free-breathing. In clinics, real-time applications should require as little manual interventions as possible. Therefore, tracking algorithms might benefit from incorporating the automatic detection of stable features to track throughout the treatment and update tracking regions when necessary. Feature detection might be based on analyzing the tracking results from an initial set of images covering a few breathing cycles. Testing this functionality is outside the CLUST challenge, as the aim is to compare tracking performances for the same landmarks.
7. Conclusion
This paper describes the results of the MICCAI 2014 challenge on liver US tracking (CLUST14), which enabled for the first time the quantitative, direct comparison of tracking methods for this application.
The challenge data included a large number of realistic sequences, which varied in length, spatial and temporal resolution, acquisition settings and US scanner.
Quantitative evaluation of all results showed a mean tracking error from 1.4 mm to 2.1 mm for 2D points, and from 2.6 mm to 4.6 mm for 3D points. Considering the median tracking results of all methods improved the mean error to 1.2 mm (2D) and 2.5 mm (3D). The segmentation task, fulfilled only by one participant, resulted in a mean Dice coefficient of 84.1%. All best approaches are comparable or better than the state-of-the-art.
Applicability for therapy guidance still requires general improvement of the 3D landmark tracking accuracy as well as reduction of tracking failures ( mm). Advances in image resolution and quality will support this task. Furthermore the diaphragm is a prominent feature in liver US images which should also be tracked.
For some sequences, the variability of the observers was particularly high, due to the difficulty in manually annotating 3D volumes. Having more observers might lead to a more reliable reference for computing the tracking results.
Most of the methods achieved, or are close to, real-time, while running on standard machines.
In conclusion, this CLUST14 challenge provided a good basis for a first comparison of US tracking methods for the liver. Its accompanying workshop facilitated lively discussions of the involved researchers. The research community can benefit from this benchmark and the CLUST challenge remains open for future participants to evaluate their method and be included in the online leader board. Future work includes the increase of the number of sequences and tracking of other available structures in the liver, e.g. the diaphragm.
Acknowledgments
ETH and MEVIS + MED acknowledge funding from the European Union’s 7th Framework Program (FP7/2007-2013) under grant agreement n° 611889 (TRANS-FUSIMO). The work of MEVIS and MEVIS + MED was supported by Fraunhofer Internal Programs under Grant n° MAVO 823 287. M A Lediju Bell acknowledges funding from the UNCF/Merck Postdoctoral Science Research Fellowship. A Cifor thanks Wellcome Trust/EPSRC Centre of Excellence in Medical Engineering—Personalized Healthcare WT088877/Z/09 and Oxford EPSRC Impact Acceleration Account for financial support.
We would like to express our sincere appreciation to the authors whose contributions to this paper have required considerable commitment of time and effort. We also thank (in alphabetical order of surnames) J Banerjee from the Biomedical Imaging Group, Erasmus MC, Rotterdam, The Netherlands; F Lindseth and S Vijayan from SINTEF Medical Technology, Trondheim, Norway; J A Schnabel from the Institute of Biomedical Engineering, University of Oxford, UK; and their colleagues for providing data and annotations. Without their help CLUST14 would not have been possible.
Footnotes
The challenge proceedings are available at http://clust14.ethz.ch/clust2014.html
The mean 3D error was estimated from MTE2D by , which assumes equal error components.
References
- Arulampalam M S, Maskell S, Gordon N, Clapp T. A tutorial on particle filters for online nonlinear/non-Gaussian Bayesian tracking. IEEE Trans. Signal Process. 2002;50:174–88. doi: 10.1109/78.978374. [DOI] [Google Scholar]
- Banerjee J, Klink C, Peters E D, Niessen W J, Moelker A, van Walsum T. Fast and robust 3D ultrasound registration—block and game theoretic matching. Med. Image Anal. 2015;20:173–83. doi: 10.1016/j.media.2014.11.004. [DOI] [PubMed] [Google Scholar]
- Banerjee J, Klink C, Peters E D, Niessen W J, Moelker A, van Walsum T. 4D liver ultrasound registration. Lecture Notes in Computer Science; Biomedical Image Registration; Berlin: Springer; 2014. pp. pp 194–202. [Google Scholar]
- Cifor A, Risser L, Chung D, Anderson E, Schnabel J. Hybrid feature-based Log-Demons registration for tumour tracking in 2D liver ultrasound images. 2012 IEEE 9th Int. Symp. on Biomedical Imaging; 2012. pp. pp 724–7. [Google Scholar]
- Cifor A, Risser L, Chung D, Anderson E, Schnabel J. Hybrid feature-based diffeomorphic registration for tumor tracking in 2D liver ultrasound images. IEEE Trans. Med. Imaging. 2013;32:1647–56. doi: 10.1109/TMI.2013.2262055. [DOI] [PubMed] [Google Scholar]
- Comaniciu D, Ramesh V, Meer P. Kernel-based object tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2003;25:564–77. doi: 10.1109/TPAMI.2003.1195991. [DOI] [Google Scholar]
- De Luca V, Tanner C, Szekely G. Speeding-up image registration for repetitive motion scenarios. 2012 IEEE 9th Int. Symp. on Biomedical Imaging; 2012. pp. pp 1355–8. [Google Scholar]
- De Luca V, Tschannen M, Szekely G, Tanner C. Medical Image Computing and Computer-Assisted Intervention. New York: Springer; 2013. A learning-based approach for fast and robust vessel tracking in long ultrasound sequences; pp. pp 518–25. (Lecture Notes in Computer Science vol 8149). [DOI] [PubMed] [Google Scholar]
- Epanechnikov V A. Non-parametric estimation of a multivariate probability density. Theory Probab. Appl. 1969;14:153–8. doi: 10.1137/1114019. [DOI] [Google Scholar]
- Farnebäck G. Image Analysis. Berlin: Springer; 2003. Two-frame motion estimation based on polynomial expansion; pp. pp 363–70. (Lecture Notes in Computer Science vol 2749). [Google Scholar]
- Fischer B, Modersitzki J. Curvature based image registration. J. Math. Imaging Vision. 2003;18:81–5. doi: 10.1023/A:1021897212261. [DOI] [Google Scholar]
- Fukunaga K, Hostetler L. The estimation of the gradient of a density function, with applications in pattern recognition. IEEE Trans. Inf. Theory. 1975;21:32–40. doi: 10.1109/TIT.1975.1055330. [DOI] [Google Scholar]
- Haber E, Modersitzki J. Medical Image Computing and Computer-Assisted Intervention (Lecture Notes in Computer Science) vol 4191. Berlin: Springer; 2006. Intensity gradient based registration and fusion of multi-modal images; pp. pp 726–33. [DOI] [PubMed] [Google Scholar]
- Harris E, Miller N, Bamber J C, Symonds-Tayler J, Evans P. Speckle tracking in a phantom and feature-based tracking in liver in the presence of respiratory motion using 4D ultrasound. Phys. Med. Biol. 2010;55:3363–80. doi: 10.1088/0031-9155/55/12/007. [DOI] [PubMed] [Google Scholar]
- Heckemann R A, Hajnal J V, Aljabar P, Rueckert D, Hammers A. Automatic anatomical brain MRI segmentation combining label propagation and decision fusion. NeuroImage. 2006;33:115–126. doi: 10.1016/j.neuroimage.2006.05.061. [DOI] [PubMed] [Google Scholar]
- Isard M, Blake A. Computer Vison. Berlin: Springer; 1996. Contour tracking by stochastic propagation of conditional density; pp. pp 343–56. (Lecture Notes in Computer Science vol 1064). [Google Scholar]
- Keall P, et al. The management of respiratory motion in radiation oncology report of AAPM task group 76. Med. Phys. 2006;33:3874–900. doi: 10.1118/1.2349696. [DOI] [PubMed] [Google Scholar]
- Kittler J, Hatef M, Duin R, Matas J. On combining classifiers. IEEE Trans. Pattern Anal. Mach. Intell. 1998;20:226–39. doi: 10.1109/34.667881. [DOI] [Google Scholar]
- König L, Kipshagen T, Rühaak J. A non-linear image registration scheme for real-time liver ultrasound tracking using normalized gradient fields. Proc. MICCAI Challenge on Liver Ultrasound Tracking; Boston, USA. 2014. [Google Scholar]
- König L, Rühaak J. A fast and accurate parallel algorithm for non-linear image registration using normalized gradient fields. 2014 IEEE 11th Int. Symp. on Biomedical Imaging; 2014. pp. pp 580–3. [Google Scholar]
- Lediju Bell M A, Byram B C, Harris E J, Evans P M, Bamber J C. In vivo liver tracking with a high volume rate 4D ultrasound scanner and a 2D matrix array probe. Phys. Med. Biol. 2012;57:1359–74. doi: 10.1088/0031-9155/57/5/1359. [DOI] [PubMed] [Google Scholar]
- Lediju M A, Byram B C, Harris E J, Evans P M, Bamber J C. 3D liver tracking using a matrix array: implications for ultrasonic guidance of IMRT. 2010 IEEE Ultrasonics Symp.; 2010. pp. pp 1628–31. [Google Scholar]
- Lübke D, Grozea C. High performance online motion tracking in abdominal ultrasound imaging. Proc. MICCAI Challenge on Liver Ultrasound Tracking; Boston, USA. 2014. [Google Scholar]
- Mansi T, Pennec X, Sermesant M, Delingette H, Ayache N. iLogDemons: a demons-based registration algorithm for tracking incompressible elastic biological tissues. Int. J. Comput. Vision. 2011;92:92–111. doi: 10.1007/s11263-010-0405-z. [DOI] [Google Scholar]
- Matthews I, Ishikawa T, Baker S. The template update problem. IEEE Trans. Pattern Anal. Mach. Intell. 2004;26:810–5. doi: 10.1109/TPAMI.2004.16. [DOI] [PubMed] [Google Scholar]
- McClelland J R, Hawkes D J, Schaeffter T, King A P. Respiratory motion models: a review. Med. Image Anal. 2013;17:19–42. doi: 10.1016/j.media.2012.09.005. [DOI] [PubMed] [Google Scholar]
- Modersitzki J. FAIR: Flexible Algorithms for Image Registration. Philadelphia, PA: Society for Industrial and Applied Mathematics; 2009. [Google Scholar]
- Mory B, Somphone O, Prevost R, Ardon R. Medical Image Computing and Computer-Assisted Intervention (Lecture Notes in Computer Science) vol 7510. Berlin: Springer; 2012. Real-time 3D image segmentation by user-constrained template deformation; pp. pp 561–8. [DOI] [PubMed] [Google Scholar]
- Ning J, Zhang L, Zhang D, Wu C. Scale and orientation adaptive mean shift tracking. Comput. Vision. 2012;6:52–61. doi: 10.1049/iet-cvi.2010.0112. [DOI] [Google Scholar]
- Preiswerk F, De Luca V, Arnold P, Celicanin Z, Petrusca L, Tanner C, Bieri O, Salomir R, Cattin P C. Model-guided respiratory organ motion prediction of the liver from 2D ultrasound. Med. Image Anal. 2014;18:740–51. doi: 10.1016/j.media.2014.03.006. [DOI] [PubMed] [Google Scholar]
- Rohlfing T, Russakoff D B, Maurer C R. Information Processing in Medical Imaging (Lecture Notes in Computer Science) vol 2732. Berlin: Springer; 2003. Expectation maximization strategies for multi-atlas multi-label segmentation; pp. pp 210–21. [DOI] [PubMed] [Google Scholar]
- Shirato H, Shimizu S, Kitamura K, Onimaru R. Organ motion in image-guided radiotherapy: lessons from real-time tumor-tracking radiotherapy. Int. J. Clin. Oncol. 2007;12:8–16. doi: 10.1007/s10147-006-0633-y. [DOI] [PubMed] [Google Scholar]
- Sinha A, Chen H, Danu D, Kirubarajan T, Farooq M. Estimation and decision fusion: a survey. Neurocomputing. 2008;71:2650–6. doi: 10.1016/j.neucom.2007.06.016. [DOI] [Google Scholar]
- Somphone O, et al. Fast myocardial motion and strain estimation in 3D cardiac ultrasound with Sparse Demons. 2013 IEEE 10th Int. Symp. on Biomedical Imaging; 2013. pp. pp 1182–5. [Google Scholar]
- Tanner C, Boye D, Samei G, Szekely G. Review on 4D models for organ motion compensation. Crit. Rev. Biomed. Eng. 2012;40:135–54. doi: 10.1615/CritRevBiomedEng.v40.i2.40. [DOI] [PubMed] [Google Scholar]
- Thirion J P. Image matching as a diffusion process: an analogy with Maxwell’s demons. Med. Image Anal. 1998;2:243–60. doi: 10.1016/S1361-8415(98)80022-4. [DOI] [PubMed] [Google Scholar]
- Vijayan S, Klein S, Hofstad E F, Lindseth F, Ystgaard B, Langø T. Motion tracking in the liver: validation of a method based on 4D ultrasound using a nonrigid registration technique. Med. Phys. 2014;41:082903. doi: 10.1118/1.4890091. [DOI] [PubMed] [Google Scholar]
- Vijayan S, Klein S, Hofstad E, Lindseth F, Ystgaard B, Langø T. Validation of a non-rigid registration method for motion compensation in 4D ultrasound of the liver. 2013 IEEE 10th Int. Symp. on Biomedical Imaging; 2013. pp. pp 792–5. [Google Scholar]
- Zhang X, Günther M, Bongers A. Medical Imaging and Augmented Reality (Lecture Notes in Computer Science) vol 6326. Berlin: Springer; 2010. Real-time organ tracking in ultrasound imaging using active contours and conditional density propagation; pp. pp 286–94. [Google Scholar]









