Skip to main content
Journal of Digital Imaging logoLink to Journal of Digital Imaging
. 2010 May 1;24(3):485–493. doi: 10.1007/s10278-010-9300-y

Retrospective Evaluation of PET-MRI Registration Algorithms

Zuyao Y Shan 1,, Sara J Mateja 2, Wilburn E Reddick 1, John O Glass 1, Barry L Shulkin 3
PMCID: PMC3092046  PMID: 20437075

Abstract

The purpose of this study is to evaluate the accuracy of registration positron emission tomography (PET) head images to the MRI-based brain atlas. The [18F]fluoro-2-deoxyglucose PET images were normalized to the MRI-based brain atlas using nine registration algorithms including objective functions of ratio image uniformity (RIU), normalized mutual information (NMI), and normalized cross correlation (CC) and transformation models of rigid-body, linear, affine, and nonlinear transformations. The accuracy of normalization was evaluated by visual inspection and quantified by the gray matter (GM) concordance between normalized PET images and the brain atlas. The linear and affine registration based on the RIU provided the best GM concordance (average similarity index of 0.71 for both). We also observed that the GM concordances of linear and affine registration were higher than those of the rigid and nonlinear registration among the methods evaluated.

Key words: Normalization, PET, MR, brain, tissue concordance

Background

Positron emission tomography (PET) has matured in recent years as a functional imaging modality that provides insight into cell metabolism in health and disease.1,2 [18F]Fluoro-2-deoxyglucose (FDG) PET images deliver quantitative data on human brain metabolism.3 Unfortunately, FDG PET data contain little anatomic information. In contrast, magnetic resonance (MR) images provide details of anatomic structure. Therefore, combining PET and MR provides important information on the structure–function relationship and permits precise anatomically based definition of a region of interest.1 The fusion of PET and MR images can be achieved by using hardware or software, that is, a dedicated system acquiring PET and MR images simultaneously or a computation algorithm fusing PET and MR images that are collected separately. Although the hardware solution provides near-perfect image registration, very few such systems are available in current clinic settings due to the cost and technology limitations. In addition, with the recent development of the multimodality and population-based atlases,4 it will be a great benefit to normalize PET information with other information such as cytoarchitectonic probability maps and molecular architectonic maps. The population-averaged standard space based on MR images is generated to allow spatial normalization of multidimensional data; for example, ICBM452 is an averaged brain atlas based on MR images of 452 healthy subjects, and cytoarchitectonic probability maps are currently adding to it (www.loni.ucla.edu/ICBM/). Therefore, registration plays an important role in PET studies.

A computation registration algorithm is typically made up of four components: an objective function, a transformation model, an optimization process, and an interpolation method. The objective function defines the quantitative measure of the spatial agreement between two images. The objective functions establish images' correspondence extrinsically (fiducial localization errors) or intrinsically. The intrinsic correspondences include feature-based methods and intensity-based methods. The representative feature-based objective functions are “head-and-hat”,5 the iterative closest point algorithm,6 and, recently, wavelet-based attribute vector.7 The representative intensity-based objective functions include cross correlation (CC),8 square intensity difference,9 ratio image uniformity (RIU),10,11 mutual information,12 and Kullback–Leibler distance.13 The transformation model defines degrees of freedom (DOF) of moving images, including rigid-body registration with 6 DOF (3 translation and 3 rotation), linear registration with 9 DOF (3 scaling plus the 6 DOF of rigid-body registration), affine registration with 12 DOF (3 shearing plus the 9 DOF of linear registration), and nonlinear registration with more than 12 DOF (dependent on transformation models used). Many nonlinear transformation models have been developed. A detailed review of registration objective functions and transformation models can be found in a previous report.14 The optimization process is a computer search for the extremum of the objective function. The interpolation method is used to resample the source images to the desired image resolution. In this study, we evaluated image fusing algorithms using five well-known registration toolkits. These toolkits use different objective functions and transformation models, and several objective functions and transformation models are implemented in each toolkit. We believe that objective functions and transformation models are the key factors that affect the accuracy of registration. Therefore, we evaluated registration algorithms in terms of objective function/transformation model rather than registration toolkit.

Although almost every registration algorithm is developed with experimental validation results, each validation experiment unfortunately has a unique data set and design. Therefore, independent quantitative and qualitative assessments of registration fidelity are essential for multimodality image fusing.14 Previous validation studies can be divided into four groups based on the methodology of selection/generation of ground truth: fiducial markers, phantoms, simulated images, or features extracted from images. A brief introduction is provided below with examples of the evaluation of PET-MR registration studies; however, registration evaluation without the involvement of PET-MR registration was not included due to the scope of this study. The first approach to generating the ground truth of registration evaluation is introducing extrinsic markers either by neurosurgery or by attaching them to the skin surface. For example, the “Retrospective Registration Evaluation Project” provided a common evaluation framework based on gold standard PET, computed tomography (CT), and MR images of nine patients undergoing neurosurgery (four fiducial markers on each patient).15 The disadvantage of validation based on fiducial markers is that markers are spatially sparse and far from the interior brain structures and, thus, do not provide local resolution and accuracy sufficient for validation.14 An alterative ground truth method is to use a physical phantom with fiducial markers. For example, eight registration algorithms were evaluated using a Hoffman brain phantom filled with 99Tcm (single-photon emission tomography [SPET]) and a phantom filled with water doped with Gd-DTPA.16 The physical phantom images share the same limitation of spatially sparse markers (four markers were used in a previous study 16). The third approach to generate ground truth is construction of simulated images with known transformations. For example, two registration algorithms were compared using 192 PET images that were simulated from six MR images.17 Validations using the features extracted from the images are often used for inter-subject registration of MR images. For example, the manually segmented brain structures have been used to evaluate the different transformation models in registration of MR images in previous reports.1820 A selective-wavelet reconstruction technique using a frequency-adaptive wavelet space threshold was also used to compare transformation models for inter-subject MR PET registrations.21

The purpose of this study was to evaluate the agreement of the spatial normalized PET images with the MR-based brain atlas using well-accepted registration algorithms. The rationale for this study was based on the following considerations. First, normalization of PET images to the brain atlas will play an important role in interpreting PET results with the development of population-based multimodal brain atlases. However, previous validations were focused on multi-modality registration of images from the same patient or simulated images from the same subject. Secondly, most of the previous comparison studies were focused on rigid-body registration, although it is reasonable to believe that nonlinear registration may not improve the agreement of normalization with consideration of the spatial resolution of PET images. We do expect affine and linear registration to provide better agreement than the previously compared rigid-body registration. In this study, nine registration algorithms with RIU,10,11,22 normalized mutual information (NMI),12 and normalized CC objective functions with rigid-body, linear, affine, and nonlinear transformation models were used to normalize head sections of whole-body PET images of 25 patients to the population-based brain atlas ICBM452. These registration algorithms are implemented in registrations toolkits of Automated Image Registration (AIR 5.0);10,11,22,23 Medical Image Processing, Analysis and Visualization (MIPAV) (mipav.cit.nih.gov); HERMES software (Hermes Medical Solutions, Sweden); and VTK CISG. The gray matter (GM) on normalized PET images was segmented using the fuzzy C-means algorithm24 implemented in MIPAV. The tissue concordance of GM on the normalized PET images and those provided with the atlas were calculated to evaluate the agreement of the spatial normalization.

Methods

Data

The head sections of whole-body FDG PET images of 25 patients treated for malignant lymphoma or Hodgkin's disease were used for the evaluation. The patient group was arbitrarily selected and consisted of 11 girls and 14 boys with a median age of 6.92 years (range, 5.08–9.0 years). The whole-body PET images were acquired for diagnosis purposes on a GE Discovery LS PET-CT scanner with a spatial resolution of Inline graphic (Fig. 1).

Fig 1.

Fig 1

An illustration of a PET image and the brain atlas used for registration evaluation. Transverse slices near the level of basal ganglia and generally showing the putamen and lateral ventricle were selected for illustration. The left image and the right image are slices from the PET and atlas data, respectively.

The brain atlas, ICBM452 (www.loni.ucla.edu/Atlases/), is a population-based brain atlas averaged from T1-weighted MR images of 452 healthy young adult brains. The space of the atlas is not based on any single subject but is an average space constructed from the average position, orientation, scale, and shear from all the individual subjects (Fig. 1).

Retrospectively using the above data for this study was approved by the institutional review board at our institution.

Registration Algorithms

Nine registration methods with RIU, NMI, or normalized CC objective functions were evaluated using rigid, linear, affine, and nonlinear transformation models. All registration algorithms were implemented as true 3D registrations. The registration methods were selected on the basis of the knowledge of and availability to the authors. This study focused on objective functions and transformation models. Therefore, we used the optimization search method provided by each registration toolkit, and trilinear interpolation was used for all the registration tests.

Four registration algorithms on the basis of the RIU objective function with rigid, linear, affine, or nonlinear transformation models were evaluated using AIR 5.0.10,11,22,23 The RIU objective function is a mean-normalized standard deviation of the ratio of the source image intensity to the target image intensity.22,23 The rigid, linear, and affine models are the standard transformation models as described in “Introduction”. A third-order polynomial with 60 parameters was used for nonlinear registration as recommended by the AIR developer (bishopw.loni.ucla.edu/AIR3/howtosubjects.html). The optimization procedure used in AIR is an iterative, univariate Newton–Raphson search.22,23

Four registration algorithms on the basis of NMI objective function with rigid, linear, affine, or nonlinear transformation models were evaluated. The NMI objective function is a measure of how well one image explains the other and is calculated as the sum of probability distributions of the source and target images divided by the joint probability distribution of the source and target images.12,25 We used the NMI-based rigid-body registration implemented in MIPAV (mipav.cit.nih.gov) with the Powell optimization search. The linear registration-based NMI used was implemented in the commercially available HERMES software (Hermes Medical Solutions) with the Powell optimization search. The affine registration used was implemented in SPM2 (www.fil.ion.ucl.ac.uk/spm/software/spm2/) with the Powell optimization search. The free form deformation (FFD) based on NMI implemented in the VTK CISG registration toolkit 26 was used as the nonlinear registration.

One registration algorithm with the normalized CC objective function and the affine transformation model implemented in MIPAV was tested. The normalized CC is a measure of the correlation between the source and target images with the assumption of a linear relationship. It is calculated as the sum of the intensity correlation (normalized by subtracting the mean) between the target and source images divided by the standard deviation.27 The Powell optimization search was used for this registration algorithm.

Evaluation

The agreement of spatial normalization was evaluated qualitatively by visual inspection of normalized PET images overlaid on the brain atlas and quantitatively by calculation of GM tissue concordance between the normalized PET images and the brain atlas. The visual inspection included visual assessment of the agreement of brain surfaces, cerebella, brain stems, and boundaries of corpus callosum and lateral ventricles between the normalized PET images and the brain atlas in orthogonal views.

The normalized PET images were segmented using fuzzy C-means 24 implemented in MIPAV. The fuzzy C-means algorithm is an unsupervised segmentation method based on fuzzy set theory and generalized K-means algorithms. It iteratively minimizes the fuzzy membership function weighted difference between voxel intensity and clusters' centers,

graphic file with name M2.gif 1

in which Inline graphic is the fuzzy membership function and m can be any number greater than 1 (m = 2 in MIPAV). The iteration stops when the difference between membership functions of all voxels in two consecutive iterations is smaller than the tolerance. The detailed implementation of the algorithm is described in the MIPAV documentation (mipav.cit.nih.gov). The validity of fuzzy C-means segmentation algorithm on PET was reported in previous studies.24,28 The normalized PET images were segmented into three classes with tolerance of 0.0001 for this study. The GM regions on the atlas were defined as voxels for which GM probability (provided with the atlas) was greater than that of white matter (WM) and cerebrospinal fluid (CSF) on the tissue probability map and greater than 0.5. The tissue concordances of GM between the normalized PET images and atlases were calculated as kappa indices:

graphic file with name M4.gif 2

where Ip and IA are normalized PET images and the brain atlas, respectively.

Results

Figure 2 shows a typical normalized PET image overlaid on the atlas using RIU with the four transformation models. Visual inspection of RIU-based transformations showed that the affine registrations provided better agreement of brain surface and matched the low FDG uptakes with the WM and ventricles better than the rigid-body registration did. Visual inspection also showed that normalized PET images from 20 of 25 patients had distorted transformations; thus, we considered that these registrations failed (Fig. 3).

Fig 2.

Fig 2

An example of spatial normalization of a PET image to the MR brain atlas using RIU with various transformation models. The normalized PET image was overlaid onto the brain atlas with blue, green, yellow, and red representing increasing FDG uptake (intensities of PET images). Orthogonal views of the same image are displayed in each column. From left to right, the normalized PET images show rigid-body, linear, affine, and nonlinear transformation models.

Fig 3.

Fig 3

A PET image that was considered a failed normalization with distortion. From left to right, orthogonal (sagittal, coronal, and transverse) views are shown.

An example of the spatial normalization of PET images to the MR brain atlas using NMI with the four transformation models is shown in Figure 4. Visual inspection showed a similar degree of mismatching among all the transformation models (better brain surface agreement with less agreement of FDG uptake, WM, and CSF or vice versa). However, no distorted normalization was observed in the nonlinear registrations.

Fig 4.

Fig 4

Spatial normalization of PET to the brain atlas using NMI with various transformation models. The normalized PET image was overlaid onto the brain atlas with blue, green, yellow, and red representing increasing FDG uptake (intensities of PET images). Orthogonal views of the same image are displayed in each column. From left to right, the normalized PET images show rigid-body, linear, affine, and nonlinear transformation models.

A comparison of spatial normalization using various objective functions with the affine registration model is shown in Figure 5. Visual inspection showed that spatial normalization using CC and RIU had better agreement of brain surface agreement than those using NMI had and that normalization using RIU matched the low FDG uptake with CSF and WM better than using NMI.

Fig 5.

Fig 5

Comparison of the spatial normalization of PET data to the brain atlas using various objective functions with affine registration. The normalized PET image was overlaid onto the brain atlas with blue, green, yellow, and red representing increasing FDG uptake (intensities of PET images). Orthogonal views of the same image are displayed in each column. From left to right, the images show the CC, RIU, and NMI objective functions.

The GM concordances of the normalized PET images and the brain atlas are summarized in Table 1. The spatial normalization using RIU had the best agreement (averaged kappa index of 0.71) among the objective functions evaluated. The linear and affine registrations had better GM agreement than rigid-body registration (most of the nonlinear registrations failed) using RIU as the objective function.

Table 1.

GM Concordance and Robustness of Registration

Registration Averaged GM concordance Standard deviation of the concordance Robustness (%)
RIU/rigid 0.60 0.06 100
RIU/linear 0.71 0.03 100
RIU/affine 0.71 0.04 100
RIU/nonlinear 20
NMI/rigid 0.59 0.07 100
NMI/linear 0.59 0.06 100
NMI/affine 0.52 0.04 96
NMI/nonlinear 0.56 0.11 92
CC/affine 0.64 0.04 100

GM concordance was calculated as the kappa index between normalized PET and brain atlas. Robustness was calculated as the percentage of registrations without obvious distortion over the total number of registrations. RIU, NMI, and CC represent the objective functions of RIU, NMI, and CC. “–” designates that the GM concordance was not evaluated or registration was not evaluated

Discussion

The linear and affine registration methods using RIU had the best agreement among the registrations we evaluated. It is well known that registration performance is affected by the level of noise and resolution of image data. Therefore, we used PET data acquired from actual clinical care rather than data generated for research purposes or simulated from MR images for the evaluation. The whole-body PET scans of these patients were acquired for lymphoma or Hodgkin's disease without any sign of central nervous system insult. We used GM concordance as the quantitative measure of spatial normalization because the data were acquired without landmarks. Although previous evaluation studies focused on rigid-body registration, the results of the co-registration portion of our study were consistent with those of previous studies.1517,29

The visual inspection was consistent with the measure of GM concordance, although visual inspection cannot produce a conclusive judgment in some cases; for example, there was no visually discerned difference in the comparison of spatial normalization using various objective functions with the affine registration model (Fig. 5). Visual inspection showed that the affine registration model had the best agreement; hence, we used it to evaluate the objective functions. The other objective functions, including count difference, shape difference, sign changes, variance, square root, 2D gradient, and 3D gradient, were compared with RIU in a previous study,16 and RIU yielded the smallest errors for SPET-MRI registration. Therefore, we did not repeat those comparisons in this study. We used GM concordance as the quantitative metric for evaluation in this study because we believe a “perfect” registration will align the brain tissues together. However, GM concordance could be affected by the segmentation, inherent resolutions, and other factors. We believe that the effect of the accuracy of segmentation on the evaluation process is negligible because there are significant intensity differences (FDG uptakes) between the GM and WM/CSF on PET images. The GM segmentation on the brain atlas is provided with the atlas, averaged from the individual brain images. Therefore, GM concordance can be used as an index for comparison between registration methods. We did not use the WM or CSF concordances because the intensities of WM and CSF are similar on PET images. Although we are not able to ensure that the evaluated algorithms stand for the state-of-art registration methodology, to the best of our knowledge, the software and algorithms evaluated in this study are widely used and well accepted in clinic and research studies.

We found that spatial normalization using RIU had the best agreement of GM concordance among the objective functions tested. Although both RIU and NMI measure the intensity correspondence between two images, RIU segments one image into partitions and maps the other image voxel intensities into each partition, whereas NMI maps two images' intensity distribution with the assumption that the GM intensity on the MR image can be mapped to the GM intensity on the PET image with a similar intensity distribution (not necessarily equivalence of voxel intensity). There is a great difference between GM and WM on FDG PET images, and variations in voxel intensity for each tissue are greater than those with other image modalities such as MR and CT. This feature enables RIU-based registration to more reliably find the global minimum because of the greater weighted sum of standard deviations, although the low resolution and partial volume effect make boundaries on PET images less sharp. We found that linear and affine registrations had better agreement than the rigid-body registration. This was expected for inter-subject registration because of inter-subject size variations. However, we were surprised to find out that registrations with nonlinear transformation models did not yield better or equal agreements than affine registration, considering that nonlinear registration was performed on the affinely normalized images. We think that the possible reason for this is that the limited information resulting from the low resolution of PET images cannot produce a “correct” transformation; that is, the distorted registration does have smaller objective function value although it does not demonstrate better “alignment” than the affine normalization.

There is no consensus on how good is good enough for registration. That determination depends on the purpose of registration and availability of other approaches. The best agreement we obtained in this study was a GM concordance of 0.71 by linear/affine registration based on RIU. We think that it is satisfactory for spatial normalization of PET data to a brain atlas, considering the fact that the GM concordance is calculated on resampled PET and brain atlas images with spatial resolution of Inline graphic, whereas the spatial resolution of the original PET image was Inline graphic. This is in agreement with the conclusions from previous validation studies.16,17,29,30 Furthermore, there are different brain templates available for various purposes of registration. For example, we believe that the MNI PET template is more appropriate if the purpose of registration is to normalize individual PET image into a common space. We used ICBM452 brain atlas based on MR images because we would like to explore the relationship between FDG uptakes and anatomic structures and cytoarchitectonic structures in the future. For the same reason, we did not normalize the PET data to the pediatric brain template.

There were several limitations to this study. First, whole-body PET scans were used rather than brain PET scans. It is to be expected that the whole-body PET images had higher noise levels and lower resolution than brain-specific images would have had. The whole-body PET data were used based on the availability of normal PET data (with the assumption that head images of the patients with malignant lymphoma or Hodgkin's disease would not be affected by their disease). Second, we tested objective functions and transformation models but not other key factors of the optimization process and interpolation method. Third, the time required for computation of spatial normalization was not evaluated. The second and third limitations were because we used publicly available image-processing toolkits. We did not code algorithms to test the optimization process and interpolation method because these two factors are more standardized than the other two factors.

Conclusion

In summary, we found that either linear or affine registration using RIU as the objective function provides the best GM concordance among the registration methods we tested. The linear and affine registration models generally yield higher GM concordance than rigid and nonlinear models.

Footnotes

Part of the study has been presented on ISMRM 17th annual meeting.

References

  • 1.Carson RE, Daube-Witherspoon ME, Herscovitch P. Quantitative Functional Brain Imaging With Postron Emission Tomography. San Diego, CA 92101: Academic Press; 1998. [Google Scholar]
  • 2.Phelps ME. PET: Molecular Imaging and Its Biological Applications. New York, NY: Springer; 2004. [Google Scholar]
  • 3.Phelps ME, Huang SC, Hoffman EJ, Selin C, Sokoloff L, Kuhl DE. Tomographic measurement of local cerebral glucose metabolic-rate in humans with (F-18)2-fluoro-2-deoxy-d-glucose—validation of method. Ann Neurol. 1979;6:371–388. doi: 10.1002/ana.410060502. [DOI] [PubMed] [Google Scholar]
  • 4.Toga AW, Thompson PM, Mori S, Amunts K, Zilles K. Towards multimodal atlases of the human brain. Nat Rev Neurosci. 2006;7:952–966. doi: 10.1038/nrn2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Pelizzari CA, Chen GTY, Spelbring DR, Weichselbaum RR, Chen CT. Accurate 3-dimensional registration of CT, PET, and/or MR images of the brain. J Comput Assist Tomogr. 1989;13:20–26. doi: 10.1097/00004728-198901000-00004. [DOI] [PubMed] [Google Scholar]
  • 6.Besl PJ, Mckay ND. A method for registration of 3-D Shapes. IEEE Trans Pattern Anal Mach Intell. 1992;14:239–256. doi: 10.1109/34.121791. [DOI] [Google Scholar]
  • 7.Xue Z, Shen DG, Davatzikos C. Correspondence detection using wavelet-based attribute vectors. Med Image Comput Comput Assist Interv—MICCAI 2003. 2003;2879:762–770. doi: 10.1007/978-3-540-39903-2_93. [DOI] [Google Scholar]
  • 8.Lemieux L, Kitchen ND, Hughes SW, Thomas DGT. Voxel-based localization in frame-based and frameless stereotaxy and its accuracy. Med Phys. 1994;21:1301–1310. doi: 10.1118/1.597403. [DOI] [PubMed] [Google Scholar]
  • 9.Hajnal JV, Saeed N, Oatridge A, Williams EJ, Young IR, Bydder GM. Detection of subtle brain changes using subvoxel registration and subtraction of serial MR-images. J Comput Assist Tomogr. 1995;19:677–691. doi: 10.1097/00004728-199509000-00001. [DOI] [PubMed] [Google Scholar]
  • 10.Woods RP, Grafton ST, Watson JDG, Sicotte NL, Mazziotta JC. Automated image registration: II. Intersubject validation of linear and nonlinear models. J Comput Assist Tomogr. 1998;22:153–165. doi: 10.1097/00004728-199801000-00028. [DOI] [PubMed] [Google Scholar]
  • 11.Woods RP, Grafton ST, Holmes CJ, Cherry SR, Mazziotta JC. Automated image registration: I. General methods and intrasubject, intramodality validation. J Comput Assist Tomogr. 1998;22:139–152. doi: 10.1097/00004728-199801000-00027. [DOI] [PubMed] [Google Scholar]
  • 12.Maes F, Collignon A, Vandermeulen D, Marchal G, Suetens P. Multimodality image registration by maximization of mutual information. IEEE Trans Med Imaging. 1997;16:187–198. doi: 10.1109/42.563664. [DOI] [PubMed] [Google Scholar]
  • 13.Gan R, Wu J, Chung ACS, Yu SCH, Wells WM. Multiresolution image registration based on Kullback–Leibler distance. Med Image Comput Comput Assist Interv—MICCAI 2004. 2004;3216:599–606. doi: 10.1007/978-3-540-30135-6_73. [DOI] [Google Scholar]
  • 14.Gholipour A, Kehtarnavaz N, Briggs R, Devous M, Gopinath K. Brain functional localization: A survey of image registration techniques. IEEE Trans Med Imaging. 2007;26:427–451. doi: 10.1109/TMI.2007.892508. [DOI] [PubMed] [Google Scholar]
  • 15.West J, Fitzpatrick JM, Wang MY, Dawant BM, Maurer CR, Kessler RM, Maciunas RJ, Barillot C, Lemoine D, Collignon A, Maes F, Suetens P, Vandermeulen D, vandenElsen PA, Napel S, Sumanaweera TS, Harkness B, Hemler PF, Hill DLG, Hawkes DJ, Studholme C, Maintz JBA, Viergever MA, Malandain G, Pennec X, Noz ME, Maguire GQ, Pollack M, Pelizzari CA, Robb RA, Hanson D, Woods RP. Comparison and evaluation of retrospective intermodality brain image registration techniques. J Comput Assist Tomogr. 1997;21:554–566. doi: 10.1097/00004728-199707000-00007. [DOI] [PubMed] [Google Scholar]
  • 16.Koole M, D'Asseler Y, Laere K, Walle R, Wiele C, Lemahieu I, Dierckx RA. MRI-SPET and SPET-SPET brain co-registration: Evaluation of the performance of eight different algorithms. Nucl Med Commun. 1999;20:659–669. doi: 10.1097/00006231-199907000-00009. [DOI] [PubMed] [Google Scholar]
  • 17.Kiebel SJ, Ashburner J, Poline JB, Friston KJ. MRI and PET coregistration—A cross validation of statistical parametric mapping and automated image registration. Neuroimage. 1997;5:271–279. doi: 10.1006/nimg.1997.0265. [DOI] [PubMed] [Google Scholar]
  • 18.Noblet V, Heinrich C, Heitz F, Armspach JP. Retrospective evaluation of a topology preserving non-rigid registration method. Med Image Anal. 2006;10:366–384. doi: 10.1016/j.media.2006.01.001. [DOI] [PubMed] [Google Scholar]
  • 19.Crum WR, Rueckert D, Jenkinson M, Kennedy D, Smith SM. A framework for detailed objective comparison of non-rigid registration algorithms in neuroimaging. Med Image Comput Comput Assist Interv—MICCAI 2004. 2004;3216:679–686. doi: 10.1007/978-3-540-30135-6_83. [DOI] [Google Scholar]
  • 20.Grachev ID, Berdichevsky D, Rauch SL, Heckers S, Kennedy DN, Caviness VS, Alpert NM. A method for assessing the accuracy of intersubject registration of the human brain using anatomic landmarks. Neuroimage. 1999;9:250–268. doi: 10.1006/nimg.1998.0397. [DOI] [PubMed] [Google Scholar]
  • 21.Dinov ID, Mega MS, Thompson PM, Woods RP, Sumners DL, Sowell EL, Toga AW. Quantitative comparison and analysis of brain image registration using frequency-adaptive wavelet shrinkage. IEEE Trans Inf Technol Biomed. 2002;6:73–85. doi: 10.1109/4233.992165. [DOI] [PubMed] [Google Scholar]
  • 22.Woods RP, Mazziotta JC, Cherry SR. MRI-PET registration with automated algorithm. J Comput Assist Tomogr. 1993;17:536–546. doi: 10.1097/00004728-199307000-00004. [DOI] [PubMed] [Google Scholar]
  • 23.Woods RP, Cherry SR, Mazziotta JC. Rapid automated algorithm for aligning and reslicing PET images. J Comput Assist Tomogr. 1992;16:620–633. doi: 10.1097/00004728-199207000-00024. [DOI] [PubMed] [Google Scholar]
  • 24.Pham DL, Prince JL. An adaptive fuzzy C-means algorithm for image segmentation in the presence of intensity inhomogeneities. Pattern Recogn Lett. 1999;20:57–68. doi: 10.1016/S0167-8655(98)00121-4. [DOI] [Google Scholar]
  • 25.Studholme C, Hill DLG, Hawkes DJ. An overlap invariant entropy measure of 3D medical image alignment. Pattern Recogn. 1999;32:71–86. doi: 10.1016/S0031-3203(98)00091-0. [DOI] [Google Scholar]
  • 26.Crum WR, Hartkens T, Hill DLG. Non-rigid image registration: Theory and practice. Br J Radiol. 2004;77:S140–S153. doi: 10.1259/bjr/25329214. [DOI] [PubMed] [Google Scholar]
  • 27.Hajnal JV, Hill DLG, Hawkes DJ. Medical Image Registration. New York: CRC Press; 2001. [Google Scholar]
  • 28.Hatt M, Rest CC, Turzo A, Roux C, Visvikis D. A fuzzy locally adaptive Bayesian segmentation approach for volume determination in PET. IEEE Trans Med Imaging. 2009;28:881–893. doi: 10.1109/TMI.2008.2012036. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Strother SC, Anderson JR, Xu XL, Liow JS, Bonar DC, Rottenberg DA. Quantitative comparisons of image registration techniques based on high-resolution MRI of the brain. J Comput Assist Tomogr. 1994;18:954–962. doi: 10.1097/00004728-199411000-00021. [DOI] [PubMed] [Google Scholar]
  • 30.West J, Fitzpatrick JM, Wang MY, Dawant BM, Maurer CR, Kessler RM, Maciunas RJ. Retrospective intermodality registration techniques for images of the head: Surface-based versus volume-based. IEEE Trans Med Imaging. 1999;18:144–150. doi: 10.1109/42.759119. [DOI] [PubMed] [Google Scholar]

Articles from Journal of Digital Imaging are provided here courtesy of Springer

RESOURCES