Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2010 Oct 1.
Published in final edited form as: Neuroimage. 2009 Apr 14;47(4):1341–1351. doi: 10.1016/j.neuroimage.2009.04.024

Implicit Reference-Based Group-wise Image Registration and Its Application to Structural and Functional MRI

Xiujuan Geng 1, Gary E Christensen 2, Hong Gu 1, Thomas J Ross 1, Yihong Yang 1
PMCID: PMC2719682  NIHMSID: NIHMS110839  PMID: 19371788

Abstract

In this study, an implicit reference group-wise (IRG) registration with a small deformation, linear elastic model was used to jointly estimate correspondences between a set of MRI images. The performance of pair-wise and group-wise registration algorithms was evaluated for spatial normalization of structural and functional MRI data. Traditional spatial normalization is accomplished by group-to-reference (G2R) registration in which a group of images are registered pair-wise to a reference image. G2R registration is limited due to bias associated with selecting a reference image. In contrast, implicit reference group-wise (IRG) registration estimates correspondences between a group of images by jointly registering the images to an implicit reference corresponding to the group average. The implicit reference is estimated during IRG registration eliminating the bias associated with selecting a specific reference image. Registration performance was evaluated using segmented T1-weighted magnetic resonance images from the Nonrigid Image Registration Evaluation Project (NIREP), DTI and fMRI images. Implicit reference pair-wise (IRP) registration—a special case of IRG registration for two images—is shown to produce better relative overlap than IRG for pair-wise registration using the same small deformation, linear elastic registration model. However, IRP-G2R registration is shown to have significant transitivity error i.e., significant inconsistencies between correspondences defined by different pair-wise transformations. In contrast, IRG registration produces consistent correspondence between images in a group at the cost of slightly reduced pair-wise RO accuracy compared to IRP-G2R. IRG spatial normalization of the fractional anisotropy (FA) maps of DTI is shown to have smaller FA variance compared with G2R methods using the same elastic registration model. Analyses of fMRI data sets with sensorimotor and visual tasks show that IRG registration, on average, increases the statistical detectability of brain activation compared to G2R registration.

Introduction

Inter-subject spatial normalization is an important process in medical image analysis and in particular brain image analysis. The traditional approach to spatially normalize a group of medical images is to use pair-wise image registration in which all images are individually registered to a target image. The problem with this approach is that it introduces a bias according to the target image used. In contrast, group-wise image registration jointly estimates the correspondences between a group of images. Many group-wise registration techniques naturally estimate transformations from individual subjects to a common coordinate space eliminating the bias associated with registering to a target image. Applications that use group-wise image registration have increased in recent years including creation of atlases (Frangi et al., 2002; Christensen et al., 2006b), building shape models (Cootes et al., 1995, 1999; Rueckert et al., 2003), motion tracking along time series (Crum et al., 2004), aging population studies (Mega et al., 2005; Pieperhoff et al., 2008) and disease population analysis (Zhang et al., 2007).

Reference-based group-wise registration is a normalization approach in which each image in the group is registered to a selected reference image. A limitation of this approach is that the selected reference may not adequately capture common structure in the population, and therefore the spatial normalization to this reference may induce bias into later analysis. To minimize the bias caused by the reference selection, some methods have been proposed (Guimond et al., 2000; Christensen et al., 2006b) to iteratively update the selected reference image based on averaging the deformed images or transformation fields.

A strategy to avoid reference selection is to simultaneously estimate transformations for each subject in the group that map them into a common space without choosing any reference image. Different criteria such as joint probability distribution (Studholme, 2003; Studholme and Cardenas, 2004; Bhatia et al., 2004), voxel-wise entropy (Zollei et al., 2005), Jensen-Shannon divergence (Wang et al., 2006), squared voxel intensity difference (Avants and Gee, 2004; Joshi et al., 2004), and the minimum description length (Twining et al., 2005) have been utilized as similarity costs to define the objective function and estimate transformations. Group-wise registration of non-scalar images, such as tensor components in diffusion tensor imaging (DTI) can also be formulated in a similar way (Zhang et al., 2007). Many of these methods compute and update the unbiased reference image by averaging the deformed images during the estimation procedure.

This paper uses an implicit reference group-wise (IRG) registration framework which avoids bias caused by reference selection and evaluates its performance with various image modalities. The method jointly estimates transformations from each image in the group to a “hidden” reference by optimizing the intensity difference of each pair of deformed images. The basis of the method is that transformations take each image to a common coordinate, therefore, the images after deformation should have similar shapes to each other, and the intensity difference at the corresponding location should be minimized. This framework is based on the work in Christensen and Johnson (2001) in terms of using the same small elastic deformation model, while expansion and modification of the similarity cost function are made to enable the method to map a set of more than two images to a common space instead of estimating transformations from one image directly to another.

The group-wise registration algorithm used in this work is similar to the large deformation group-wise registration algorithm used in Joshi et al. (2004). In contrast, the algorithm used in this work assumes a small deformation linear elastic model and uses the Fourier series to parametrize the deformation field. We show in the discussion section that minimizing the similarity cost function used in this paper is equivalent to minimizing the similarity function used in Joshi et al. (2004).

A set of annotated T1-weighted MRI data (Christensen et al., 2006a) were used to evaluate the performance of the IRG registration method. Results show that the implicit reference method produced smaller registration errors compared to reference based methods that directly register each image to a reference. Moreover, experiments were designed to compare the directly estimated transformations from one image to another and the composed transformations from each image to the implicit reference, and results indicate that the improvement of the implicit reference method is partly due to the composition of two small deformations.

This paper demonstrates the potential of IRG registration for group analyses of DTI and functional MRI (fMRI). DTI enables the non-invasive study of white matter integrity and anatomical connectivity (Basser and Pierpaoli, 1996). Fractional anisotropy (FA), a scalar index of fiber directionality derived from DTI data sets, was used for image normalization and analyses. The performance of various group-wise registration methods was compared, and the results show that the implicit reference technique, with the small deformation elastic model, provides better transformations compared to reference-based methods in terms of smaller within-group variance after registration.

Functional MRI techniques provide non-invasive study of neural activity detected by blood oxygen level dependent (BOLD) or other signals (Ogawa et al., 1990). The accuracy of spatial normalization is crucial for group analyses of functional activation. Most current fMRI normalization techniques follow a standard approach that utilizes a lower order registration model under a reference-based framework. A motivation of this work was to examine whether improved spatial normalization would enhance the statistical detectability of brain activation (Miller et al., 2005). The implicit reference method and reference-based methods were applied to a set of fMRI data sets with sensorimotor and visual tasks. Results demonstrate that the smaller within-group variance using the implicit reference method resulted in improved statistics for brain activation detection.

Methods

Pair-wise (PW) image registration

We focus on estimating diffeomorphic transformations between two or more images under a small deformation model (Christensen et al., 1997). Let Ii represent an image volume with voxel dimensions of D1 × D2 × D3. The Eulerian transformation from the coordinate system of Ii to that of Ij can be parameterized as a set of vectors in Euclidean space, and described as a function hij : Ω → Ω, where Ω = {(x1, x2, x3)|0 ≤ x1D1; 0 ≤ x2D2; 0 ≤ x3D3}. Under the transformation hij, Ii(x) is deformed to Ii(hij(x)).

Conventional pair-wise image registration methods define the registration problem as finding the transformation hij from image Ii to Ij by minimizing an objective function such as

C(Ii(hij),Ij)=D(Ii(hij),Ij)+R(hij)

where D is a similarity cost function and R is a regularization term to penalize transformations with large and unsmooth distortion. In this paper, PW will be used to refer to pair-wise registration with a sum of squared differences similarity cost function denoted as D(I,J)=ΩI(x)J(x)2dx and linear elastic regularization function as in Eq. 1. The PW registration algorithm used in this paper is equivalent to the unidirectional, small deformation, linear elastic registration method described in Christensen and Johnson (2001). Alternatively, the pair-wise image registration problem can be formulated in a symmetric fashion to minimize the asymmetry caused by structural differences between the images being registered (Christensen and Johnson, 2001; Leow et al., 2007) and can use many different similarity cost functions (Studholme, 2003; Studholme and Cardenas, 2004; Bhatia et al., 2004; Zollei et al., 2005; Wang et al., 2006; Joshi et al., 2004).

Implicit reference pair-wise (IRP) image registration

Implicit reference pair-wise (IRP) registration is a special case of IRG registration (see below) when only two images are registered. Instead of directly estimating transformations between images Ii and Ij, IRP registration estimates transformations hiR from Ii to a common reference space halfway between the images. Similar approaches with large deformation registration model were proposed in previous studies (Avants and Gee, 2004; Avants et al., 2008; Beg and Khan, 2007). The similarity cost becomes D(Ii(hiR), Ij(hjR)), and the total cost function is defined by

C(hiR,hjR)=D(Ii(hiR),Ij(hjR))+R(hiR)+R(hjR).

For this paper, the sum of squared differences was used for D and R was chosen to be a linear elastic regularization function as in Eq. 1.

Group-to-Reference (G2R) image registration

Traditional spatial normalization is accomplished by group-to-reference (G2R) registration in which a group of images are registered pair-wise to a target or reference image. Throughout the rest of the paper, the terms PW-G2R and IRP-G2R will refer to G2R registration computed using the PW and IRP registration algorithm, respectively. G2R registration allows images from different individuals to be compared in a standard reference frame, i.e., the image space of the target or reference image. The drawback of G2R image registration is that the results are biased by factors such as the shape of the structures in the target image and noise in the image.

Implicit reference group-wise (IRG) image registration

The implicit reference group-wise (IRG) image registration problem can be stated as: estimate the N > 2 transformations hiR from each image Ii in a population to an unknown common space by minimizing the summation of the intensity difference between each pair of deformed images. The unknown common space is defined as the implicit reference space. The objective function for this registration problem can be stated as follows:

C=σij,i<jΩIi(hiR(x))Ij(hjR(x))2dx+ρiΩL(uiR(x))2dx, (1)

where hiR(x) = uiR(x) + x is the transformation from image i to the implicit reference R, uiR(x) is the displacement field from i to R, L(uiR(x)) is a linear differential operator regularization constraint, and σ and ρ are weighting parameters. In this work, we chose L to be the linear elasticity operator with the form Lu(x)=α2u(x)β(u(x))+γu(x), where =[x1,x2,x3] and 2==[2x12+2x22+2x32] (Christensen and Johnson, 2001). Each displacement field uiR(x) is a 3×1 vector-valued function parameterized by the 3D Fourier series uiR(x)=k=0d1μi,kej2πkx/d where j=1,k=[k1,k2,k3]T,d1=[d11,d21,d31]T,x/d=[x1/d1,x2/d2,x3/d3]T and μi,k=[μi,k,1,μi,k,2,μi,k,3]T. The basis coefficients μi,k = ai,k + jbi,k are initialized to be 0s, and determined by gradient descent using

ai,kn+1=ai,kn+C(μi,kn)ai,k,andbi,kn+1=bi,kn+C(μi,kn)bi,k, (2)

where is the step size, and the superscript n corresponds to the value of the parameters at the nth iteration. The estimation details can be found in Christensen and Johnson (2001). The weighting parameter σ of the similarity cost was set to be 1, and the weighting parameter ρ of the regularization constraint was set to be 0.00025 through out the optimization procedure. A spatial and frequency multi-resolution procedure was used to estimate the full resolution registration to avoid local minima. Different levels of resolutions were applied for different image modalities with various voxel dimensions. At each resolution level, the initial and final number of harmonics and the number of iterations were set differently as described in Table 1. When the number of harmonics, r, is set, the triple summation of uiR over k where k = [0 ≤ kd − 1] is replaced by k = [0 ≤ kr − 1,dr − 1 ≤ kd − 1]. The calculation of forward and reverse fast Fourier transforms (FFT) of hiR dominates the computational time at each iteration. Despite the double summation in the similarity cost term, the computational cost is still approximately linear to the number of images, since the number of transformations is equal to the number of images. For example, this algorithm takes about 3 hours to register 15 images with voxel dimensions of 80 × 95 × 75, while it takes about 6 hours to register 30 images with the same dimensions on a 2.4 GHz AMD Opteron Linux system.

Table 1.

Multi-resolution IRG registration parameters used to register T1-weighted MRI, DTI and fMRI images with different voxel dimensions.

voxel dims resolution initial num of harmonics final num of harmonics iterations
256 × 300 × 256 1/8 1 5 500
1/4 5 9 500
1/2 9 13 250
1 13 17 50
80 × 95 × 75 1/4 1 5 500
1/2 5 9 500
1 9 13 250
54 × 64 × 50 1 5 9 500

Fig. 1 illustrates the IRG registration method for mapping images to a common space and mapping any pair of images. The transformation hiR and the inverse of hjR are concatenated to compute the transformation between a pair of images Ii and Ij. The inverse transformation was computed using the method described in Christensen and Johnson (2001).

Figure 1.

Figure 1

Implicit reference group-wise (IRG) registration method framework. Transformations hiR from each image to an implicit space are estimated; the transformation hij between every pair of images is obtained by concatenating transformations, hij(x)=hiR(hjR1(x)).

Fig. 2 illustrates the IRG registration method applied to a population of four 2D shapes that consisted of a circle, two ellipses, and a square. It is shown that the transformations bring each shape to the average shape of all the shapes and the deformed shapes are similar to each other.

Figure 2.

Figure 2

IRG registration of four 2D phantom shapes. The top row shows the original four shapes that were registered with the IRG method. The middle row shows the deformed images produced by applying the estimated transformation hiR to image i above it. The images in the middle row correspond to the implicit reference image. The bottom row shows the transformation hiR applied to a rectangular grid.

Evaluation using NIREP

In order to evaluate the performance of registration algorithms, it is necessary to have a wide collection of consistently annotated image data sets and evaluation statistics. The NA0 database from the registration evaluation package of the Non-Rigid Image Registration Evaluation Project (NIREP) (Christensen et al., 2006a) was used to evaluate the IRG registration method. The NA0 database consists of MRI data from 16 subjects that include gray matter segmentations (Damasio, 2005; Allen et al., 2002, 2003). The images were initially segmented with Brainvox (Frank et al., 1997). The segmentations, including gray and white matter, were then restricted to the gray matter by applying the gray matter segmentation generated using the approach described in Grabowski et al. (2000) to avoid arbitrary boundaries in white matter. The final regions of interest (ROIs) include gray matter regions in the frontal lobe: frontal pole, superior frontal gyrus, middle frontal gyrus, inferior frontal gyrus, orbital frontal gyrus, precentral gyrus; parietal lobe: postcentral gyrus, superior parietal lobule, inferior parietal lobule; temporal lobe: temporal pole, superior temporal gyrus, infero-temporal region, parahippocampal gyrus; occipital lobe; cingulate gyrus; and insula. The 16 data sets were rigidly rotated and translated using the anterior commissure, posterior commissure, and inter-hemispheric fissure as landmarks.

The relative overlap (RO), which is also called the Jaccard or Tanimoto coefficient, was used to evaluate registration performance. The RO is computed for each ROI and measures how well the two ROIs agree with each other. The RO is defined as the number of voxels in the intersection of two ROIs divided by the number of voxels in the union. Stated mathematically, the RO of the kth ROI when image Ij is registered to Ii is defined as:

ROijk=x=1MROIik(x)ROIjk(hji(x))x=1MROIik(x)ROIjk(hji(x)), (3)

where ROIik(x) returns 1 if x is in the kth ROI in Ii, and returns 0 if is not, ∩ and ∪ denote AND and OR operations respectively, and M is the number of voxels in the image. Note that hji is estimated directly for pair-wise (PW) registration and obtained by hji=hjR(hiR1) for IRP and IRG registrations.

The average relative overlap (ARO) of the kth ROI using Ii as reference is defined as:

AROik=1N1j=1,jiNROijk, (4)

where N is the number of images in the group being averaged. The ARO using implicit reference method is defined as:

ARORk=1N×(N1)j=1,jiNi=1Nx=1MROIik(hiR(x))ROIjk(hjR(x))x=1MROIik(hiR(x))ROIjk(hjR(x)). (5)

Note that ARORk is defined in the implicit reference space while AROik is defined in the coordinate system of image Ii. Care must be taken when comparing ARO values that are computed in different coordinate systems due to coordinate system specific biases such as different sizes and shapes of the ROIs between different coordinate systems.

The transitivity error (Christensen and Johnson, 2003; Christensen et al., 2006a) of a set of transformations that register a group of images is a measure of consistency between the correspondences defined by the transformations. Ideally, transformations that define correspondence between three images should project a point from image A to B to C to A back to the original position. The transitivity error for a set of transformations is defined as the squared error difference between the composition of the transformations between three images and the identity map. The cumulative transitivity error (CTE) with respect to template image k is computed as

CTEk(x)=1(N1)(N2)i=1ikNj=1jijkNhki(hij(hjk(x)))x, (6)

where N is the number of images in the population. The average CTE of each ROI for different methods were computed for pair-wise, IRP and IRG registration. Transformations hijs were directly estimated using pair-wise registration, and were composed by hiR and hjR1 using IRP and IRG registration.

Application to diffusion tensor images

Thirty healthy subjects were recruited as part of a protocol approved by the Institutional Review Board (IRB) of the National Institute on Drug Abuse Intramural Research Program and provided written informed consent. All data were scanned on a Siemens 3T Allegra scanner. An EPI-based spin-echo pulse sequence was used to acquire diffusion-weighted MRI images. For each subject, 35 axial images were prescribed to cover the whole brain with a 128×128 in-plane matrix at a resolution of 1.719×1.719×4mm3. Besides the non-diffusion-weighted reference image, 12 directions were used to apply the diffusion-sensitive gradients with a b-factor of 1000s/mm2. For EPI, TR/TE = 5000/87ms, BW= 1700Hz/Pixel, and NEX=4. Three-dimensional T1-weighted anatomical images were also acquired.

Each structural image was automatically normalized to Talairach space using the AFNI software package (Cox, 1996). The 12 diffusion weighted images (DWIs) were aligned to their corresponding structural images for correction of motion and image artifacts using mutual information based affine registration (Jenkinson and Smith, 2001), provided by the FSL package (Smith et al., 2004). Normalization matrices were applied to DWIs, and FA images were then calculated from the diffusion tensor. After the preprocessing, all FA images have a 2 × 2 × 2mm3 voxel resolution and a 80 × 95 × 70 matrix.

A comparison of IRG and pair-wise G2R registration was performed to evaluate registration of FA images. Two different non-rigid registration models were applied for G2R registration: the B-spline based free form deformation (FFD) model (Rueckert et al., 1999) used in FSL and the small deformation elastic (SDE) model utilized in this work. The FFD method with a cross correlation similarity measure and the default parameters for FA registration in FSL was used: full resolution registration with a control point spacing of 20 mm resulting a 9 × 10 × 8 mesh of control points. The parameters used in the SDE model are listed in Table 1 for the voxel dimensions of 80 × 95 × 75. The reference was chosen in two different ways. One was to use the standard FA image provided by FSL, which is the average of 58 well-aligned high quality FA images with a 1 × 1 × 1mm3 resolution. The other was to select the “most representative” image. After obtaining all transformations (30 × 29 = 870) using FFD, the image with the smallest amount of average warping, which was necessary to align all other images to it, was selected as the “most representative” image. Including the IRG method, a total of 5 different non-rigid group-wise registration approaches were used: G2R registration using the standard image as reference with the FFD model and the SDE model; G2R registration using the “most representative” image as reference with the FFD model and the SDE model; and IRG registration with the SDE model. The FFD registration was used as a benchmark to evaluate the performance of the IRG method instead of a like-for-like comparison of the FFD and SDE models.

The standard deviation of deformed FA maps were used to measure the registration performance. Although the FA statistics are unknown in normal subjects, it is reasonable to assume that FA values consist of separate distributions for white matter and gray matter. For misaligned structures, this mixture of distributions with unequal means would produce larger sample variance than a single distribution for well aligned structures. Therefore, the cross subject variance is a reasonable indicator of how well the image structures are aligned.

Applications to fMRI studies

Twenty-nine different healthy controls provided written informed consent and were scanned on the same scanner using the same IRB approved protocol as the DTI study. The functional data were acquired using a single-shot gradient-echo EPI sequence with a TR of 4 s and a TE of 27 ms. Thirty-nine contiguous 4-mm thick oblique slices (30° axial to coronal) were prescribed to cover the whole brain with a 64 × 64 in-plane matrix at a resolution of 3.44 × 3.44 × 4mm3. Subjects performed a block-design finger tapping task cued by a flashing checkerboard, which started with 20s “off” (watching a cross on the screen without moving fingers) and followed by 7 cycles of 20s “on” (watching a flashing checkerboard on the screen and moving fingers) and 20s “off” states. High resolution T1-weighted anatomical images were also acquired with a TR of 2.5 s, a TE of 4.38 ms, a FA of 7°, and a voxel size of 1 × 1 × 1mm3.

EPI data preprocessing included slice-timing correction, motion correction and linear detrending. All functional data were spatially normalized to the standard Talairach space using affine registration with a resampled resolution of 3×3×3mm3. Spatial smoothing with a 6-mm Gaussian kernel was performed to increase spatial signal to noise ratio (SNR) after affine alignment and before the non-rigid group registration. All data preprocessing were conducted in AFNI.

Both IRG and pair-wise G2R registration were applied to the affine aligned EPI data. A 3D volume (i.e., at the fourth time point) from the EPI time series was selected from each subject as the registration input. The transformation fields obtained from the volume was then applied to all volumes along the time series. The results from affine registration were used as the baseline. For G2R registration, each subject in the group was selected as the reference, and the rest of the subjects were registered to it, resulting in 29 groups of reference-based normalized images. Including the implicit reference registration, there were a total of 31 groups of registration approaches (1 affine, 29 G2R and 1 implicit reference registration). After registration, across-subject image intensity standard deviation maps were computed for all registration groups. No further smoothing was applied after non-rigid registration. The general linear model was then exploited for analyzing the functional activation, and the linear regression coefficient β maps were obtained from each deformed 4D fMRI data.

For each group of the registered data sets, a one-sample t-test was applied to the β maps to test if the activation coefficients were significantly different from zero. A threshold of t > 5.3 with a cluster size greater than 54 mm3 (pcorrected < 0.01) was used to generate group functional activation maps. Four activated regions were selected as ROIs including visual cortex, left sensorimotor, right sensorimotor cortices and supplementary motor area (SMA). Common ROIs were obtained by taking the intersection of the corresponding 31 ROIs. The average of intensity standard deviation, β weights and the corresponding t-statistics on the four common ROIs were computed and compared.

Results

NIREP evaluation

Fig.3 shows intensity averages of registered images using the IRG algorithm (bottom row), rigid G2R registration (top top) and the PW group-to-reference (PW-G2R) registration in which MRI images 2-16 were registered to image 1 (middle row). The average intensity images produced by the PW-G2R and IRG registrations are much sharper than that produced by rigid G2R registration, showing that both approaches produce similar registration results. Note that the registered images in Fig.3(b) are shown in the coordinate system of the reference image whereas the registered images in Fig.3(c) are shown in the coordinate system of the implicit reference or group average (see Discussion).

Figure 3.

Figure 3

G2R and IRG registration results. Row (a) shows 8 of the 16 MRI images from the NA0 database after rigid G2R registration and the average intensity of the 16 images after rigid registration. Row (b) shows the corresponding 8 deformed images after PW-G2R registration using the first image as the reference and the average intensity computed from the 16 images in the coordinate space of image 1. Row (c) shows the corresponding 8 deformed images after IRG registration and the average of the 16 IRG registered images in the coordinate system of the implicit-reference space.

Fig. 4 shows the voxel-wise overlap produced by mapping the ROIs from 15 of the NA0 data sets to the 16th target reference image. The reference image and the ROIs associated with this reference coordinate system are displayed in Fig. 4(a) and (b) respectively. The other 15 MRI images have similar ROIs that were mapped to the reference coordinate system. The images in Fig. 4(c) show the number of ROIs that agree with the target ROI for rigid, PW, and IRP registration. This figure shows that PW registration is an improvement over rigid registration and IRP registration is an improvement over PW registration. Moreover, visualization of the overlap shows where the registration methods are more and less accurate with respect to the shape and complexity of the reference MRI image.

Figure 4.

Figure 4

Region of Interest (ROI) overlap comparison between rigid, PW, and IRP registration. Panel (a) shows the T1-weighted MRI reference image; panel (b) shows the gray matter ROIs associated with the reference image; and panel (c) shows the overlap of the ROIs produced by mapping the other 15 MRI data sets to the reference image using rigid, PW, and IRP registration. The overlap of the the ROIs are show for a range of 0.5 to 1.0 where 0.5 corresponds to 50% agreement and 1.0 corresponds to 100% agreement.

Fig. 5 compares IRG to G2R image registration performance using the ARO statistic for 32 ROIs. The G2R image registration was computed using the PW and IRP registration algorithms and will be referred to as PW-G2R and IRP-G2R, respectively. Due to the bias associated with picking a reference image, registration was repeated 16 times for PW-G2R and IRP-G2R registration using a different image (1-16) as the reference. Producing the PW-G2R and IRP-G2R results required computing and storing 2×16×15 = 480 transformations, i.e., all the pair-wise registrations between the 16 images for both the PW and IRP methods. This is in contrast to the 16 transformations computed using the IRG approach. The PW-G2R, IRP-G2R, and IRG approaches used the same small deformation elastic model and same set of parameters (see Table 1 for the voxel dimensions of 256 × 300 × 256).

Figure 5.

Figure 5

Comparison of (a) IRG to PW-G2R registration and (b) IRG to IRP-G2R registration methods by average relative overlap (ARO) for the 32 region of interests (ROIs) in the NIREP NA0 evaluation database. For each ROI, the square represents the ARO after rigid alignment, the star denotes the ARO using IRG registration, and the 16 “X”-shaped points denote to AROs using G2R registrations, where each color corresponds to a reference-based registration using one of the 16 images as the reference.

For each of the 32 ROIs, the ARO was measured in the implicit reference space for the IRG method and in the reference image coordinate system for G2R registration. In general, it is difficult to compare the ARO values across different spaces due to varying sizes and shapes of the ROIs. However, for this experiment, the ROIs are roughly the same shape and size in each of the coordinate systems making for a less than perfect but still meaningful comparison of the methods.

The results in Fig. 5 show that the IRG registration always outperforms the PW-G2R registration with respect to the ARO independent of the reference used. However, the IRG registration does not always outperform the IRP-G2R registration with respect to ARO. The reason for this is that the IRP registration is a special case of the IRG registration in which only two images are registered. In some cases, adding additional images to the registration helps improve the registration by giving additional information to get out of local minima. In other cases, adding additional images may produce additional local minima or other errors that reduce registration performance. Given that the ROIs have small and complex shapes (Geng, 2007), the relative overlap values are reasonable.

Fig. 6 shows a comparison between the pair-wise registrations produced by PW, IRP, and IRG registration with respect to RO and transitivity error (TE). The purpose of this analysis was to examine the trade-offs between estimating pair-wise registrations directly using PW and IRP registration and estimating pair-wise registration indirectly using IRG registration. The transformations hij from image Ii to Ij were computed directly for PW registration and computed by composing transformations using the formula hij=hiR(hjR1) for IRP and IRG registration approaches. Note that the implicit reference for each IRP registration is different and can be thought of roughly as the coordinate system halfway between the coordinate systems of images Ii and Ij. In contrast, the implicit reference for the IRG method can be thought of as the coordinate system corresponding to the average shape of the group. The RO calculated using Eq.3 and its standard deviation were compared for each ROI. The top bar chart of Fig. 6 shows that there is no significant difference of the average ROs between PW and IRG registration for registering two images together. For 9 out of 32 ROIs, the PW registration produced slightly larger RO values, while for the rest ROIs, the IRG registration generated larger values. In contrast, IRP outperforms the IRG method by 5% – 15% suggesting that IRP registration is better to use for pair-wise registration than either PW or IRG registration.

Figure 6.

Figure 6

Comparison of pair-wise transformations produced by the PW, IRP, and IRG registration methods with respect to (a) relative overlap (RO) and (b) average transitivity error (ATE). Average and standard deviation of RO and ATE are plotted for each ROI corresponding to 16 × 15 pair-wise transformations in the population.

The bottom bar chart in Fig.6 compares the average transitivity error (ATE) for each ROI for the PW-G2R, IRP-G2R, and IRG methods. The ATE was computed using Eq. 6. Results show that the ATE range is from 2 to 10 voxels for the PW-G2R and IRP-G2R registrations whereas the ATE for IRG registration are all smaller than 0.04 voxel. IRG registration produces much less transitivity error compared to the PW-G2R and IRP-G2R registration indicating that IRG generates a consistent set of transformations between a group of images.

To summarize, results shown in Fig. 6 suggest that IRP registration produces better relative overlap than IRG for pair-wise registration at the cost of significant transitivity error for IRP-G2R registration, i.e., significant inconsistencies between correspondences defined by different pair-wise transformations. In contrast, IRG registration produces consistent correspondence between images in a group at the cost of slightly reduced pair-wise RO accuracy compared to IRP-G2R. Thus, it is better to use IRP registration if the task is to register two images, but it is better to use IRG registration if the task is to register a group of images. Furthermore, the IRP registration should be favored over PW registration when registering two images.

DTI data sets

The mean and standard deviation of the deformed FA maps using different registration methods are displayed in Fig. 7. The average standard deviation (ASD) of FA values in the whole brain and a white matter ROI are presented in Table 2. The white matter ROI was generated by thresholding the average deformed FA images obtained from each of the five methods at 0.3 and taking the overlap region of the five white matter masks. As shown in Table 2, the B-spline FFD model reduced the ASD approximately 30 percent compared to the affine model, and the SDE model reduced ASD approximately 10 percent compared to the FFD model. Within the SDE model, the IRG method further reduced the ASD 6 percent in white matter as well as in the entire brain compared with the G2R registration using either the standard image or the “most representative” image as the reference. Note, since the SDE model with the parameters used in this work generates transformations with more degrees of freedom compared to the FFD model with the default parameters in FSL (3 × 13 × 13 × 13 vs. 3 × 9 × 10 × 8), the reduced ASD does not necessarily indicate that the SDE model is superior to the FFD model. The results give a sense of how good the IRG registration is compared to a standard method in a software package with a large number of users.

Figure 7.

Figure 7

Comparison of average and standard deviation of deformed FA images after different registration methods. Top row includes average deformed FA images after (a) affine registration, (b) G2R registration using standard image as reference with B-spline free form deformation (FFD) model, (c) G2R registration using “most representative” image as reference with FFD model, (d) G2R registration using standard image as reference with small deformation elastic (SDE) model, (e) G2R registration using “most representative” image as reference with SDE model, and (f) IRG registration with SDE model. Bottom row contains standard deviation of FA images after registrations in the same order as in the top row.

Table 2.

Average standard deviation of deformed FA maps in white matter and whole brain under different registration methods.

Ave std dev Affine FFD SDE
std ref sel ref std ref sel ref IRG
white matter ROI .09690 .05997 .06450 . 05278 .05306 .05006
whole Brain .06685 .04659 .04860 .04125 .04147 .03887

Note: “std ref” refers to the G2R registration using the standard FA image provided by FSL as the reference; “sel ref” refers to the G2R registration selecting the “most representative”image as the reference. The number of degrees of freedom used in SDE model is 3 × 13 × 13 × 13, whereas it is 3 × 9 × 10 × 8 in the FFD model with default parameter setup in FSL.

fMRI data sets

The functional activation maps are shown in Fig. 8. Compared with affine alignment, the non-rigid elastic registration using both G2R and IRG registration methods provide larger β values in the left and right sensorimotor cortices and visual cortex. Fig.8(d) contains the log-Jacobian map of a typical transformation from one image to the implicit reference. The log Jacobian map shows that the IRG method generated large deformations in some locations indicating that affine registration for normalizing fMRI data may not sufficiently account for local variation.

Figure 8.

Figure 8

Functional activation maps with different registration methods: (a) affine alignment; (b) G2R registration with one of the group images as reference; (c) IRG registration; (d) log-Jacobian of a transformation from one image to the implicit reference. All activation maps are overlaid on the average of the deformed EPI data sets using IRG registration. (a), (b) and (c) are color coded by β values.

Fig.9(a) shows the ASD of the image intensity in the whole brain and in each ROI under different registration methods. The G2R registration reduced the ASD 30 percent, on average, in the whole brain compared to the affine alignment. The IRG registration further reduced the ASD approximately 2 percent compared to the average performance of the G2R registration in the whole brain. The IRG registration provides a consistent improvement in terms of smaller ASD in the four ROIs as well. Because fMRI data have lower resolution and less structural information compared to structural data such as T1-weighted and DTI images, the improvement of IRG registration using fMRI data is smaller than that using structural data.

Figure 9.

Figure 9

The average of (a) intensity standard deviation, (b) the β weights, and (c) t-statistics on different regions. In each panel, the first bar displays the measurement using affine alignment, the second bar shows the average measurements using 29 G2R registrations (with box plot on the top showing the median, the 5th, 25th, 75th, and 95th percentiles and outliers) and the third bar displays the measurement using IRG registration.

Fig.9(b) and (c) show the average of β weights and t-statistics in the four ROIs under different registration methods. Compared with the affine alignment, the elastic registration methods increased the average β and the t-statistics in each ROI. The IRG registration further increased the β values in the regions of right and left sensorimotor cortices and visual cortex, and the t-statistics in all ROIs compared with the G2R registration.

Discussion

Analysis of reference-based and implicit reference registration errors

The registration results of T1-weighted anatomical images, DTI and fMRI images in this study consistently showed that the implicit reference registration method produced smaller registration errors. This can be explained by the following theoretical analysis. Starting with the registration of two images, the similarity objective function for the implicit reference method is defined as D(I1(h1R), I2(h2R)), and the similarity cost for the reference-based method can be written as D(I1(h12), I2), which is equal to D(I1(h12), I2(hid)), where hid is the identity map from I2 to I2. The optimization of the former cost function searches the optimal transformations h1R and h2R to minimize the similarity cost between I1(h1R) and I2(h2R). The optimization of the latter cost function searches the optimal transformation h12 that minimizes the similarity cost between I1(h12) and I2(hid). Therefore, the minimizer of the latter can be considered as a local minimizer of the former, which means that the similarity error between I1(h12) and I2 is no smaller than the error between I1(h1R) and I2(h2R). Similar analysis can be extended to the registration of more than two images.

Comparison of PW, IRP, and IRG registration

The results in Fig.6 suggest that IRP registration produces better pair-wise registration than using the PW approach with respect to relative overlap. To see why this may be the case, consider that the IRP transformation from Ii to Ij is produced by composing the two transformations hiR with hRj=hjR1, where the implicit reference IR is roughly halfway between Ii and Ij. One reason for the better performance could be that there is less error for smaller deformations. In this case, hiR and hRj have less error combined than the transformation hij since both hiR and hRj have roughly half the shape differences to accommodate compared hij. Another reason for the better performance could be that estimating hiR and hjR solves the registration problem from both directions as opposed to one direction. Solving the registration from both directions could help reduce the number of local minima and may reduce the impact of getting stuck in a local minima since only half the distance between the shapes in the two images is accommodated.

When considering more than two images, the implicit reference is similar to the average of all images, and its shape is likely to be different from the average shape of Ii and Ij. As the population average diverges in shape from the average shape of Ii and Ij, the registration performance of the IRG transformations from Ii to Ij may be similar or even worse than direct pair-wise registration. The registration performance is similar if Ii or Ij has similar shape as the mean and worse when the population average is very different from the average shape of Ii and Ij. To see this, assume that Ii is similar in shape to the population mean, then the transformation from Ii to the reference is close to the identity mapping and the transformation from the reference to Ij will be close to the direct pair-wise registration between Ii and Ij. Now, assume that the population average is different than the average shape of Ii and Ij and that the registration error is proportional to the shape differences. In this case, the error in registering Ii to the reference is added to the error registering the reference to Ij, which could have more error than directly registering Ii to Ij. However, in terms of computation cost, the pair-wise method estimates and stores N(N - 1) transformations, while the IRG methods only computes N transformations.

An experiment using the 16 images from the NA0 NIREP database was used to test the hypothesis that the IRG performance of the composed transformations hiR(hjR1) is affected by the distance between the reference and the average of the two images. The distance of ROIk of two images was defined to be 1RO(ROI1k,ROI2k) where 0 means perfect overlap and 1 means no overlap. We also made a crude simplification that for ROIk, the distance between IR and the average of I1 and I2 to be 112(RO(ROIRk,ROI1k)+RO(ROIRk,ROI2k)). Assume the registration input has three images. To reduce computations, one image was fixed from the 16 images, and two more images were randomly chosen from the rest of the images. A total of (152) registrations were computed. In each registration, we computed the distance between IR and the average shape of Ii and Ij, and the distance between Ii and deformed Ij which measures the performance of hiR(hjR1). After least square linear regression, the β coefficient between the two distances is 0.175 with p < 0.001, indicating that the accuracy of hiR(hjR1) is correlated to the distance between the reference image and the average of Ii and Ij.

Inverse consistency and transitivity

A set of transformations that define correspondences between a group of images should satisfy desirable properties such as inverse consistency and transitivity (Christensen and Johnson, 2003; Geng, 2007). Based on the implicit reference registration method, the transformations between any two images Ii and Ij can be calculated as

hij=hiRhjR1,andhji=hjRhiR1.

Theoretically, hij=hji1 by construction for the implicit reference registration method implying no inverse consistency error. However, in practice there are small errors due to computing the inverse and composing the transformations such that the implicit reference methods have some small inverse consistency error.

Theoretically, the IRG registration has no transitivity error. To see this, let Ii, Ij, and Ik represent three images in the group such that

hij=hiRhjR1,hjk=hjRhkR1,andhik=hiRhkR1.

Combining equations gives

hikhkj=hiRhkR1(hkRhjR1)=hiRhjR1=hij

which shows theoretically no transitivity error. In practice, the IRG registration has transitivity error due to errors in inverting the transformations and composing the transformations. Results in Fig. 6(b) show that the transitivity error of IRG is less than 0.04 and much less than other registration methods.

Inverse consistency and transitivity are necessary but not sufficient properties for a set of transformations to define meaningful correspondences between images in a group. For example, a set of identity transformations have zero inverse consistency and transitivity errors but they provide meaningless correspondences. Likewise, a set of transformations can be constructed to have near zero inverse consistency and transitivity error using the method described in Skrinjar and Tagare (2004). The drawback of their approach is that correspondence errors present in the original transformations are still present in the newly constructed set of transformations.

Improvement of group analysis

The motivation of this work was to find a registration framework that estimates “good” correspondences between a group of images and improves the sensitivity of group analysis. The result that the IRG registration reduces across-subject variation of FA images suggests that the sensitivity in detecting white matter alterations between populations, as reflected by FA changes at a group level, should be improved by more accurate registration methods. When the IRG registration was applied to fMRI data sets, the increase of the regression coefficient β and t-statistics reflected the enhancement of the functional signal contrast (signal intensity during task vs. resting) and the detection power respectively. Since the registration method only deforms the shape of the input images but does not change the signal intensity, the resulting larger signal contrast can be attributed to a better alignment of the fMRI images across subjects in the group. Based on the assumption that brain structures correspond well to specific functions, it is reasonable to expect that stronger activation signals at a group level should be observed when the corresponding structures are aligned more accurately across all subjects in the group. The stronger t-statistics on the activated regions may be due to the larger average β values or the smaller group variance of β values. The smaller β variance within the group may be caused by the better alignment of the group subjects. The IRG registration provides the potential to improve group analyses of DTI and fMRI data in terms of better sensitivity and detectability.

The IRG registration in the current work assumes that the intensity values at each voxel follow a fixed Gaussian distribution with equal mean and variance across the population. This assumption may not always hold since the distribution of tissues at a particular voxel is not always Gaussian and the intensities at a voxel may not follow the same distribution across the population. Further improvement may be achieved if the intensity distribution used by the similarity term is characterized from the population, i.e., using the sum of voxel-wise entropies (Zollei et al., 2005) as the similarity term. Moreover, this modification will allow registering images from different modalities as well.

Similarity Cost Function Comparison

The similarity cost function in Eq. 1 is similar to the similarity cost CAtlas=iΩIi(hi(x))I(x)2dx with I(x)=1NiIi(hi(x)) used in the unbiased atlas construction method developed by Joshi et al. (2004). At first glance the cost functions appear to be different since Eq. 1 computes the difference between pairs of images while the cost function CAtlas computes the difference between each image and the mean image I(x). However, CAtlas can be manipulated in the following manner:

CAtlas=iΩIi(hiR(x))1Nj(Ij(hjR(x)))2dx=iΩ1N(NIi(hiR(x))jIj(hjR(x)))2dx=1N2iΩj,ij(Ii(hiR(x))Ij(hjR(x)))2dx=2N2iΩj,i<j(Ii(hiR(x))Ij(hjR(x)))2dx. (7)

Notice that the similarity cost function in Eq. 7 differs from the cost function in Eq. 1 only in the order of the square and summation operators and the constant 2/N2. This implies that the gradients of the cost functions are the same (up to transformation model), thus the gradient descent optimizer will result in the same minimization. Another difference between the registration algorithm used in this work and that in Joshi et al. (2004) is due to the transformation parametrization and how the transformations were regularized. In this work the transformations were parameterized using a Fourier series and regularized using a small deformation linear elastic model where as the work presented in Joshi et al. (2004) used a large deformation model.

The choice of using a small or large deformation model depends on the application. In general, a large deformation registration approach provides more degrees of freedom for registration than a small deformation model at the cost of increased computation and parameters. Increasing the model complexity and computational requirements may not always be needed to answer a particular scientific question.

Conclusions

An implicit reference group-wise (IRG) registration method was used in which transformations from each image in the group to a hidden reference were estimated using a small deformation linear elastic registration model. The performance of pair-wise and group-wise registration algorithms was evaluated for spatial normalization of T1-weighted MRI from NIREP, DTI and fMRI data. Results with the NA0 evaluation database from NIREP show that implicit reference pair-wise (IRP) registration produces better relative overlap (RO) than implicit reference group-wise (IRG) for pair-wise registration. However, IRP-G2R registration was shown to have a significantly larger transitivity error than IRG registration, i.e., IRP-G2R registration had significant inconsistencies between correspondences defined by different pair-wise transformations compared to IRG registration. In contrast, IRG registration produced consistent correspondence between images in a group at the cost of slightly reduced pair-wise RO accuracy compared to IRP-G2R. Thus, it is better to use IRP registration if the task is to register two images, but it is better to use IRG registration if the task is to register a group of images. Furthermore, the IRP registration should be favored over PW registration when registering two images. IRG spatial normalization of the fractional anisotropy (FA) maps of DTI were shown to have smaller FA variance compared with G2R methods using the same elastic registration model. Analyses of fMRI data sets with sensorimotor and visual tasks showed that IRG registration, on average, increases the statistical detectability of brain activation compared to G2R registration.

Acknowledgments

This work was supported in part by the Intramural Research Program of the National Institute on Drug Abuse (NIDA), National Institute of Health (NIH) and NIH grant EB004126.

Footnotes

Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

References

  1. Allen JS, Damasio H, Grabowski TJ. Normal neuroanatomical variation in the human brain: an MRI-volumetric study. Am J Phys Anthropol. 2002;118:341–58. doi: 10.1002/ajpa.10092. [DOI] [PubMed] [Google Scholar]
  2. Allen JS, Damasio H, Grabowski TJ, Bruss J, Zhang W. Sexual dimorphism and asymmetries in the gray white composition of the human cerebrum. NeuroImage. 2003;18:880–894. doi: 10.1016/s1053-8119(03)00034-x. [DOI] [PubMed] [Google Scholar]
  3. Avants B, Gee JC. Geodesic estimation for large deformation anatomical shape averaging and interpolation. NeuroImage. 2004;23:S139–S150. doi: 10.1016/j.neuroimage.2004.07.010. [DOI] [PubMed] [Google Scholar]
  4. Avants BB, Epstein CL, Grossman M, Gee JC. Symmetric diffeomorphic image registration with cross-correlation: Evaluating automated labeling of elderly and neurodegenerative brain. Medical Image Analysis. 2008;12(1):26–41. doi: 10.1016/j.media.2007.06.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  5. Basser PJ, Pierpaoli C. Microstructural and physiological features of tissues elucidated by quantitative-diffusion-tensor MRI. Journal of Magnetic Resonance Series B. 1996;111:209–219. doi: 10.1006/jmrb.1996.0086. [DOI] [PubMed] [Google Scholar]
  6. Beg MF, Khan A. Symmetric data attachment terms for large deformation image registration. IEEE Trans Med Imaging. 2007;26(9):1179–1189. doi: 10.1109/TMI.2007.898813. [DOI] [PubMed] [Google Scholar]
  7. Bhatia KK, Hajnal JV, Puri BK, Edwards AD, Rueckert D. Consistent groupwise non-rigid registration for atlas construction. Biomedical Imaging: Macro to Nano; IEEE International Symposium on; 2004. pp. 908–911. [Google Scholar]
  8. Christensen GE, Geng X, Kuhl JG, Bruss J, Grabowski TJ, Allen JS, Pirwani IA, Vannier MW, Damasio H. Introduction to the non-rigid image registration evaluation project (NIREP). 3rd International Workshop on Biomedical Image Registration. LNCS 4057; Springer-Verlag; 2006a. pp. 128–135. [Google Scholar]
  9. Christensen GE, Johnson HJ. Consistent image registration. IEEE Trans Med Imaging. 2001;20(7):568–582. doi: 10.1109/42.932742. [DOI] [PubMed] [Google Scholar]
  10. Christensen GE, Johnson HJ. Invertibility and transitivity analysis for nonrigid image registration. Journal of Electronic Imaging. 2003;12(1):106–117. [Google Scholar]
  11. Christensen GE, Johnson HJ, Vannier MW. Synthesizing average 3D anatomical shapes. NeuroImage. 2006b;32:146–158. doi: 10.1016/j.neuroimage.2006.03.018. [DOI] [PubMed] [Google Scholar]
  12. Christensen GE, Joshi SC, Miller MI. Volumetric transformation of brain anatomy. IEEE Trans on Med Imaging. 1997 December;16(6):864–877. doi: 10.1109/42.650882. [DOI] [PubMed] [Google Scholar]
  13. Cootes T, Beeston C, Edwards G, Taylor C. A unified framework for atlas matching using active appearance models. In: Kuba A, Samal M, editors. Information Processing in Medical Imaging. LNCS 1613. Springer-Verlag; Berlin: 1999. pp. 322–333. [Google Scholar]
  14. Cootes T, Taylor C, Cooper D, Graham J. Active shape models—their training and application. Computer Vision and Image Understanding. 1995;61(1):38–59. [Google Scholar]
  15. Cox RW. AFNI: Software for analysis and visualization of functional magnetic resonance neuroimages. Computers and Biomedical Research. 1996;29:162–173. doi: 10.1006/cbmr.1996.0014. [DOI] [PubMed] [Google Scholar]
  16. Crum WR, Hartkens T, Hill DLG. Non-rigid image registration: theory and practice. The British Journal of Radiology. 2004;77:140–153. doi: 10.1259/bjr/25329214. [DOI] [PubMed] [Google Scholar]
  17. Damasio H. Human brain anatomy in computerized images. 2nd. Oxford University Press; New York: 2005. [Google Scholar]
  18. Frangi AF, Rueckert D, Schnabel JA, Niessen WJ. Automatic construction of multiple-object three-dimensional statistical shape models: Application to cardiac modeling. IEEE Trans Med Imaging. 2002;21(9):1151–1166. doi: 10.1109/TMI.2002.804426. [DOI] [PubMed] [Google Scholar]
  19. Frank R, Damasio H, Grabowski T. Brainvox: an interactive, multimodel, visualization and analysis system for neuroanatomical imaging. NeuroImage. 1997;5:13–30. doi: 10.1006/nimg.1996.0250. [DOI] [PubMed] [Google Scholar]
  20. Geng X. Ph D thesis. Department of Electrical and Computer Engineering, The University of Iowa; Iowa City, IA: 2007. Transitive inverse-consistent image registration and evaluation. 52242. [Google Scholar]
  21. Grabowski T, Frank R, Brown NSC, Damasio H. Validation of partial tissue segmentation of single-channel magnetic resonance images of the brain. NeuroImage. 2000;12:640–656. doi: 10.1006/nimg.2000.0649. [DOI] [PubMed] [Google Scholar]
  22. Guimond A, Meunier J, Thirion J. Average brain models: A convergence study. Computer Vision and Image Understanding. 2000;77(2):192–210. [Google Scholar]
  23. Jenkinson M, Smith S. A global optimisation method for robust affine registration of brain images. Medical Image Analysis. 2001;5(2):143–156. doi: 10.1016/s1361-8415(01)00036-6. [DOI] [PubMed] [Google Scholar]
  24. Joshi S, Davis B, Jomier M, Gerig G. Unbiased diffeomorphic atlas construction for computational anatomy. NeuroImage. 2004;23:S151–S160. doi: 10.1016/j.neuroimage.2004.07.068. [DOI] [PubMed] [Google Scholar]
  25. Leow AD, Yanovsky I, Chiang MC, Lee AD, Klunder AD, Lu A, Becker JT, Davis SW, Toga AW, Thompson PM. Statistical properties of jacobian maps and the realization of unbiased large-deformation nonlinear image registration. IEEE Trans Med Imaging. 2007;26 doi: 10.1109/TMI.2007.892646. [DOI] [PubMed] [Google Scholar]
  26. Mega MS, Dinov ID, Mazziotta JC, Manese M, Thompson PM, Lindshield C, Moussai J, Tran N, Olsen K, Zoumalan CI, Woods RP, Toga AW. Automated brain tissue assessment in the elderly and demented population: Construction and validation of a sub-volume probabilistic brain atlas. NeuroImage. 2005;26:1009–1018. doi: 10.1016/j.neuroimage.2005.03.031. [DOI] [PubMed] [Google Scholar]
  27. Miller MI, Beg MF, Ceritoglu C, Stark C. Increasing the power of functional maps of the medial temporal lobe by using large deformation diffeomorphic metric mapping. Proceedings of the National Academy of Sciences. 2005;102(27):9685–9690. doi: 10.1073/pnas.0503892102. [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Ogawa S, Lee T, Nayak A, Glynn P. Oxygenation-sensitive contrast in magnetic resonance image of rodent brain at high magnetic fields. Magnetic Resonance in Medicine. 1990;14(1):68–78. doi: 10.1002/mrm.1910140108. [DOI] [PubMed] [Google Scholar]
  29. Pieperhoff P, Homke L, Schneider F, Habel U, Shah NJ, Zilles K, Amunts K. Deformation field morphometry reveals age-related structural differences between the brains of adults up to 51 years. Journal of Neuroscience. 2008;28(4):828–842. doi: 10.1523/JNEUROSCI.3732-07.2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  30. Rueckert D, Frangi A, Schnabel J. Automatic construction of 3-D statistical deformation models of the brain using nonrigid registration. IEEE Transactions on Medical Imaging. 2003;22(8):1014–1025. doi: 10.1109/TMI.2003.815865. [DOI] [PubMed] [Google Scholar]
  31. Rueckert D, Sonoda L, Hayes C, Hill D, Leach M, Hawkes D. Nonrigid registration using free-form deformations: application to breast MR images. IEEE Transactions on Medical Imaging. 1999;18(8):712–721. doi: 10.1109/42.796284. [DOI] [PubMed] [Google Scholar]
  32. Skrinjar OM, Tagare H. Symmetric, transitive, geometric deformation and intensity variation invariant nonrigid image registration. Proceedings of the 2004 IEEE International Symposium on Biomedical Imaging: From Nano to Macro. 2004;1:920–923. [Google Scholar]
  33. Smith SM, Jenkinson M, Woolrich MW, Beckmann CF, Behrens TE, Johansen-Berg H, Bannister PR, Luca MD, Drobnjak I, Flitney DE, Niazy RK, Saunders J, Vickers J, Zhang Y, Stefano ND, Brady JM, Matthews PM. Advances in functional and structural MR image analysis and implementation as FSL. NeuroImage. 2004;23(S1):208–219. doi: 10.1016/j.neuroimage.2004.07.051. [DOI] [PubMed] [Google Scholar]
  34. Studholme C. Simultaneous population based image alignment for template free spatial normalisation of brain anatomy. In: Maintz T, Vannier M, Gee J, editors. 2nd International Workshop on Biomedical Image Registration. LNCS 2717. Springer-Verlag; Berlin: 2003. pp. 81–90. [Google Scholar]
  35. Studholme C, Cardenas V. A template free approach to volumetric spatial normalization of brain anatomy. Pattern Recognition Letters. 2004;25:1191–1202. [Google Scholar]
  36. Twining CJ, Cootes T, Marsland S, Petrovic V, Schestowitz R, Taylor CJ. Information Processing in Medical Imaging. Springer-Verlag; Berlin: 2005. A unified information-theoretic approach to groupwise non-rigid registration and model building; pp. 1–13. [DOI] [PubMed] [Google Scholar]
  37. Wang F, Vemuri BC, Rangarajan A. Proceedings CVPR 2006. IEEE Computer Society Conference on Computer Vision and Pattern Recognition. IEEE Computer Society; 2006. Groupwise point pattern registration using a novel CDF-based Jensen-Shannon divergence; pp. 1283–1288. [DOI] [PMC free article] [PubMed] [Google Scholar]
  38. Zhang H, Avants BB, Yushkevich PA, Woo JH, Wang S, McCluskey LF, Elman LB, Melhem ER, Gee JC. High-dimensional spatial normalization of diffusion tensor images improves the detection of white matter differences: An example study using amyotrophic lateral sclerosis. IEEE Trans Med Imaging. 2007;26 doi: 10.1109/TMI.2007.906784. [DOI] [PubMed] [Google Scholar]
  39. Zollei L, Learned-Miller E, Grimson E, Wells W. Efficient population registration of 3D data. ICCV 2005, Computer Vision for Biomedical Image Applications; Beijing. 2005. [Google Scholar]

RESOURCES