Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2011 Jan 11.
Published in final edited form as: Med Image Comput Comput Assist Interv. 2010;13(Pt 2):684–691. doi: 10.1007/978-3-642-15745-5_84

Groupwise Registration by Hierarchical Anatomical Correspondence Detection

Guorong Wu 1, Qian Wang 1,2, Hongjun Jia 1, Dinggang Shen 1
PMCID: PMC3018804  NIHMSID: NIHMS217225  PMID: 20879375

Abstract

We present a novel feature-based groupwise registration method to simultaneously warp the subjects towards the common space. Due to the complexity of the groupwise registration, we resort to decoupling it into two easy-to-solve tasks, i.e., alternatively establishing the robust correspondences across different subjects and interpolating the dense deformation fields based on the detected sparse correspondences. Specifically, several novel strategies are proposed in the correspondence detection step. First, attribute vector, instead of intensity only, is used as a morphological signature to guide the anatomical correspondence detection among all subjects. Second, we detect correspondence only on the driving voxels with distinctive attribute vectors for avoiding the ambiguity in detecting correspondences for non-distinctive voxels. Third, soft correspondence assignment (allowing for adaptive detection of multiple correspondences in each subject) is also presented to help establish reliable correspondences across all subjects, which is particularly necessary in the beginning of groupwise registration. Based on the sparse correspondences detected on the driving voxels of each subject, thin-plate splines (TPS) are then used to propagate the correspondences on the driving voxels to the entire brain image for estimating the dense transformation for each subject. By iteratively repeating correspondence detection and dense transformation estimation, all the subjects will be aligned onto a common space simultaneously. Our groupwise registration algorithm has been extensively evaluated by 18 elderly brains, 16 NIREP, and 40 LONI data. In all experiments, our algorithm achieves more robust and accurate registration results, compared to a groupwise registration method and a pairwise registration method, respectively.

1 Introduction

Registration of a population data has received more and more attention in recent years due to its importance in population analysis [1-4]. Since groupwise registration method is able to register all images without explicitly selecting the template, it can avoid bias in template selection and thus becomes attractive to the precise analysis of population data, compared to the pairwise registration methods. However, it is complicated for groupwise registration of multiple images simultaneously.

Although groupwise registration can be achieved by exhausting pairwise registrations between all possible subject combinations in the population [4], this type of method suffers from very heavy computation. Recently, more favorable approaches were proposed to align all subjects simultaneously by following the groupwise concept explicitly. Specifically, Joshi et al. [1] proposed to perform groupwise registration by iteratively (1) registering all subjects to the group mean image, and (2) constructing the group mean image as the Fréchet mean of all registered images. Also, Learned-Miller [3] proposed a congealing method to jointly warp the subjects towards a hidden common space by minimizing the sum of stack entropies in the population. Balci et al. [2] further extended the congealing method to non-rigid image registration by modeling the transformations with B-Splines. However, as the method itself is intensity-based, it is intrinsically insufficient to establish good anatomical correspondences across images. Furthermore, although the groupwise registration can be solved through steepest descent optimization [2, 3], it is unfortunately sensitive to local minima. Also, since the cost function is estimated based on only ~1% randomly sampled voxels (regardless of their morphological importance), the registration performance could be seriously affected.

To the best of our knowledge, the issue of anatomical correspondence in groupwise registration, which is very critical to measure the inter-subject difference, has not been well addressed in the literature. In this paper, we propose a novel feature-based groupwise registration method for achieving robust anatomical correspondence detection. Specifically, we formulate our groupwise registration by alternatively (1) estimating the sparse correspondences across all subjects and (2) interpolating the dense transformation field based on the established sparse correspondences.

In Step (1), we use attribute vector, instead of intensity only, as a morphological signature to help guide correspondence detection. Furthermore, the robustness of correspondence detection based on attribute vectors is achieved in two ways. First, we only detect correspondences for the most distinctive voxels, called as driving voxels, in the brain images, and then use their detected correspondences to guide the transformations of the nearby non-driving voxels. Second, multiple correspondences are allowed to alleviate the ambiguities particularly in the beginning of registration, and these one-to-many correspondences are gradually restricted to one-to-one correspondence with progress of registration in order to achieve accuracy for the final registration results. It is worth noting that this soft assignment strategy is also applied to all subjects in the population, where the contributions from different subjects are dynamically controlled through the registration. In Step (2), TPS is utilized to interpolate the dense transformation fields based on the sparse correspondences.

We have compared the performance of our groupwise registration with the congealing method [2] and the pairwise HAMMER registration algorithm [5, 6] by evaluation on 18 elderly brains, 16 NIREP dataset with 32 manually delineated ROIs, and 40 LONI dataset with 54 manually labeled ROIs. Experimental results show that our method can achieve the best performance.

2 Methods

Given a group of subjects S = {Si|i = 1 … N}, the goal of the groupwise registration is to find a set of transformations G = {gi|gi(x) = x + hi(x), x = (x1, x2, x3) ∈ ℜ3, i = 1, ⋯, N} that are able to transform each subject towards a hidden common space with its individual displacement hi(x). Fig. 1 schematically illustrates the concept of our groupwise registration. To connect each pair of subjects through the common space, the inverse transformation fields G1={gi1gi1(x)=x+hi1(x),x=(x1,x2,x3)3,i=1,,N} should be calculated as in Fig. 1. Thus the composite transformation gj1gi can be used to warp subject Si to Sj, and also gi1gj can be used to warp subject Sj to Si (as shown in the right panel of Fig. 1). In the following, we use i to index the subject under consideration and j for any other subject except Si. Also, we call G the set of forward transformations and G−1 the set of backward transformations, which enter and leave the common space, respectively. In the following, we will present the energy function in Section 2.1, and then provide a solution to groupwise registration in Section 2.2.

Fig. 1.

Fig. 1

The schematic illustration of the proposed groupwise registration algorithm. All subjects in the group are connected by the forward transformations gi (i.e., red solid arrows) to the common space (i.e., a purple circled region), and by the backward transformations gi1 (i.e., blue dashed arrows) coming from the common space. The right panel shows the composite transformations bridging subjects Si and Sj.

2.1 Energy Function in Groupwise Registration

As pointed in Fig. 1, all subjects will agglomerate to the hidden common space by following the simultaneously estimated transformation fields. To identify the anatomical correspondence among all subjects, we propose using attribute vector as a morphological signature of each voxel x for guiding the correspondence detection. Without loss of generality, the geometric moment invariants of white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF) are calculated from a neighborhood of each voxel x for defining its attribute vector a(x) [5]. Further, we hierarchically select distinctive voxels (with distinctive attribute vectors) as driving voxels in the image [5], by adaptively setting thresholds on the attribute vectors. Here, the driving voxels are represented as {xi,m|i = 1, …, N,m = 1, …, Mi}, where Mi is the number of the driving voxels in subject Si. In our groupwise registration, we establish the sparse correspondences only on the driving voxels due to their distinctiveness, and let these driving voxels steer the dense transformation, by considering them as control points in TPS-based interpolation [7].

The overall energy function of groupwise registration can be defined to minimize the differences of attribute vectors on the corresponding locations across different subjects and also preserve the smoothness of estimated transformation fields:

E(G)=Σi=1NΣxSiΣj=1,jiNai(x)aj(gj1°gi(x))2+Σi=1NLgi2, (1)

where L is an operator to compute the bending energy of transformation field gi.

However, directly optimizing E(G) is usually intractable. Thus, we introduce the sparse correspondence fields F = {fi(x)|x ∈ ℜ3, i = 1, …, N} in our method. Here, each fi is a set of correspondence vectors defined for subject Si: for each driving voxel xi,m in subject Si, it gives the latest estimated corresponding location (pointing to the common space), while for each non-driving voxel, it keeps the previous estimated transformation. As the result, the energy function in Eq. 1 becomes:

E(F,G)=Σi=1NΣxSi{Σj=1,jiNai(x)aj(gj1°fi(x))2+fi(x)gi(x)2}+Σi=1NLgi2, (2)

The advantage of introducing F is that it decouples the complicated optimization problem into two simple-to-solve sub-problems, i.e., alternatively (SP1): estimating the correspondence field fi via correspondence detection; and (SP2): interpolating the dense transformation gi with regularization on G.

Estimating the Correspondence Field F (SP1)

In this step, we take advantage of the driving voxels to establish the correspondence on each driving voxel xi,m of subject Si by inspecting each candidate in a neighborhood n1, w.r.t. each of other subjects Sj one by one. For evaluating each candidate, several useful strategies are employed here to achieve robust correspondences. First, not only the voxelwise but also the regionwise difference on attribute vectors is proposed by computing the distance of each pair of corresponding attribute vectors within a neighborhood n2. Second, multiple spatial correspondences are allowed on each driving voxel xi,m by introducing a spatial assignment πi,mj,v to indicate the likelihood of the true correspondence v w.r.t. subject Sj. Also, we use τi,mj to describe the likelihood of subject Sj being selected as a reference image for correspondence detection of xi,m of subject Si, according to the similarity of local morphology between Si and Sj.

Therefore, by fixing G in Eq. 2, the new energy function E1 in this step can be defined as:

E1(π,τ,F)=Σi=1NΣm=1MiEcorr(πi,m,τi,m,fi)+Σi=1NΣm=1MiEfuzzy(πi,m,τi,m)+Σi=1NΣm=1Mifi(xi,m)gi(xi,m)2Ecorr(πi,m,τi,m,fi)=Σj=1Nτi,mjΣvn1(gi(xi,m))πi,mj,v[gi(xi,m)v2+d(ai(xi,m),fi,gj1)]Efuzy(πi,m,τi,m)=r1πi,mj,vlogπi,mj,v+r2τi,mjlogτi,mj (3)

where d(ai(xi,m),fi,gj1)=Σzn2(xi,m)a(z)a(gj1(fi(z)))2 measures the regionwise difference of attribute vector ai(xi,m) and its corresponding counterpart in Sj w.r.t. the current estimated correspondence fi and the previous obtained inverse transformation field gj1. There are totally three terms in the energy function E1. The first term Ecorr measures the matching discrepancy on each driving voxel xi,m, where the criteria in evaluating the candidate v w.r.t. subject Sj are: 1) the spatial distance between gi(xi,m) and v in the common space should be as small as possible according to the ICP principle [8]; 2) not only the candidate location v but also its neighborhood should have the similar attribute vectors based on the measurement d(ai(xi,m),fi,gj1).

Soft assignment is very important for brain image registration to reduce the risk of mismatching, particularly in the beginning of registration. All voxels in the search neighborhood n1 have the chance to become correspondence candidate, but their contributions to the true correspondence vary according to the matching criteria. To increase the registration accuracy and specificity, it is also necessary to evolve to oneto-one correspondence in the end of registration. Therefore, the second term Efuzzy(πi,m,τi,m) is used to dynamically control the soft assignment by requiring the entropy of πi,m and τi,m gradually to decrease with progress of registration.

The third term in E1 ensures that the correspondence field F be close to the previous estimated transformation field G, by minimizing the difference between each pairs of fi and gi.

Interpolating the Dense Transformation Field G (SP2)

After updating the correspondence in each driving voxel, the energy function in this step is given as:

E2(G)=Σi=1NΣm=1Mi{fi(xi,m)gi(xi,m)2}+Σi=1NLgi2. (4)

By regarding the driving voxels as control points in each subject, TPS interpolation can be used to estimate the optimal gi that fits the transformation on xi,m to fi(xi,m) and reaches the minimal bending energy (the second term) [7,9].

2.2 Implementation for Groupwise Registration

In our method, we alternatively optimize SP1 and SP2 in each round of registration. In SP1, the explicit solutions of πi,mj,v and τi,mj are obtained by letting E1πi,mj,v=0 and E1τi,mj=0:

E1τi,mj=0:πi,mj,v=c1exp(ξvr1),πi,mj,vπi,mj,vΣvn2(gi(xi,m))πi,mj,v (5)
πi,mj=c2exp(Σvn2(gi(xi,m))πi,mj,vξvr2),τi,mjτi,mj/Σj=1Nτi,mj (6)

where ξv=[gi(xi,m)v2+d(ai(xi,m),fi,gj1)] denotes the bracket part in Ecorr(πi,m, τi,m, fi), and c1 and c2 are constants. After πi,mj,v and τi,mj are determined by Eqs. 5 and 6 on each xi,m, the correspondence fi(xi,m) is updated by moving it to the mean location of all candidates under the guidance of πi,mj,vτi,mj:

f^i(xi,m)=Σj=1NΣvn2(gi(xi,m))τi,mjπi,mj,vv. (7)

It is worth noting that we introduce a multi-resolution strategy to implement our proposed groupwise registration method for fast and robust registration. Specifically, in each resolution, the size of the search neighborhood n1 decreases gradually with the progress of registration, for achieving more specific detection of correspondences. Moreover, in the initial state of registration, only a small set of voxels with distinctive features, such as those locating at ventricular boundaries, sulcal roots and gyral crowns, are selected as driving voxels. After that, more and more driving voxels are added to drive the registration and eventually all voxels in the brains join the groupwise registration.

By taking the driving voxels as control points, TPS is used to interpolate the dense transformation field after we repeat the calculations of Eqs. 5~7 for each driving voxel xi,m. To avoid the cumbersome inversion of large matrix in TPS (proportional to the number of control points), we perform TPS interpolation in overlapping blocks (32 × 32 × 32) and also down-sample the driving voxels in each block.

3 Experiments

In our experiments, we have extensively evaluated the performances of our groupwise registration method in atlas building and ROI labeling. For comparison, we use the congealing groupwise registration method [2] with its available codes, http://www.insight-journal.org/browse/publication/173. To demonstrate the advantage of groupwise registration over pairwise registration, the registration results by a pairwise registration method, namely HAMMER [5], are also provided.

To demonstrate the group overlap of labeled brain regions after registration, we specifically vote a reference by assigning each voxel with a tissue label that is the majority of all tissue labels at the same location from all aligned subjects. Then, the overlap ratio between each of the registered label images and the voted reference can be calculated. Here, we use the Jaccard Coefficient metric as the overlap ratio to measure the alignment of the two regions (A and B) with the same label, defined as:

J(A,B)=|AB||AB|. (8)

18 Elderly Brain Images

18 elderly brain images, each with 256 × 256 × 124 voxels and the resolution of 0.9375 × 0.9375 × 1.5mm3, are used in this experiment. The group mean images produced by the congealing method and our groupwise registration method are both displayed in Fig. 2. Through visual inspection, the group mean of our method is sharper and gains better contrast (especially around ventricles) than that of the congealing method, indicating better performance of our registration method. The overlap ratios on WM, GM, and VN, as well as the overall overlap ratio on the whole brain, by our method and the congealing method, are provided in Table 1. It can be observed that our method achieves better results than the congealing method in each tissue type. On the other hand, it is interesting to compare the performance between groupwise and pairwise registrations of these 18 brain images. In HAMMER-based pairwise registration, 5 out of 18 subjects are randomly selected as the templates to align all other 17 remaining subjects. The average overlap ratios produced by these five different templates, as well as the standard deviations, are shown in the first row of Table 1, which verify again the power of our groupwise registration in consistently registering the population data.

Fig. 2.

Fig. 2

The groupwise registration results by the congealing method and our method. It can be observed that our group mean image is much sharper than that by the congealing method, indicating a more accurate and consistent registration by our method.

Table 1.

Overall overlap ratios of WM, GM, and VN by pairwise HAMMER algorithm, congealing method, and our groupwise registration method.

WM GM VN Overall
Pairwise HAMMER 63.86% (±3.87%) 57.25% (±2.18%) 76.51% (±3.70%) 65.64% (±3.15%)
Congealing Method 59.68% 51.09% 70.61% 59.43%
Our Method 75.81% 63.61% 81.16% 73.52%

NIREP Data and LONI Data

In this experiment, we employ pairwise HAMMER, congealing method, and our groupwise registration method to align 16 NIREP data (with 32 manual ROIs) and LONI40 dataset (with 54 manual ROIs). Table 2 shows the average overlap ratios on these two datasets by the three registration methods. Obviously, our groupwise registration method achieves the most accurate registration results among all three registration methods. In particular, Fig. 3 shows the performance of registration accuracy at the left and the right precentral gyri of 16 NIREP brains by the two groupwise methods. The average overlap ratio is 45.95% by congealing and 57.34% by our method. The brighter color indicates the higher consistency of registration across different subjects, while the darker color means the poor alignment. Again, our method achieves much better alignment.

Table 2.

Overall overlap ratios of the aligned ROIs in NIREP and LONI datasets by pairwise HAMMER algorithm, congealing method, and our groupwise registraion method.

Pairwise HAMMER Congealing Method Our Method
NIREP (32 ROIs) 56.58% 52.07% 61.52%
LONI (54 ROIs) 54.12% 60.60% 67.02%

Fig. 3.

Fig. 3

3D renderings of the aligned left and right precentral gyri by the congealing method and our groupwise registration method.

4 Conclusion

In this paper, we have presented a new feature-guided groupwise registration method and also demonstrated its applications in atlas building and population data analysis. Specifically, by taking advantage of the driving voxels (with distinctive features) automatically detected from all images, we develop a feature-based groupwise registration method by alternatively estimating the correspondences on the driving voxels and updating the dense transformation fields by TPS. Extensive experiments have been performed to compare the performance of our method with that of the congealing method and the pairwise HAMMER algorithm. All experimental results show that our method can achieve the best performance.

5 References

  • 1.Joshi S, Davis B, Jomier M, Gerig G. Unbiased diffeomorphic atlas construction for computational anatomy. NeuroImage. 2004;23:151–160. doi: 10.1016/j.neuroimage.2004.07.068. [DOI] [PubMed] [Google Scholar]
  • 2.Balci SK, Golland P, Shenton M, Wells WM. Free-Form B-spline Deformation Model for Groupwise Registration; Workshop on Open-Source and Open-Data for 10th MICCAI; 2007. pp. 105–121. [Google Scholar]
  • 3.Learned-Miller EG. Data driven image models through continuous joint alignment. Pattern Analysis and Machine Intelligence, IEEE Transactions on. 2006;28:236–250. doi: 10.1109/TPAMI.2006.34. [DOI] [PubMed] [Google Scholar]
  • 4.Shattuck DW, Mirza M, Adisetiyo V, Hojatkashani C, Salamon G, Narr KL, Poldrack RA, Bilder RM, Toga AW. Construction of a 3D probabilistic atlas of human cortical structures. NeuroImage. 2008;39(3):1064–1080. doi: 10.1016/j.neuroimage.2007.09.031. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Shen D, Davatzikos C. HAMMER: Hierarchical attribute matching mechanism for elastic registration. IEEE Transactions on Medical Imaging. 2002;21(11):1421–1439. doi: 10.1109/TMI.2002.803111. [DOI] [PubMed] [Google Scholar]
  • 6.Shen D, Davatzikos C. Very high resolution morphometry using mass-preserving deformations and HAMMER elastic registration. NeuroImage. 2003;18(1):28–41. doi: 10.1006/nimg.2002.1301. [DOI] [PubMed] [Google Scholar]
  • 7.Bookstein FL. Principal Warps: Thin-Plate Splines and the Decomposition of Deformations. IEEE Transactions on Pattern Analysis and Machine Intelligence. 1989;11(6):567–585. [Google Scholar]
  • 8.Besl P, McKay N. A Method for Registration of 3-D Shapes. I IEEE Transactions on Pattern Analysis and Machine Intelligence. 1992;14(2):239–256. [Google Scholar]
  • 9.Chui H, Rangarajan A. A new point matching algorithm for non-rigid registration. Computer Vision and Image Understanding. 2003;89:114–141. [Google Scholar]

RESOURCES