Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2020 Jun 17.
Published before final editing as: IEEE Trans Image Process. 2018 Dec 17:10.1109/TIP.2018.2884563. doi: 10.1109/TIP.2018.2884563

A New Multi-Atlas Registration Framework for Multimodal Pathological Images Using Conventional Monomodal Normal Atlases

Zhenyu Tang 1, Pew-Thian Yap 2, Dinggang Shen 3,*
PMCID: PMC6579720  NIHMSID: NIHMS1520720  PMID: 30571622

Abstract

Using multi-atlas registration (MAR), information carried by atlases can be transferred onto a new input image for the tasks of region of interest (ROI) segmentation, anatomical landmark detection, and so on. Conventional atlases used in MAR methods are monomodal and contain only normal anatomical structures. Therefore, the majority of MAR methods cannot handle input multimodal pathological images, which are often collected in routine image-based diagnosis. This is because registering monomodal atlases with normal appearances to multimodal pathological images involves two major problems: (1) missing imaging modalities in the monomodal atlases, and (2) influence from pathological regions. In this paper, we propose a new MAR framework to tackle these problems. In this framework, a deep learning based image synthesizers are applied for synthesizing multimodal normal atlases from conventional monomodal normal atlases. To reduce the influence from pathological regions, we further propose a multimodal low-rank approach to recover multimodal normal-looking images from multimodal pathological images. Finally, the multimodal normal atlases can be registered to the recovered multimodal images in a multi-channel way. We evaluate our MAR framework via brain ROI segmentation of multimodal tumor brain images. Due to the utilization of multimodal information and the reduced influence from pathological regions, experimental results show that registration based on our method is more accurate and robust, leading to significantly improved brain ROI segmentation compared with state-of-the-art methods.

Keywords: Image registration, multimodal image, pathological brain image, image synthesis, low-rank image recovery

I. Introduction

USING multi-atlas registration (MAR), information carried by atlases can be transferred onto a new input image for tasks such as segmentation [1]–[3], landmark detection [4], and surgical planning [5]. Atlases adopted in MAR methods are generally obtained from monomodal normal images (e.g., T1-weighted MR images of normal brains) [6]–[9]. Thus, the majority of MAR methods cannot be applied to multimodal pathological images. This is because registering monomodal normal atlases to multimodal pathological images involves two major problems: (1) missing imaging modalities in the monomodal atlases, and (2) influence from pathological regions because of lacking correspondence between pathological and normal regions. However, multimodal pathological images are often collected in routine image-based diagnosis. For example, T1-weighted, T1 contrast-enhanced (T1c), T2-weighted and FLAIR MR images are usually acquired for patients with brain tumors.

One way to deal with the problem of missing imaging modalities is by discarding the modalities that are not available in the monomodal atlases, but at the cost of losing useful information. An alternative is to use cross-modal similarity metrics such as mutual information (MI) [10] and normalized mutual information (NMI) [11] for cross-modality registration. However, unlike monomodal similarity metrics, such as the mean squared difference of image intensity, local anatomical similarity cannot be directly or efficiently calculated with MI or NMI [12].

To reduce the influence from pathological regions, Brett et al. [13], [14] proposed a cost function masking (CFM) strategy to exclude pathological regions from registration, and registration is driven only by normal regions. Gooya et al. [15], [16] synthesized pathological regions (i.e., tumors) in a normal brain atlas to make the brain atlas similar to the input tumor brain image, facilitating subsequent registration. Liu et al. [17], [18] used a method called Low-Rank plus Sparse matrix Decomposition (LRSD) [19] to recover normal-looking brain image from the input tumor brain image. Then typical image registration methods [20]–[24] can be used to register normal brain atlases to the recovered normal-looking image.

In this paper, we propose a new MAR framework to solve the two aforementioned problems in traditional MAR methods, which have not been well addressed in the literature. Our MAR framework can handle multimodal pathological images based on conventional monomodal normal atlases. In our framework, deep learning based image synthesizers are introduced to produce multimodal normal atlases from conventional monomodal normal atlases. Then based on the resulting multimodal normal atlases, a multimodal low-rank method is further proposed to recover a multimodal normal-looking image from the input multimodal pathological image. Finally, the multimodal normal atlases are registered to the recovered multimodal image in a multi-channel way, where each channel operates one modality guided by monomodal image similarity metric. In our MAR framework, the multimodal low-rank recovery and the multi-channel registration are iteratively refined until convergence. We demonstrate the effectiveness of our MAR framework via segmentation of brain region of interest (ROI) in multimodal tumor brain images. The experimental results indicate that segmentation is significantly improved compared with state-of-the-art methods.

II. Methods

Since brain tumor is a common brain disease, and multimodal MR tumor brain images are usually captured for patients with suspicious brain cancer syndromes, our MAR framework is formulated for multimodal MR tumor brain images. By default, each multimodal MR tumor brain image has four modalities: T1, T1c, T2 and FLAIR, denoted as T={TT1,TT1c,TT2,TFLAIR}. The monomodal brain atlases adopted in our framework are N T1-weighted MR normal brain images, denoted as Ai,T1, i = 1, ..., N. An overview of our framework is shown in Fig. 1.

Fig. 1.

Fig. 1.

Overview of our multi-atlas registration framework for multimodal pathological brain images.

Our MAR framework consists of three main components. The first component contains CycleGAN [25] based image synthesizers, which synthesize the brain atlases of missing modalities, i.e., Ai,T1c, Ai,T2 and Ai,FLAIR, from Ai,T1, i = 1, ..., N, resulting in the multimodal brain atlases Ai={Ai,T1,Ai,T1c,Ai,T2,Ai,FLAIR}, i = 1, ..., N. This component is needed to be executed only once. The second component uses a low-rank method called multimodal SCOLOR (Spatially COnstrained LOw-Rank), which is an improvement of our previously proposed method [26], to recover multimodal normal-looking image T^={T^T1,T^T1c,T^T2,T^FLAIR} from the input multimodal tumor brain images T based on the information provided by Ai, i = 1, ..., N. The third component involves multi-channel image registration of the multimodal brain atlases Ai, i = 1, ..., N to the recovered multimodal image T^, where each channel operates one modality using monomodal image similarity metric. The multimodal SCOLOR based image recovery and the multi-channel image registration components are iteratively proceeded to refine the image recovery and registration results until convergence (illustrated by the two blue arrows in Fig. 1). Details on each component are presented next.

A. CycleGAN based image synthesizers

We use CycleGAN [25] to get multimodal brain atlases by synthesizing missing modalities from available monomodal brain atlases. CycleGAN is a special kind of generative adversarial network (GAN). Conventional GAN [27] consists of a generator G : XY which is responsible for synthesizing image yY from image xX and a discriminator D which is responsible for discriminating the synthesized images G(x), xX, from the images in Y. G is typically a fully convolutional network (FCN) [28], and D is a convolutional neural network (CNN) [29]. Given a set of training data X = {x1, ..., xM} (from the reference modality) and Y = {y1, ..., yM} (from the desired modality), the objective function of GAN is

LGAN(G,D,X,Y)=minGmaxD(EyiY(log(D(yi)))+ExiX(log(1D(G(xi))))). (1)

To minimize (1), G should produce synthetic images G(x) that cannot be discriminated by D from images in Y. On the other hand, D tries to reject all synthetic images produced by G, i.e., D(G(x)) = 0. The optimization process consists of two steps. In the first step, the discriminator D is updated by stochastic gradient ascent

1Mi=1M(log(D(yi))+log(1D(G(xi)))). (2)

In the second step, the generator G is updated by stochastic gradient descent

1Mi=1M(log(1D(G(xi)))). (3)

These two steps are iterated until convergence, i.e., the discriminator D cannot differentiate the synthetic images G(x) from the images in Y. GAN requires paired training data to make G(xi) have consistent image content with xi. However, paired normal brain images of different modalities are usually unavailable. Therefore, we use CycleGAN, which only require unpaired training data. The CycleGAN can be regarded as a combination of two GANs. Given a set of unpaired training data X˜={x˜1,,x˜M} of the reference modality and Y˜={y˜1,,y˜M} of the desired modality, One GAN (G1 and D1) is created for X˜Y˜, and the other GAN (G2 and D2) is created for Y˜X˜. The objective function of CycleGAN is defined as

LCycGAN(G1,G2,D1,D2,X˜,Y˜)=minG1,G2maxD1,D2(LGAN(G1,D1,X˜,Y˜)+LGAN(G2,D2,Y˜,X˜)+Lcyc(G1,G2)) (4)

where LGAN(G1,D1,X˜,Y˜) and LGAN(G2,D2,Y˜,X˜) are two GANs defined in (1), λ is a weighting factor, and Lcyc(G1,G2) is the cycle consistency loss which is defined as

Lcyc(G1,G2)=Ey˜iY˜(G1(G2(y˜i))y˜i1)+Ex˜iX˜(G2(G1(x˜i))x˜i1). (5)

The cycle consistency loss encourages G1(G2(y˜i))=y˜i and G2(G1(x˜i))=x˜i, therefore it prevents images in X˜ from mapping to random images in Y˜.

We construct three CycleGAN image synthesizers, each of which is used to synthesize the brain atlases of one of the three missing modalities (i.e., T1c, T2 and FLAIR) from available T1-weighted brain atlases. Three training datasets, each containing images from the available modality (i.e., T1) and one of the desired modalities (i.e., one of T1c, T2 and FLAIR) are used to train the image synthesizers. Ideally, these training datasets should be collected from normal brain images. However, normal MR brain images with all four modalities are rare. For example, the T1c modality is only used for tumor cases. Therefore, we create these three training image datasets based on a tumor brain image dataset called BRATS2015 [30]. In BRATS2015, the multimodal MR image of each subject has T1, T1c, T2 and FLAIR modalities. For each modality, 1000 image slice sections containing only normal brain regions, are extracted from 100 subjects in BRATS2015. Each image slice section has three consecutive 2D image slices in an MR image. Based on the resulting four sets of image slice sections, the three training datasets, i.e., (T1, T1c), (T1, T2) and (T1, FLAIR), can be created and used to train each CycleGAN based image synthesizer. In the testing stage, for each T1-weighted MR normal brain atlas Ai,T1, i = 1, ..., N, every image slice section in Ai,T1 is proceeded by the three CycleGAN based image synthesizers to produce synthetic image slice sections of respective modalities. Then center slices of the resulting synthetic image slice sections of each modality are stacked into the final synthetic brain atlas of the corresponding modality. In this way, the multimodal brain atlases of four modalities Ai={Ai,T1,Ai,T1c,Ai,T2,Ai,FLAIR}, i = 1, ... , N can be obtained. Fig. 2 shows an example of a multimodal brain atlas, which has T1, T1c, T2 and FLAIR modalities. It is worth noting that, except the T1 modality which is real, the other modalities are synthetic.

Fig. 2.

Fig. 2.

Example of a multimodal brain atlas of T1, T1c, T2 and FLAIR modalities. The T1 modality is real, and other modalities are produced by the CycleGAN based image synthesizers. A 2D slice at the same level is used to represent the brain atlas of each modality.

B. Multimodal SCOLOR based image recovery

We represent a set of MR brain images of the same modality as a matrix I=[I1,,IW]H×W with each image in a column. W is the number of images, and H is the number of voxels in each image. Conventional low-rank methods, e.g., Low-Rank plus Sparse matrix Decomposition (LRSD) [19], can recover a low-rank matrix I^=[I^1,,I^W]H×W from I by minimizing the following function:

minI^(II^1+λI^*). (6)

The first term of (6) is the residual error constraint, which is a L1 norm of residual error II^, encouraging I^ to be close to I. The second term is the low-rank constraint, which is a nuclear norm of I^, preventing I^ having large rank. If images in I contain tumor regions, the rank of I will be larger than the case that I is comprised of normal brain images. This is because tumors usually have inconsistent image appearances and locations. Therefore, tumor regions in I are recovered by normal-looking brain regions in I^ to make I^ satisfy the low-rank constraint. In this way, if an MR brain image IiI contains tumors, I^iI^ is its corresponding recovered image of normal-looking brain. To achieve effective recovery of tumor regions, the residual error constraint has to be relaxed to allow large difference between I and I^, i.e., large residual error II^. Since the residual error constraint is equally imposed on the whole images in I regardless of tumor regions or normal brain regions, effective recovery of tumor regions is always at the cost of seriously distorted normal brain regions in recovered images.

To effectively recover tumor regions and avoid introducing distortion to normal brain regions in the recovered image, in our previous study, we proposed a low-rank method called spatially constrained low-rank (SCOLOR) [26] for image recovery. Unlike conventional low-rank methods, a tumor mask is introduced to impose different residual error constraints on tumor regions and normal brain regions. Specifically, weak residual error constraint is imposed on tumor regions for effective recovery, whereas strong residual error constraint is imposed on normal brain regions for good preservation. The objective function of SCOLOR is

minI^,C(1C)(II^)F2+λI^*+α1hH1wWPhwChw+β1h,kH1wWWhw,kw|ChwCkw| (7)

where 1H×W, 1hw = 1,1 ≤ hH, 1 ≤ wW, CH×W is the tumor mask whose elements Chw are equal to 0 (normal brain regions) or 1 (tumor regions). Therefore, only the residual error of normal brain regions is restricted by the first item (Frobenius norm). The second item is the nuclear norm which is the same as the conventional low-rank methods, e.g., LRSD. The third and fourth items are constraint and regularization terms for the tumor mask C. Specifically, PH×W in the constraint term is the probability map of normal brain regions, and it has large values in normal brain regions and small values in tumor regions. C is constrained by P, which will be discussed later. W(H×W)×(H×W) is an adjacent matrix, and its element Whw,kw = 1 means Chw and Ckw are adjacent to each other, otherwise Whw,kw = 0. So the regularization term encourages adjacent elements in C have the same label (i.e., 1 or 0). In this paper, for each element in C, its adjacent elements are set to be within its 1 voxel radius (i.e., 26-voxel neighborhood) in the same image (3D space), which corresponds to the same column of C. Optimization of (7) is composed of two steps. In the first step, I^ is updated with fixed tumor mask C, then (7) becomes

minI^(1C)(II^)F2+λI^*. (8)

(8) is a matrix completion problem which can be solved by soft impute method [31]. Based on the result of (8), the probability map of normal brain region P can be obtained by

Phw=1W11lWlwe(rhw(|II^|)rhl(|II^|))22 (9)

where rhw(|II^|) is the average absolute residual error in 3×3×3 image patch centered at the hth voxel (i.e., row) in the wth image (i.e., column) of absolute residual error |II^|. It is worth noting that both rhw(|II^|) and rhl(|II^|) are calculated in 3D image space where each column of |II^| is reformed as a 3D image. The definition of P is based on the observation that tumor regions usually have inconsistent positions and big values in |II^|, while normal brain regions usually have consistent positions and small values in |II^|. Therefore, the value of P is larger in normal brain regions than in tumor regions. In the second step, C is updated with fixed I^, then (7) can be rewritten as

minC1hHlwW(αPhw(IhwI^hw)2)Chw+β1h,kHlwWWhw,kw|ChwCkw|+ϵ (10)

where ϵ is a constant. Element in C, i.e., Chw, tends to be 0 (normal brain regions) when Phw is larger and 1 (tumor regions) when Phw is smaller. This is because large Phw indicates normal brain region, and the corresponding square residual error (IhwI^hw)2 is usually small. In this way, the first item of (10), i.e., (αPhw(IhwI^hw)2), is of high probability to be positive, which encourages Chw to be 0. While small Phw means tumor region, whose corresponding square residual error is usually large, making the first item of (10) be negative. So Chw is encouraged to be 1. (10) is in the form of Markov random field and can be solve by graph cut method [32]. These two steps are iterated until convergence, for obtaining recovered images I^. It is worth noting that, in the first iteration, all elements in C are set to 0. Therefore, SCOLOR is identical to conventional low-rank methods. As the iteration goes, C is refined based on the square residual error from previous iteration. Since the objective function of SCOLOR (7) decreases in each step and has a low bound, the convergence of SCOLOR is always guaranteed.

Original SCOLOR proposed in [26] accepts monomodal images only. In this paper, we improve the original SCOLOR to make it work in the context of multimodal images. The new SCOLOR is denoted as multimodal SCOLOR. In our MAR framework, the multimodal SCOLOR is responsible for recovering the multimodal normal-looking image T^={T^T1,T^T1c,T^T2,T^FLAIR} from the input multimodal tumor brain image T={TT1,TT1c,TT2,TFLAIR} based on the information of multimodal brain atlases Ai={Ai,T1,Ai,T1c,Ai,T2,Ai,FLAIR}, i = 1, ..., N. Specifically, the objective function of the multimodal SCOLOR is defined as

minI^mod,C1|mod|mod((1C)(ImodI^mod)F2+λI^mod*)+α1hH1wWPhwChw+β1h,kH1wWWhw,kw|ChwCkw| (11)

where input matrices Imod=[Tmod,A1,mod,,AN,mod]H×(N+1), mod ∈ {T1, T1c, T2, FLAIR}, |mod| is the number of possible modalities, and here |mod| = 4. Similar to the original SCOLOR defined in (7), optimization of (11) consists of two steps: one is the matrix completion problem which can be solved by soft impute method

minI^mod1|mod|mod((1C)(ImodI^mod)F2+λI^mod*), (12)

and the other is the graph cut problem

minC(1hHlwN+1(αPhw1|mod|mod(Ihw,modI^hw,mod)2)Chw+β1h,kHlwN+1Whw,kw|ChwCkw|+ϵ) (13)

where Ihw,mod and I^hw,mod are values at the hth row and wth column of the input matrix and the low-rank matrix of each modality, respectively. P′ is the probability map of normal brain region, which is defined by

Phw=1W11lWlwemod(rhw(|ImodI^mod|)rhl(|ImodI^mod|))22|mod|. (14)

The two steps of the multimodal SCOLOR are iterated until convergence, and each modality in the recovered multimodal normal-looking image T^={T^T1,T^T1c,T^T2,T^FLAIR} is at the first column of the corresponding low-rank matrix I^mod=[T^mod,A^1,mod,,A^N,mod], mod ∈ {T1, T1c, T2, FLAIR}. It is worth noting that, the first column of Imod is the tumor brain image, and the rest columns are normal brain atlases. Thus, only the first column of C is needed to be calculated using (13), and the values for the rest columns of C are all 0, i.e., normal brain regions.

The core of SCOLOR is the tumor mask C, by which tumor regions can be effectively recovered without distorting normal brain regions in the recovered image. In essence, the accuracy of the tumor mask heavily depends on the intensity difference between tumor regions and normal brain regions in the input images. If tumor regions have discriminative image intensity from normal brain regions, (IhwI^hw)2 in (10) and 1|mod|mod(Ihw,modI^hw,mod)2, mod ∈ {T1, T1c, T2, FLAIR} in (13) will be large, making Chw be prone to 1 (tumor regions). Comparing with (10) proposed in original SCOLOR, which is based on mono-modality, (13) utilizes multi-modality to calculate the tumor mask. Usually, tumor regions can be more easily discriminated from normal brain regions using multi-modality than mono-modality. Therefore, the resulting tumor mask calculated based on multi-modality could be more accurate and reliable than that using mono-modality. With improved tumor mask, tumor regions can be more effectively recovered and normal brain regions can be better preserved in the resulting recovered images.

C. Multi-channel image registration

At this stage, we have the multimodal brain atlases Ai={Ai,T1,Ai,T1c,Ai,T2,Ai,FLAIR}, i = 1, ..., N from the Cycle-GAN based image synthesizers and the recovered multimodal image of normal-looking brain T^={T^T1,T^T1c,T^T2,T^FLAIR} from the multimodal SCOLOR based image recovery. Thus, each atlas Ai can be registered to T^ in a multi-channel way, where each channel operates one modality. We choose SyN [20] to do multi-channel registration, which uses the monomodal similarity metric mean squared difference (MSD). The objective function of the multi-channel SyN is

mindAiT^(mod{T1,T1c,T2,FLAIR}1HxΩ(T^mod(x)Ai,moddAiT^(x))2) (15)

where dAiT^ is the deformation field warping Ai to T^, Ω is the whole image, and H is the number of voxels in an image.

In our MAR framework, the multimodal SCOLOR based image recovery and the multi-channel image registration components are iteratively proceeded to mutually refine their results until convergence. Specifically, in the tth iteration, the original multimodal brain atlases Ai, i = 1, ..., N are warped using the deformation fields dAiT^t1, i = 1..., N produced by the multi-channel image registration in the previous iteration, i.e., AidAiT^t1, i=1,,N(T^t1 is the recovered multimodal image in the previous iteration). The warped multimodal brain atlases and the original input multimodal tumor brain image T={TT1,TT1c,TT2,TFLAIR} can then be combined into new input matrices Imodt= [Tmod,A1,moddA1T^t1,,AN,moddANT^t1], mod ∈ {T1, T1c, T2, FLAIR}, which are proceeded by the multimodal SCOLOR based image recovery to get the new recovered multimodal image T^t. Then the original multimodal brain atlases Ai, i = 1, ..., N are registered to the new recovered multimodal image T^t by the multi-channel image registration. The resulting deformation fields dAiT^t, i = 1..., N are used to warp the original multimodal brain atlases Ai, i = 1, ..., N again for the next iteration.

It is worth noting that in the first iteration (t=1), the multimodal brain atlases in Imodt are aligned to the input multimodal tumor brain image T by affine transformation on T1 modality, i.e., Imodt=[Tmod,A1,moddA1,T1TT1,,AN,moddAN,T1TT1], mod ∈ {T1, T1c, T2, FLAIR}, where dAi,T1TT1, i = 1, ..., N means the affine transformation from Ai,T1 to TT1. It is clear that images cannot be well aligned using affine transformation only. Therefore large modification in Imodt is needed in the multimodal SCOLOR based image recovery to meet the low rank criterion, which results in a relatively low quality recovered multimodal image T^t. Such kind of T^t could also make the subsequent multi-channel image registration inconvenience. As the iteration goes, images in Imodt get well aligned. Therefore, little modification in Imodt is required to make I^modt satisfy the low-rank criterion, making the recovered multimodal image T^t has more consistent normal brain regions with T in each modality. In turn, with improved recovered multimodal image T^t, images in Imodt can get further aligned. The iteration stops when T^t is stable or changes little.

III. Results

We evaluate our MAR framework via multi-atlas segmentation (MAS) of brain ROI of multimodal tumor brain images. Specifically, a label fusion stage is added after our MAR framework to get the final segmentation result from registered atlases for each input image. The accuracy of the final segmentation result mainly depends on the registration quality between the atlases and the input image. Both synthetic and real multimodal tumor brain images are used to evaluate our method. Each multimodal tumor brain image contains four modalities: T1, T1c, T2 and FLAIR. In addition, other four different MAS methods, which use different MAR frameworks, are also tested. Particularly, the first MAS method uses conventional MAR framework, which registers each brain atlas to the input image by typical image registration methods, hence it can only handle normal brain images, and we denote it as ORI+MAS. The second MAS is based on the MAR framework which adopts a well-known strategy called cost function masking (CFM) [13] for registering normal brain atlases to pathological brain images, and we denoted it as CFM+MAS. The third one integrates a conventional low-rank method i.e., LRSD (defined in (6)) to handle registration of tumor brain images, and we denote it as LRSD+MAS. The last one is similar as LRSD+MAS except that the low-rank method is SCOLOR (defined in (7)) instead of LRSD, and we denoted it as SCOLOR+MAS. LRSD+MAS and SCOLOR+MAS have similar processing flow as our method, where image recovery and registration of normal brain atlases to the recovered image are iterated until convergence. The difference is that our method works with multimodal images while monomodal images are used in LRSD+MAS and SCOLOR+MAS. Note that, these four MAS methods can only handle monomodal brain images which have the same modality as the adopted brain atlases. For CFM+MAS, the tumor mask used in original CFM is delineated manually, but for the sake of fairness, the tumor mask used in CFM+MAS is the same as the tumor mask calculated in SCOLOR+MAS. The brain atlases used in all methods under comparison are T1-weighted MR normal brain images from LPBA40 [33], and each brain atlas carries 54 manually labeled brain ROIs. Therefore, except our method, ORI+MAS, CFM+MAS, LRSD+MAS and SCOLOR+MAS can only use T1 modality of the multimodal tumor brain images. The registration algorithm used in our method is multi-channel SyN, and the other four MAS methods under comparison use single-channel SyN. All the MAS methods under comparison adopt joint label fusion [9] to do the label fusion. Table I gives the summary of all MAS methods under comparison. The multimodal tumor brain images are preprocessed by affine transformation [34] with reference to MNI152 [35] of 182×218×182 voxels (1×1×1 mm3 voxel) and histogram matching [36] with reference to a multimodal brain atlas used in our method (i.e., generated by the Cycle-GAN based image synthesizers).

Table I.

List of all methods under comparision in the experiment.

Method name Atlas modality Input modality Tumor strategy Registration Iterative process Label fusion
ORI+MAS monomodal (T1) monomodal (T1) not applied SyN (single channel) no joint label fusion
CFM+MAS monomodal (T1) monomodal (T1) cost function masking SyN (single channel) no joint label fusion
LRSD+MAS monomodal (T1) monomodal (T1) low-rank recovery (LRSD) SyN (single channel) yes joint label fusion
SCOLOR+MAS monomodal (T1) monomodal (T1) low-rank recovery (SCOLOR) SyN (single channel) yes joint label fusion
Our method monomodal (T1) multimodal (T1,T1c,T2, FLAIR) low-rank recovery (multimodal SCOLOR) SyN (multi-channel) yes joint label fusion

A. Parameter tuning

Like most low-rank based methods [17], [26], [37], [38], in LRSD+MAS, SCOLOR+MAS and our method, the parameter λ, which is used to balance the residual error constraint and the low-rank constraint, is heuristically determined in the experiment. Actually, the recovered images are relatively insensitive to λ, and the final segmentation results of our method keep a relatively consistent quality within a large range of λ. Particularly, for all the testing images including synthetic and real tumor brain images, λ is set to 800 in LRSD+MAS and 40 in SCOLOR+MAS and our method. In section III-D, we will show the impact of λ on the final segmentation results of SCOLOR+MAS and our method. Parameters α and β in SCOLOR (see (10)) and multimodal SCOLOR (see (13)) are used to control the balance between the constraint and the regularization terms of the tumor mask C. In the experiment, we set α = 0.08 and β = 1 for SCOLOR in SCOLOR+MAS, and α = 0.1 and β = 1 for multimodal SCOLOR in our method. In section III-E, we will show the impact of α and β on the resulting tumor mask C in SCOLOR+MAS and our method.

B. Evaluation of synthetic tumor brain images

The synthetic multimodal tumor brain images are produced based on LPBA40 image dataset which consists of 40 T1-weighted MR normal brain images of different subjects. For each test, one image from LPBA40 is selected to make the synthetic multimodal tumor brain image, and the rest are used as the available monomodal brain atlases. Specifically, a synthetic multimodal normal brain image is first produced from the selected image by CycleGAN based image synthesizers. Then tumor(s) in a real multimodal tumor brain image, which is randomly selelcted from BRATS2015 [30], are inserted into the synthetic multimodal normal brain image. The mass effect of the inserted tumor(s) is also considered. Particularly, since images are aligned with MNI152 by affine transformation in the preprocessing stage, tumor regions in the real multimodal tumor brain image are directly used to replace the image contents at the same location of the synthetic multimodal normal brain image. The mass effect of the inserted tumor(s) to the surrounding normal tissues is simulated by adding deformation fields at the boundaries of the inserted tumor(s) (with perpendicular direction to the tumor boundaries and diffused by a Gaussian kernel with σ = 3.0) to deform the surrounding normal tissues. The magnitude of the resulting deformation vectors at the boundary of the inserted tumor(s) is around 3 mm. An example of a synthetic multimodal tumor brain image is shown in Fig. 3.

Fig. 3.

Fig. 3.

An example of a synthetic multimodal tumor brain image.

Totally, 40 synthetic multimodal tumor brain images, each of which contains tumor(s) from different real multimodal tumor brain image in BRATS2015, are made and tested in our experiment. Details of the 40 synthetic multimodal tumor brain images are listed in the supplement. Since we have the ground truth of tumor-free image and the corresponding 54 manually labeled brain ROIs for each synthetic multimodal tumor brain image, recovery quality (for LRSD+MAS, SCOLOR+MAS and our method) and segmentation accuracy are evaluated. The recovery quality is quantified by recovery error ratio which is defined as

Erc=xΩ|T^mod(x)Gmod(x)|xΩ|Gmod(x)|, (16)

where Ω stands for the whole image, T^mod(x) and Gmod(x) are image intensities at position x in the recovered image and the ground truth of tumor-free image of certain modality, respectively. Since the real tumor-free images are T1-weighted MR images, and only T1-weighted recovered images can be produced by LRSD+MAS and SCOLOR+MAS, the recovery error ratio is evaluated using T1 modality, i.e., mod = T1.

Fig. 4 shows some recovered images in LRSD+MAS, SCOLOR+MAS and our method. Specifically, Fig. 4 (a) shows an input synthetic multimodal tumor brain image. The square residual errors of each modality using SCOLOR, i.e., (ImodI^mod)2, and multi-modality using multimodal SCOLOR, i.e., mod(ImodI^mod)2|mod|, mod ∈ {T1, T1c, T2, FLAIR}, are shown in Fig. 4 (b). The resulting tumor masks are presented in Fig. 4 (c). Since the tumor mask C of multimodal SCOLOR in our method is calculated based on the square residual error of multi-modality, the resulting tumor mask is much closer to the ground truth than SCOLOR in SCOLOR+MAS which is based on the square residual error of T1 modality. Consequently, the recovered image of our method contains more effectively recovered tumor regions and better preserved normal brain regions than both LRSD+MAS and SCOLOR+MAS, especially in the regions marked by red circles in Fig. 4 (d).

Fig. 4.

Fig. 4.

(a) Example of input synthetic multimodal tumor brain image; (b) The square residual errors of each modality using SCOLOR and multi-modality using multimodal SCOLOR; (c) The tumor masks calculated by SCOLOR in SCOLOR+MAS and multimodal SCOLOR in our method. Since the multimodal SCOLOR utilizes multimodal information, the tumor mask is much closer to the ground truth than SCOLOR which is based on T1-weighted MR images only; (d) The recovered image of our method contains more effectively recovered tumor regions and better preserved normal brain regions than LRSD+MAS and SCOLOR+MAS, especially in the regions marked by red circles.

Details of the evaluation result of recovery error ratio using LRSD+MAS, SCOLOR+MAS and our method are shown in Fig. 5 (a). The recovery error ratios of the three methods decrease after each iteration. This is because as atlases and synthetic tumor brain images get well aligned after each iteration, less modification in the input image matrix is required to make the resulting recovered images satisfy the low-rank constraint. Due to the tumor mask in SCOLOR and multimodal SCOLOR, normal brain regions are well preserved, making the recovered images of SCOLOR+MAS and our method closer to their ground truth tumor-free images than LRSD+MAS. In contrast, normal brain regions are usually distorted in the recovered images using LRSD. Therefore, recovery error ratios of SCOLOR+MAS and our method are smaller than LRSD+MAS after each iteration. Since the multimodal SCOLOR used in our method works with multimodality where complementary information of tumor regions of different modalities is used, the tumor mask calculated in the multimodal SCOLOR is more accurate than the SCOLOR used in SCOLOR+MAS. Furthermore, as different modalities may contain exclusive features of the same anatomical structure, images could be better registered in multi-channel way than single-channel way. As a result, the recovery error ratio of our method is the smallest. For each synthetic brain tumor image, LRSD+MAS needs less than 7 iteration to reach to convergence, while less than 4 iterations are required for SCOLOR+MAS and our method.

Fig. 5.

Fig. 5.

(a) Recovery error ratios of 40 synthetic tumor brain images after each iteration of LRSD+MAS, SCOLOR+MAS and our method; (b) Average of 40 whole-brain Dice indices of segmented synthetic tumor brain images after each iteration using ORI+MAS, CFM+MAS, LRSD+MAS, SCOLOR+MAS, our method and TF+MAS. TF+MAS applies the same MAS method as ORI+MAS to segment original tumor-free images in LPBA40. ORI+MAS, CFM+MAS and TF+MAS have no iterative process, so their average Dice indices after each iteration are constant.

Segmentation results are evaluated by calculating the Dice index [39] of the whole brain (54 brain ROIs) between each segmented result and the corresponding ground truth, which is defined as

i=154ViGTVGT2|ΩisegΩiGT||Ωiseg|+|ΩiGT|, (17)

where VGT and ViGT are the volume (number of voxel) of the whole brain and the ith brain ROI in the ground truth, respectively. Ωiseg and ΩiGT are the segmented ith brain ROI and its corresponding ground truth, respectively. Fig. 5 (b) shows the average of 40 whole-brain Dice indices of segmented synthetic tumor brain images using ORI+MAS, CFM+MAS, LRSD+MAS, SCOLOR+MAS, and our method after each iteration. Moreover, original tumor-free images of LPBA40 are also tested using the same MAS method as ORI+MAS, which is denoted as TF+MAS in Fig. 5 (b). Obviously, TF+MAS achieves the highest average whole-brain Dice index of all methods under comparison. Because of distorted normal brain regions in recovered images using LRSD, the average whole-brain Dice index after the first iteration of LRSD+MAS is even lower than ORI+MAS. But as the iteration goes in LRSD+MAS, atlases get aligned with input tumor brain images, which improve both the recovered images and the registration results, eventually outperforming ORI+MAS after the second iteration. The averages and standard deviations of the 40 whole-brain Dice indices after the final iteration of all methods under comparison are ORI+MAS (0.821±0.015), CFM+MAS (0.828±0.017), LRSD+MAS (0.83±0.018), SCOLOR+MAS (0.834±0.017), our method (0.847±0.012) and TF+MAS (0.853±0.01). The Wilcoxon signed rank test [40] is adopted to do statistical significance test over the 40 whole-brain Dice indices after the final iteration of ORI+MAS, CFM+MAS, LRSD+MAS, SCOLOR+MAS and our method. The p values between our method and other four methods are all 3.569×10−8, indicating that our method is better than the other four methods with statistical significance.

C. Evaluation of real tumor brain images

Totally, 50 real multimodal tumor brain images of different subjects are randomly selected from BRATS2015 as testing images. Details of the 50 real multimodal tumor brain images are listed in the supplement. 40 T1-weighted MR normal brain images from LPBA40 are used as brain atlases in all methods under comparison. For each testing image, LRSD+MAS requires no more than 6 iterations to reach to convergence, whereas SCOLOR+MAS and our method need no more than 4 iterations. Due to the lack of tumor-free images, recovery error ratio cannot be calculated, and the recovery quality is evaluated by visual inspection. Benefiting from multimodal SCOLOR, the recovered images in our method have better visual quality than LRSD+MAS and SCOLOR+MAS. Fig. 6 shows an example of a real multimodal tumor brain image and its corresponding recovered T1-weighted MR images using LRSD+MAS, SCOLOR+MAS and our method after the final iteration. It is clear that, in the square residual error using multimodal SCOLOR, high value elements cover most of the tumor regions than the square residual error of each modality using SCOLOR (see Fig. 6 (b)). Therefore, as Fig. 6 (c) shows, the tumor mask calculated by multimodal SCOLOR in our method (based on the square residual error of multi-modality) is much closer to the ground truth than SCOLOR in SCOLOR+MAS (based on the square residual error of T1 modality). As a result, the recovered image of our method contains more effectively recovered tumor regions and better preserved normal brain regions than LRSD+MAS and SCOLOR+MAS, especially in the regions marked by red circles in Fig. 6 (d).

Fig. 6.

Fig. 6.

(a) Example of input real multimodal tumor brain image; (b) The square residual errors of each modality using SCOLOR and multi-modality using multimodal SCOLOR; (c) The tumor masks calculated by SCOLOR in SCOLOR+MAS (based on the square residual error of T1 modality) and multimodal SCOLOR in our method (based on the square residual error of multi-modality); (d) The recovered image of our method contains more effectively recovered tumor regions and better preserved normal brain regions than LRSD+MAS and SCOLOR+MAS, especially in the regions marked by red circles.

Segmentation results are evaluated by calculating Dice indices of grey matter (GM), whiter matter (WM), and cerebrospinal fluid (CSF) between the segmented result of each method and the ground truth. Specifically, we use SPM12 [41] to coarsely segment GM, WM and CSF from the 50 tumor brain images and the 40 brain atlases. Then these segmented results are further revised by an expert and regarded as the ground truth. Fig. 7 shows the Dice indices of GM, WM and CSF of the segmentation results using ORI+MAS, CFM+MAS, LRSD+MAS, SCOLOR+MAS and our method after each iteration. It is worth noting that tumor regions are ignored in the calculation of Dice index. Our method outperforms other methods in GM, WM and CSF after each iteration. LRSD+MAS gets the lowest Dice indices of GM, WM, and CSF after the beginning iterations because of distorted normal brain regions in the recovered images.

Fig. 7.

Fig. 7.

Average Dice indices of segmented GM, WM, and CSF of 50 real MR tumor brain images using ORI+MAS, CFM+MAS, LRSD+MAS, SCOLOR+MAS, and our method after each iteration.

Table II shows average Dice indices and standard deviations of GM, WM, and CSF of the final segmentation results using ORI+MAS, CFM+MAS, LRSD+MAS, SCOLOR+MAS, and our method. Statistical significance test is perform over these evaluation results (50 Dice indices of GM, WM, and CSF for each method) using Wilcoxon signed rank test. Our method outperforms all the other methods in GM, WM and CSF with statistical significance (p < 0.05).

Table II.

Average Dice index and standard deviation of the final sigmentation result obtained by each method under comparision.

GM WM CSF
ORI+MAS 0.639±0.050 0.698±0.040 0.566±0.053
CFM+MAS 0.657±0.045 0.712±0.034 0.571±0.047
LRSD+MAS 0.657±0.038 0.711±0.036 0.573±0.038
SCOLOR+MAS 0.669±0.044 0.719±0.036 0.580±0.046
Our method 0.701±0.035* 0.748±0.030* 0.593±0.041*

Asterisk indicates the evaluation result is better than the other four methods with statistical significance (P<0.05 in Wilcoxon signed rank test).

D. Impact of λ on the segmentation result

The 40 synthetic and 50 real tumor brain images are used to evaluate the impact of the parameter λ in SCOLOR and multimodal SCOLOR on the segmentation result of SCOLOR+MAS and our method. Specifically, we test SCOLOR+MAS and our method using λ from 10 to 100 with step size of 10. For each λ, we calculate the Dice indices of the corresponding segmentation results of SCOLOR+MAS and our method. Fig. 8 shows the evaluation result. It is clear to see that as the λ increases, the Dice indices of segmentation results using both SCOLOR+MAS and our method are first increased because of the recovered tumor regions and then decreased because of the distorted normal brain regions using large λ, i.e., over strength low-rank constraint. The segmentation results of both SCOLOR+MAS and our method are good and stable using λ from 30 to 60. Furthermore, since multimodal SCOLOR and multi-channel registration are used in our method, the segmentation result of our method is more stable than SCOLOR+MAS. For example, the largest differences of the average whole-brain Dice index of segmented 40 synthetic tumor brain images using λ within the range of 30–60 are 0.004 (SCOLOR+MAS) and 0.001 (our method), and the largest differences of the average Dice indices of GM, WM, and CSF of segmented 50 real tumor brain images using λ within the range of 30–60 are 0.01, 0.013, and 0.005 (SCOLOR+MAS), and 0.002, 0.003, and 0.002 (our method). Therefore, we set λ to 40 for SCOLOR+MAS and our method in the experiment as mentioned before.

Fig. 8.

Fig. 8.

Average whole-brain Dice indices of 40 segmented synthetic tumor brain images (left) and average Dice indices of GM, WM and CSF of 50 real tumor brain images (right) using different λ in SCOLOR+MAS (top) and our method (bottom).

E. Impact of α and β on tumor mask

For both SCOLOR used in SCOLOR+MAS and multimodal SCOLOR used in our method, the tumor mask C, which is used to discriminate tumor regions and normal brain regions, plays an important role. For tumor regions defined in C (values of corresponding elements in C are 1), strong recovery is performed, while relatively weak recovery is used for normal brain region defined by C (values of corresponding elements in C are 0). Therefore, with accurate tumor mask, recovered images of effectively recovered tumor regions and well preserved normal brain regions can be produced. According to the objective functions of C, i.e., (10) (SCOLOR) and (13) (multimodal SCOLOR), the result of tumor mask C is influenced by parameters α and β which control the strength of the constraint and regularization terms imposed on C. In our experiment, we evaluate the impact of α and β on the resulting tumor mask C after the final iteration of SCOLOR+MAS and our method. Particularly, we evaluate the resulting tumor masks for the 90 tumor brain images (40 synthetic and 50 real tumor brain images) using different value of α and β. For each tumor brain image, its corresponding manually segmented tumor regions are available and used as the ground truth. We change α and β separately and calculate the average Recall and Precision [42] of the resulting tumor masks according to the ground truth. Evaluation result is shown in Fig. 9. High Recall means most of tumor regions are correctly defined in the tumor mask, and tumor regions can be effectively recovered. High Precision means most tumor regions defined by the tumor mask are correct, and normal brain regions can be well preserved in the recovered image. As shown in Fig. 9 left, when α becomes bigger, Recall is decreased and Precision is increased. Similar results can be observed when β becomes bigger as shown in Fig. 9 right, except that Recall and Precision are relatively more stable within a large range of β than that of α. It is clear that high Recall and Precision of the resulting tumor masks could produce good recovered images, which could enhance the subsequent image registration as well as the final segmentation result. Therefore, a trade-off between Recall and Precision has to be made. In our experiment, we found that for recovered images, ineffectively recovered tumor regions could degrades the registration quality and the final segmentation result more than low-rank distorted normal brain regions of comparable region size. So as α or β increase, the registration quality and the final segmentation result would be first improved as the shrunken low-rank distorted normal brain regions, which are much larger than the real tumor regions at the beginning (i.e., high Recall and low Precision). Then as the low-rank distorted normal brain regions shrink to a comparable size as the ineffectively recovered tumor regions (i.e., similar Recall and Precision), the registration quality and the final segmentation result would be degraded. So as aforementioned, in the experiment, α is set to 0.08 (SCOLOR+MAS) and 0.1 (our method), and β is set to 1 for both SCOLOR+MAS and our method. The corresponding average Recall and Precision of the 90 tumor images after the final iteration of SCOLOR+MAS and our method are (0.701, 0.670) and (0.833, 0.785), respectively. It is clear that, the average Recall and Precision of our method are higher than SCOLOR+MAS. This is because tumor regions are more discriminative using information from multimodal images than monomodal images. Table III shows the average Recall, Precision and Dice index of the tumor masks of 40 synthetic and 50 real brain images after the final iteration of SCOLOR+MAS and our method in detail.

Fig. 9.

Fig. 9.

Average Recall (left) and Precision (right) of resulting tumor masks after the final iteration of SCOLOR+MAS (top) and our method (bottom) using different α and β.

Table III.

Evaluation result of tumor masks calculated by SCOLOR in SCOLOR+MAS (S) and multimodal SCOLOR in our method (O).

Recall Precision Dice
Synthetic tumor brain images 0.73±0.07 (S) 0.70±0.17 (S) 0.70±0.10 (S)
0.90±0.05* (O) 0.83±0.12* (O) 0.86±0.08* (O)

Real tumor brain images 0.68±0.12 (S) 0.65±0.18 (S) 0.64±0.10 (S)
0.78±0.10* (O) 0.75±0.13* (O) 0.76±0.10* (O)

Asterisk means statistical significance (P<0.05 in Wilcoxon signed rank test).

IV. Conclusion

We proposed a new multi-atlas registration (MAR) framework for input multimodal pathological images using conventional monomodal normal atlases. Our framework systematically solved two challenging problems in existing MAR methods: one is the missing modality in monomodal normal atlases, and the other is the registration of normal atlases to pathological images. Specifically, by using deep learning (CycleGAN) based image synthesizers, desired modalities can be produced from available monomodal normal atlases. With the proposed multimodal SCOLOR based image recovery, influence of pathological regions in registration can be effectively reduced. Finally, information of multimodal atlases and recovered images is fully utilized by multi-channel image registration. In our experiment, synthetic and real tumor brain images were used to evaluate our method. Our method showed better performance than state-of-the-art methods in terms of both image recovery quality and segmentation accuracy (i.e., registration accuracy between atlases and input images).

Theoretically, input images of any possible number and types of modalities can be handled by our framework using conventional monomodal normal atlases. Moreover, images containing pathological regions which have similar properties as the brain tumor (e.g., relatively large region, inconsistent location across subjects and different image appearance from normal regions) can be proceeded by our framework.

Supplementary Material

supplement

Acknowledgment

This work was supported in part by NIH grants (AG053867, EB006733, EB008374) and National Natural Science Foundation of China (No. 61502002).

Footnotes

This paper has supplementary downloadable material available at http://ieeexplore.ieee.org., provided by the author. The material includes details of tumor brain images used in the experiment. Contact dgshen@med.unc.edu for further questions about this work.

Contributor Information

Zhenyu Tang, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC, USA and also the School of Computer Science and Technology, Anhui University..

Pew-Thian Yap, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC, USA..

Dinggang Shen, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC 27599, USA and also Department of Brain and Cognitive Engineering, Korea University, Seoul 02841, Republic of Korea..

References

  • [1].Iglesias JE and Sabuncu MR, “Multi-atlas segmentation of biomedical images: a survey,” Medical image analysis, vol. 24, no. 1, pp. 205–219, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [2].Wu G, Wang Q, Zhang D, Nie F, Huang H, and Shen D, “A generative probability model of joint label fusion for multi-atlas based brain segmentation,” Medical Image Analysis, vol. 18, no. 6, pp. 881–890, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [3].Xue Z, Shen D, and Davatzikos C, “Classic: consistent longitudinal alignment and segmentation for serial image computing.” Neuroimage, vol. 30, no. 2, pp. 388–399, 2005. [DOI] [PubMed] [Google Scholar]
  • [4].Jacinto H, Valette S, and Prost R, “Multi-atlas automatic positioning of anatomical landmarks,” Journal of visual communication and image representation, vol. 50, pp. 167–177, 2018. [Google Scholar]
  • [5].Smit N, Lawonn K, Kraima A, DeRuiter M, Sokooti H, Bruckner S, Eisemann E, and Vilanova A, “Pelvis: Atlas-based surgical planning for oncological pelvic surgery,” IEEE transactions on visualization and computer graphics, vol. 23, no. 1, pp. 741–750, 2017. [DOI] [PubMed] [Google Scholar]
  • [6].Lotj¨ onen JM, Wolz R, Koikkalainen JR, Thurfjell L, Walde-¨ mar G, Soininen H, Rueckert D, A. D. N. Initiative et al. , “Fast and robust multi-atlas segmentation of brain magnetic resonance images,” Neuroimage, vol. 49, no. 3, pp. 2352–2365, 2010. [DOI] [PubMed] [Google Scholar]
  • [7].Jia H, Yap P-T, and Shen D, “Iterative multi-atlas-based multi-image segmentation with tree-based registration,” NeuroImage, vol. 59, no. 1, pp. 422–430, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [8].van der Lijn F, den Heijer T, Breteler MM, and Niessen WJ, “Hippocampus segmentation in mr images using atlas registration, voxel classification, and graph cuts,” Neuroimage, vol. 43, no. 4, pp. 708–720, 2008. [DOI] [PubMed] [Google Scholar]
  • [9].Wang H, Suh JW, Das SR, Pluta JB, Craige C, and Yushkevich PA, “Multi-atlas segmentation with joint label fusion,” IEEE transactions on pattern analysis and machine intelligence, vol. 35, no. 3, pp. 611–623, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [10].Maes F, Collignon A, Vandermeulen D, Marchal G, and Suetens P, “Multimodality image registration by maximization of mutual information,” IEEE transactions on medical imaging, vol. 16, no. 2, pp. 187–198, 1997. [DOI] [PubMed] [Google Scholar]
  • [11].Studholme C, Hill DL, and Hawkes DJ, “An overlap invariant entropy measure of 3d medical image alignment,” Pattern recognition, vol. 32, no. 1, pp. 71–86, 1999. [Google Scholar]
  • [12].Chen M, Carass A, Jog A, Lee J, Roy S, and Prince JL, “Cross contrast multi-channel image registration using image synthesis for mr brain images,” Medical image analysis, vol. 36, pp. 2–14, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [13].Brett M, Leff AP, Rorden C, and Ashburner J, “Spatial normalization of brain images with focal lesions using cost function masking,” Neuroimage, vol. 14, no. 2, pp. 486–500, 2001. [DOI] [PubMed] [Google Scholar]
  • [14].Andersen SM, Rapcsak SZ, and Beeson PM, “Cost function masking during normalization of brains with focal lesions: still a necessity?” Neuroimage, vol. 53, no. 1, pp. 78–84, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [15].Gooya A, Biros G, and Davatzikos C, “Deformable registration of glioma images using em algorithm and diffusion reaction modeling,” IEEE transactions on medical imaging, vol. 30, no. 2, pp. 375–390, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [16].Gooya A, Pohl KM, Bilello M, Cirillo L, Biros G, Melhem ER, and Davatzikos C, “Glistr: glioma image segmentation and registration,” IEEE transactions on medical imaging, vol. 31, no. 10, pp. 1941–1954, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [17].Liu X, Niethammer M, Kwitt R, Singh N, McCormick M, and Aylward S, “Low-rank atlas image analyses in the presence of pathologies,” IEEE transactions on medical imaging, vol. 34, no. 12, pp. 2583–2591, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [18].Liu X, Niethammer M, Kwitt R, McCormick M, and Aylward S, “Low-rank to the rescue–atlas-based analyses in the presence of pathologies,” in International Conference on Medical Image Computing and Computer-Assisted Intervention Springer, 2014, pp. 97–104. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [19].Peng Y, Ganesh A, Wright J, Xu W, and Ma Y, “Rasl: Robust alignment by sparse and low-rank decomposition for linearly correlated images,” IEEE transactions on pattern analysis and machine intelligence, vol. 34, no. 11, pp. 2233–2246, 2012. [DOI] [PubMed] [Google Scholar]
  • [20].Avants BB, Epstein CL, Grossman M, and Gee JC, “Symmetric diffeomorphic image registration with cross-correlation: evaluating automated labeling of elderly and neurodegenerative brain,” Medical image analysis, vol. 12, no. 1, pp. 26–41, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [21].Thirion J-P, “Image matching as a diffusion process: an analogy with maxwell’s demons,” Medical image analysis, vol. 2, no. 3, pp. 243–260, 1998. [DOI] [PubMed] [Google Scholar]
  • [22].Rueckert D, Sonoda LI, Hayes C, Hill DL, Leach MO, and Hawkes DJ, “Nonrigid registration using free-form deformations: application to breast mr images,” IEEE transactions on medical imaging, vol. 18, no. 8, pp. 712–721, 1999. [DOI] [PubMed] [Google Scholar]
  • [23].Xue Z and Shen DC, “Statistical representation of high-dimensional deformation fields with application to statistically constrained 3d warping,” Medical Image Analysis, vol. 10, no. 5, pp. 740–751, 2006. [DOI] [PubMed] [Google Scholar]
  • [24].Yang J, Shen D, Davatzikos C, and Verma R, “Diffusion tensor image registration using tensor geometry and orientation features,” in Medical Image Computing and Computer-Assisted Intervention – MICCAI 2008, Metaxas D, Axel L, Fichtinger G, and Szekely G, Eds.´ Berlin, Heidelberg: Springer Berlin Heidelberg, 2008, pp. 905–913. [DOI] [PubMed] [Google Scholar]
  • [25].Zhu J-Y, Park T, Isola P, and Efros AA, “Unpaired image-to-image translation using cycle-consistent adversarial networks,” arXiv preprint arXiv:1703.10593, 2017. [Google Scholar]
  • [26].Tang Z, Cui Y, and Jiang B, “Groupwise registration of mr brain images containing tumors via spatially constrained low-rank based image recovery,” in International Conference on Medical Image Computing and Computer-Assisted Intervention Springer, 2017, pp. 397–405. [Google Scholar]
  • [27].Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, and Bengio Y, “Generative adversarial nets,” in Advances in neural information processing systems, 2014, pp. 2672–2680. [Google Scholar]
  • [28].Long J, Shelhamer E, and Darrell T, “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 3431–3440. [DOI] [PubMed] [Google Scholar]
  • [29].LeCun Y, Bengio Y, and Hinton G, “Deep learning,” nature, vol. 521, no. 7553, p. 436, 2015. [DOI] [PubMed] [Google Scholar]
  • [30].Menze BH, Jakab A, Bauer S, Kalpathy-Cramer J, Farahani K, Kirby J, Burren Y, Porz N, Slotboom J, Wiest R et al. , “The multimodal brain tumor image segmentation benchmark (brats),” IEEE transactions on medical imaging, vol. 34, no. 10, pp. 1993–2024, 2015. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [31].Mazumder R, Hastie T, and Tibshirani R, “Spectral regularization algorithms for learning large incomplete matrices,” Journal of machine learning research, vol. 11, no. Aug, pp. 2287–2322, 2010. [PMC free article] [PubMed] [Google Scholar]
  • [32].Boykov Y, Veksler O, and Zabih R, “Fast approximate energy minimization via graph cuts,” IEEE transactions on pattern analysis and machine intelligence, vol. 23, no. 11, pp. 1222–1239, 2001. [Google Scholar]
  • [33].Shattuck DW, Mirza M, Adisetiyo V, Hojatkashani C, Salamon G, Narr KL, Poldrack RA, Bilder RM, and Toga AW, “Construction of a 3d probabilistic atlas of human cortical structures,” Neuroimage, vol. 39, no. 3, pp. 1064–1080, 2008. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • [34].Jenkinson M and Smith S, “A global optimisation method for robust affine registration of brain images,” Medical image analysis, vol. 5, no. 2, pp. 143–156, 2001. [DOI] [PubMed] [Google Scholar]
  • [35].Evans AC, Collins DL, Mills S, Brown E, Kelly R, and Peters TM, “3d statistical neuroanatomical models from 305 mri volumes,” in Nuclear Science Symposium and Medical Imaging Conference, 1993., 1993 IEEE Conference Record IEEE, 1993, pp. 1813–1817. [Google Scholar]
  • [36].Nyul LG, Udupa JK, and Zhang X, “New variants of a method´ of mri scale standardization,” IEEE transactions on medical imaging, vol. 19, no. 2, pp. 143–150, 2000. [DOI] [PubMed] [Google Scholar]
  • [37].Zhou X, Yang C, and Yu W, “Moving object detection by detecting contiguous outliers in the low-rank representation,” IEEE transactions on pattern analysis and machine intelligence, vol. 35, no. 3, pp. 597–610, 2013. [DOI] [PubMed] [Google Scholar]
  • [38].Dong W, Shi G, Li X, Ma Y, and Huang F, “Compressive sensing via nonlocal low-rank regularization,” IEEE transactions on image processing, vol. 23, no. 8, pp. 3618–3632, 2014. [DOI] [PubMed] [Google Scholar]
  • [39].Dice LR, “Measures of the amount of ecologic association between species,” Ecology, vol. 26, no. 3, pp. 297–302, 1945. [Google Scholar]
  • [40].Woolson R, “Wilcoxon signed-rank test,” Wiley encyclopedia of clinical trials, 2008. [Google Scholar]
  • [41].Penny WD, Friston KJ, Ashburner JT, Kiebel SJ, and Nichols TE, Statistical parametric mapping: the analysis of functional brain images. Elsevier, 2011. [Google Scholar]
  • [42].Powers DM, “Evaluation: from precision, recall and f-measure to roc, informedness, markedness and correlation,” 2011. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

supplement

RESOURCES