Abstract
Magnetic resonance (MR) protocols rely on several sequences to assess pathology and organ status properly. Despite advances in image analysis, we tend to treat each sequence, here termed modality, in isolation. Taking advantage of the common information shared between modalities (an organ’s anatomy) is beneficial for multi-modality processing and learning. However, we must overcome inherent anatomical misregistrations and disparities in signal intensity across the modalities to obtain this benefit. We present a method that offers improved segmentation accuracy of the modality of interest (over a single input model), by learning to leverage information present in other modalities, even if few (semi-supervised) or no (unsupervised) annotations are available for this specific modality. Core to our method is learning a disentangled decomposition into anatomical and imaging factors. Shared anatomical factors from the different inputs are jointly processed and fused to extract more accurate segmentation masks. Image misregistrations are corrected with a Spatial Transformer Network, which non-linearly aligns the anatomical factors. The imaging factor captures signal intensity characteristics across different modality data and is used for image reconstruction, enabling semi-supervised learning. Temporal and slice pairing between inputs are learned dynamically. We demonstrate applications in Late Gadolinium Enhanced (LGE) and Blood Oxygenation Level Dependent (BOLD) cardiac segmentation, as well as in T2 abdominal segmentation. Code is available at https://github.com/vios-s/multimodal_segmentation.
Keywords: Multimodal segmentation, disentanglement, Magnetic Resonance Imaging
I. Introduction
IN medical imaging multiple acquisitions of the same subject capture complementary information. Within Magnetic Resonance (MR), different pulse sequences (modalities) attenuate different tissue characteristics to identify anatomical and functional information and produce images of different contrasts. Automatic segmentation of such data remains important and relies on accurate annotations, which however can be expensive to acquire, thus resulting in scarce or imperfectly annotated datasets. To overcome this, many solutions have been proposed for example with the use of unlabelled data [1].
We approach this problem with a method that leverages data from a secondary modality and is based on disentangled representations. Our method is designed to address challenges of multimodal data that include differences in signal intensities, lack of annotated data, and anatomical and temporal misalignments due to varying spatial resolutions or due to moving organs, as in the case of dynamic heart imaging.
Multimodal learning allows capturing of information present in one modality (e.g. the anatomy) for use in another modality that has greater contrast. As a motivating example, myocardial segmentation in LGE is challenging, since LGE nulls myocardial signal to accentuate signal originating from myocardial infarction. In fact, in clinical practice, analysis of LGE is typically combined with cine-MR [3].
A naive way to propagate knowledge between modalities would be co-registration. This has been successful in the brain (see Section II-B) but precise multimodal registration needs modality independent metrics [4] and the brain remains static within an imaging session, whereas the heart is constantly moving. Furthermore, multimodal cardiac data are often inconsistent both in the number of images (different slices, cardiac phases, and perhaps more penalising resolution differences, e.g. slice thickness) and in the number of annotations. In addition, some sequences are static (LGE) and others dynamic (cine-MR). These challenges necessitate solutions that alleviate misregistrations and can pair input images.
A. Overview of the proposed approach
We propose a mechanism that can propagate knowledge for segmentation and can learn with and without annotations using a reconstruction objective as self-supervision. Moreover, our approach co-registers data within an anatomical representation space, becoming thus robust to variations in imaging contrast. Our 2D approach, Disentangle Align and Fuse Network (DAFNet), achieves the above by mapping multimodal images of the same subject into disentangled anatomy and modality factors.1 At inference time, as Figure 1 shows, DAFNet fuses the disentangled anatomy factors to combine multimodal information and extract segmentation masks.
Anatomy factors are represented as categorical feature maps with each category corresponding to input pixels that are ideally spatially similar and hence belong to the same anatomical part. This promotes semantic consistency and helps learn spatial correspondences between anatomical parts from different modalities. Modality factors encode pixel intensities in a smooth multivariate Gaussian manifold as per the Variational Autoencoder (VAE) [6]. Anatomy factors are used by a segmentation network to obtain masks, whereas a decoder re-entangles both factors to achieve image reconstruction.
Learning a disentangled representation is encouraged by minimising the information capacity of each factor respectively: thresholding the anatomy factors, which results in binarised discrete feature channels, prevents the storage of image intensity information as continuous values, which impairs disentanglement [7], whereas the variational objective minimises information in the modality factors [8]. Disentanglement is also influenced by the decoder design, either through inductive biases (see discussion in Section III-D and evaluation in Section V-H), or through learning constraints (see cross-modal decoding of one anatomy in the modality of another in Section III-D, similar to [5]).
However, a disentangled representation with modality invariant anatomy factors is not enough for multimodal learning. A multimodal representation also requires similar anatomy factors across modalities, such that each feature channel corresponds to the same anatomical region for any modality. This is achieved by weight sharing in the anatomy encoders, as well as by shared segmentation and decoder networks. These constraints implicitly create common anatomy semantics, which are essential when no labels, but only images, are available for one of the modalities. In this case we project all images to the common anatomical space, where a single segmentation network is trained with supervision only on the annotated modalities. Finally, when input data are not paired a new loss term in the cost function selects the most “informative” multimodal pairs by comparing anatomy factors.2
Our contributions are summarised as follows:
We propose a 2D method for learning disentangled representations of anatomy and modality factors in multimodal medical images for segmentation.
We demonstrate the importance of semantic anatomy factors, achieved through the model design, since they allow learning registration and fusion operators.
We propose a loss term in the cost function that enables the selection of the most “informative” multimodal pairs.
We demonstrate our method’s robustness over other approaches with extensive experiments on several datasets, in cardiac MRI and abdominal segmentation.
We show that our model works both on single-modal and multimodal inference, and that it outperforms other variants when trained with different amount (semi-supervised) or zero annotations for one of the modalities.
We discuss decoder designs using FiLM [10] and SPADE [11] respectively and evaluate disentanglement by estimating the dependence between anatomy and modality factors with distance correlation [12].
II. Related work
Multimodal machine learning is a research area that involves diverse modalities. While in computer vision modalities might refer to heterogeneous sources of information, such as text and images, here, as common in the medical domain, we restrict to different image acquisitions. We consider multimodal learning as combining information from different images, present at training or more critically at inference time. Using image translation or domain adaptation as augmentation strategy to reconcile lack of annotated data also enables training, although not inference, with multimodal images (we term this multimodal inference).3
We review work on disentangled representations, the main focus of our method, and also on multimodal medical imaging, split in spatially registered or unregistered inputs. We further discuss differences between other methods and ours regarding the representation semantics and the approach to fusion, in particular when having misaligned inputs and limited annotations. We highlight that currently no work exists that is able to simultaneously achieve multimodal fusion from unregistered data for segmentation, be robust to the number of training annotations, and be applied to single or multimodal inference.
A. Disentangled representation learning
Disentangling content from style for style transfer is gaining popularity in computer vision with many examples, such as [5]. In medical imaging, disentangled representations have been used for semi-supervised cardiac segmentation [7], [21], multi-task learning [7], [22], lung nodule synthesis [23], and multimodal registration [24]. Disentangling multimodal images has also been used for liver segmentation with domain adaptation [25] albeit without information fusion, and for brain tumour segmentation [26] although using registered images.
For anatomical features to be useful in clinical tasks, they need to be semantic and quantifiable. This is not guaranteed in disentanglement techniques used in style transfer [5] and in medical segmentation [25], [26] that do not impose restrictions on the content features. Semantic representations have recently been pursued in computer vision in the form of feature masks [27] or by learning geometry with landmarks [28]. Instead, we disentangle quantifiable anatomical features so that they are useful in segmentation, whereas interpretability is promoted with explicit design constraints (Section III-A).
B. Multimodal learning with registered images
Early work on multimodal deep learning concatenated co-registered images in different input channels to improve MR brain segmentation [29]. Robustness to missing modalities [30] was achieved with encoders per modality that mapped images to modality invariant features. Common features with multiple encoders was also proposed for cross-modal classification [31].
Another aspect of multimodal learning is information fusion. Most commonly, fusion is performed on latent features [29], [30], although fusion at multiple levels can be achieved with densely connected layers [32] to exploit multiscale correlations. Furthermore, cross-modal convolutions are used to weigh each modality’s contribution [33]. Finally, attention modules and residual learning focus on specific regions for brain MRI segmentation [34]. In contrast to the above, we take advantage of the strictly defined anatomy factors and use a max fusion operator to combine all distinct features.
C. Multimodal learning with unregistered images
Unfortunately, misalignment between multimodal images remains common. Correcting small misalignments is possible with a Spatial Transformer Network (STN) [2] applied on features [30]. Alternatively, encoder-decoder setups can be trained with multimodal data to learn common features. An exploration of different setups [35] showed that separate encoders and decoders sharing their last and first layer respectively, achieve the highest performance.
In cardiac image analysis approaches are limited. Multiple inputs can be combined by adapting segmentation masks with contour models [36], [37]. Alternatively, reducing the field of view (to patch level) and ensembling (fusing results from several atlases) can alleviate the effect of errors [38]. Recently, simultaneous segmentation and registration of multimodal cardiac MR, models the joint distribution as Multivariate Mixtures [39]. Multimodal images can also be used as samples of the same data distribution to form an expanded dataset [40]. Finally, multimodal registration, although susceptible to errors, can create “noisy” labels [41]. Our method is the first to learn a suitable representation for segmentation, co-registration, and information fusion, but in a semi-supervised setting.
III. Methodology
Here we describe DAFNet, a multi-component 2D model for multimodal and semi-supervised segmentation that is illustrated (at training time) in Figure 2. Initially, input images are encoded into anatomy and modality factors (Section III-A). Anatomy factors are aligned (with a STN) and fused (Section III-B), and participate in segmentation losses (Section III-C). Training also employs image reconstruction (Section III-D) and modality reconstruction losses (Section III-E). Finally, a multimodal pairing loss allows to dynamically learn how to pair image sources (Section III-F). Below we detail the individual components, as well as the employed cost functions.
A. Encoding into anatomy and modality factors
Given modality i ∈ {M1,…, Mn} with samples xi ∈ Xi, where is the set of images, and H and W are the height and width respectively, the encoding process achieves a disentanglement of anatomy and modality factors. Anatomy factors si are tensors produced by encoders dedicated to each modality i: si = fanatomy (xi|θi), where θi are the encoder parameters. The encoder architecture is modelled after the U-Net [42] and is shown in Figure 3a. To reduce model parameters and encourage a common anatomical representation among the multimodal data, we employ weight sharing in the decoder of each U-Net. Thus, parameters θi are split into individual parameters ϕi of the encoding path, and shared parameters ρ of the decoding path: si = fanatomy(xi|ϕi, ρ).
An anatomy factor is represented as a binary tensor that is also a one-hot encoding in the channel dimension. A binary anatomy factor cannot store different pixel intensities of an image as continuous values and this promotes the factorisation process [7]. Furthermore, this one-hot encoding encourages a particular image region to appear in a single channel. More formally, si ∈ {0, 1}H×W×C, s.t. . Two example anatomy factors produced by a cine and a LGE image can be seen in Figure 4.
Although binary, the anatomy factors are not segmentations. Instead they must be useful in segmentation and image decoding tasks described in the following sections.4
Divergence loss LKL:
Our model also decomposes images into a modality factor zi, which aims to capture intensity and texture characteristics of the input data. The modality factors are vectors produced by a single stochastic encoder, which, given an image sample xi and its anatomy factor si, learns aprobability distribution q(zi|xi, si). In order to encourage a smooth space and also minimise the encoded information [8], the posterior distribution is optimised to follow a multivariate Gaussian prior, , using the K L–divergence and the re-parameterisation trick [6]:
The modality encoder is shown in Figure 3b and predicts the mean and standard deviation of a Gaussian, which are used to draw a random sample vector zi = fmodality (xi, si).
B. Alignment and fusion of the anatomy factors
Following factor encoding, two anatomy factors si and sj of modalities i and j respectively, are aligned using non-linear registration. Given a grid of 5 × 5 control points, a STN (architecture in Figure 3c) first predicts the grid’s offsets. Then, thin plate spline interpolates the surface passing through the control points to register sj with si. The result of alignment , is a deformed anatomy factor corresponding to sj and vice versa ( corresponds to si). We optimise STN with gradients in image space (self-supervised cost of Section III-D) and with the segmentation cost of Section III-C, since we aim to align segmentation masks.5
During inference the deformed anatomies are combined to produce a fused representation, containing all unique and shared features that are present in the constituent anatomy factors. Since they are spatially aligned, a pixel wise operation such as the pixel-wise max is able to preserve all encoded features. More formally, and . One benefit of max-fusion is that it is invariant to the number of inputs and is therefore directly applicable in cases with more than two modalities.
C. Segmentation
Given an anatomy factor si, a simple neural network (architecture in Figure 3d) infers a corresponding segmentation mask mi = h(si), s.t. mi ∈ Mi := {0, 1}H×W×L, where Mi is the set of masks of modality i and L is the number of segmentation classes. The segmentation network is common for all modalities, is applied to the deformed and fused anatomies of Section III-B and is optimised as follows.
Supervised loss Lsup:
Given a set of images paired with masks (xi, mi), a supervised cost is defined as a weighted sum of the differentiable Dice loss and Cross Entropy (CE):
where α and β control the contribution of each loss. The cross entropy and differentiable Dice are respectively defined as:
where h, w, and c refer to the height, width and channel, and ph,w,l is the probability for a pixel belonging to class l.
Adversarial loss :
An unsupervised segmentation cost is also defined with a mask discriminator DM, modelled after LS-GAN [43]. The adversarial objective given real masks sampled from all modalities m ~ Mi, i ∈ {1, 2,… n} is:
where the discriminator is adversarially trained against the segmentation network. The discriminator’s architecture consists of 4 convolutional layers followed by LeakyReLU and a final single neuron layer and uses Spectral Normalisation [44] to stabilise training. In both segmentation costs, the anatomy factors si come either from the input images directly or are the result of the alignment step of a secondary j modality: . In the latter case, the gradients produced by the segmentation cost are back-propagated to the STN module to learn its parameters.6
Segmentation losses help learn better anatomy factors that separate the anatomies of interest in respective channels (see myocardium and left ventricle in Figure 4). If supervision is not available for modality i, training is performed with the adversarial loss . In addition, the deformed anatomy factor participates in the supervised segmentation of another modality j. The latter encourages anatomy factors to be similarly represented across modalities and thus enables unsupervised segmentation of modality i.
D. Decoding
The anatomy factors are further decoded into an output image of a style dictated by a modality factor zi: yi = g(si, zi). This factor entanglement can be performed with different decoders, which also indirectly influence the type of disentanglement, or in other words the type of information captured by the anatomy and modality factors. We investigate two decoder architectures based on FiLM [10] and SPADE [11].
The input of the FiLM-based decoder (Figure 3e) is the anatomy factors, which, after a series of convolutions, are conditioned by z samples. These are used to predict a scale and an offset parameter and , which modulate each intermediate feature map , where H, W and C are the height, width and number of channels respectively: FiLM(F|γ, β) = F ʘ γ + β.
We also consider a SPADE-based decoder (Figure 3f), which has been demonstrated to generate texture details on synthetic images given segmentation masks. The input to this decoder is a z sample, which is processed by a series of convolutional layers and conditioned by the anatomy factor defining the output “shape”. An Instance Normalisation layer with parameters μ and σ is firstly applied to a feature map , which is then modulated by tensors Γ and B (same size as F) .
Self-supervised costs Lrec and :
The decoders are trained to reconstruct the input with the ℓ1 loss:
In addition, synthesis of realistic images is encouraged with an image discriminator DI, i for a modality i, modelled after LS-GAN. This defines an adversarial loss:
As in the segmentation case, si is an encoding of image xi or a deformed encoding of another image . When si = fanatomy (xi|θi), the model acts as an auto-encoder. This is critical to allow the use of non-annotated images and thus enable semi-supervised learning. In the case where , the backpropagated gradients train the STN module and also aid the factorisation process (since the “style” of the output image as specified by zi, corresponds to modality i and not j).
The decoder also participates in the following loss that promotes disentanglement.
E. Reconstruction of the modality factor loss
In order to encourage disentanglement, and also avoid posterior collapse of the modality factor, we reconstruct the modality factor of a synthetic image. This prevents the decoder from ignoring the z-factors and only use the anatomy factors for image synthesis. We minimise the following loss:
where z is a sample from a unit Gaussian and y is the synthetic image produced by this z sample. Encouraging the use of modality factors by the decoder is further achieved by cross-reconstructing a deformed anatomy in a modality dictated by the corresponding z-factor, as described in Section III-D.
F. Non-expert pairing
Better multimodal fusion and STN registration will be achieved by multimodal image pairs {xi, xj} that are more similar in terms of their spatial and temporal positions. In cases where the multimodal images are not expertly paired, we can automatically measure anatomical similarities with an optional cost, which directly compares the anatomy factors, and “selects” only the most informative image pairs.
During training, and given an image xi and a set of k candidate images from modality , the multimodal segmentation and reconstruction losses for a sample xi are weighted accordingly by k weights, s.t. :
where , and . By weighting the loss functions, the STN module does not need to learn deformations for all pairs, as well as prevents it from trying to match slices with different anatomical content.
Due to the semantics of the anatomy factors, and the fact that they are categorical, we can directly evaluate their overlap in terms of the Dice score. The Dice for each pair, becomes the input to a small neural network of two fully connected layers that outputs the weights, and is similar to the temperature scaling technique proposed for calibrating classification outputs [45]. At inference time, the most accurate mi segmentation is produced from the weighted sum of the fusion with different slices :
This optional weighting of the cost function is only used in unpaired data, and as shown in experiment V-C converges to the same result as manual pairing.
IV. Experimental setup
A. Training details
The model is trained with a multi-component loss function A higher Lsup weight encourages separation of segmentation classes. A reduced LK L weight prevents posterior collapse, in which the decoder ignores the z factor; however, an even lower LK L would not well approximate the Gaussian prior leading to a non-smooth intensity manifold. Number of s channels and z dimensions are set to C=8 and nz=8 respectively, as in [7].
Code is written in Keras (https://keras.io) and is available at https://github.com/vios-s/multimodal_segmentation along with pseudocode of the training process. We train with Adam (learning rate 10−4) and use Stochastic Weight Averaging [46] to reliably compare methods. Training takes approximately 12 hours on a Titan-X GPU and inference of one image takes 50 ms. Quantitative evaluation is performed on 3-fold cross-validation, where the training, validation and test sets correspond to 70%, 15% and 15% of the data volumes.
B. Data
Experiments use three multimodal datasets of a source and a target modality that have been rescaled to [−1, 1].
For LGE segmentation, we use cine-MR and LGE data of 28 patients [47], acquired at Edinburgh Royal Infirmary (ERI), with spatial resolution 1.562mm2/pixel, and slice thickness 9mm. End diastolic myocardial contours are provided. The image size is 192 × 192 pixels. The number of segmented images is 358 (for each of cine-MR, LGE).
To evaluate robustness on different medical data, we use abdominal T1-dual inphase and T2-SPIR images from CHAOS [48] dataset, for T2 segmentation. Images of 20 subjects with liver, kidneys and spleen segmentations are acquired by a 1.5T Philips MRI scanner, which produces 12-bit DICOM images of 256 × 256 resolution. We resample to an x-y spacing of 1.89mm, and crop to 192 × 192 pixels. In total, there are 1594 images.
Finally, we evaluate BOLD segmentation with a dataset (shorthand BOLD) of cine-MR and CP-BOLD images [49] of 10 (mechanically ventilated) subjects, with an in-plane resolution of 1.25mm × 1.25mm, acquired at baseline and severe ischemia (inflicted as controllable stenosis of the left-anterior descending coronary artery (LAD) on a 1.5T Espree (Siemens Healthineers). The image acquisition is at short axis view, covering the midventricle, and is performed using cine-MR and a flow and motion compensated CP-BOLD sequence, where each sequence is applied one after the other in the protocol in separate breath-holds. The pixel resolution is 192 × 114. In total there are 129 cine-MR and 264 CP-BOLD images with expert segmentations from all cardiac phases.
C. Baseline and benchmark methods
We consider the following baselines, which assume source masks being available at inference time. If predicted masks were used e.g. the result of a U-Net, an additional confounder would be introduced. Thus, we report numbers with ground truth masks for a bias-free estimate, which albeit is elevated.
A lower bound computes the Dice score between real masks of two modalities, which is also a measure of misalignment of the multimodal data. This is referred to as “copy” and can be used for segmenting a target modality without annotations from the target modality.
This lower bound can be improved after registering the multimodal images and applying the registration field to the source masks. The deformation field is calculated by affine registration using mutual information, followed by symmetric diffeomorphic using cross-correlation [50]. This is referred to as “register” baseline and can also be used without annotations of the target modality. “Copy” and “register” are common in clinical evaluation.
Finally, we implement a version of a non-coupled active contour model akin to the one in [37]. We initialised the contour using the “copy” above. Then via grid search, we found optimal contour length, smoothness, and stepping hyperparameters as: for ERI [0.5, 0.15, 0.7], BOLD [0.01, 0.15, 0.7] and CHAOS [0.5, 0.15, 0.7], respectively.
We also consider the following deep learning benchmarks.
As a supervised benchmark we train a UNet on annotated data of the target modality and refer to it as UNet-single. We further re-train a UNet on mixed training data of all modalities to evaluate its capability of concurrently handling multimodal data and refer to it as UNet-multi.
We train SDNet [7] with full or semi supervision on data of the target modality and refer to it as SDNet-single. We also train SDNet by mixing multimodal data, as demonstrated in [7], and refer to it as SDNet-multi.
We get two final benchmarks by training MUNIT [5] for image translation. The first uses MUNIT to translate images from source to target modality [17], and the second translates multimodal images to a domain invariant space [25]. In both cases, segmentation is performed post-hoc with a UNet on the combined data. We refer to these approaches as Translation and DADR respectively.
Finally, we implement DualStream [35], the most recent Deep Learning based method for handling multimodal data which does not require registered data.
For all benchmarks we use the architectures of the original papers, resulting in a comparable number of parameters. For UNet, SDNet, MUNIT, DualStream and DAFNet these are approximately in millions 35, 35, 26, 56 and 52, respectively.
V. Experiments and discussion
Sections V-A and V-B present segmentation results, assuming a source modality that always contains annotations during training. The source modality is cine-MR for ERI and BOLD datasets, and T1 for CHAOS. The target modality is LGE, BOLD and T2 for ERI, BOLD and CHAOS, respectively. Unless explicitly specified, DAFNet uses a FiLM-based decoder, and we report test Dice of the fused anatomies. We evaluate the effects of input pairing (Section V-C), registration (Section V-D), and a SPADE-based decoder (Section V-G). Then, Section V-H evaluates disentanglement of each decoder design. Where appropriate, bold font denotes the best (on average) method and an asterisk (*) denotes statistical significance of paired t-tests (p < 0.05 assessed via permutations) comparing with the second best (to avoid multiple comparisons).
A. Multimodal segmentation: full and zero supervision setting
The prime contribution of our work is the ability to learn and infer in a multimodal setting. Thus, we first demonstrate that multiple inputs at training and inference time benefit segmentation. Table I presents test Dice scores on three datasets for DAFNet and the benchmarks of Section IV-C. Two setups are evaluated, assuming either that annotations are available for the target modality or not.
TABLE I:
100% target annotations | |||||||||
---|---|---|---|---|---|---|---|---|---|
Methods | Train | Test | Masks in test | ERI | BOLD | CHAOS | |||
LGE | cine | BOLD | cine | T2 | T1 | ||||
copy | – | multi | Yes | 6706 | 6706 | 8001 | 8001 | 7110 | 7110 |
register | – | multi | Yes | 6807 | 6705 | 8104 | 8405 | 7007 | 7305 |
AC | – | multi | Yes | 6615 | 6613 | 6802 | 7205 | 6522 | 6522 |
UNet | single | single | No | 7804 | 8508 | 9101 | 8901 | 8517 | 8605 |
SDNet | single | single | No | 8003 | 8409 | 8903 | 8804 | 8316 | 8508 |
UNet | multi | single | No | 8103 | 8308 | 8903 | 8802 | 8515 | 8803 |
SDNet | multi | single | No | 8005 | 8605 | 8902 | 8703 | 8511 | 8801 |
DualStream | multi | single | No | 8006 | 8609 | 8909 | 8802 | 8516 | 8509 |
Translation | multi | single | No | 7906 | 8405 | 8306 | 8802 | 8309 | 8706 |
DADR | multi | single | No | 7905 | 8306 | 8804 | 8602 | 8416 | 7222 |
DAFNet | multi | single | No | 8203 | 8602 | 8801 | 9102 | 8317 | 8801 |
DAFNet | multi | multi | No | 8203 | 8402 | 9101 | 9101 | 8701 | |
0% target annotations | |||||||||
Methods | Train | Test | Masks in test | ERI | BOLD | CHAOS | |||
LGE | cine | BOLD | cine | T2 | T1 | ||||
copy | – | multi | Yes | 6706 | 6706 | 8001 | 8001 | 7110 | 7110 |
register | – | multi | Yes | 6807 | 6705 | 8104 | 8405 | 7007 | 7305 |
AC | n/a | multi | Yes | 6615 | 6613 | 6802 | 7205 | 6522 | 6522 |
UNet | single | single | No | n/a | n/a | n/a | n/a | n/a | n/a |
SDNet | single | single | No | n/a | n/a | n/a | n/a | n/a | n/a |
UNet | multi | single | No | 3823 | 6812 | 6823 | 8505 | – | – |
SDNet | multi | single | No | 6118 | 7307 | 8003 | 8503 | 5109 | 6313 |
DualStream | multi | single | No | 3823 | 6812 | 6823 | 8505 | – | – |
Translation | multi | single | No | 3723 | 6113 | 6110 | 7407 | – | 4511 |
DADR | multi | single | No | 4619 | 6313 | 6811 | 8501 | – | 4917 |
DAFNet | multi | single | No | 7206 | 7805 | 7802 | 8203 | 7212 | 7406 |
DAFNet | multi | multi | No | 7604 | 8602 | 7106 |
In the 100% case, training with multiple inputs improves accuracy, even when multimodal data simply constitute an augmented dataset. When segmenting the target modality, the usage of multiple inputs at inference time by DAFNet, obtains similar Dice as other benchmarks, but considerably reduces the standard deviation, such as in CHAOS from 11% to 5%.
In the 0% case the (learned) benchmark methods do not produce accurate target segmentations for all datasets. As expected, models trained only on the source modality learn modality-specific features, and as such cannot generalise to the unseen target modality. DAFNet on the other hand, consistently maintains a better average and smaller variance by leveraging information from the source modality. This is due to the aligning (in an embedding sense) of the multimodal representations in the anatomy space, which allows the shared segmentor trained with supervision on the source, to also segment the target modality with “zero” supervised examples.
We then exchange source and target and report the cine-MR and T1 Dice by training new models where appropriate. The CP-BOLD sequence is very similar to cine, showing anatomy, but has elevated T2 contrast (the BOLD effect) [49]. In addition, these data are acquired in controlled experiments, with the cine-MR and BOLD images acquired one after the other in the protocol. Thus, for the BOLD dataset all methods perform well in the 0% case of cine-MR segmentation, with multiple inputs further improving DAFNet performance. On the contrary, segmenting LGE (and T2) is more difficult. The Dice is overall lower than the cine-MR in the single-output DAFNet with the difference being bigger in the 0% case. As a result, this hurts the multi-output cine-MR Dice. This is expected since the benefit of multimodal segmentation comes when one modality is easier to segment.7 Therefore LGE benefits when considering cine-MR images, but the contrary would only be beneficial in cine-MR reconstruction problems, e.g. in the presence of motion artefacts [51].
B. Semi-supervised segmentation
Here we evaluate the sensitivity of all methods on different amounts of ground truth annotations available during training. Table II presents the average (across all labels) cross-validation test set Dice score. Exemplar test results are shown in Figure 5. The number of images for both source and target modalities are fixed, but the amount of target annotations varies. Sampling the amount of annotations is performed on a subject-level, to avoid having a mixture of annotated and non-annotated images of the same subject in the training set. The DAFNet results correspond to using multiple inputs at inference time.
TABLE II:
ERI: Target LGE Dice (Source cine) | |||
---|---|---|---|
Method | 50% | 25% | 12.5% |
copy | 6706 | 6706 | 6706 |
register | 6807 | 6807 | 6807 |
AC | 6615 | 6615 | 6615 |
UNet-single | 7612 | 6614 | 5121 |
SDNet-single | 7604 | 6909 | 5418 |
UNet-multi | 7608 | 6711 | 5019 |
SDNet-multi | 7604 | 7307 | 6419 |
DualStream | 7603 | 6113 | 4423 |
Translation | 7507 | 6714 | 6214 |
DADR | 7705 | 6611 | 5719 |
DAFNet | |||
BOLD: Target BOLD Dice (Source cine) | |||
Method | 50% | 25% | 12.5% |
copy | 8001 | 8001 | 8001 |
register | 8104 | 8104 | 8104 |
AC | 6802 | 6802 | 6802 |
UNet-single | 7917 | 5927 | 4929 |
SDNet-single | 8403 | 6817 | 6414 |
UNet-multi | 8703 | 7517 | 7213 |
SDNet-multi | 8607 | 8503 | 8003 |
DualStream | 8601 | 5826 | 4928 |
Translation | 8402 | 7906 | 4726 |
DADR | 8702 | 7901 | 7115 |
DAFNet | 8701 | 8603 | |
CHAOS: Target T2 Dice (Source T1) | |||
Method | 50% | 25% | 12.5% |
copy | 7110 | 7110 | 7110 |
register | 7007 | 7007 | 7007 |
AC | 6522 | 6522 | 6522 |
UNet-single | 8017 | 7615 | 7217 |
SDNet-single | 8214 | 7716 | 7514 |
UNet-multi | 8415 | 7916 | 7516 |
SDNet-multi | 8411 | 8013 | 7809 |
DualStream | 8119 | 7816 | 7516 |
Translation | 8107 | 7511 | 7010 |
DADR | 8411 | 7714 | 7411 |
DAFNet | 8405 | 8203 |
Average Dice for all methods is comparable when the number of annotations is high, although DAFNet achieves the lowest variance. With a reducing number of annotations, the performance of the competing methods is also reduced with a simultaneous increase in the variance. DAFNet maintains good results and robustness to edge cases, as evidenced by the small variance achieved throughout all setups.
C. Effect of pair matching
The results of Sections V-A and V-B correspond to expertly paired multimodal inputs. Here, we evaluate the sensitivity of DAFNet on unpaired multimodal images, as well as the effect of the automated pairing cost proposed in Section III-F.
We randomly shuffle the multimodal pairs by two positions, with the shuffled pairs differing up to two spatial slices within a 3D volume.8 We measure the LGE segmentation Dice score on ERI data when using 100% and 0% LGE annotations. We thus compare our automated method with expert pairing (upper bound) and a random shuffle (lower bound). Table III presents the results of copy method, as well as of DAFNet evaluated with both cine-MR and LGE inputs.
TABLE III:
Pair matching | copy | DAFNet 0% | DAFNet 100% |
---|---|---|---|
expert | 6706 | 7404 | 8203 |
automated | n/a | 7106 | 8003 |
random | 4416 | 6508 | 7706 |
Shuffling the multimodal pairs decreases the copy performance considerably. In both cases, automated matching of candidate pairs based on the semantics of the anatomy factors proves effective in ignoring distant slices (in the volume), with results very closely approaching the ones achieved by expert pairing. To show how our model learns appropriate weights, the evolution of weights across training epochs is shown in Figure 6, in which w1 corresponds to the closest pair converged to a probability of one early on in training.
During inference, a “soft” segmentation mask is produced as a weighted sum between each weight with its corresponding mask. However, this converges to using the prediction of the “closest” pair, as evidenced by Figure 6.
D. Effect of STN
We assess the need for a registration module with an ablated model. We compare the accuracy of a fused segmentation that is obtained with and without the STN module. Two DAFNet models are compared, trained on ERI data with 100% and 0% LGE annotations. The mean Dice without the STN is measured to be 75 ± 6% and 71 ± 6% respectively. This is lower than the Dice of DAFNet with STN that is 82 ± 3% and 74 ± 4%. Furthermore, in the 100% case the difference is statistically significant at the 1% level. Thus, clearly registration helps.
An example anatomy alignment is shown in Figure 7. Although not a perfect alignment of the images is required, the left ventricle and myocardium of the cine-MR have deformed to match the corresponding LGE (marked in red boxes).
E. Ablation study on cost components
We assess the contribution of cost components in the fused segmentation on ERI with 100% and 0% LGE annotations. We evaluate the effect of losses involved in reconstruction, the adversarial training losses on masks () and images (), and the modality factor reconstruction loss ().9 Table IV shows that best results are achieved with all cost components. Removing the disentanglement and reconstruction losses significantly reduces performance at 0% supervision, which conclusively shows their role in learning without annotations.10 In addition, the data-driven reconstruction cost (via the image discriminator) contributes by encouraging more accurate synthesis, which helps learn better anatomical representations.
TABLE IV:
LKL | Lrec | DAFNet 0% | DAFNet 100% | |||
---|---|---|---|---|---|---|
— | — | ✓ | — | — | 6607 | 8103 |
✓ | ✓ | — | ✓ | ✓ | 7202 | 8003 |
✓ | ✓ | ✓ | — | ✓ | 6505 | 7117 |
✓ | ✓ | ✓ | ✓ | — | 7104 | 8103 |
✓ | ✓ | ✓ | ✓ | ✓ | 7404 | 8203 |
F. Ablation on factor sizes C and nz
In all experiments, factor sizes are set to C=8, and nz=8. C is determined experimentally, so that there is enough capacity for all segmentation classes and background anatomy. A large C does not affect segmentation, and the redundant capacity is ignored, see “empty” channels of Figure 4. This is confirmed with ablated models with C=4 or C=16 trained with 100% annotations on ERI and CHAOS. The ERI model achieves 82± 2% for both setups, the same as when C=8. The CHAOS model achieves 74 ± 12% and 85 ± 5% for C=4 and C=16, respectively. The performance significantly drops when C=4, since there is not enough capacity.
Size nz is determined according to our previous [7], and related work [5]. We experimented with nz=4 and nz=16 and 100% annotations on ERI and CHAOS. We find no effect on segmentation accuracy. However, nz affects the information capacity, approximated by the average variance [52], of each z-dimension, where smaller variance implies higher informativeness. With nz = 16, the lowest variance is 0.63, the first 8 dimensions have an average of 0.86 and the remaining 8 an average of 0.95. For nz=8, the variance ranges between 0.47 and 0.80, and for nz=4, between 0.43 and 0.60. Admittedly, lower nz results in higher information content in each dimension, thus large nz seems redundant in this setup.
G. Effect of decoder design on segmentation accuracy
The modular design of DAFNet permits incorporation of components with different designs. We evaluate segmentation accuracy achieved by two decoder architectures: FiLM and SPADE. Specifically, we train a SPADE-based DAFNet on ERI and CHAOS and compare with the FiLM-based DAFNet for 100% and 0% annotations.
With 100% annotations, the SPADE-based DAFNet achieves 82 ± 3% and 85 ± 5% on ERI and CHAOS respectively, identical to the Dice achieved by FiLM. With 0% annotations, the SPADE-based DAFNet achieves 73 ± 4% and 75±7%, whereas FiLM-based results are 74±4% and 74±3% respectively on ERI and CHAOS.
We can conclude that the regularising effect of the reconstruction process on extracting segmentations, is similar with both decoder variants. However, different decoder designs influence the way the anatomy and modality factors interact to produce a synthetic image. We explore this next.
H. Evaluating disentanglement
Even though FiLM and SPADE decoders do not result in evident differences in segmentation accuracy, they produce synthetic images of different quality (Figure 8). Since the anatomy factors contain flat regions, FiLM-based conditioning with scalar parameters tends to produce images with less texture details than SPADE-based conditioning.
Here, we assess the information retained in the modality factors and characterise the achieved disentanglement. This is a challenging problem not addressed in existing literature: all assume vector latent variables (e.g. BetaVAE score [6]). In DAFNet, and typically in content/style disentanglement, the factors of variation are not of the same dimensionality, with the anatomy being spatial. For the experiments below, we use models trained on CHAOS with 100% T2 annotations to assess (dis)entanglement using classification tests, factor arithmetics, and a proposed metric of independence of random variables.
1). Modality classification:
On the premise that the common modality encoder correctly extracts modality features, a classifier should detect the modality type, given just the z-factor. We assess this hypothesis, by training a logistic regression classifier to predict whether different z-factors correspond to T1 or T2 images. The classifier’s accuracy is 99% and 97% for FiLM and SPADE, respectively, on a test set of three subjects.
We further evaluate whether specific dimensions in z capture the modality type by repeating the experiment for each dimension. In the FiLM model, the 2nd dimension achieves 100% accuracy, whereas the rest vary between 54% and 64%. Similarly, in the SPADE model the 7th dimension achieves 97% accuracy vs. 42% and 63% of the others.
2). Modality factor arithmetics:
We qualitatively examine the information retained in each dimension of vector z with latent space arithmetics. The likelihood of the modality factor approximates a Gaussian prior, and therefore interpolating in the range [−3, 3] covers the probability space. Figure 9 shows synthetic images arranged in a grid; images of each row are produced by interpolating the values of a single dimension of z, with the remaining ones fixed. The final two columns highlight affected regions by calculating the per-pixel Pearson correlation, as well as the difference, Δimage, between the synthetic images at extreme z values −3 and 3, respectively.
Both decoders have one z-dimension with a global effect (2 and 7 respectively) controlling the “modality” type. This finding is inline with the classification results above. Furthermore, some dimensions of the FiLM decoder, e.g. the 8th, appear focused on specific regions, whereas the dimensions of the SPADE decoder produce more diffused correlation images. The latter likely relates to SPADE generating texture better.
3). Disentanglement metric:
We propose distance correlation [12] as a metric of factor independence, which is invariant to the input variable dimensionality, and can also detect linear and non-linear associations. While it has been used before for reducing data leakage [53], we use distance correlation for measuring (dis)entanglement. Distance correlation is equal to
where dCov(s, z) is the distance covariance of s and z, and dVar(.) is the distance variance respectively. Given n random samples sk and zk with k ∈ [1, n], the distance covariance is the product of two distance matrices (one for each variable) averaged by n2, where each distance matrix d(.) is double centred by subtracting the mean row, the mean column and the overall mean from each element . The distance variance is then dVar2(s) = dCov2(s, s), and dVar2(z) = dCov2(z, z).
The distance correlation between s and z values from a FiLM-based model is dCor = 0.55, whereas the equivalent for a SPADE-based model is dCor = 0.78. This suggests that factors obtained by the FiLM decoder are more independent and therefore the FiLM-based model is more disentangled. Although distance correlation cannot explicitly evaluate the type of information in each variable, this result can be explained intuitively by the decoder design. The SPADE decoder allows more flexibility to the z factors as evidenced in the synthetic images, which contain more texture, but also in the diffused correlation images of Figure 9b, implying a higher anatomical correlation (and higher entanglement) between factors.
VI. Conclusion
We have presented a method for multimodal learning and specifically multimodal segmentation, that is robust to the requirement for registered and paired input images. This has been made possible by disentangling images into semantic anatomy factors that are consistently represented across modalities and modality factors that model the intensity variability of the multimodal inputs into a smooth latent space.
We have proposed DAFNet, which, to the best of our knowledge, is the first work that enables multimodal segmentation by aligning disentangled anatomical representations and can be trained with zero annotations for one of the modalities. We presented the benefit of multimodal (over single-modal) learning in cardiac and abdominal segmentation, where we achieve high accuracy and low variance by fusing anatomical information of different modalities. We further demonstrated robustness to misalignments between multimodal images (achieved by a spatial transformer network) and robustness to the quality of the multimodal pair matching (with an optional pair weighting), both made possible by comparing the anatomy factors. Finally, we made a first step in evaluating the quality of the content/style disentanglement using the distance correlation.
The significance of our work lies in the potential for the use of disentangled representations in other challenging problems of medical research. Future directions include the learning of further factorisations suitable for medical data, capturing pathological information and specific artefacts for instance, as well as a theoretical characterisation of the disentangling process and precise quantification of the type of information that is captured by each factor. Admittedly this is more complex in content/style disentanglement than in vectorised latent spaces for which metrics have been recently suggested [54].
Acknowledgments
This work was supported by US National Institutes of Health (1R01HL136578-01) and used resources from the Edinburgh Compute and Data Facility. S.A. Tsaftaris is also supported by a Canon Medical / Royal Academy of Engineering Research Chair (RCSRF1819\8\25).
Footnotes
In computer vision these are typically referred to as content and style factors, respectively [5]. However, as we detail in Section II-A in medical applications such disentanglement has more stringent requirements.
We improve our preliminary work [9] as follows: (1) we reduce model parameters and encourage common anatomy semantics by employing weight sharing in the anatomy encoders; (2) we introduce a cost that reduces the need for expert pairing of multimodal inputs; (3) we design and evaluate another decoder; (4) we propose the use of distance correlation to assess disentanglement; (5) we generalise to non-cardiac datasets.
For completeness we mention few recent methods. Image translation is proposed with cycle consistency [13]–[16] or disentanglement [17] losses. In domain adaptation, multimodal images are related with different augmentations [18], histogram matching [19] or adversarial losses [20].
In fact, segmentations cannot be directly used in the decoding task when these contain multiple anatomical structures, as this causes tension between the segmentation and decoding tasks. As an example, consider papillary muscles, which here are part of left ventricle segmentations. Synthetic images cannot contain papillary muscles unless they are separated in the anatomy encoding (Figure 4). We break this tension with a segmentation network, which combines different channels of an anatomy factor to produce a segmentation.
We avoid direct comparison of si and , since they are binary and thus different small deformations might generate the same error.
We omit the use of as input to the segmentation network to avoid backpropagating gradients both to the STN and the jth anatomy encoder, which might result in the STN not achieving a good convergence.
Indeed, cine-MR is designed to show anatomical information, whereas LGE to highlight infarcted myocardium.
Similar results can be obtained by shuffling the different cardiac phases in the cine-MR temporal stack.
We do not include ablations of the supervised segmentation (Lsup) because DAFNet is not fully unsupervised.
We repeat this experiment using a variant without a segmentor, where the anatomy encoder directly predicts segmentations that are aligned by the STN. This performs significantly worse than results of Table IV, showing the importance of an intermediate representation in registration.
Contributor Information
Agisilaos Chartsias, School of Engineering, The University of Edinburgh, Edinburgh, EH93JL, UK.
Giorgos Papanastasiou, Edinburgh Imaging Facility QMRI; Centre for Cardiovascular Science, Edinburgh, EH164TJ, UK.
Chengjia Wang, Edinburgh Imaging Facility QMRI; Centre for Cardiovascular Science, Edinburgh, EH164TJ, UK.
Scott Semple, Edinburgh Imaging Facility QMRI; Centre for Cardiovascular Science, Edinburgh, EH164TJ, UK.
David E. Newby, Edinburgh Imaging Facility QMRI; Centre for Cardiovascular Science, Edinburgh, EH164TJ, UK
Rohan Dharmakumar, Cedars Sinai Medical Center Los Angeles CA 90048, USA.
Sotirios A. Tsaftaris, School of Engineering, The University of Edinburgh, Edinburgh, EH93JL, UK; The Alan Turing Institute, London, NW12DB, UK, School of Engineering, The University of Edinburgh, Edinburgh, EH93JL, UK.
References
- [1].Tajbakhsh N, Jeyaseelan L, Li Q, Chiang JN, Wu Z, and Ding X, “Embracing imperfect datasets: A review of deep learning solutions for medical image segmentation,” Med Image Anal, vol. 63, p. 101693, 2020. [DOI] [PubMed] [Google Scholar]
- [2].Jaderberg M, Simonyan K, Zisserman A, and Kavukcuoglu K, “Spatial transformer networks,” in NIPS, 2015, pp. 2017–2025. [Google Scholar]
- [3].Kim HW, Farzaneh-Far A, and Kim RJ, “Cardiovascular magnetic resonance in patients with myocardial infarction: current and emerging applications,” J Am Coll Cardiol, vol. 55, no. 1, pp. 1–16, 2009. [DOI] [PubMed] [Google Scholar]
- [4].Sotiras A, Davatzikos C, and Paragios N, “Deformable medical image registration: A survey,” IEEE Trans Med Imag, vol. 32, no. 7, pp. 1153–1190, July 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [5].Huang X, Liu M-Y, Belongie S, and Kautz J, “Multimodal unsupervised image-to-image translation,” in ECCV, 2018, pp. 172–189. [Google Scholar]
- [6].Kingma DP and Welling M, “Auto-encoding variational bayes,” in ICLR, 2014. [Google Scholar]
- [7].Chartsias A, Joyce T, Papanastasiou G, Semple S, Williams M, Newby DE et al. “Disentangled representation learning in cardiac image analysis,” Med Image Anal, vol. 58, p. 101535, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [8].Alemi AA, Fischer I, Dillon JV, and Murphy K, “Deep variational information bottleneck,” ICLR, 2017. [Google Scholar]
- [9].Chartsias A, Papanastasiou G, Wang C, Stirrat C, Semple S, Newby D et al. “Multimodal cardiac segmentation using disentangled representation learning,” in STACOM. Cham: Springer, 2020, pp. 128–37. [Google Scholar]
- [10].Perez E, Strub F, de Vries H, Dumoulin V, and Courville AC, “FiLM: Visual reasoning with a general conditioning layer,” in AAAI, 2018, pp. 3942–51. [Google Scholar]
- [11].Park T, Liu M-Y, Wang T-C, and Zhu J-Y, “Semantic image synthesis with spatially-adaptive normalization,” in CVPR, 2019, pp. 2337–46. [Google Scholar]
- [12].Székely GJ, Rizzo ML, and Bakirov NK, “Measuring and testing dependence by correlation of distances,” Ann Stat, vol. 35, no. 6, pp. 2769–2794, 2007. [Google Scholar]
- [13].Campello VM, Martín-Isla C, Izquierdo C, Petersen SE, Ballester MAG, and Lekadir K, “Combining multi-sequence and synthetic images for improved segmentation of late gadolinium enhancement cardiac MRI,” in STACOM. Cham: Springer, 2020, pp. 290–99. [Google Scholar]
- [14].Huo Y, Xu Z, Moon H, Bao S, Assad A, Moyo TK et al. “SynSeg-Net: Synthetic segmentation without target modality ground truth,” IEEE Trans Med Imag, vol. 38, no. 4, pp. 1016–1025, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [15].Zhang Z, Yang L, and Zheng Y, “Translating and segmenting multimodal medical volumes with cycle-and shape-consistency generative adversarial network,” in CVPR, 2018, pp. 9242–9251. [Google Scholar]
- [16].Tao X, Wei H, Xue W, and Ni D, “Segmentation of multimodal myocardial images using shape-transfer GAN,” in STACOM. Cham: Springer, 2020, pp. 271–279. [Google Scholar]
- [17].Chen C, Ouyang C, Tarroni G, Schlemper J, Qiu H, Bai W et al. “Unsupervised multi-modal style transfer for cardiac MR segmentation,” in STACOM. Cham: Springer, 2020, pp. 209–219. [Google Scholar]
- [18].Ly B, Cochet H, and Sermesant M, “Style data augmentation for robust segmentation of multi-modality cardiac MRI,” in STACOM. Cham: Springer, 2020, pp. 197–208. [Google Scholar]
- [19].Liu Y, Wang W, Wang K, Ye C, and Luo G, “An automatic cardiac segmentation framework based on multi-sequence MR image,” in STACOM. Cham: Springer, 2020, pp. 220–227. [Google Scholar]
- [20].Chen C, Dou Q, Chen H, Qin J, and Heng P-A, “Synergistic image and feature adaptation: Towards cross-modality domain adaptation for medical image segmentation,” in AAAI, vol. 33, 2019, pp. 865–72. [Google Scholar]
- [21].Chartsias A, Joyce T, Papanastasiou G, Semple S, Williams M, Newby D et al. , “Factorised spatial representation learning: application in semi-supervised myocardial segmentation,” in MICCAI. Cham: Springer, 2018, pp. 490–498. [Google Scholar]
- [22].Meng Q, Pawlowski N, Rueckert D, and Kainz B, “Representation disentanglement for multi-task learning with application to fetal ultrasound,” in PIPPI. Cham: Springer, 2019, pp. 47–55. [Google Scholar]
- [23].Liu S, Gibson E, Grbic S, Xu Z, Setio AAA, Yang J et al. “Decompose to manipulate: Manipulable object synthesis in 3D medical images with structured image decomposition,” arXiv:1812.01737, 2018. [Google Scholar]
- [24].Qin C, Shi B, Liao R, Mansi T, Rueckert D, and Kamen A, “Unsupervised deformable registration for multi-modal images via disentangled representations,” in IPMI. Cham: Springer, 2019, pp. 249–61. [Google Scholar]
- [25].Yang J, Dvornek NC, Zhang F, Chapiro J, Lin M, and Duncan JS, “Unsupervised domain adaptation via disentangled representations: Application to cross-modality liver segmentation,” in MICCAI. Cham: Springer, 2019, pp. 255–63. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [26].Chen C, Dou Q, Jin Y, Chen H, Qin J, and Heng P-A, “Robust multimodal brain tumor segmentation via feature disentanglement and gated fusion,” in MICCAI. Cham: Springer, 2019, pp. 447–56. [Google Scholar]
- [27].Ma L, Jia X, Georgoulis S, Tuytelaars T, and Gool LV, “Exemplar guided unsupervised image-to-image translation,” in ICLR, 2019. [Google Scholar]
- [28].Wu W, Cao K, Li C, Qian C, and Loy CC, “TransGaGa: Geometry-aware unsupervised image-to-image translation,” in CVPR, 2019, pp. 8012–21. [Google Scholar]
- [29].Havaei M, Davy A, Warde-Farley D, Biard A, Courville A, Bengio Y et al. “Brain tumor segmentation with deep neural networks,” Med Image Anal, vol. 35, pp. 18–31, 2017. [DOI] [PubMed] [Google Scholar]
- [30].Joyce T, Chartsias A, and Tsaftaris SA, “Robust multi-modal MR image synthesis,” in MICCAI. Cham: Springer, 2017, pp. 347–55. [Google Scholar]
- [31].van Tulder G and de Bruijne M, “Learning cross-modality representations from multi-modal images,” IEEE Trans Med Imag, vol. 38, no. 2, pp. 638–648, 2019. [DOI] [PubMed] [Google Scholar]
- [32].Dolz J, Gopinath K, Yuan J, Lombaert H, Desrosiers C, and Ayed IB, “HyperDense-Net: A hyper-densely connected CNN for multimodal image segmentation,” IEEE Trans Med Imag, vol. 38, no. 5, pp. 1116–1126, 2019. [DOI] [PubMed] [Google Scholar]
- [33].Tseng K-L, Lin Y-L, Hsu W, and Huang C-Y, “Joint sequence learning and cross-modality convolution for 3D biomedical segmentation,” in CVPR, 2017, pp. 6393–6400. [Google Scholar]
- [34].Chen H, Qi Y, Yin Y, Li T, Liu X, Li X et al. “MMFNet: A multi-modality MRI fusion network for segmentation of nasopharyngeal carcinoma,” Neurocomputing, vol. 394, pp. 27–40, 2020. [Google Scholar]
- [35].Valindria V, Pawlowski N, Rajchl M, Lavdas I, Aboagye E, Rockall A et al. “Multi-modal learning from unpaired images: Application to multi-organ segmentation in CT and MRI,” in WACV, 2018, pp. 547–56. [Google Scholar]
- [36].Liu J, Xie H, Zhang S, and Gu L, “Multi-sequence myocardium segmentation with cross-constrained shape and neural network-based initialization,” Comput Med Imaging Graph, vol. 71, pp. 49–57, 2019. [DOI] [PubMed] [Google Scholar]
- [37].Wei D, Sun Y, Chai P, Low A, and Ong SH, “Myocardial segmentation of late gadolinium enhanced MR images by propagation of contours from cine MR images,” in MICCAI. Berlin, Heidelberg: Springer, 2011, pp. 428–35. [DOI] [PubMed] [Google Scholar]
- [38].Zhuang X and Shen J, “Multi-scale patch and multi-modality atlases for whole heart segmentation of MRI,” Med Image Anal, vol. 31, pp. 77–87, 2016. [DOI] [PubMed] [Google Scholar]
- [39].Zhuang X, “Multivariate mixture model for myocardial segmentation combining multi-source images,” IEEE Trans Pattern Anal Mach Intell, vol. 41, no. 12, pp. 2933–2946, 2019. [DOI] [PubMed] [Google Scholar]
- [40].Wang X, Yang S, Tang M, Wei Y, Han X, He L et al. “SK-Unet: an improved u-net model with selective kernel for the segmentation of multi-sequence cardiac MR,” in STACOM. Cham: Springer, 2020, pp. 246–53. [Google Scholar]
- [41].Roth H, Zhu W, Yang D, Xu Z, and Xu D, “Cardiac segmentation of LGE MRI with noisy labels,” in STACOM. Cham: Springer, 2020, pp. 228–36. [Google Scholar]
- [42].Ronneberger O, Fischer P, and Brox T, “U-Net: Convolutional networks for biomedical image segmentation,” in MICCAI. Cham: Springer, 2015, pp. 234–41. [Google Scholar]
- [43].Mao X, Li Q, Xie H, Lau RYK, Wang Z, and Smolley SP, “On the effectiveness of least squares generative adversarial networks,” IEEE Trans Pattern Anal Mach Intell, vol. 41, no. 12, pp. 2947–2960, 2019. [DOI] [PubMed] [Google Scholar]
- [44].Miyato T, Kataoka T, Koyama M, and Yoshida Y, “Spectral normalization for generative adversarial networks,” ICLR, 2018. [Google Scholar]
- [45].Guo C, Pleiss G, Sun Y, and Weinberger KQ, “On calibration of modern neural networks,” in ICML, 2017, pp. 1321–1330. [Google Scholar]
- [46].Izmailov P, Podoprikhin D, Garipov T, Vetrov D, and Wilson AG, “Averaging weights leads to wider optima and better generalization,” in UAI, 2018, pp. 876–85. [Google Scholar]
- [47].Stirrat CG, Alam SR, MacGillivray TJ, Gray CD, Dweck MR, Raftis J et al. , “Ferumoxytol-enhanced magnetic resonance imaging assessing inflammation after myocardial infarction,” Heart, vol. 103, no. 19, pp. 1528–1535, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [48].Kavur AE, Gezer NS, Bari M, Conze P-H, Groza V, Pham DD et al. “CHAOS Challenge–Combined (CT-MR) Healthy Abdominal Organ Segmentation,” arXiv:2001.06535, 2020. [DOI] [PubMed] [Google Scholar]
- [49].Tsaftaris SA, Zhou X, Tang R, Li D, and Dharmakumar R, “Detecting myocardial ischemia at rest with cardiac phase–resolved blood oxygen level–dependent cardiovascular magnetic resonance,” Circ Cardiovasc Imaging, vol. 6, no. 2, pp. 311–19, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [50].Tustison NJ, Yang Y, and Salerno M, “Advanced normalization tools for cardiac motion correction,” in STACOM. Cham: Springer, 2015, pp. 3–12. [Google Scholar]
- [51].Oksuz I, Clough J, Ruijsink B, Puyol-Antón E, Bustin A, Cruz G et al. , “Detection and correction of cardiac mri motion artefacts during reconstruction from k-space,” in MICCAI. Cham: Springer, 2019, pp. 695–703. [Google Scholar]
- [52].Burgess CP, Higgins I, Pal A, Matthey L, Watters N, Desjardins G et al. , “Understanding disentangling in β-vae,” NIPS Workshop on Learning Disentangled Representations, 2017. [Google Scholar]
- [53].Vepakomma P, Gupta O, Dubey A, and Raskar R, “Reducing leakage in distributed deep learning for sensitive health data,” ICLR, 2019. [Google Scholar]
- [54].Do K and Tran T, “Theory and evaluation metrics for learning disentangled representations,” ICLR, 2020. [Google Scholar]