Skip to main content
Elsevier Sponsored Documents logoLink to Elsevier Sponsored Documents
. 2020 Feb 1;206:116324. doi: 10.1016/j.neuroimage.2019.116324

An automated framework for localization, segmentation and super-resolution reconstruction of fetal brain MRI

Michael Ebner a,b,∗,1, Guotai Wang c,a,b,∗∗,1, Wenqi Li d,b, Michael Aertsen e, Premal A Patel f,a, Rosalind Aughwane g,a, Andrew Melbourne b,h, Tom Doel a, Steven Dymarkowski e, Paolo De Coppi i, Anna L David g,a,j, Jan Deprest j,g,a, Sébastien Ourselin b, Tom Vercauteren b,a,j
PMCID: PMC7103783  PMID: 31704293

Abstract

High-resolution volume reconstruction from multiple motion-corrupted stacks of 2D slices plays an increasing role for fetal brain Magnetic Resonance Imaging (MRI) studies. Currently existing reconstruction methods are time-consuming and often require user interactions to localize and extract the brain from several stacks of 2D slices. We propose a fully automatic framework for fetal brain reconstruction that consists of four stages: 1) fetal brain localization based on a coarse segmentation by a Convolutional Neural Network (CNN), 2) fine segmentation by another CNN trained with a multi-scale loss function, 3) novel, single-parameter outlier-robust super-resolution reconstruction, and 4) fast and automatic high-resolution visualization in standard anatomical space suitable for pathological brains. We validated our framework with images from fetuses with normal brains and with variable degrees of ventriculomegaly associated with open spina bifida, a congenital malformation affecting also the brain. Experiments show that each step of our proposed pipeline outperforms state-of-the-art methods in both segmentation and reconstruction comparisons including expert-reader quality assessments. The reconstruction results of our proposed method compare favorably with those obtained by manual, labor-intensive brain segmentation, which unlocks the potential use of automatic fetal brain reconstruction studies in clinical practice.

Keywords: Fetal MRI, Deep learning, Super resolution, Convolutional neural network, Brain localization, Segmentation

1. Introduction

Fetal Magnetic Resonance Imaging (MRI) has become increasingly important in prenatal diagnosis as a complementary tool to ultrasound, for its advantages in demonstrating pathologies in soft tissues, that may not be apparent or cannot be accurately characterized on prenatal ultrasonography. To mitigate the effect of fetal (and maternal) motion, fast imaging methods such as Single-Shot Fast Spin Echo (SSFSE) are used to acquire thick, low-resolution stacks of 2D slices that can largely freeze in-plane motion (Saleem, 2014). With motion commonly occurring in between slice acquisitions, this generally results in motion-corrupted stacks of slices in multiple orientations with poor 3D image integrity and resolution. In order to assess and quantify fetal brain development and pathology, it is highly desirable to reconstruct a single isotropic, high-resolution volume of the fetal brain in standard anatomical planes from multiple low-resolution stacks acquired in different views.

Currently existing reconstruction toolkits generally rely on an approach that iteratively operates motion correction and super-resolution reconstruction (SRR) (Rousseau et al., 2006; Gholipour et al., 2010; Kuklisova-Murgasova et al., 2012; Kainz et al., 2015b). Since the position and orientation of the fetal brain vary significantly between different patients in relation to maternal structures, localizing the fetal brain and obtaining a segmented mask to exclude the surrounding tissues is crucial to achieve accurate motion correction. Current motion-correction approaches typically employ rigid registration with the assumption that the brain has rigid and surrounding tissues non-rigid motion patterns. Thus, localization and segmentation can help to clearly delineate the brain region so that rigid motion correction becomes meaningful. At present, this usually requires manual localization of the fetal brain and uses manual or semi-automatic methods to obtain fetal brain masks, which is laborious and time consuming.

With localized or segmented fetal brain masks, multiple motion-corrupted stacks of low-resolution 2D slices can be reconstructed into a single high-resolution 3D brain volume. There are two main challenges associated with the high-resolution volume reconstruction step. First, the inter-slice motion can lead to inconsistent appearance in neighboring slices. This is mainly due to the fact that the SSFSE sequence acquires fetal MR images in an interleaved fashion to reduce the scan time and avoid slice cross-talk artifacts (Gholipour et al., 2014). An M-interleaved scanning leads to M sub-stacks that are temporally sequential but spatially interleaved, where M is usually set as 2 or 3. The motion pattern within each sub-stack is relatively consistent and smooth while that between sub-stacks can be inconsistent, as shown in Fig. 1(b). Moreover, motion during image acquisition can lead to various types of artifacts such as in-plane image blur, slice crosstalk and spin-history artifacts that can considerably affect the image quality of individual slices (Gholipour et al., 2014) as visualized in Fig. 1(c)–(e). Second, robustness against outlier slices characterized by either misregistration or image artifacts is key for a high-fidelity high-resolution reconstruction framework (Gholipour et al., 2010; Kuklisova-Murgasova et al., 2012). However, for previously presented approaches, no complete outlier slice rejection is achieved in Gholipour et al. (2010), and the method of Kuklisova-Murgasova et al. (2012); Kainz et al. (2015b) relies on multiple hyperparameters to be tuned in order to achieve optimal reconstructions while both require time-consuming optimization methods due to their resulting non-convex problem formulation.

Fig. 1.

Fig. 1

Three example stacks of MRI of fetuses with spinal bifida (a)–(c) with gestational ages of 24, 24 and 29 weeks, respectively. (a) Has a consistent appearance with small inter-slice motion. (b) Shows motion between two interleaved sub-stacks. (c) Illustrates artifact-affected slices with two such ‘outlier’ slices shown in (d) and (e).

Previous reconstruction frameworks have shown the potential benefit of estimating high-resolution 3D visualizations of the fetal brain (Rousseau et al., 2006; Gholipour et al., 2010; Kuklisova-Murgasova et al., 2012; Kainz et al., 2015b). However, a larger cohort study has not yet been performed to demonstrate their effectiveness on pathological cases where high-resolution 3D brain MRI reconstructions could add important information for more accurate diagnosis and clinical management. One indication for fetal MRI is spina bifida, where MRI plays a role in characterizing the spinal lesion as well as the associated brain changes (Ovaere et al., 2015; Aertsen et al., 2019). In open spina bifida (myelomeningocoele and myeloschisis), a fault in the development of the spinal cord and surrounding vertebrae leaves a gap in the spine, allowing the spinal cord and nerve tissue to bulge through a defect on the fetus's back. Because of a suction gradient by leakage of cerebrospinal fluid at the lesion, the hindbrain descends through the base of the skull where the spinal cord exits (a condition termed a Chiari II malformation). This may be associated with excessive accumulation of fluid in the brain ventricles (ventriculomegaly), as shown in Fig. 2. In these cases, high-resolution 3D reconstructions would aid more accurate measurements, currently performed on low-resolution 2D stacks (Aertsen et al., 2019), and help characterize associated brain changes, ruling out those that are prognostically important.

Fig. 2.

Fig. 2

Comparison of a normal fetus and a fetus with open spina bifida showing a Chiari II malformation with ventriculomegaly. Image courtesy of UZ Leuven.

Furthermore, high-resolution visualization of the brain in standard radiological anatomical planes is highly desirable for clinical assessment. The availability of orthogonal planes of the brain in axial, coronal and sagittal views are of utmost importance for the quality of the examination but are difficult to achieve for fetal MRI in practice due to unpredictable fetal motion (Prayer et al., 2017). True orthogonal planes are vital for adequate measurements and anatomical recognition which improves the detection of pathology, and prevents potential left-right confusion of hemispheres. It also facilitates cross-sectional comparisons of different patient cohorts, and can be especially useful for longitudinal comparisons to visualize anatomical differences of the brain associated with either natural changes or clinical intervention, such as surgery. Thus, template-space reconstructions help to make use of the obtained high-resolution 3D brain visualizations in an optimal way. However, previously presented methods for automatic brain reconstruction in a template space were designed for normal, or mildly pathological, cases only (Tourbier et al., 2017; Gholipour et al., 2017) and suffer from low robustness when dealing with pathological brains such as encountered with spina bifida.

Hence, we hypothesize that a fully automatic reconstruction pipeline based on automatic fetal brain localization, segmentation and robust reconstruction and template-space alignment steps is favorable to achieve efficient and accurate fetal brain reconstructions for potential clinical translation.

This work is a substantial extension of our preliminary conference publication (Ebner et al., 2018), where we presented a novel framework for automatic fetal brain localization, segmentation and reconstruction in the standard radiological anatomical planes. We proposed a novel CNN-based fetal brain segmentation method in a coarse-to-fine fashion to reduce false positives and applied the proposed framework for high-resolution volume reconstruction of MR images of fetuses with spina bifida. Additionally, we introduced an approach for effective complete outlier rejection as alternative to the methods in Gholipour et al. (2010); Kuklisova-Murgasova et al. (2012), that relies on a single hyperparameter only and retains a linear least-squares formulation which can be solved efficiently.

In this paper, we give a more detailed description of the framework and make substantial extensions in three aspects. First, we propose an additional fast template space alignment method that is robust to large brain morphology changes such as encountered in spina bifida. Second, we further show the superiority of our automatic localization, segmentation and reconstruction methods by comparing them with different variants and state-of-the-arts. Third, we validate the proposed framework with a larger dataset of images from both normal fetuses and fetuses with spina bifida (Aertsen et al., 2019). Experimental results show that our framework can achieve comparable reconstruction output to that of manual segmentation-based reconstruction, and it outperforms existing fetal brain extraction and reconstruction methods on different cohorts of fetuses.

2. Related works

2.1. Fetal brain localization and segmentation

Extracting fetal brain from fetal MRI usually serves as a prerequisite step for high-resolution volume reconstruction of the fetal brain. Anquez et al. (2009) proposed an automated fetal brain segmentation method by localizing the fetal eyes and then segmenting the neighboring skull bone content, which can lead to a poor performance when the inter-slice motion is large. Taleb et al. (2013) used a template-based method to generate fetal brain masks. It obtains a region of interest (ROI) based on the intersection of multiple scans of the same patient, and then registers the ROI to an age-specific template. Tourbier et al. (2017) used template-to-slice block matching and deformable slice-to-template registration for automatic fetal brain localization and segmentation. It achieved good performance at the cost of a very long computational time up to several hours. Kainz et al. (2014) proposed to localize the fetal brain by voxel classification using rotation invariant volume descriptors. Keraudren et al. (2013) used bundled Scale-Invariant Feature Transform (SIFT) features to fit a 3D bounding box of the fetal brain where prior knowledge of the fetal brain development was used to define size and shape constraints for robust localization. Keraudren et al. (2014) extended that method for fetal brain segmentation with image-specific online learning based on Random Forests. It is limited by hand-crafted features and inefficiency during inference. Additionally, some deep learning-based object detection methods such as R-CNN (Girshick et al., 2014) and YOLO (Redmon et al., 2016) are designed for detection of thousands of objects from natural 2D RGB images. However, applying them to fetal brain localization from 3D MRI needs further investigation.

Recently, deep learning with CNNs has been used for fetal brain segmentation from fetal MRI. Rajchl et al. (2016) used a fully convolutional neural network (FCN) for fetal brain segmentation under distributed weak supervision. Salehi et al. (2018) used a 2D U-Net (Ronneberger et al., 2015) for slice-by-slice fetal brain segmentation. These methods predict the fetal brain mask directly without a localization step and, compared with previous methods, are more efficient at test time in terms of computational time. However, they can easily cause false positives and show poor performance for challenging cases.

2.2. Fetal brain reconstruction

Rousseau et al. (2006) proposed a slice-to-volume registration (SVR) method for fetal brain reconstruction based on semi-automatic segmentation results. It consisted of three steps: motion correction, volume reconstruction and contrast correction. In the motion correction step, each low-resolution stack is globally aligned first followed by a hierarchical slice package motion correction approach based on the temporal slice interleave. Iterative reconstructions are used as reference for motion correction which were obtained by using scattered interpolation with a narrow Gaussian kernel as the point spread function (PSF). A contrast correction step is used to correct the local relative intensity distortion between the low-resolution stacks. Jiang et al. (2007) used multilevel scattered B-spline interpolation for the reconstruction task that requires sufficient samples to allow full representation of the structure to be reconstructed. Kim et al. (2010) proposed a reconstruction-free registration approach that relies on a slice intersection motion correction (SIMC) method that directly co-aligns multiple stacks of 2D slices which was followed by a single Gaussian-weighted averaging step for the volumetric reconstruction. Gholipour et al. (2010) formulated the volumetric reconstruction step as a super-resolution reconstruction problem that allowed a minimum error representation of the obtained high-resolution volume, whereby a slice acquisition model was used. For the super-resolution problem, it used a robust M-estimation formulation that minimizes a Huber's error function to reduce the influence of potential outliers. Kuklisova-Murgasova et al. (2012) built on this idea and proposed a reconstruction method with complete outlier rejection that can entirely exclude identified misregistered or corrupted voxels and slices using expectation-maximization (EM)-based robust statistics. Additional intensity matching was used to compensate inconsistent scaling factors and bias fields of acquired slices. Kainz et al. (2015b) developed a GPU-accelerated implementation of Kuklisova-Murgasova et al. (2012), and proposed to automatically select the stack with least motion as the reference stack. Tourbier et al. (2017) proposed a fully automated reconstruction pipeline including template-space alignment step for the high-resolution visualization in the standard radiological anatomical planes, but presented gestational age-matching to select the template from the normal brain atlas for healthy and mildly pathological cases only. Hou et al. (2017) utilized CNNs to predict the initial transformation parameters of SVR in the motion correction step. McDonagh et al. (2017) proposed a context-sensitive upsampling method based on CNNs to improve the resolution of each low-resolution stack, and then used the upsampled low-resolution stacks as the inputs of an SVR-based 3D reconstruction method. Alansary et al. (2017) proposed a patch-to-volume registration (PVR) framework to reconstruct the whole uterus by splitting each slice into smaller patches used for rigid motion correction. However, apart from the much higher computational requirements, this leads to overall non-rigid motion correction and, thus, suboptimal outcomes for rigidly moving regions such as the fetal brain. Hou et al. (2017, 2018) utilized CNNs to predict the initial transformation parameters of SVR in the motion correction step to achieve more robust initialisations for the slice-to-volume registration step.

Thus, the existing methods have either focused on automatic segmentation of the fetal brain without demonstrating their utility in the context of automatic fetal brain reconstruction for a larger patient cohort, or investigated the reconstruction problem using manually or semi-automatically segmented fetal brain masks. Moreover, there are no appropriate studies showing the performance of pipelining independently developed methods. Therefore, to the best of our knowledge, existing works have not fully solved all these issues which are vital for a robust, consistent, fully-automated reconstruction framework that allows for clinical translation.

3. Methods

An overview of our proposed fully automatic framework for fetal brain reconstruction is depicted in Fig. 3. We first use a CNN to automatically localize the fetal brain region in each input low-resolution stack and obtain a 3D bounding box of the fetal brain. Within the bounding box, we use another CNN to automatically generate a fine segmentation of the fetal brain. The automatic high-resolution volume reconstruction stage includes the two-step iterative SVR and outlier-robust SRR step followed by a fast and robust standard anatomical template space alignment step. For the outlier-robust SRR, we propose a novel outlier rejection method by defining a similarity measurement to remove outlier slices and frame the SRR problem as a linear least-squares formulation that can be solved efficiently. For the template-registration step, we propose a rigid registration approach based on symmetric block-matching between the SRR and a brain-volume-matched template that is initialized by the rigid alignment of the respective principal brain axes (PBA). The three stages of automatic localization, segmentation and reconstruction are detailed in Sections 1, 2, 3.3, respectively. All implementations are available as open-source packages.2

Fig. 3.

Fig. 3

The proposed fully automatic framework for fetal brain MRI reconstruction to obtain high-resolution (HR) visualizations in standard anatomical planes from multiple low-resolution (LR) input stacks. The automatic localization, segmentation and reconstruction parts are detailed in Figs. 4, Figs. 5 and 6, respectively.

3.1. Localization based on coarse segmentation

Differently from traditional top-down object localization methods using sliding window classification (Criminisi et al., 2009) or bounding box regression (Gauriau et al., 2015; He et al., 2017), we use a bottom-up strategy for fetal brain localization based on a coarse segmentation by a CNN with pixel-level prediction. The advantage of such a localization method is that it gives an explainable support for the localization result and is well-suited for single-object localization. To reduce computational requirements, we apply the CNN at a down-sampled version of an input low-resolution stack, as shown in Fig. 4. We refer to this network for the fetal brain localization task as Loc-Net.

Fig. 4.

Fig. 4

The proposed fetal brain localization method using a CNN (Loc-Net) to obtain a coarse segmentation followed by 3D bounding box fitting.

The proposed framework is theoretically amenable to different CNN models. However, as the input low-resolution stack has a large inter-slice spacing and is potentially corrupted by motion between neighboring slices, it is more compelling to use a 2D CNN than a 3D CNN. We choose the 2D P-Net (Wang et al., 2018) for its compactness and efficiency. It consists of six blocks of convolution layers with dilated convolution (Yu and Koltun, 2016) to preserve resolution for dense prediction. The first five blocks have 2, 2, 3, 3 and 3 convolution layers respectively and they have dilation parameters of 1, 2, 4, 8 and 16, respectively. The convolution channel number for these layers is 64. Multi-scale features from these five blocks are concatenated and fed into the 6-th block which is a voxel-wise classifier with 1 × 1 convolution. A softmax layer is used to obtain probability-like outputs.

Let I denote a stack of slices and Ii represent the i-th 2D slice of I. To reduce the inference time and memory consumption for the localization task, we down-sample Ii to a given size, i.e., 96×96, obtaining I'i. We keep the through-plane resolution the same as the input stack, and use I' to denote the stack of down-sampled 2D slices. As shown in Fig. 4, to get a 3D bounding box of the fetal brain in a stack, we first use the Loc-Net to obtain a segmentation of I' by stacking the 2D segmentations, i.e., a coarse segmentation.

With the coarse segmentation of the down-sampled stack I', we employ two post-processing steps to reduce segmentation noise and obtain a smoother result. First, we use a 3D morphological closing and opening operation on the result of Loc-Net. Then we select the largest connected 3D component as the post-processed coarse segmentation of the fetal brain and fit a 3D bounding box to the component as the localization result in I'. The final localization result for input I is obtained by rescaling the bounding box to the original space of I and expanding by a margin of 5 mm.

3.2. Fine segmentation

After the localization, we further obtain a fine segmentation of the fetal brain from I with a second CNN that is referred to as Seg-Net. It works on the ROI of the localization result of I to reduce false positives of the dense prediction. Similar to the localization step, we use the 2D P-Net structure (Wang et al., 2018) for the fine segmentation rather than a 3D network considering the inter-slice spacing and motion.

Due to the change in appearance of the fetal brain at different gestational ages and as a consequence of the presence of pathologies such as spina bifida, it is challenging to achieve robust segmentation results. We propose a multi-scale loss function for training to improve the performance of fine segmentation. The commonly adopted logistic loss and Dice loss functions for image segmentation use a sum of pixel-wise losses (Sudre et al., 2017) and only penalize prediction errors at the finest scale, without considering the relationship between neighboring pixels at a larger scale. This potentially leads to noisy and spatially inconsistent segmentations. In contrast, dealing with the image in the scale-space representation helps to achieve more robust results, as shown by previous works inspired by the scale-space theory (Lindeberg, 1994; Hu et al., 2018).

We propose a training loss function across multiple scales as depicted in Fig. 5. Let Y represent the pixel-wise probability prediction of an image given by a segmentation CNN and G denote the corresponding pixel-wise probabilistic ground truth. The loss function l(Y,G) measures the similarity between Y and G and guides the network to obtain a segmentation as close as possible to the ground truth. It is commonly defined as a pixel-wise function for segmentation tasks. For example, the Dice loss function is defined as (Milletari et al., 2016; Sudre et al., 2017):

lDice(Y,G)=12iNyigiiNyi2+iNgi2 (1)

where N is the number of pixels. yi represents the probability of pixel i being the foreground given by Y and gi represents that probability given by G. Let s be a scale index (s{1,2,}), and Ys and Gs be the downscaled versions of Y and G at scale s, respectively. Then the proposed multi-scale loss function is

L(Y,G)=1Ss=1Sl(Ys,Gs), (2)

where S is the total number of scales. Thus, the loss function L(Y,G) is the average of l(Ys,Gs) across multiple scales. When s=1, Ys is the same as Y, and when s>1, Ys is a downscaled version of Y. Ys can be obtained by down-sampling Y or applying max-pooling on Y. However, both methods cause the obtained Ys to contain little contextual information. In contrast, Gaussian smoothing and average-pooling summarize the prediction of a local patch for more contextual information. Since average-pooling is more efficient and more straightforward to implement than Gaussian smoothing, we use average-pooling for the downscaling. Let Plavg() denote the average-pooling operation. We set the pooling kernel size as 2×2 with stride size 2×2. Therefore, Plavg() averages every neighboring 2×2 pixels. Plavg() is used recursively to down-scale Y and G for larger scales s:

Ys={Yfor ​s=1Plavg(Ys1)for ​s>1 (3)

With a larger s, Ys and Gs encode the prediction and the ground truth at a higher level with more non-local information. Therefore, L(Y,G) not only penalizes the pixel-wise difference between Y and G, but also encourages their similarity at multiple non-local scales. In this paper, we use the Dice loss function as the loss function at each scale due to its good performance in dealing with imbalanced classes, i.e., l(Ys,Gs)=lDice(Ys,Gs). We set the total number of scales S as 4, as shown in Fig. 5.

Fig. 5.

Fig. 5

The proposed fetal brain segmentation method using a CNN (Seg-Net) that works on the localization result. We propose to use a multi-scale loss function to train Seg-Net.

3.3. Robust high-resolution volume reconstruction framework

The steps of the high-resolution volume reconstruction stage are shown in Fig. 6. We briefly list them here and further detail the main contributions in Sections 3.3.1, 3.3.2. For a set of low-resolution stacks of 2D slices acquired in multiple orientations, we preprocess the images using the bias field correction method (Tustison et al., 2010). Using a volume-to-volume registration based on symmetric block-matching (Modat et al., 2014), all stacks are rigidly aligned with an automatically chosen target stack (more details in 4.2). Based on the brain segmentation of the target stack, all remaining, volumetrically-aligned, stacks are intensity corrected using a linear regression with the masked target stack voxel intensities serving as reference values. An initial high-resolution volume is obtained by applying a scattered data approximation (SDA) scheme on the low-resolution stacks that uses an efficient discrete implementation of Nadaraya-Watson kernel regression (Vercauteren et al., 2006; Ebner et al., 2017). It is based on nearest neighbor sampling onto a regular grid followed by a subsequent Gaussian blurring operation for each single slice. Similarly, SDA is used to obtain a brain mask high-resolution volume from the individual low-resolution stack masks. Then, an updated high-resolution volume is obtained through a two-step iterative registration-reconstruction approach (Rousseau et al., 2006; Gholipour et al., 2010). In each iteration, the rigid registration step registers the slices to the high-resolution volume constructed in the previous iteration for motion correction constrained by the respective slice and high-resolution brain mask. Subsequently, the reconstruction step constructs a high-resolution volume and brain mask from the aligned slices and segmentations, respectively. After reconstruction in the subject's space, we rigidly align the high-resolution volume to a spatiotemporal atlas of normal brains (Gholipour et al., 2017) to obtain the reconstruction in the standard anatomical planes.

Fig. 6.

Fig. 6

The proposed outlier-robust high-resolution volume reconstruction method for fetal brain MRI. As part of a two-step motion-correction/volumetric reconstruction cycle, we propose an effective robust SRR method for complete outlier rejection that relies on a single hyperparameter only and retains a linear least-squares formulation. A fast template-space alignment, which is robust also for pathological brains, is achieved by using a principal brain axes (PBA)-initialized rigid volume-to-template registration based on symmetric block-matching.

3.3.1. Outlier-robust super-resolution reconstruction

After each SVR step of the two-step registration-reconstruction iteration i, an SRR step is used to recover the most likely high-resolution volume xi that satisfies the slice acquisition model (Gholipour et al., 2010)

yki=Akixi+eki (4)

where yki is the k-th slice in a stack. Aki represents the image acquisition process including rigid transformation, slice selection, blurring according to the PSF, and down-sampling. eki denotes the vector of observed noise. The intensity of each voxel in a low-resolution slice is therefore influenced by a certain neighborhood of this voxel within a high-resolution volume x given by the assumed PSF that is specific to the slice profile of the MR acquisition (Liang and Lauterbur, 2000). For SSFSE sequences, a common approximation is given by a slice-aligned 3D Gaussian function that depends on the in- and through-plane resolution of the low-resolution slice (Jiang et al., 2007; Kuklisova-Murgasova et al., 2012). The position and orientation of the slice (and PSF) within the volume is estimated in the rigid SVR step. In order to prevent misregistered or artifact-corrupted outlier slices from affecting the reconstructed high-resolution volume, we propose a robust SRR with outlier rejection in a maximum a-posteriori (MAP) formulation:

xi:=argminx0(kKβi12ykiAkix22+α2||x||22) (5)

where α0 denotes a regularization parameter, the differential operator, and Kβi a set of indices of inlier slices

Kβi:={1kK:Sim(yki,Akixi1)β} (6)

that demonstrate high agreement with their simulated counterparts projected from the previous SRR iterate using a similarity measure Sim and a threshold β>0. Thus, slices with a value of Sim() lower than β are regarded as outliers and rejected in the SRR step. More complex SRR models have been proposed in addition to the MAP formulation including modifying (5) to rely on robust M-estimator (Gholipour et al., 2010) and total variation formulations (Tourbier et al., 2015). However, while they substantially increase the computational cost, in our experience, they tend to show little improvement in the obtained reconstruction quality in case of appropriate motion correction of SSFSE-like sequences (Ebner et al., 2019). Assuming a fixed Kβi, we obtain a convex SRR problem with complete outlier rejection in a linear least-squares formulation which we solve using matrix-free operations (Diamond and Boyd, 2015; Ebner et al., 2017). We use a dedicated linear least-squares solver to deal with this large linear system whereby positivity is enforced by clipping negative values.

Furthermore, we create an high-resolution brain mask by applying the fast SDA approach on the motion-corrected inlier slice masks which is used for both motion correction and the labelling of inlier-slices in (6) in the subsequent iteration.

3.3.2. Reconstruction in standard radiological anatomical planes

Obtaining the high-resolution fetal brain reconstructions in the standard radiological anatomical planes can facilitate brain studies and is typically favored for the clinical assessment by clinicians. To define the template space, we deployed the recently presented spatiotemporal atlas3 (Gholipour et al., 2017) which was constructed from 81 normal fetuses scanned between 19 and 39 weeks of gestation. Rigid registration can be used to align the subject-space SRR to a template. However, given the substantial morphological differences in brain volume and shape between pathological and normal fetuses, a direct registration approach is likely to get stuck in local minima leading to an incorrect template space alignment. To avoid this problem, we propose to use an initialized transformation that is based on the rigid alignment of fetal brain masks only. Using principal component analysis, we first rigidly align the principal brain axes (PBA) of the template and high-resolution brain masks whereby the template is selected based on brain-volume matching. Following the PBA-based initialization, we perform a 3D rigid registration based on block-matching (Modat et al., 2014). For increased robustness, all four permutations of the right-handed bases of principal eigenvectors are tested and the best registration transform is selected as determined by the normalized mutual information similarity between warped SRR and template.

After the 3D rigid registration, we use an additional SRR step to obtain the high-resolution volume in the template space, considering that the resampling process during the 3D rigid registration may affect the image quality.

4. Experiments and results

4.1. Data

The automatic reconstruction framework was applied to routinely acquired clinical images of fetuses with normal brains, yet scanned for other anomalies, and fetuses with spina bifida (SB) that were scanned at University Hospitals KU Leuven between March 2011 and August 2016 as reported in Aertsen et al. (2019). Access to anonymized images was facilitated through the GIFT-Cloud platform for data sharing (Doel et al., 2017). For normal fetuses, 134 stacks from 37 individuals were scanned at the gestational age (GA) of 27.30±4.11 weeks (“normal” group A). 32 fetuses with spina bifida were scanned before fetal surgical closure at a GA of 23.06±1.64 weeks (“pre-surgery” group B1), and 16 of them were additionally scanned after fetal surgical closure at a GA of 25.69±1.21 weeks (“post-surgery” group B2). Details of the dataset are summarized in Table 1. The distribution of the GAs for the experimental data set is shown in Fig. 8. Due to a local 1.5T scanner upgrade from Siemens Sonata to Siemens Aera in the hospital during the scanning period, two different scanners were involved in our study but the acquisition protocol remained unchanged with identical parameter settings. For each study, 3 to 9 SSFSE stacks in at least three different orientations were collected with pixel size 0.39 mm–1.48 mm and slice thickness 2.50 mm–4.40 mm. All images were acquired with no slice overlap nor gap using an echo time of 133 ms and a repetition time of 1000 ms. For the purpose of testing the robustness of our proposed framework, we used all available SSFSE stack acquisitions and therefore also kept heavily motion- and artifact-corrupted stacks and also images where brains were only partially scanned.

Table 1.

Information of the datasets used for experiments. The number of available subjects and stacks are listed for training, validation and testing, respectively. Gestational age (GA) is stated as mean and standard deviation.

Group A Group B1 Group B2
Pathology Normal Spina bifida (pre-surgical) Spina bifida (post-surgical)
Subjects (26, 4, 7) (12, 4, 16) (0, 0, 16)
Stacks (78, 12, 44) (36, 12, 119) (0, 0, 105)

Fig. 8.

Fig. 8

Distribution of gestational age in the experimental fetal image set.

For the fetal brain detection and segmentation set-up, 78 stacks of 26 patients from Group A and 36 stacks of 12 patients from Group B1 were used for training, and 12 stacks of 4 patients from Group A and 12 stacks of 4 patients from Group B1 were used for validation. The remaining images were used for testing, as shown in Table 1. Manual segmentations of the fetal brains on the 2D slices were used as the ground truth for the segmentation task, and the bounding box of the manual segmentation was extended by 5 mm to be used as the ground truth for fetal brain localization. We normalized the intensity of each stack by its mean and standard deviation.

4.2. Implementation details

Our CNNs were implemented in TensorFlow4 using NiftyNet5 (Li et al., 2017; Gibson et al., 2018). For the training of Seg-Net, we set the number of scales S to 4 in (2) and employed Dice loss as the loss function used in each individual scale. The detection/segmentation experiments were implemented with an NVIDIA GeForce GTX 980 GPU. For both Loc-Net and Seg-Net, we used Adaptive Moment Estimation (Adam) (Kingma and Ba, 2015) for training, with initial learning rate 103, batch size 10, weight decay 107 and 10k iterations at which point the performance on the validation set stopped to increase.

The high-resolution volume reconstruction algorithm was applied to the testing data of Group A (normal), B1 (pre-surgical spina bifida) and B2 (post-surgical spina bifida) cases using the obtained automatic segmentation results to guide the rigid SVR step. All stacks were preprocessed using the bias field correction method N4ITK described in Tustison et al. (2010) by using the implementation of Insight Toolkit (ITK).6 The reconstruction was performed with three cycles of the two-step motion correction and robust volumetric reconstruction iterations, which was found to be sufficient for algorithmic convergence (Inline Supplementary Fig. S3). For automatic target stack selection, we empirically chose the stack with (estimated) brain volume closest to 70% of the median brain volume using the automatically segmented (Seg-Net) brain masks. This approach was chosen for simplicity to define a target stack that shows good brain coverage but avoids stacks with unrealistically high brain volume estimates due to false-positive segmentations or heavy motion-corruption. However, and as shown in the results section, this may still lead to a target stack with substantial motion artifacts suitable to test the robustness of our algorithm also for challenging cases. The orientation of the subject space is defined by the target stack whereby the reconstruction grid is obtained by extending the bounding box of the union of the volumetrically aligned stack masks by 10 mm in each direction. Based on this target stack, all remaining, volumetrically aligned, stacks were intensity corrected using linear regression where the masked target stack voxel intensities served as reference values. For the respective volume-to-volume (and volume-to-template) rigid registrations we deployed the symmetric block-matching algorithm RegAladin that is based on normalized cross-correlation (NCC) as part of NiftyReg7 (Modat et al., 2014). The implementation of the SDA approach was based on the Young & Van Vliet recursive Gaussian smoothing filter8(Vidal-Migallon et al., 2013). We empirically chose a standard deviation of 1 for both the high-resolution volume and brain mask high-resolution volume SDAs. For the rigid slice-to-volume motion correction steps we used ITK with NCC, whereby both the individual slice mask and the current high-resolution mask volume iterate were used to constrain the registration. Similarly, for the volume-to-template rigid registration we used the final high-resolution mask volume to estimate the PBA-based initialization transform, and to constrain the block-matching-based 3D rigid registration together with the template mask for achieving template-space alignment. Experiments were performed to investigate the sensitivity of the proposed outlier-robust SRR method to the outlier-threshold β and the input fetal brain masks (Inline Supplementary Figs. S1 and S2). By choosing NCC as the similarity measurement function Sim() in (6), a good balance between conservative slice retention and effective outlier rejection was found for β=0.8. For the experiments, we set the threshold value β to be 0.5,0.65 and 0.8 per iteration to account for increasing accuracy in (5), respectively, whereby the slice similarities were evaluated only for the slice-projected high-resolution mask volume voxels. The matrix-free implementation of the forward operator Aki in (5) (and its adjoint) was done by extending the resampling operator in ITK to allow for oriented Gaussian filtering9 representing the oriented PSF kernel whereby the SciPy LSMR algorithm as dedicated linear-least squares solver was used for efficient numerical minimization of (5). The isotropic resolution of the high-resolution volume was set to match the final template space resolution of 0.8 mm (Gholipour et al., 2017) for both subject and template space reconstructions. The regularization parameter α=0.01 was determined by visual assessments supported by L-curve studies (Hansen, 2001). We reconstructed the entire field of view for both subject and template spaces from the brain-motion corrected slices to provide anatomical context beyond the brain.10 Our Python code, including both the automatic brain segmentation tool fetal_brain_seg11 and the outlier-robust SRR framework NiftyMIC,12 is publicly available.

4.3. Localization results

For the choice of network structure of Loc-Net, we compared 2D P-Net with 2D U-Net, 3D P-Net (Wang et al., 2018) and 3D U-Net (Çiçek et al., 2016) to investigate whether 2D or 3D networks are more suitable for uncorrected fetal MR image stacks. We also compared our coarse segmentation-based localization method with a modified YOLO (Redmon et al., 2016) that uses a CNN to predict the 2D coordinate and size of a fetal brain bounding box and the associated foreground probability in each slice directly. We employed the CNN structure used by Redmon et al. (2016) and changed the foreground class number to 1 for our task. These networks were all implemented using NiftyNet. In addition, we compared Loc-Net using these network structures with the method described in Keraudren et al. (2014) 13 that is based on classification of image regions using SIFT features and combined with prior knowledge of brain size and shape based on gestational age.

Fig. 7 shows the fetal brain localization results for three cases from Group A, B1 and B2 respectively. In the first case, the centroid of the bounding box obtained by Keraudren et al. (2014) is close to that of the ground truth. However, the size of the localization result is larger than that of the ground truth. In contrast, the result of our Loc-Net with 2D P-Net matches better with the ground truth. It can also be observed that the result of 2D P-Net is better than that of the other three networks. In the second case, the in-plane visualization shows that similar results are achieved by the 2D and 3D networks. However, the through-plane visualization shows that 2D P-Net and 2D U-Net outperform their 3D counterparts.

Fig. 7.

Fig. 7

Visual comparison of different methods for fetal brain localization. The three rows show examples from Group A (controls), B1 (pre-surgical spina bifida), and B2 (post-surgical spina bifida), respectively. Column 1–6: in-plane. Column 7–12: through-plane. Yellow: ground truth. Green: detection results.

Quantitative evaluations of these localization methods are shown in Fig. 10. We calculated the Intersection over Union (IoU) score and centroid distance between the localized 3D bounding box and the localization ground truth. Fig. 10 shows that 2D P-Net outperforms 2D U-Net when used as Loc-Net, and both of them achieve better localization accuracy than their 3D counterparts. Our Loc-Net with 2D P-Net achieved average IoUs of 86.54%, 84.74% and 83.67% for these the groups of fetuses, respectively, and it outperformed Keraudren et al. (2014) and the modified YOLO. The stack-level runtime of our proposed localization method was 2.35 ± 1.02s, and the corresponding time of Keraudren et al. (2014) was 15.03 ± 3.54s.

Fig. 10.

Fig. 10

Quantitative evaluation of different methods for fetal brain localization.

4.4. Segmentation results

We compared using 2D P-Net and 2D U-Net as the Seg-Net. As a baseline, both of them were trained with the Dice loss with a scale s=1, i.e. at the input resolution. Then we trained these networks with the proposed multi-scale (ML) Dice loss. These four variants are referred to as Seg-Net (2D P-Net), Seg-Net (2D U-Net), Seg-Net (2D P-Net) ​+ ​ML, Seg-Net (2D U-Net) ​+ ​ML respectively. All of them take the same localization result of Loc-Net (2D P-Net) as input for a fair comparison. We also compared them with the method of Salehi et al. (2018) that applies 2D U-Net to the whole input image for segmentation without a localization stage. We followed their implementation available onlin e14 and trained the model from scratch with our training images. Fig. 12 shows the effect of the number of scales used in our multi-scale loss function on the segmentation performance. The results are based on validation images from Group A and B1. We found that the segmentation performance was improved when the scale number increased from 1 to 4. Using 5 scales did not lead to additional improvement in the segmentation accuracy. The training of the P-Net using single-scale and multi-scale loss functions took 12k and 10k iterations or early stop when there was no further performance improvement on the validation set, respectively.

Fig. 12.

Fig. 12

Fetal brain segmentation performance obtained by our multi-scale loss function using different number of scales S. The results are based on validation images from Group A and B1.

Fig. 9 presents a visual comparison of the different fetal brain segmentation methods. It shows that the method of Salehi et al. (2018) tends to generate false positives in tissues surrounding the fetal brain. In contrast, the variants of Seg-Net based on localization results achieve more accurate segmentation with reduced false positives. The results of Seg-Net (2D P-Net) and Seg-Net (2D U-Net) have some under-segmentations and unsmoothed contours. By training with the proposed multi-scale loss function, their corresponding results are more spatially consistent and accurate.

Fig. 9.

Fig. 9

Visual comparison of different methods for fetal brain segmentation. The three rows show examples from Group A (controls), B1 (pre-surgical spina bifida), and B2 (post-surgical spina bifida), respectively. Column 1–5: in-plane. Column 6–10: through-plane. Yellow: ground truth. Green: segmentation results.

Fig. 11 shows quantitative evaluations of these segmentation methods for the fetal brain. We calculated the Dice score and Hausdorff distance between the segmentation results and the pixel-level ground truth. It shows that all the Seg-Net variants outperformed Salehi et al. (2018). Seg-Net (2D P-Net) ​+ ​ML achieved average Dice scores of 93.21%, 93.87% and 92.94% for Group A, B1 and B2 respectively, and it significantly outperformed Seg-Net (2D P-Net) that was trained without ML. The total runtime (forward pass time) of our CNN-based localization and segmentation steps was 3.65s ± 1.34s for one stack including preprocessing and image I/O.

Fig. 11.

Fig. 11

Quantitative evaluation of different methods for fetal brain segmentation.

4.5. Outlier-robust SRR results

For the experiments, we computed the high-resolution volume reconstructions using various methods: 1) the automatic localization results based on Loc-Net (2D P-Net), 2) the automatic fine segmentation results obtained by Seg-Net (2D P-Net), and 3) manual segmentation results. These three variants are denoted as SRR (L), SRR (S) and SRR (M), respectively. Additionally, we applied the state-of-the-art toolkit (Kainz et al., 2015a)15 as described in Kainz et al. (2015b) using the manual segmentations as input masks, denoted as Kainz et al. (M).16

All of the 39 cases of the groups A (7), B1 (16) and B2 (16) were used for analysis as at least one of the reconstruction methods provided a successful reconstruction in the subject-space. In Table 2 we demonstrate the effectiveness of the proposed template-space alignment step. For comparison purposes, we also provided the number of successful template-space alignments using NiftyReg, FLIRT17(Jenkinson et al., 2002) and their respective compositions. Overall, only two cases failed at the template-space alignment step for SRR (S) (one each for B1 and B2) for our proposed method whereas all alignments were successful for SRR (M). All seven cases of group A were successfully reconstructed and aligned. Therefore, the success rate of our proposed framework for all the groups was 37/39 (and 39/39 when the template-space alignment is not considered). The poor performance of SRR (L) for the template space alignment step can be explained by the obtained, rectangular-shaped high-resolution brain volume masks, which leads to non-informative PBA-initialization and volume-to-template registration mask constraints. The failed template-space alignments were manually initialized for SRR (S) so that in total 39 cases were available for the following evaluations. Some visual comparisons of the obtained SRR (S) in the template space along with the high-resolution mask reconstructions for different input masks are provided in Fig. 13.

Table 2.

Assessment of the robustness of the proposed template-space alignment approach. The comparison shows the number of successful template space alignments based on a total of 39 cases with 7 normal (group A), 16 pre-surgical and 16 post-surgical spina bifida cases (groups B1 and B2). A template space alignment was considered successful if a correct alignment in the standard anatomical planes was confirmed visually. FLIRT is based on correlation ratio as similarity measure whereas NiftyReg uses symmetric block-matching based on NCC. Generally, NiftyReg achieves a more robust alignment given a sufficiently good initial alignment. Using our proposed principal brain axes (PBA)-initialized block-matching registration step, a very robust template-space alignment without a failure case can be achieved even for pathological brains. The SRR (S) with the overlaid SRR (L)/(S)/(M) brain masks for the failed B1 case is shown in Fig. 13.

SRR (L)
SRR (S)
SRR (M)
A B1 B2 A B1 B2 A B1 B2
NiftyReg 0 0 0 1 0 0 1 0 0
FLIRT 0 0 0 4 0 4 3 0 4
FLIRT PBA-init 0 0 0 7 0 4 4 0 3
NiftyReg PBA-init 2 0 0 7 15 15 7 16 16
Total number of cases 7 16 16 7 16 16 7 16 16

Figure 13.

Figure 13

Comparison of SRR (S) with overlaid SRR (L)/(M)/(S) high-resolution masks obtained using either the manual masks (SRR (M); blue colour), the automatic segmentations by Seg-Net (SRR (S); differences to SRR (M) in green colour) or the localization results by Loc-Net (SRR (L); differences to SRR (M)/(S) in red colour). The respective visualizations of SRR (S) were obtained by reconstructing the entire template-space field of view using the brain-motion corrected slice transformations transformed into the template space. The last row shows the only B1-case that failed in the template-space alignment step for SRR (S), see Table 2; the final alignment was obtained after manual re-initialization of the volume-to-template registration.

Fig. 14 presents a visual comparison of the obtained SRRs for a B1 (pre-surgical spina bifida) and an A (normal) case in the subject space. Despite the challenging target stacks due to intra-stack motion, in-plane blur and intensity artifact corruption, successful reconstructions were obtained using the proposed outlier-robust SRR method. In particular, the high-resolution visualizations for SRR (M) and SRR (S) appear visually almost indistinguishable.

Fig. 14.

Fig. 14

Qualitative comparison of reconstruction methods in the subject space. Visual comparisons of different reconstruction methods for a B1 (left) and an A (right) case where challenging target stacks were (automatically) selected. Additional visualizations associated with the For the group A case (b), additional visualizations are provided to assess the outlier-rejection performance (Fig. 16) and for template space comparisons (Inline Supplementary Fig. S6). Dilated SRR (M) masks were used for visual cropping. SRR (M) without outlier rejection (OR) presents various artifacts. Similarly, the localization masks as used for SRR (L) lead to poor reconstruction outcomes despite the use of outlier rejection. The outlier-robust results SRR (M) and the proposed SRR (S) based on manual and automated brain masks, respectively, provide successful reconstructions and are, visually, almost indistinguishable. Green arrows indicate artifacts in SRR (M) without OR that are eliminated using our proposed OR method. Red arrows show differences between our proposed method and Kainz et al. (M).

In the supplementary material, additional experiments are summarized to investigate the influence of the intensity correction steps (bias field correction and linear intensity correction) on the obtained reconstruction outcomes (Inline Supplementary Fig. S4). The results underline that both the bias field correction and subsequent linear intensity correction steps lead to statistically significant improvements towards more coherent intensity values of the obtained volumetric reconstructions. Furthermore, we investigated the success rate of the proposed template-space alignment method for different template space selections including the volume-matched template (proposed) and the age-matched template (Tourbier et al., 2017). The results indicate a higher robustness for the proposed, volume-matching approach (Inline Supplementary Table S1).

To investigate the performance of our proposed outlier-rejection method, the number of slice rejections relative to the amount of motion present in the stacks was analyzed (Fig. 15). Whereas relatively few slices are rejected for the majority of stacks with estimated little or moderate slice motion, a higher slice-rejection rate can be observed for stacks associated with higher estimated motion. However, a few stacks show a very high number of slice rejections despite a relatively small average motion of non-rejected slices with the most extreme sample shown in Fig. 16. This comparison underlines that the outlier-rejection method is able to successfully detect and reject artifact-corrupted slices while keeping slices with good in-plane quality for the final volumetric reconstruction step. It is worth noting that the stack in Fig. 16 served as the target stack for the SRR algorithm for that case. Regardless of this poor reference, high-quality reconstructions were obtained with their visualizations in subject and template spaces shown in Fig. 14(b) and Inline Supplementary Fig. S6, respectively.

Fig. 15.

Fig. 15

Histogram relating the number of slice rejections with the average slice motion per stack. The mean values of the 2-norm of translation tx,ty,tz (mm) and rotation rx,ry,rz (degree) parameters of the non-rejected slices for each individual stack after the final motion correction iteration i=3 for SRR (S) for all 39 cases are shown. The stack associated with the sample in the upper-left corner is shown in Fig. 16.

Fig. 16.

Fig. 16

Stack associated with the upper-left corner in Fig. 15 showing substantial in-plane artifacts with relatively moderate slice motion for the non-rejected slices. Red crosses mark the slices that were automatically rejected by the proposed outlier-robust SRR (S) algorithm (only the slices covering the brain are shown; additional six, automatically segmented slices outside the brain were successfully rejected too). The NCC slice similarities Sim(yki,Akixi1)<βi, at the time of rejection at iteration i{1,2,3} with (β1,β2,β3)=(0.5,0.65,0.8) are shown in addition. Thus, the outlier-rejection method is able to successfully detect and reject artifact-corrupted slices while keeping slices with good in-plane quality for the final volumetric reconstruction step. It is worth noting that this stack served as the target stack for the SRR algorithm. Successful reconstructions in subject and template spaces for that case are shown in Fig. 14b and Inline Supplementary Fig. S6, respectively.

For quantitative evaluations of the obtained reconstruction outcomes, we used Sim(yki,Akixi) after the final SVR-SRR iteration (i=3) to measure the similarities between the motion-corrected slices of the input low-resolution stacks and their corresponding simulated slices from the reconstructed high-resolution volume. We present structural similarity index measure (Wang et al., 2004) (SSIM) and peak-signal-to-noise ratio (PSNR) measurements for the comparisons here. Alternative similarity measures (NCC; normalized mutual information, NMI; root mean squared error, RMSE; mean absolute error, MAE) were also generated and are presented in Inline Supplementary Fig. S5 for the sake of manuscript conciseness. Fig. 17 illustrates that all methods apart from SRR (S) produce statistically significant differences compared to SRR (M) in terms of measured slice similarities. Thus, SRR (S) and SRR (M) appear of similar volumetric self-consistency as quantified by the similarities between motion-corrected and respectively projected high-resolution volume slices.

Fig. 17.

Fig. 17

Quantitative comparison of different reconstruction methods based on Sim(yki,Akixi) after the final SVR-SRR iteration (i=3) in terms of SSIM and PSNR. A * denotes a significant difference compared to SRR (M) within each group based on Kruskal-Wallis with post-hoc Dunn tests (p<0.05). Thus, SRR (S) and SRR (M) appear of similar volumetric self-consistency as quantified by the similarities between motion-corrected and respectively projected high-resolution volume slices.

In absence of a ground-truth of the high-resolution volume, an additional subjective quality assessment in a clinical context was made. Two pediatric radiologists (MA and PP) assessed all reconstructions in the template-space side-by-side blinded to the reconstruction methods. The high-resolution masks obtained by SRR (M) were used for a consistent visual cropping of the reconstructions. The radiologists gave scores of three metrics on the results: 1) anatomical clarity of the cerebellar structure (CS), the cerebral aqueduct and the interhemispheric fissure (CAIF) and the longitudinal cerebral fissure (LCF) in the range of [0, 4], 2) SRR quality against introduced artifactual structures and edge uncertainty in the range of [0, 2], and 3) radiologists’ preference in the range of [0, 2]. A higher score for each metric indicates a better reconstruction. The evaluation results are summarized in Fig. 18 (a more detailed comparison of the individual scores is provided in the Inline Supplementary Figs. S9 and S10). It shows that SRR (S) and SRR (M) achieved high-quality reconstruction results that are subjectively almost indistinguishable. Moreover, both SRR (S) and SRR (M) were consistently preferred over Kainz et al. (M) by the radiologists which can be explained by the high anatomical clarity and SRR quality achieved by our proposed reconstruction framework. The comparison against Kainz et al. (M) confirms the effectiveness of our proposed outlier-robust SRR framework which is also illustrated in Fig. 14, Fig. 19, Fig. 20 and in the Inline Supplementary Figs. S6–S8.

Fig. 18.

Fig. 18

Summary of clinical evaluation. Two radiologists performed a qualitative assessment of the obtained high-resolution reconstructions regarding anatomical clarity, SRR quality and subjective preference involving 39 cases. A higher score indicates a better outcome. For anatomical clarity scores indicate how well CS, CAIF and LCF are visualized in each image with ratings 0 (structure not seen), 1 (poor depiction), 2 (suboptimal visualization; image not adequate for diagnostic purposes), 3 (clear visualization of structure but reduced tissue contrast; image-based diagnosis feasible), and 4 (excellent depiction; optimal for diagnostic purposes). SRR quality is a combined average score of individual visible artifacts and blur scores with ratings 0 (lots of artifacts/blur) to 2 (no artifact/blur). Radiologists' preference ranks subjectively from the least (0) to the most preferred (2) reconstruction. A * denotes a significant difference compared to SRR (M) based on a Wilcoxon signed-rank test (p<0.05). The results underline that SRR (M)/(S) represent high-quality reconstructions with high anatomical clarity that are visually indistinguishable and were subjectively preferred over Kainz et al. (M) by the two radiologists.

Fig. 19.

Fig. 19

Qualitative comparison of reconstruction methods in the template space. The comparison shows the template space reconstructions of a group B2 subject (post-surgical SB, GA=27 weeks) based on 7 low-resolution input stacks. An original stack (linearly resampled) with resolution of 0.472×3 mm3 is provided for reference. Red arrows show anatomical differences between SRR (S) and Kainz et al. (M).

Fig. 20.

Fig. 20

Qualitative comparison of reconstruction methods in the template space. The comparison shows the template space reconstructions of a group B2 subject (post-surgical SB, GA=26 weeks) based on 4 low-resolution input stacks. An original stack (linearly resampled) with resolution of 0.742×3 mm3 is provided for reference. Green arrows indicate the rejection of the final intensity-artifacted slice of the original stack using the outlier-threshold β=0.85. Red arrows show anatomical differences between SRR (S) and Kainz et al. (M) in direct comparison with the original stack.

For practical purposes, it is important to understand at which input data scenarios the proposed reconstruction framework can produce high-resolution 3D reconstructions with high anatomical fidelity. Using the case that is associated with the highest number of nine available input stacks, six different input data configurations were tested to evaluate the achievable reconstruction quality based on, (i) three approximately sagittal (“3s”, three stacks); (ii) five approximately sagittal (“5s”, five stacks); (iii) three approximately coronal and two approximately sagittal (“3c+2s”, five stacks); (iv) one approximately axial, one approximately coronal, and one approximately sagittal (“1a+1c+1s”, three stacks); (v) one approximately axial, two approximately coronal, and two approximately sagittal (“1a+2c+2c”, five stacks); and (vi) all available data, i.e. one approximately axial, three approximately coronal, and five approximately sagittal (“1a+3c+5s”, nine stacks). A qualitative comparison of the obtained template-space reconstructions using SRR (S) alongside the comparison of their quantitative similarity scores against “1a+3c+5s” (all data) based on NCC, SSIM and RMSE is provided in Fig. 21 (similar comparisons for other cases are shown in Inline Supplementary Figs. S11 and S12). For the quantitative evaluation, only masked voxels associated with the obtained high-resolution volume mask of “1a+3c+5s” were considered. The comparisons underline that at least three stacks in three different orientations are required in order to get reconstructions that show high anatomical detail in all three anatomical planes. In particular, the experiments illustrate that reconstructions based on three approximately orthogonal stacks can lead to better reconstructions compared to using five stacks acquired in only two orientations. Increasing the number of stacks per orientation can further increase the reconstruction quality.

Fig. 21.

Fig. 21

Comparison of obtained reconstructions in the template space for six different input data configurations using the case with the highest number of nine available input stacks (B1 subject, pre-surgical SB, GA=25 weeks). The horizontal axis for the quantitative comparisons is sorted in ascending order based on the NCC outcome, whereby “1a+3c+5s” constrained by its mask was used as reference. Using at least three stacks in three different orientations leads to a high anatomical detail in all three anatomical planes. Increasing the number of stacks per orientation can further increase the reconstruction quality. Additional comparisons for other cases are shown in Inline Supplementary Figs. S11 and S12.

Using our non-optimized implementation on a single computer with four CPUs, the typical processing time for SRR (S) was approximately 13min for the subject-space reconstruction, i.e. the computation of the two-step iterative motion-correction and volumetric reconstruction steps, and approximately 11min for the template-space reconstruction, i.e. the combined template-space alignment and volumetric reconstruction from motion-corrected slices. The reconstruction times for SRR (M) were comparable. For Kainz et al. (M) the subject-space reconstructions took approximately 6min on average.

5. Discussion and conclusion

Our automated pipeline for fetal brain reconstruction in MR imaging benefits from deep learning-based localization and segmentation where a CNN-based coarse segmentation is proposed for robust localization and a multi-scale loss function for a fine segmentation of the fetal brain. Compared with Keraudren et al. (2014), our localization method does not need prior information such as shape and size of the fetal brain and it achieved superior performance in less time. Different from Salehi et al. (2018), which takes a whole image as the input of a CNN, our segmentation method follows a coarse-to-fine approach. The benefit is three-fold: 1) it rejects false positives outside the localization result, 2) the training data for Seg-Net are local regions around the fetal brain and therefore they have less imbalance between the foreground and background, and 3) Seg-Net has lower memory requirements and is more efficient than working on the whole image. However, it requires training of two networks independently, and may be improved by joint training or adopting attention mechanisms (Oktay et al., 2018; He et al., 2017) in the future. Minimizing our proposed multi-scale loss function encourages a segmentation to be close to the ground truth at multiple non-local levels, and helps to obtain more spatially consistent results as shown in Fig. 9. Our testing set consisted of a relatively large dataset of 268 images characterized by a wide range of resolution and appearance differences. Results show that our localization and segmentation method is robust against images with bias field, different patient groups and gestational ages (Fig. 10, Fig. 11). However, it would be of interest to further investigate its ability to generalize to a wider range of MR acquisition parameters and gestational ages (e.g., fetuses in the third trimester).

Moreover, we present an alternative reconstruction framework that includes a novel outlier-rejection method for robust SRR. In contrast to Gholipour et al. (2010) and Kuklisova-Murgasova et al. (2012), our formulation leads to a simple, yet effective, linear least-squares problem with a single hyperparameter whose unique solution can be solved for very efficiently using dedicated least-squares solvers. Despite its simplicity, this outlier-rejection method using a single threshold parameter value was shown to allow for a remarkably robust elimination of outliers for most cases (Fig. 16, Fig. 20). We demonstrate that our proposed outlier-robust framework can produce high-quality high-resolution visualizations from highly heterogeneous and challenging clinical data with results superior to the state-of-the-art toolkit Kainz et al. (M) (Kainz et al., 2015b; Kuklisova-Murgasova et al., 2012). This involved stacks with multiple image resolutions per case including high slice thicknesses between 2.5 mm and 4.4 mm which can be severely affected by substantial motion and intensity artifacts. In particular, we show that high-fidelity reconstructions with clear tissue boundary definitions can be achieved even in case a corrupted target stack is selected (Fig. 14). Anecdotal evidence showed that the target stack selection method proposed by Kainz et al. (2015b), although slower, seems to be slightly better in choosing higher-quality stacks as initial reference than the automatic approach we presented. Even though our reconstruction framework showed good performance for our entire experimental dataset, a higher-quality initial reference may be beneficial for other cases in practice. To ease future work on this, it will be integrated in our open-source reconstruction software framework NiftyMIC.

Limitations of the comparison against Kainz et al. (M) include that the publicly available method19 only allows to specify a single mask for the target stack whereas our method can take advantage of using separate masks for all input stacks. This allows a more accurate motion correction for our method and can therefore lead to higher quality high-resolution reconstructions. A potential exclusion of heavy motion- or artifact-corrupted stacks is likely to improve the SRR quality further for cases that performed less satisfying (e.g. Inline Supplementary Fig. S8). An automatic exclusion criteria could be based on a motion score similar to the one presented in Kainz et al. (2015b). Or, more generally, a stack-specific automatic scoring system could be devised in addition to the outlier-rejection mechanism in order to prioritize stacks based on usability.

Additionally, we present a fast template space alignment method for high-resolution visualizations in the standard radiological anatomical planes that is robust to large brain morphology changes such as encountered in spina bifida. Further robustness to more substantial false-positives in the brain mask high-resolution volume (Fig. 13) could be achieved by using robust principal component analysis (Candès et al., 2011; Parikh and Boyd, 2014) to estimate the principal brain axes. Faster computational times for the high-resolution volume reconstructions can be achieved by more efficient multi-core or GPU implementations including an, in principle, trivially parallelizable computation of the, currently, sequentially performed rigid slice-to-volume registrations. In cases with a large inter-slice motion, the extracted bounding boxes will be larger than bounding boxes that tightly fit brains as captured on non-motion corruption stacks. This can lead to an increased computational time of the subsequent volumetric reconstruction step due to a larger region of interest provided to the algorithm. However, since we use the segmented brain mask for the reconstruction, the final reconstruction quality is less likely to be affected by this. Moreover, and as shown in our experiments (Fig. 14, Fig. 18), our method displays remarkable robustness even in such cases.

We investigated the impact of different input data configurations on our proposed automatic reconstruction pipeline. Based on these experiments, we conclude that at least three approximately orthogonal stacks are required for our SRR framework to obtain high-resolution reconstructions with high anatomical detail in all three anatomical planes. In particular, if only one or two orthogonal orientations are available, the obtained reconstruction quality is generally of inferior quality even if more than three stacks are used (Fig. 21). Using more stacks for each of the three (approximately) orthogonal orientations can further increase the reconstruction quality due to the improved recovery of partial voluming effects using the SRR formulation (5). Similar conclusions were drawn in Ebner et al. (2019) by performing controlled experiments in the context of super-resolution for upper abdominal SSFSE sequences. In our experience, at least five stacks in three approximately orthogonal orientations are desirable for fetal MRI. However, depending on the degree of motion corruption and image artifacts more stacks may be needed due to the potentially higher rate of required slice rejections.

Limitations of this study include that fetal MRI SSFSE sequences were acquired using two different scanners at a single imaging center. It would be of interest to investigate the performance of the proposed automatic segmentation and reconstruction method for fetal images using multiple scanners and imaging centers. In principle, however, the same volumetric reconstruction model appears promising for a range of other applications and anatomies provided a suitable parametrization of the PSF is available to define the slice acquisition model (4) for the used 2D MRI sequence. In particular, the outlier-robust reconstruction framework has shown good results also for other types of sequences such as upper abdominal single-shot T2-weighted (Ebner et al., 2019) and fetal resting-state functional MRI sequences (Sobotka et al., 2019). Similarly, the P-Net has also demonstrated good performance on other structures such as the placenta and the fetal lungs in our previous work (Wang et al., 2018). Beyond structural MRI, it would be of interest to test the applicability of our framework to obtain fully automated reconstructions for functional fetal MRI (Rutherford et al., 2019).

In conclusion, we present a fully automated, and publicly available,18 framework for fetal brain MRI localization, segmentation and super-resolution reconstruction. Our experiments with fetuses with normal brain anatomy as well as fetuses with brain changes associated with spina bifida show that the proposed pipeline produces automatic reconstructions that are comparable to manual segmentation-based reconstructions, therefore, effectively eliminating the need of any manual intervention. In the future, we aim to apply this framework to quantify the impact of spina bifida surgical closure by measuring the resolution of the Chiari type II malformation and the degree of ventriculomegaly.

Declaration of competing interest

WL was employed by King's College London during most of the preparation of this work and was employed by the company Nvidia for the final editing and proofreading of the manuscript. SO is a founder and shareholder of BrainMiner Ltd, UK.

Acknowledgements

This work was supported by the Wellcome Trust [WT101957, WT97914, 203145Z/16/Z, 203148/Z/16/Z, 210182/Z/18/Z], Engineering and Physical Sciences Research Council (EPSRC) [NS/A000027/1, NS/A000049/1, NS/A000050/1, EP/L016478/1], and the National Institute for Health Research University College London Hospitals Biomedical Research Centre (NIHR BRC UCLH/UCL High Impact Initiative). JD is supported by the Great Ormond Street Hospital Charity fund. MA is supported by a stipend of the Klinische Onderzoeks en Opleidings-Raad UZ Leuven. TV is supported by a Medtronic / Royal Academy of Engineering Research Chair [RCSRF1819∖7∖34].

Footnotes

2

The automatic localization and segmentation framework FETAL_BRAIN_SEG is available at https://github.com/gift-surg/fetal_brain_seg and is integrated in the outlier-robust SRR framework NiftyMIC (https://github.com/gift-surg/NiftyMIC).

10

As we only correct for rigid brain motion, the surrounding maternal/fetal tissue reconstruction may be of limited anatomical accuracy.

11

https://github.com/gift-surg/fetal_brain_seg, version from 1 Feb 2019 was used.

12

https://github.com/gift-surg/NiftyMIC, version 0.7.3 was used.

15

https://github.com/bkainz/fetalReconstruction, version from 20 Jan 2017 was used.

16

For the reconstructions, we used the CPU version for both SVR and SRR as the GPU-accelerated approach tends to produce blurrier SRR outcomes in our experience. Therefore, the resulting benchmark method reduces, in principle, to the approach as described in Kuklisova-Murgasova et al. (2012). However, given that we use the implementation as provided by Kainz et al. (2015a) we stick to the notation “Kainz et al. (M)”.

19

The automatic localization and segmentation framework FETAL_BRAIN_SEG is available at https://github.com/gift-surg/fetal_brain_seg and is integrated in the outlier-robust SRR framework NiftyMIC (https://github.com/gift-surg/NiftyMIC).

Appendix A

Supplementary data to this article can be found online at https://doi.org/10.1016/j.neuroimage.2019.116324.

Contributor Information

Michael Ebner, Email: michael.ebner.14@ucl.ac.uk.

Guotai Wang, Email: guotai.wang@uestc.edu.cn.

Appendix A. Supplementary data

The following are the Supplementary data to this article:

Multimedia component 1
mmc1.pdf (2.6MB, pdf)

References

  1. Aertsen M., Verduyckt J., De Keyzer F., Vercauteren T., Van Calenbergh F., De Catte L., Dymarkowski S., Demaerel P., Deprest J. Reliability of MR imaging-based posterior fossa and brain stem measurements in open spinal dysraphism in the era of fetal surgery. Am. J. Neuroradiol. 2019;40:191–198. doi: 10.3174/ajnr.A5930. http://www.ajnr.org/lookup/doi/10.3174/ajnr.A5930 [DOI] [PMC free article] [PubMed] [Google Scholar]
  2. Alansary A., Rajchl M., McDonagh S.G., Murgasova M., Damodaram M., Lloyd D.F.A., Davidson A., Rutherford M., Hajnal J.V., Rueckert D., Kainz B. PVR: patch-to-volume reconstruction for large area motion correction of fetal MRI. IEEE Trans. Med. Imaging. 2017;36:2031–2044. doi: 10.1109/TMI.2017.2737081. http://arxiv.org/abs/1611.07289 http://ieeexplore.ieee.org/document/8024032/ arXiv:1611.07289. [DOI] [PMC free article] [PubMed] [Google Scholar]
  3. Anquez J., Angelini E.D., Bloch I. 2009 IEEE International Symposium on Biomedical Imaging: from Nano to Macro. IEEE; 2009. Automatic segmentation of head structures on fetal MRI; pp. 109–112.http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=5192995 [Google Scholar]
  4. Candès E.J., Li X., Ma Y., Wright J. Robust principal component analysis? J. Assoc. Comput. Mach. 2011;58:1–37. http://portal.acm.org/citation.cfm?doid=1970392.1970395 [Google Scholar]
  5. Çiçek Ö., Abdulkadir A., Lienkamp S.S., Brox T., Ronneberger O. 3D U-net: learning dense volumetric segmentation from sparse annotation. In: Ourselin S., Joskowicz L., Sabuncu M.R., Unal G., Wells W., editors. Medical Image Computing and Computer-Assisted Intervention – MICCAI 2016. Springer International Publishing; Cham: 2016. pp. 424–432.http://link.springer.com/10.1007/978-3-319-46723-8{∖_}49 [Google Scholar]
  6. Criminisi A., Shotton J., Bucciarelli S. MICCAI Workshop on Probabilistic Models for Medical Image Analysis. 2009. Decision forests with long-range spatial context for organ localization in CT volumes; pp. 69–80.https://www.microsoft.com/en-us/research/wp-content/uploads/2016/02/Criminisi{∖_}MICCAI{∖_}PMMIA{∖_}2009{∖_}.pdf [Google Scholar]
  7. Diamond S., Boyd S. 2015 IEEE International Conference on Computer Vision (ICCV) IEEE; 2015. Convex optimization with abstract linear operators; pp. 675–683.http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=7410441Ωhttp://stanford.edu/{∼}boyd/papers/abs{∖_}ops.htmlΩhttps://arxiv.org/abs/1506.00760v2Ωhttp://ieeexplore.ieee.org/document/7410441/ [Google Scholar]
  8. Doel T., Shakir D.I., Pratt R., Aertsen M., Moggridge J., Bellon E., David A.L., Deprest J., Vercauteren T., Ourselin S. GIFT-Cloud: a data sharing and collaboration platform for medical imaging research. Comput. Methods Progr. Biomed. 2017;139:181–190. doi: 10.1016/j.cmpb.2016.11.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
  9. Ebner M., Chouhan M., Patel P.A., Atkinson D., Amin Z., Read S., Punwani S., Taylor S., Vercauteren T., Ourselin S. Point-spread-function-aware slice-to-volume registration: application to upper abdominal MRI super-resolution. In: Zuluaga M.A., Bhatia K., Kainz B., Moghari M.H., Pace D.F., editors. Reconstruction, Segmentation, and Analysis of Medical Images. RAMBO 2016. Springer International Publishing; 2017. pp. 3–13.http://link.springer.com/10.1007/978-3-319-52280-7%7b%5c%5f%7d1 volume 10129 of Lecture Notes in Computer Science. [Google Scholar]
  10. Ebner M., Patel P.A., Atkinson D., Caselton L., Firmin L., Amin Z., Bainbridge A., De Coppi P., Taylor S.A., Ourselin S., Chouhan M.D., Vercauteren T. Super-resolution for upper abdominal MRI: acquisition and post processing protocol optimization using brain MRI control data and expert reader validation. Magn. Reson. Med. 2019;82:1905–1919. doi: 10.1002/mrm.27852. https://onlinelibrary.wiley.com/doi/abs/10.1002/mrm.27852 [DOI] [PMC free article] [PubMed] [Google Scholar]
  11. Ebner M., Wang G., Li W., Aertsen M., Patel P.A., Aughwane R., Melbourne A., Doel T., David A.L., Deprest J., Ourselin S., Vercauteren T. Medical Image Computing and Computer-Assisted Intervention – MICCAI 2018. Springer; 2018. An automated localization, segmentation and reconstruction framework for fetal brain MRI; pp. 313–320.https://link.springer.com/chapter/10.1007/978-3-030-00928-1{∖_}36http://link.springer.com/10.1007/978-3-030-00928-1{∖_}36 [Google Scholar]
  12. Gauriau R., Cuingnet R., Lesage D., Bloch I. Multi-organ localization with cascaded global-to-local regression and shape prior. Med. Image Anal. 2015;23:70–83. doi: 10.1016/j.media.2015.04.007. [DOI] [PubMed] [Google Scholar]
  13. Gholipour A., Estroff J.A., Barnewolt C.E., Robertson R.L., Grant P.E., Gagoski B., Warfield S.K., Afacan O., Connolly S.A., Neil J.J., Wolfberg A., Mulkern R.V. Fetal MRI: a technical update with educational aspirations. Concepts Magn. Reson. 2014;43:237–266. doi: 10.1002/cmr.a.21321. [DOI] [PMC free article] [PubMed] [Google Scholar]
  14. Gholipour A., Estroff J.A., Warfield S.K. Robust super-resolution volume reconstruction from slice acquisitions: application to fetal brain MRI. IEEE Trans. Med. Imaging. 2010;29:1739–1758. doi: 10.1109/TMI.2010.2051680. http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=5482022 [DOI] [PMC free article] [PubMed] [Google Scholar]
  15. Gholipour A., Rollins C.K., Velasco-Annis C., Ouaalam A., Akhondi-Asl A., Afacan O., Ortinau C.M., Clancy S., Limperopoulos C., Yang E., Estroff J.A., Warfield S.K. A normative spatiotemporal MRI atlas of the fetal brain for automatic segmentation and analysis of early brain growth. Sci. Rep. 2017;7:476. doi: 10.1038/s41598-017-00525-w. http://www.nature.com/articles/s41598-017-00525-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  16. Gibson E., Li W., Sudre C., Fidon L., Shakir D.I., Wang G., Eaton-Rosen Z., Gray R., Doel T., Hu Y., Whyntie T., Nachev P., Modat M., Barratt D.C., Ourselin S., Cardoso M.J., Vercauteren T. NiftyNet: a deep-learning platform for medical imaging. Comput. Methods Progr. Biomed. 2018;158:113–122. doi: 10.1016/j.cmpb.2018.01.025. http://arxiv.org/abs/1709.03485 http://linkinghub.elsevier.com/retrieve/pii/S0169260717311823 arXiv:1709.03485. [DOI] [PMC free article] [PubMed] [Google Scholar]
  17. Girshick R., Donahue J., Darrell T., Malik J. 2014 IEEE Conference on Computer Vision and Pattern Recognition. IEEE; 2014. Rich feature hierarchies for accurate object detection and semantic segmentation; pp. 580–587.http://ieeexplore.ieee.org/document/6909475/ arXiv:1311.2524. [Google Scholar]
  18. Hansen P.C. Computational Inverse Problems in Electrocardiology (Advances in Computational Bioengineering) vol. 5. WIT press Holland; 2001. The L-curve and its use in the numerical treatment of inverse problems; pp. 119–142. [Google Scholar]
  19. He K., Gkioxari G., Dollar P., Girshick R. The IEEE International Conference on Computer Vision (ICCV) 2017. Mask R-CNN. [Google Scholar]
  20. Hou B., Alansary A., McDonagh S., Davidson A., Rutherford M., Hajnal J.V., Rueckert D., Glocker B., Kainz B. Medical Image Computing and Computer Assisted Intervention – MICCAI 2017. Springer International Publishing; 2017. Predicting slice-to-volume transformation in presence of arbitrary subject motion; pp. 296–304.http://arxiv.org/abs/1702.08891http://link.springer.com/10.1007/978-3-319-66185-8{∖_}34 arXiv:1702.08891. [Google Scholar]
  21. Hou B., Khanal B., Alansary A., McDonagh S., Davidson A., Rutherford M., Hajnal J.V., Rueckert D., Glocker B., Kainz B. 3-D reconstruction in canonical Co-ordinate space from arbitrarily oriented 2-D images. IEEE Trans. Med. Imaging. 2018;37:1737–1750. doi: 10.1109/TMI.2018.2798801. https://ieeexplore.ieee.org/document/8295121/ [DOI] [PMC free article] [PubMed] [Google Scholar]
  22. Hu Y., Modat M., Gibson E., Li W., Ghavami N., Bonmati E., Wang G., Bandula S., Moore C.M., Emberton M., Ourselin S., Noble J.A., Barratt D.C., Vercauteren T. Weakly-supervised convolutional neural networks for multimodal image registration. Med. Image Anal. 2018;49:1–13. doi: 10.1016/j.media.2018.07.002. https://linkinghub.elsevier.com/retrieve/pii/S1361841518301051 arXiv:1807.03361. [DOI] [PMC free article] [PubMed] [Google Scholar]
  23. Jenkinson M., Bannister P., Brady M., Smith S. Improved optimization for the robust and accurate linear registration and motion correction of brain images. Neuroimage. 2002;17:825–841. doi: 10.1016/s1053-8119(02)91132-8. http://www.ncbi.nlm.nih.gov/pubmed/12377157 https://linkinghub.elsevier.com/retrieve/pii/S1053811902911328 [DOI] [PubMed] [Google Scholar]
  24. Jiang S., Xue H., Glover A., Rutherford M., Rueckert D., Hajnal J.V. MRI of moving subjects using multislice snapshot images with volume reconstruction (SVR): application to fetal, neonatal, and adult brain studies. IEEE Trans. Med. Imaging. 2007;26:967–980. doi: 10.1109/TMI.2007.895456. http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=4265750 [DOI] [PubMed] [Google Scholar]
  25. Kainz B., Keraudren K., Kyriakopoulou V., Rutherford M., Hajnal J.V., Rueckert D. 2014 IEEE 11th International Symposium on Biomedical Imaging (ISBI) IEEE; 2014. Fast fully automatic brain detection in fetal MRI using dense rotation invariant image descriptors; pp. 1230–1233.http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=6868098 [Google Scholar]
  26. Kainz B., Steinberger M., Wein W., Kuklisova-Murgasova M., Malamateniou C., Keraudren K., Torsney-Weir T., Rutherford M., Aljabar P., Hajnal J.V., Rueckert D. Fast volume reconstruction from motion corrupted 2D slices. 2015. https://github.com/bkainz/fetalReconstruction/ [DOI] [PMC free article] [PubMed]
  27. Kainz B., Steinberger M., Wein W., Kuklisova-Murgasova M., Malamateniou C., Keraudren K., Torsney-Weir T., Rutherford M., Aljabar P., Hajnal J.V., Rueckert D. Fast volume reconstruction from motion corrupted stacks of 2D slices. IEEE Trans. Med. Imaging. 2015;34:1901–1913. doi: 10.1109/TMI.2015.2415453. http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=7064742 http://ieeexplore.ieee.org/document/7064742/ [DOI] [PMC free article] [PubMed] [Google Scholar]
  28. Keraudren K., Kuklisova-Murgasova M., Kyriakopoulou V., Malamateniou C., Rutherford M., Kainz B., Hajnal J., Rueckert D. Automated fetal brain segmentation from 2D MRI slices for motion correction. Neuroimage. 2014;101:633–643. doi: 10.1016/j.neuroimage.2014.07.023. http://linkinghub.elsevier.com/retrieve/pii/S1053811914005953 [DOI] [PubMed] [Google Scholar]
  29. Keraudren K., Kyriakopoulou V., Rutherford M., Hajnal J.V., Rueckert D. Medical Image Computing and Computer-Assisted Intervention. vol. 8149. MICCAI 2013; 2013. Localisation of the brain in fetal MRI using bundled SIFT features; pp. 582–589.http://link.springer.com/10.1007/978-3-642-40811-3{∖_}73 LNCS. [DOI] [PubMed] [Google Scholar]
  30. Kim K., Habas P., Rousseau F., Glenn O., Barkovich A., Studholme C. Intersection based motion correction of multislice MRI for 3-D in utero fetal brain image formation. IEEE Trans. Med. Imaging. 2010;29:146–158. doi: 10.1109/TMI.2009.2030679. http://www.ncbi.nlm.nih.gov/pubmed/19744911 http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=5233790 arXiv:NIHMS150003. [DOI] [PMC free article] [PubMed] [Google Scholar]
  31. Kingma D.P., Ba J.L. International Conference on Learning Representations. 2015. Adam: a method for stochastic optimization.http://arxiv.org/abs/1412.6980 arXiv:1412.6980. [Google Scholar]
  32. Kuklisova-Murgasova M., Quaghebeur G., Rutherford M.A., Hajnal J.V., Schnabel J.A. Reconstruction of fetal brain MRI with intensity matching and complete outlier removal. Med. Image Anal. 2012;16:1550–1564. doi: 10.1016/j.media.2012.07.004. http://linkinghub.elsevier.com/retrieve/pii/S1361841512000965 [DOI] [PMC free article] [PubMed] [Google Scholar]
  33. Li W., Wang G., Fidon L., Ourselin S., Cardoso M.J., Vercauteren T. Vol. 3565. 2017. 2017. pp. 348–360.http://link.springer.com/10.1007/978-3-319-59050-9{∖_}28 (On the compactness, efficiency, and representation of 3D convolutional networks: brain parcellation as a pretext task). [Google Scholar]
  34. Liang Z.P., Lauterbur P.C. Wiley-IEEE Press; 2000. Principles of Magnetic Resonance Imaging: A Signal Processing Perspective.http://ieeexplore.ieee.org/xpl/bkabstractplus.jsp?bkn=5264284 [Google Scholar]
  35. Lindeberg T. Scale-space theory: a basic tool for analyzing structures at different scales. J. Appl. Stat. 1994;21:225–270. https://www.tandfonline.com/doi/full/10.1080/757582976 [Google Scholar]
  36. McDonagh S., Hou B., Alansary A., Oktay O., Kamnitsas K., Rutherford M., Hajnal J.V., Kainz B. Molecular Imaging, Reconstruction and Analysis of Moving Body Organs, and Stroke Imaging and Treatment: Fifth International Workshop, CMMI 2017, Second International Workshop, RAMBO 2017, and First International Workshop, SWITCH 2017, Held in Conjunction. Springer International Publishing; 2017. Context-sensitive super-resolution for fast fetal magnetic resonance imaging; pp. 116–126.http://arxiv.org/abs/1703.00035http://link.springer.com/10.1007/978-3-319-67564-0{∖_}12 arXiv:1703.00035. [Google Scholar]
  37. Milletari F., Navab N., Ahmadi S.A. 2016 Fourth International Conference on 3D Vision (3DV) IEEE; 2016. V-net: fully convolutional neural networks for volumetric medical image segmentation; pp. 565–571.http://ieeexplore.ieee.org/document/7785132/ arXiv:1606.04797. [Google Scholar]
  38. Modat M., Cash D.M., Daga P., Winston G.P., Duncan J.S., Ourselin S. Global image registration using a symmetric block-matching approach. J. Med. Imaging. 2014;1 doi: 10.1117/1.JMI.1.2.024003. http://medicalimaging.spiedigitallibrary.org/article.aspx?doi=10.1117/1.JMI.1.2.024003 024003–5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  39. Oktay O., Schlemper J., Folgoc L.L., Lee M., Heinrich M., Misawa K., Mori K., McDonagh S., Hammerla N.Y., Kainz B., Glocker B., Rueckert D. Medical Imaging with Deep Learning – MIDL 2018. 2018. Attention U-net: learning where to look for the pancreas.http://arxiv.org/abs/1804.03999 arXiv:1804.03999. [Google Scholar]
  40. Ovaere C., Eggink A., Richter J., Cohen-Overbeek T.E., Van Calenbergh F., Jansen K., Oepkes D., Devlieger R., De Catte L., Deprest J.A. Prenatal diagnosis and patient preferences in patients with neural tube defects around the advent of fetal surgery in Belgium and holland. Fetal Diagn. Ther. 2015;37:226–234. doi: 10.1159/000365214. https://www.karger.com/Article/FullText/365214 [DOI] [PubMed] [Google Scholar]
  41. Parikh N., Boyd S. 2014. Proximal Algorithms. Foundations and Trends® in Optimization 1; pp. 127–239.http://www.nowpublishers.com/articles/foundations-and-trends-in-optimization/OPT-003http://web.stanford.edu/{∼}boyd/papers/prox{∖_}algs.htmlΩhttp://stanford.edu/{∼}boyd/admm.html [Google Scholar]
  42. Prayer D., Malinger G., Brugger P.C., Cassady C., De Catte L., De Keersmaecker B., Fernandes G.L., Glanc P., Gonçalves L.F., Gruber G.M., Laifer-Narin S., Lee W., Millischer A.E., Molho M., Neelavalli J., Platt L., Pugash D., Ramaekers P., Salomon L.J., Sanz M., Timor-Tritsch I.E., Tutschek B., Twickler D., Weber M., Ximenes R., Raine-Fenning N. ISUOG Practice Guidelines: performance of fetal magnetic resonance imaging. Ultrasound Obstet. Gynecol. 2017;49:671–680. doi: 10.1002/uog.17412. [DOI] [PubMed] [Google Scholar]
  43. Rajchl M., Lee M.C.H., Schrans F., Davidson A., Passerat-Palmbach J., Tarroni G., Alansary A., Oktay O., Kainz B., Rueckert D. 2016. Learning under Distributed Weak Supervision. arXiv preprint arXiv:1606.01100 arXiv:1606.01100. [Google Scholar]
  44. Redmon J., Divvala S., Girshick R., Farhadi A. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) IEEE; 2016. You only look once: unified, real-time object detection; pp. 779–788.http://ieeexplore.ieee.org/document/7780460/ [Google Scholar]
  45. Ronneberger O., Fischer P., Brox T. Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015. 2015. U-net: convolutional networks for biomedical image segmentation; pp. 234–241.http://link.springer.com/10.1007/978-3-319-24574-4{∖_}28 [Google Scholar]
  46. Rousseau F., Glenn O.A., Iordanova B., Rodriguez-Carranza C., Vigneron D.B., Barkovich J.A., Studholme C. Registration-based approach for reconstruction of high-resolution in utero fetal MR brain images. Acad. Radiol. 2006;13:1072–1081. doi: 10.1016/j.acra.2006.05.003. http://linkinghub.elsevier.com/retrieve/pii/S1076633206002753 [DOI] [PubMed] [Google Scholar]
  47. Rutherford S., Sturmfels P., Angstadt M., Hect J., Wiens J., van den Heuval M., Scheinost D., Thomason M., Sripada C. 2019. Automated brain masking of fetal functional MRI; pp. 1–34.https://www.biorxiv.org/content/early/2019/07/13/525386 bioRxiv. [DOI] [PMC free article] [PubMed] [Google Scholar]
  48. Saleem S.N. Fetal MRI: an approach to practice: a review. J. Adv. Res. 2014;5:507–523. doi: 10.1016/j.jare.2013.06.001. http://linkinghub.elsevier.com/retrieve/pii/S2090123213000805 [DOI] [PMC free article] [PubMed] [Google Scholar]
  49. Salehi S.S.M., Hashemi S.R., Velasco-Annis C., Ouaalam A., Estroff J.A., Erdogmus D., Warfield S.K., Gholipour A. 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018) 2018. Real-time automatic fetal brain extraction in fetal MRI by deep learning; pp. 720–724.http://arxiv.org/abs/1710.09338https://ieeexplore.ieee.org/document/8363675/ arXiv:1710.09338. [Google Scholar]
  50. Sobotka D., Licandro R., Ebner M., Schwartz E., Ourselin S., Kasprian G., Prayer D., Langs G. Workshop on Paediatric and Perinatal Imaging. MICCAI; 2019. Reproducibility of functional connectivity estimates in motion corrected fetal fMRI. [Google Scholar]
  51. Sudre C.H., Li W., Vercauteren T., Ourselin S., Jorge Cardoso M. Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support. vol. 10553. DLMIA 2017; 2017. Generalised Dice overlap as a deep learning loss function for highly unbalanced segmentations; pp. 240–248.http://link.springer.com/10.1007/978-3-319-67558-9http://link.springer.com/10.1007/978-3-319-67558-9{∖_}28 [DOI] [PMC free article] [PubMed] [Google Scholar]
  52. Taleb Y., Schweitzer M., Studholme C., Koob M.M., Dietemann J.L., Rousseau F.F. 2013. Automatic Template-Based Brain Extraction in Fetal MR Images. (OHBM) [Google Scholar]
  53. Tourbier S., Bresson X., Hagmann P., Thiran J.P., Meuli R., Cuadra M.B. An efficient total variation algorithm for super-resolution in fetal brain MRI with adaptive regularization. Neuroimage. 2015;118:1–14. doi: 10.1016/j.neuroimage.2015.06.018. http://linkinghub.elsevier.com/retrieve/pii/S1053811915005157 [DOI] [PubMed] [Google Scholar]
  54. Tourbier S., Velasco-Annis C., Taimouri V., Hagmann P., Meuli R., Warfield S.K., Bach Cuadra M., Gholipour A. Automated template-based brain localization and extraction for fetal brain MRI reconstruction. Neuroimage. 2017;155:460–472. doi: 10.1016/j.neuroimage.2017.04.004. http://linkinghub.elsevier.com/retrieve/pii/S1053811917302938 [DOI] [PMC free article] [PubMed] [Google Scholar]
  55. Tustison N.J., Avants B.B., Cook P.A., Zheng Y., Egan A., Yushkevich P.A., Gee J.C. N4ITK: improved N3 bias correction. IEEE Trans. Med. Imaging. 2010;29:1310–1320. doi: 10.1109/TMI.2010.2046908. http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=5445030 [DOI] [PMC free article] [PubMed] [Google Scholar]
  56. Vercauteren T., Perchant A., Malandain G., Pennec X., Ayache N. Robust mosaicing with correction of motion distortions and tissue deformations for in vivo fibered microscopy. Med. Image Anal. 2006;10:673–692. doi: 10.1016/j.media.2006.06.006. http://linkinghub.elsevier.com/retrieve/pii/S1361841506000429 [DOI] [PubMed] [Google Scholar]
  57. Vidal-Migallon I., Commowick O., Pennec X., Dauguet J., Vercauteren T. GPU and CPU implementation of Young - van vliet's recursive Gaussian smoothing filter. Insight Journal (ITK) 2013 http://hdl.handle.net/10380/3425 [Google Scholar]
  58. Wang G., Li W., Zuluaga M.A., Pratt R., Patel P.A., Aertsen M., Doel T., David A.L., Deprest J., Ourselin S., Vercauteren T. Interactive medical image segmentation using deep learning with image-specific fine tuning. IEEE Trans. Med. Imaging. 2018;37:1562–1573. doi: 10.1109/TMI.2018.2791721. http://arxiv.org/abs/1710.04043 http://ieeexplore.ieee.org/document/8270673/ https://ieeexplore.ieee.org/document/8270673/ arXiv:1710.04043. [DOI] [PMC free article] [PubMed] [Google Scholar]
  59. Wang Z., Bovik A., Sheikh H., Simoncelli E. Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 2004;13:600–612. doi: 10.1109/tip.2003.819861. http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=1284395 http://scikit-image.org/docs/dev/auto{∖_}examples/plot{∖_}ssim.html [DOI] [PubMed] [Google Scholar]
  60. Yu F., Koltun V. International Conference on Learning Representations. ICLR 2016; 2016. Multi-scale context aggregation by dilated convolutions.http://arxiv.org/abs/1511.07122 arXiv:1511.07122. [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Multimedia component 1
mmc1.pdf (2.6MB, pdf)

RESOURCES