Skip to main content
NIHPA Author Manuscripts logoLink to NIHPA Author Manuscripts
. Author manuscript; available in PMC: 2022 Oct 24.
Published in final edited form as: Proc Int Soc Magn Reson Med Sci Meet Exhib Int Soc Magn Reson Med Sci Meet Exhib. 2021 May;29:1362.

Learning-based automatic field-of-view positioning for fetal-brain MRI

Malte Hoffmann 1,2, Daniel C Moyer 3, Lawrence Zhang 3, Polina Golland 3, Borjan Gagoski 1,4, P Ellen Grant 1,4, André JW van der Kouwe 1,2
PMCID: PMC9592155  NIHMSID: NIHMS1842940  PMID: 36284600

Introduction

One of the unique challenges of fetal-brain MRI is the arbitrary orientation of the anatomy relative to the device axes, whereas clinical assessment requires standard sagittal, coronal and axial views relative to the fetal brain.1,2 Unfortunately, fetal motion limits acquisitions to thick slices which preclude retroactive resampling to provide standard planes (Figure 1). Throughout the session, technologists repeat acquisitions while incrementally adjusting the field of view (FOV), deducing the head pose from the previous stack until they obtain appropriately oriented images.

Figure 1.

Figure 1.

Fetal brain MRI with incorrect anatomical orientation. Resampling across slices is typically not viable due to thick slices, spin-history artifacts, and frequent changes in position. Critically, clinical assessment requires standard anatomical views of the brain, for example, in the axial plane.

Building on our prior work on landmark detection3 and automated prescription4, we present learning-driven FOV positioning and demonstrate automatic acquisition of standard anatomical planes of the fetal brain in-vivo.

Method

The system is fully integrated with a 3-T Siemens Skyra scanner, used to acquire all presented data with a flexible 60-channel body array. A laptop connected to the local sub-network is required. All communications are automatic.

Setup

A deep neural network delineates the fetal brain and eyes in a full-uterus scout, from which the head pose is derived. A laptop receives the scout via TCP and returns coordinates to the target sequence (Figure 2). First, a scout is acquired and provided to the network for landmark detection. Second, the system resamples this image in the anatomical frame and displays it on the laptop, enabling the user to review anatomical alignment and repeat the scout if needed. Third, the user starts the target sequence, selecting sagittal, coronal or axial slices, which are positioned automatically.

Figure 2.

Figure 2.

Automatic field-of-view prescription. First, the operator acquires a rapid full-uterus scout. Second, an external laptop receives the scout via TCP and hosts the network that detects the fetal brain and eyes. Third, we construct an anatomical basis from these landmarks. Finally, the target sequence receives the anatomical frame and acquires sagittal, coronal or axial slices of the brain according to the operator’s selection. All communications are automatic, and the procedure can be repeated as needed to acquire different views or to respond to fetal head-pose changes.

Pulse sequences

The scout is a stack of 90 interleaved 3-mm GE-EPI slices with 128×128 matrix, (3 mm)2 resolution, no gap, TR/TA 6000 ms, TE 30 ms, FA 90°, 5/8 partial Fourier. The anatomical scan is a stack of 35 interleaved 3-mm HASTE5 slices with 256×256 matrix, (1.25 mm)2 resolution, no gap, TA 1:05 min, TR/TE 1800/100 ms, FA 160°, 5/8 partial Fourier, GRAPPA 2.

Landmark detection

Brain and eye masks are estimated from input images using a U-Net6 (Figure 3) trained beforehand on a manually annotated dataset including 105 EPI stacks from 78 fetuses (parameters as above, except TR/TE 7000/36 ms). At each training step, images are augmented with random rotations and translations (0–10 voxels). We fit network parameters using a cross-entropy loss for each landmark and MSE for their barycenters. Brain-mask losses are weighted five times higher than losses for the eyes. Barycentric losses are scaled by 10−4 to maintain stability.

Figure 3.

Figure 3.

Neural network used for landmark detection (standard U-Net). The U-Net is composed of four “down-sample” and four “up-sample” blocks, each of which contains a sequence of two k-channel 3×3×3 convolutions with ReLU activations followed by down or up-sampling by a factor of 2, as shown on the left. The outputs of the “down-sample” blocks are concatenated with the outputs of the corresponding “up-sample” blocks prior to feeding into the next block. Output masks are computed via sigmoid activations after the final convolution.

Anatomical frame

From the barycenters {cb, cl, cr} of the brain and eyes, we derive an orthonormal left-posterior-superior basis F = {ul, up, us}.3 An initial right-left axis u˜l=(clcr)clcr1 passes through the eyes. The mid-point cm = (cl + cr)/2 and cb define an anterior-posterior axis u˜p=(cbcm)cbcm1, and we let u˜s=u˜l×u˜p. As u˜l and u˜p may not be orthogonal, we use ul=u˜p×u˜s. The plane defined by {cb, cl, cr} is usually tilted from the axial plane by θ ≈ −30° about u˜l: we choose as basis F=Rl(θ)(ulu˜pu˜s), where Rl is a rotation matrix about ul.

Field of view

We prescribe slices in the anatomical frame by centering the FOV on cb and selecting the slice-normal vector from {ul, up, us} depending on the user’s selection. The remaining vectors define the phase and read-out directions: to minimize potential wrapping4 we choose as vulnerable phase direction the vector with the smaller component along the maternal head-foot axis z.

Results

Landmark detection

We test offline landmark detection on a held-out validation set of 10 EPI stacks that do not include the training subjects (Figure 4). Dice scores are (93.1±1.1)%, (56.1±30.1)%, and (41.0±25.0)% for the brain, left and right eye, respectively. The detection takes (4.8±0.6) sec on a laptop with a 2.8-GHz quad-core Intel i7-7700HQ CPU and 16 GB RAM. This reduces to (0.3±0.1) sec on an Nvidia V100-SXM2 GPU with 32 GB memory.

Figure 4.

Figure 4.

Offline mask-detection evaluation for the fetal brain and eyes. The network robustly masks the brain, while the eyes are detected less accurately, likely due to their smaller size.

In-vivo evaluation

We perform automatic slice prescription in two healthy fetuses at 32 and 31 weeks’ gestation (Figure 5). Landmarks are robustly identified for the first fetus. For the second, we repeat several scouts until the network detects the eyes, due to fetal motion and wrapping.

Figure 5.

Figure 5.

In-vivo evaluation in fetuses at 32 and 31 weeks’ gestation. The fetal brain is arbitrarily oriented relative to the planes of the MRI scanner. The network detects the brain, left (red) and right (green) eyes from the full-uterus scout, oriented along the device axes (sagittal to the mother). From these landmarks, we derive the head pose. The target sequence automatically acquires standard sagittal, coronal and axial views of the brain. We repeat the scout before each anatomical acquisition. The second fetus required several attempts due to substantial subject motion.

Discussion

Network robustness

Although the detection is less reliable in the second fetus, the real-time feedback enables the operator to repeat the scout until the head-pose is identified. We plan to improve the detection by increasing the training set and augmentation, e.g. with random image histogram scaling.

Scout acquisition

We will develop an optimized ~1-second scout with lower spatial resolution and a larger FOV, thus reducing between-slice motion and phase-wrap artifacts which impinge on the landmark detection.

Other applications

Auto-FOV positioning may prove useful in other applications of MRI, such as musculoskeletal MRI, where many structures are obliquely oriented to the standard planes. For example, capturing the thin articular cartilage of the hip and ankle can be challenging with 2D sequences,7 which remain the preferred clinical acquisition. Training networks for these body parts will extend our framework to aid technologists in prescribing difficult views.

Conclusion

We propose a framework for automated FOV and slice prescription of fetal-brain MRI and demonstrate its utility in-vivo. The modular design makes it extendable for users interested in other anatomy demanding oblique imaging.

Synopsis.

Unique challenges of fetal-brain MRI include successful acquisition of standard sagittal, coronal and axial views of the brain, as motion precludes acquisition of coherent orthogonal slice stacks. Technologists repeat scans numerous times by manually rotating slice prescriptions but inaccuracies in slice placement and intervening motion limit success. We propose a system to automatically prescribe slices based on the fetal-head pose as estimated by a neural network from a fast scout. The target sequence receives the head pose and acquires slices accordingly. We demonstrate automatic acquisition of standard anatomical views in-vivo.

Acknowledgements

The authors thank Paul Wighton and Robert Frost for sharing code, and Andrew Hoopes for figure design inspiration. This work uses human subject data and is formally reviewed by the institution’s Human Research Committee. Support was provided in part by NIH grants NICHD K99 HD101553, U01 HD087211, R01 HD100009, R01HD093578 and NIBIB NAC P41 EB015902.

References

  • 1.Gholipour A et al. Fetal MRI: A technical update with educational aspirations. Concepts Magn. Reson. Part A 2014;43(6):237–266. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Saleem SN et al. Fetal MRI: An approach to practice: A review. J. Adv. Res 2014;5(5):507–523. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Hoffmann M et al. Improved, rapid fetal-brain localization and orientation detection for auto-slice prescription. ISMRM, Montreal, QC, Canada. 2019. [Google Scholar]
  • 4.Hoffmann M et al. Fast, automated slice prescription of standard anatomical planes for fetal brain MRI. ISMRM, Paris, France. 2018. [Google Scholar]
  • 5.Kiefer B et al. Image acquisition in a second with half-Fourier acquired single shot turbo spin echo. JMRI. 1994;4(86):86–87. [Google Scholar]
  • 6.Ronneberger O et al. U-Net: Convolutional Networks for Biomedical Image Segmentation. CoRR. 2015;234–241. [Google Scholar]
  • 7.Naraghi A and White LM. Three-dimensional MRI of the musculoskeletal system. AJR. 2012;199(3):W283–W293. [DOI] [PubMed] [Google Scholar]

RESOURCES