Skip to main content
Radiation Oncology (London, England) logoLink to Radiation Oncology (London, England)
. 2020 May 11;15:104. doi: 10.1186/s13014-020-01528-0

Clinical implementation of MRI-based organs-at-risk auto-segmentation with convolutional networks for prostate radiotherapy

Mark H F Savenije 1,2,#, Matteo Maspero 1,2,, Gonda G Sikkes 1, Jochem R N van der Voort van Zyp 1, Alexis N T J Kotte 1, Gijsbert H Bol 1, Cornelis A T van den Berg 1,2
PMCID: PMC7216473  PMID: 32393280

Abstract

Background

Structure delineation is a necessary, yet time-consuming manual procedure in radiotherapy. Recently, convolutional neural networks have been proposed to speed-up and automatise this procedure, obtaining promising results. With the advent of magnetic resonance imaging (MRI)-guided radiotherapy, MR-based segmentation is becoming increasingly relevant. However, the majority of the studies investigated automatic contouring based on computed tomography (CT).

Purpose

In this study, we investigate the feasibility of clinical use of deep learning-based automatic OARs delineation on MRI.

Materials and methods

We included 150 patients diagnosed with prostate cancer who underwent MR-only radiotherapy. A three-dimensional (3D) T1-weighted dual spoiled gradient-recalled echo sequence was acquired with 3T MRI for the generation of the synthetic-CT. The first 48 patients were included in a feasibility study training two 3D convolutional networks called DeepMedic and dense V-net (dV-net) to segment bladder, rectum and femurs. A research version of an atlas-based software was considered for comparison. Dice similarity coefficient, 95% Hausdorff distances (HD95), and mean distances were calculated against clinical delineations. For eight patients, an expert RTT scored the quality of the contouring for all the three methods. A choice among the three approaches was made, and the chosen approach was retrained on 97 patients and implemented for automatic use in the clinical workflow. For the successive 53 patients, Dice, HD95 and mean distances were calculated against the clinically used delineations.

Results

DeepMedic, dV-net and the atlas-based software generated contours in 60 s, 4 s and 10-15 min, respectively. Performances were higher for both the networks compared to the atlas-based software. The qualitative analysis demonstrated that delineation from DeepMedic required fewer adaptations, followed by dV-net and the atlas-based software. DeepMedic was clinically implemented. After retraining DeepMedic and testing on the successive patients, the performances slightly improved.

Conclusion

High conformality for OARs delineation was achieved with two in-house trained networks, obtaining a significant speed-up of the delineation procedure. Comparison of different approaches has been performed leading to the succesful adoption of one of the neural networks, DeepMedic, in the clinical workflow. DeepMedic maintained in a clinical setting the accuracy obtained in the feasibility study.

Keywords: Prostate cancer, Radiotherapy, Magnetic resonance imaging, MR-only treatment planning, Delineation, Contouring, Segmentation, Artificial intelligence, Deep learning

Background

Structure delineation is a necessary, yet time-consuming manual procedure in radiotherapy. Consistent and accurate delineation of organs-at-risk (OARs) and target structures for prostate patients is vital when performing dose escalation and treating patients with highly conformal plans [1]. Traditionally, computed tomography (CT) has been used for radiotherapy simulation and structure delineation [2]. In the last few decades, magnetic resonance imaging (MRI) has found its way for radiotherapy simulation as it provides superior soft-tissue contrast compared to CT [3, 4], thus enabling more accurate delineation of target regions and critical structures compared to CT [57].

The manual segmentation of anatomical structures is a time-consuming process [8]. Besides, with the advent of MR-guided radiotherapy [911], the accuracy and speed of delineations become the weakest link [12] that hinders the possibilities of online adaptive radiotherapy by being responsible for longer fraction time [13].

To automatically perform delineations of target and OARs for patients affected by prostate cancer, various methods have been developed over the past years. For example, three-dimensional (3D) deformable model surface [14], organ-based modelling [15], and atlas-based solutions [16, 17] have been demonstrated. For all these methods, the time required to perform segmentation is in the order of minutes, if not hours, which is excessive to enable online adaptive treatments. To obviate this limitation, currently in online treatments only the target delineations and the OARs in the vicinity of the target (e.g. within a ring of 3-5 cm) are adjusted due to the excessive time needed for OARs segmentation [1820].

Recently, deep learning has been proposed to speed-up and automatise automatic segmentation obtaining promising results [8, 21, 22]. Deep learning is a branch of artificial intelligence and machine learning that involves the use of neural networks to generate a hierarchical representation of the input data to achieve a specific task without the need of hand-engineered features [23, 24].

Many studies focused on target delineations [8] reaching mean dice similarity coefficients compared to manual delineations in the range 0.82-0.95 [2531]. Automatic delineation of OARs is also a crucial aspect to achieve full online adaptive radiotherapy and to possibly save time to manual contouring.

In this study, we aim at investigating the feasibility of convolutional neural network-based automatic OARs delineation on MRI. A preliminary retrospective study was conducted to select a suitable network architecture and prepare for clinical implementation. After having chosen the most suitable convolutional network and performing clinical implementation, performances of automatic deep learning-based OARs delineation from our clinic are presented.

Material and methods

Patient data collection

Patients diagnosed with intermediate and high-risk prostate cancer undergoing MR-only radiotherapy [32] in the period between June 2018, and January 2020 were included in the study. Further inclusion criteria were: the presence of four gold fiducial markers for position verification and absence of hip implants. The patients were also scanned with a specific radio-frequency spoiled gradient-recalled echo (SPGR) sequence that will be described in more detail further on. The clinical exclusion criteria for MR-only radiotherapy were: patients with more than four positive lymph-nodes (N1, as on PET-CT or after pelvic lymph-nodes dissection), life expectancy <10 years (as from WHO >3), prior pelvic irradiation, IPSS >20, presence of prostatitis, active Crohn’s disease, colitis ulcerosa or diverticulitis, an anastomotic bowel in the high dose region and patients undergoing trans-rectal prostate resection less than three months before treatment. With the application of these exclusion criteria, a total of 150 patients that were included in this study and treated with external beam radiotherapy.

For all patients, 3T MRI (Ingenia MR-RT, v 5.3.1, Philips Healthcare, the Netherlands) was acquired after requesting the patients to empty their bladder and drink 200-300 ml of water one hour before the acquisition. Patients were positioned on a vendor-provided flat table using a knee support cushion (lower extremity positioning system, without adjustable FeetSupport, MacroMedics BV, the Netherlands). Patients were tattooed at the MRI with the aid of a laser system (Dorado3, LAP GmbH Laser Applikationen, Germany) to facilitate treatment positioning. Also, MR-visible markers (PinPoint for Image Registration 128, Beekley Medical, USA) were used to identify the set-up location on MRI. MR images were acquired using anterior and posterior phased array coils (dS Torso and Posterior coils, 28 channels, Philips Healthcare, the Netherlands). Two in-house-built bridges supported the anterior coil to avoid skin contour deformation.

OARs were contoured on Dixon images [33] obtained with a dual-echo three-dimensional (3D) Cartesian radio-frequency SPGR sequence. For each patient, in-phase (IP), water (W), and fat (F) images [34] (Fig. 1) were reconstructed as in [35]. Dixon images were generated as part of a proprietary solution (MRCAT, rev. 257, Philips Healthcare, Finland) that enabled MR-based dose calculation for patients with prostate cancer [36, 37]. The imaging parameters, reported in Table 1, were locked by the vendor; therefore, they were stable through the whole study. Radiotherapy technicians (RTTs) with dedicated experience in contouring delineated bladder, rectum and femurs using IP, W and F Dixon images. The OARs delineations were approved or revised by a radiation oncologist. Besides, the radiation oncologist delineated the target structures. The delineation indications followed RTOG guidelines [38] requiring that the rectum was delineated from the outer part of the sphincter (anus) until the sigmoid fold (expected length of the rectum was 10-15 cm), as described in [39], with the sphincter delineated as a separate structure. The bladder was entirely delineated, while the femurs were delineated in the whole FOV of the image. In the case of regional radiotherapy, the bowel bag was also included.

Fig. 1.

Fig. 1

Transverse view of in-phase (IP), water (W) and fat (F) images for a patient (69 yo) diagnosed with T2b cancer. Note the large portion of void space surrounding the patient body. Cropping has been applied as preprocessing to remove such void regions

Table 1.

Image parameters of the sequences used for the OARs contouring. The term FOV refers to the field-of-view, while AP to anterior-posterior and LR to right-left

Imaging parameters Value
TE1/TE2/TR [ms] 1.2/2.5/3.9
Flip Angle [ ] 10
FOV [cm3] 55.2x55.2x30
Acquisition Matrix 324x324x120
Reconstruction Matrix 528x528x120
Reconstructed Voxel [mm3] 1x1x2.5
Bandwidth [Hz/px] 1072
Readout direction AP
Phase direction RL
Geometry correction 3D
Acquisition time 2 min 17 s

expressed in terms of anterior-posterior (AP), right-left (RL) and superior-inferior directions

Study design

The first 48 patients (treated until January 2019) were included in a feasibility study training two state-of-the-art 3D convolutional networks called DeepMedic [40] and dense V-net (dV-net) [41] (“Networks architecture, image processing and training” section). Three-fold cross-validation was performed, splitting the patients in 32/16 for train/validation. The network hyperparameters were optimised on the first fold and maintained for the other two folds. For example, the number of epochs was chosen considering the loss function in the validation set by performing early stopping when loss function did not decrease after five consecutive epochs.

The performance of the networks was compared against a research version of commercial software based on multi-atlases and deformable registration and against the clinically used delineations (“Evaluation” section).

This preliminary study enabled us to choose among the three automatic methods. The preferred approach was retrained on 97 patients that were imaged and treated until August 2019; it was implemented for automatic use in the clinical workflow. The performances of the implemented model were reported on the 53 successive consecutively treated patients. A schematic overview of the study design is presented in Fig. 2.

Fig. 2.

Fig. 2

Schematic of the study design representing the timeline and the number of patients included. Also, the length and the number of patients for the preliminary study, the training of the final model and the patients used for testing the clinical implementation are reported

Networks architecture, image processing and training

Three-dimensional network architectures were chosen to investigate performance differences considering as perceptive field the whole volumes or smaller patches. In particular, DeepMedic [40] was the network chosen to perform patch-based training, while dV-net [41] was chosen to perform training on whole volumes. The two architectures, which will be described in detail in the next sections, required similar pre-processing. Three channels were used as input: IP, W and F images. The OARs that were considered as target are: bladder, rectum, right and left femur; they were decoded as masks with values from 1 to 4 without overlapping each others. To increase the amount of contextual information, the CTV was also decoded with a value of 5, which means that the networks also output CTV. Note that CTV was not considered in our study given that CTV delineation is clinically based on a different MRI, i.e. T2-weighted turbo spin-echo sequences [42]. The networks were trained on a graphical processing unit (GPU) Tesla P100 (NVIDIA Corporation, USA) with 16 GB of memory. To allow the whole volume to fit on the GPU, the IP, W and F images were initially cropped with 90 voxels at the borders of the anterior-posterior and lateral directions obtaining matrices of 348x348x120 voxels. Note that an observer controlled the presence of femurs within the FOV. Also, the image intensity of IP, W and F were clipped at their respective 99.9 percentile per each patient volume. Images were subsequently divided by the standard deviation (σ), and then a fixed value of 1 was subtracted.

After training and inference of the networks, the delineations were post-processed generating four binary volumes. Morphological operations of closure and hole filling by one voxel were applied. The largest 3D connected region was selected for each delineated structure. These operations were performed to remove possible small-sized delineations that may have been found by the networks.

DeepMedic

The DeepMedic [40] implementation employed was provided by the Kamnitsas et al.1 in Tensorflow v1.7. The model employed a three-pathway architecture for multi-resolution processing of 3D patches. A low, medium and high-resolution pathway with receptive fields of 853, 513, 173 voxels were employed with each pathway consisting of 11-layers. A fully connected network (FCN) was used for combining the pathways and post-processing, as presented by Kamnitsas et al. [40]. Note that the size of the receptive fields has been modified compared to the original implementation.

The training configuration was kept as the original, with learning rate = 0.001, Adam optimiser with momentum = 0.6, epochs = 35, batch size = 10 and 1 and 2 regularisations 2 weighted with factor 0.000001 and 0.0001, respectively. The configuration file is reported in the Supplementary Material. All the OARs were equally sampled during training enforcing that the patches considered in each epoch contains the four OARs the same amount of times. Also, as in Kamnitsas et al. [40], volumetric dice similarity coefficient was adopted as the loss function. Data augmentation was applied in terms of random shifts and rescaling perturbation of the intensity (I) by the following: I=(I+s)∗m, where s and m where Gaussian distributed with μ=0, 1 and σ=0.05, 0.01, respectively. For training, DeepMedic made use of about 9 GB of GPU memory.

Dense v-net

The dV-net implementation provided in NiftyNet was employed3. It consisted of a 3D U-Net with a sequence of three downsampling and dense upsampling feature strided stacks with skip connections to propagate higher resolution information to the final segmentation. Dilated convolutions were employed to reduce the number of features [41].

The training configuration was kept as the original, with learning rate= 0.001, Adam optimiser with momentum = 0.6, batch size = 6, 2 regularisation (weight = 0.001) and epoch = 25. The configuration file is reported in the Supplementary Material. Dice was adopted as loss function, and data augmentation was applied in terms of elastic deformation, as implemented within NiftyNet. For training, dV-net made use of about 16 GB of GPU memory.

Evaluation

Preliminary study

The first 48 patients treated between June 2018 and August 2019 were included in a preliminary study to compare the performance of the two networks and atlas-based approach to the delineation used during clinical treatment planning.

The advanced medical imaging registration engine (ADMIRE, research version 1.13.5, Elekta AB, Sweden) was the software considered; ADMIRE is based on multi-atlases [43, 44] and gradient-free dense mutual information deformable registration [45]. In particular, the rectum was delineated based on the F image, bladder and femurs were delineated based on IP images using an atlas of 9 patients that were previously acquired with the same sequence. ADMIRE took about 10 to 15 minutes to generate automatic contouring on a Tesla K20c GPU (NVIDIA Corporation, USA) with 6 GB of memory.

Performances of the three automatic approaches were evaluated in terms of (volumetric) dice similarity coefficients (DSC), 95% boundary Hausdorff distances (HD95) [46], mean surface distance (MSD) against clinical delineations. All the metrics were calculated using Plastimatch4, except for the surface distance, which was calculated as from https://github.com/deepmind/surface-distance. In particular, violin plots [47] representing the mean, σ, 95% percentile and the probability distribution were obtained for the three metrics. Also, Wilcoxon signed-rank tests were conducted among the three evaluation metrics with a confidence interval of 0.05.

For a subset of 8 patients, an RTT with five years of experience in contouring scored the quality of the delineations for all three methods. The delineations were classified from zero to three, which corresponds to clinically acceptable, small modifications, large modifications, or clinically unacceptable contours. In total, the RTT scored 96 delineations. The percentage of each score over all the contours was reported for the three methods and visualised in a pie chart. Also, the most challenging structures (structures with an average score ≥2) were reported for each method.

Clinical implementation

After a choice was made among the three automatic approaches, the best performing network was retrained for the first 97 patients that were included up to August 2019. The hyperparameters were identical to the preliminary study. The network was implemented for clinical use complying with the medical device regulation (MDR 2017/745)5. Quantitative evaluation was perfomed in terms of DSC, HD95 and MSD for the 53 consecutive patients undergoing MR-only radiotherapy from August 2019 to January 2020. The delineations adopted for clinical use, i.e. delineated by RTTs and approved or re-adjusted by a radiation oncologist, were considered as reference. Also, surface dice similarity coefficient (SDSC) [48] was calculated6 to enable comparison with previous work [49]. Besides, the performance of the network clinically implemented was compared with the performance of the same network obtained during the preliminary study.

Results

Timing performance

The inference time of the network was about 60 s for DeepMedic and approximately 4 s for dV-net using the full resolution images of 328x328x120 voxels on GPU. ADMIRE generated contours in approximately 14 min on GPU.

Preliminary study

Figure 3 represents the violin plots for DSC, HD95 and the MSD. One can observe that performances were higher for both the networks compared to ADMIRE. For the bladder, no significant differences were observed between the networks, but significant differences were observed between the networks and ADMIRE. For the rectum, no significant differences were observed among the three automatic methods. When considering the femurs, DeepMedic outperformed both dV-net and ADMIRE. For example, for the right femur, the mean (±σ) HD95 was 2.2 ±1.4, 2.5 ±1.8 and 3.2 ±1.4 mm for DeepMedic, dV-net and ADMIRE, respectively.

Fig. 3.

Fig. 3

Violin plots representing the mean (white dot), σ (black vertical rectangle), 95% percentile (black vertical line) and the probability distribution for the dice similarity coefficient (DSC, top) and 95% Hausdorff distance (HD95, middle) and surface distance (bottom) for the OARs against clinical contours in among the preliminary study. The statistical significance of the Wilcoxon signed-rank test is reported as well as the mean(±σ) of each metric. The asterisks represent p ≤0.05 (∗), p ≤0.01 (∗∗) and p ≤0.001 (∗∗∗)

The qualitative scoring by an RTT expert (Fig. 4) demonstrated that delineations from DeepMedic required fewer adaptations, followed by dV-net and then ADMIRE. Specifically, the expert RTT stated that, for all the structures, the number of delineations that were acceptable or needed small adjustment was 81%, 59% and 3% for DeepMedic, dV-net and ADMIRE, respectively. For both the networks, the rectum followed by bladder were indicated as the most challenging structures, while for ADMIRE, the bladder followed by rectum and femurs (same scoring) were the structures considered as the most challenging (score ≥ 2).

Fig. 4.

Fig. 4

Pie chart reporting the percentage of the qualitative scoring performed by the expert RTT for each auto-segmentation method

Clinical implementation

On the basis of the preliminary analysis, we decided to implement DeepMedic for our clinic. Clinical implementation was performed in August 2019.

The performance of DeepMedic in the preliminary study and after clinical implementation are presented in Table 2. After retraining DeepMedic and testing on the successive patients, the performances slightly improved. For example, it can be observed that, on average, the performance of DSC, HD95 and MSD after retraining the network on a more extensive set was ameliorated by 0.01-0.03, 1.2-1.4 mm and 0.1-0.4 mm, respectively. Delineations obtained with DeepMedic for a patient in the test set are presented in Fig. 5.

Table 2.

Comparison of performance between the preliminary study (PS) and after the clinical implementation (Clinic) for DeepMedic in terms of (volumetric) dice similarity coefficient (DSC), 95% Hausdorff distance (HD95) and mean surface distance (MSD)

Site DSC HD95 MSD
PS Clinic PS Clinic PS Clinic
[mm] [mm] [mm] [mm]
Bladder 0.95 ±0.03 0.96 ±0.02 3.8 ±3.4 2.5 ±1.1 1.0 ±0.6 0.6 ±0.3
Rectum 0.85 ±0.07 0.88 ±0.05 8.3 ±5.0 7.4 ±4.4 2.1 ±1.1 1.7 ±0.8
FemurL 0.96 ±0.01 0.97 ±0.01 2.2 ±1.4 1.6 ±0.5 0.6 ±0.2 0.5 ±0.1
FemurR 0.96 ±0.01 0.97 ±0.01 1.9 ±0.4 1.5 ±0.6 0.6 ±0.1 0.5 ±0.1

Fig. 5.

Fig. 5

Example of in-phase MRI after cropping along with segmentations (OARs) obtained with DeepMedic (contours) versus clinical segmentations (filled contours) in the transverse (left), coronal (centre) and sagittal (right) view for a patient in the test. For this patient, average performance was obtained in terms of DSC: 0.96, 0.86, 0.97 and 0.97 for bladder, rectum, and femurs, respectively. Note that DeepMedic also outputs CTV, but it was not considered for clinical evaluation

Also, the SDSC was calculated for several threshold, τ= 0.5, 1, 1.5, 2 and 3 mm as reported in Fig. 6. The mean (±σ) DSCS was 0.98 ±0.03, 0.92 ±0.05, 0.989 ±0.008 and 0.997 ±0.003 for τ=2 mm for bladder, rectum, left and right femur, respectively.

Fig. 6.

Fig. 6

Boxplots for each structure of surface Dice similarity coefficient (SDSC) as a function of threshold (τ) for the 53 patients after clinial implementation. The data is plotted for the range of τ from sub-pixel (0.5 mm) to above the voxel size (3 mm). Box plots are shown with an inter-quartile range from 25 to 75% with the horizontal line representing mean value. Upper and lower whisker represent the 2.5 and 97.5 percentiles

Discussion

The use of MRI for prostate radiotherapy delineation is becoming increasingly common among radiotherapy departments [50]. MRI are used to plan radiotherapy [32, 51]. Besides, use of MRI is also accelerated by the adoption of new advancements in linear accelerator technology, whereby daily MR imaging in treatment position is possible [911].

In this study, we demonstrated that deep learning-based approaches can utilise MRI to automatically segment OARs achieving high conformality. Also, a convolutional network has been implemented for clinical use, demonstrating the capability to maintain the performances obtained in a preliminary study.

Table 3 compiles previous work based on the use of convolutional networks and a selection of conventional approaches [16, 17, 52] for OARs delineation in the pelvic area. One can notice that CT-based segmentation [5355] achieved mean DSC in the range 0.88-0.95 for prostate, rectum and bladder. Also, MRI-based segmentation [27, 49, 56] achieved mean DSC in the range 0.82-0.95. This study seems to outperform previous studies in almost all the metrics (in bold in the Table) except for the rectum, as obtained by Kazemifar et al. [54] and the HD95 and MSD as obtained by Kazemifar et al. and Dong et al. [56]. Comparing the results of automated contouring methods should be done with caution. For example, the guidelines used for clinical delineation may be different, and the impact of inter-observer variability on deep learning-based methods is not generally investigated [57]. In this sense, our study is novel given that a comparison of approaches based on CNNs to an atlas-based method is presented.

Table 3.

Overview of the performance of automatic OARs delineations based on MRI and CT subdivided in convolutional network-based and conventional approaches. The number of patients included in the study (Pts), the imaging modality, a brief description of the method and metrics as dice similarity coefficient (DSC), 95% boundary Hausdorff distance (HD95) and mean surface distance (MSD) were reported for each study. HD95 and MSD are expressed in mm

Study Pts Modality Method(s) Bladder Rectum FemurL FemurR
DSC DSC DSC DSC
HD95 HD95 HD95 HD95
MSD MSD MSD MSD
Convolutional network-based
Men2017 [53] 218/60 CT 2D 0.92 0.93 0.92
dilated
VGG-16
Feng2018 [27] 30/10 MRI Multi-task 0.952 ±0.007 0.88 ±0.03
residual
2D FCN
Kazemifar2018 [54] 51/9/20 CT 2D 0.95 ±0.04 0.92±0.06
U-net 0.4±0.6 0.2±0.3
1.1 ±0.8a 0.8±0.6a
Balagopal2018 [55] 108/28 CT 2D U-net 0.95 ±0.02 0.84 ±0.04 0.96 ±0.03 0.95 ±0.01
mean + 3D U-net 17.0 ±14.6 4.9 ±3.9
4 models (ResNeXT) 0.5 ±0.7 0.8 ±0.7
Dong2019 [56] 140x5 + MRI 3D Cycle-GAN 0.95 ±0.03 0.89 ±0.04
+ deep attention 6.81 ±9.25 10.84 ±15.59
U-net 0.52±0.22 0.92 ±1.03
Elguindi2019 [49] 40/10/50 MRI 0.93 ±0.04 0.82 ±0.05
DeepLabV3+
0.92 ±0.1b 0.87 ±0.07b
This study 97/53 MRI 3D 0.96±0.02 0.88 ±0.05 0.97±0.01 0.97±0.01
multi-scale 2.5 ±1.1 7.4 ±4.4 1.6±0.5 1.5±0.5
DeepMedic 0.6±0.3 1.7 ±0.8 0.5±0.1 0.5±0.1
0.98 ±0.03c 0.92 ±0.05c 0.989±0.008c 0.997±0.003c
Conventional
LaMacchia2012 [16] 5 CT ABAS 2.0 0.93 ±0.03 0.77 ±0.07 0.94 ±0.04 0.94 ±0.04
VelocityAI 2.6.2 0.72 ±0.15 0.75 ±0.04 0.92 ±0.02 0.92 ±0.03
MIM 5.1.1 0.93 ±0.02 0.87 ±0.05 0.94 ±0.02 0.94 ±0.01
Dowling2015 [17] 39 MRI multi-atlas 0.86 ±0.12 0.84 ±0.06 0.91 ±0.03
voting
diffeomorphic reg 5.1 ±4.6 2.4 ±1.0 1.5 ±0.5
Delpon2016 [52] 10/10 CT Mirada 0.76 ±0.12 0.73 ±0.07 0.89 ±0.05 0.91 ±0.03
15 ±9 10 ±3 0.2 ±6.4 8.1 ±5.6
MIM 0.80 ±0.14 0.75 ±0.07 0.89 ±0.08 0.92 ±0.02
14.0 ±6.3 9.9 ±3.4 9.9 ±7.9 8.2 ±5.3
ABAS 0.81 ±0.13 0.75 ±0.09 0.91 ±0.04 0.92 ±0.02
13.6 ±7.9 9.9 ±4.4 8.6 ±6.9 8.5 ±6.1
SPICE 0.76 ±0.26 0.68 ±0.12 0.70 ±0.05 0.72 ±0.03
9.2 ±11.7 13 ±5 29.7 ±9.0 30 ±6.5
Raystation 0.59 ±0.15 0.49 ±0.12 0.91 ±0.03 0.92 ±0.02
28.5 ±13.1 16.5 ±3.7 8.8 ±7.2 6.4 ±5.0

training/(validation)/test; + indicating x... cross-fold validation; a mean surface Hausdorff distance; b,c surface dice similarity coefficient as in [48] with τ=3 or 2 mm, respectively

In this study, a qualitative assessment by a manual observer has been presented. Unfortunately, it has not been recorded whether the overall time for the delineation has been reduced. Previous studies investigated this aspect [58] when introducing deep learning-based techniques in their clinic. Also, it is unclear whether the performance of the network may further improve when a dataset larger than 97 patients is used for training. This may be an object of future research.

The time necessary for automatic delineation on full FOV is within a minute. Such time-scale can be of interest for conventional radiotherapy and for MR-guided treatments. On the one hand, for conventional radiotherapy, fast automatic OAR segmentation may facilitate the reducing delays in the start of the treatments that may lead to hampered clinical outcomes [59]. On the other hand, for online adaptive MR-guided radiotherapy, fast OAR segmentation may relieve clinicians from dedicating effort in OARs segmentation while facilitating the delineation of the target [60]. Currently, it has been reported that about 5-10 min is necessary for the for delineation in an online setting [19]. The time frame reported in our work may facilitate online adaptive radiotherapy, especially with an integrated automatic workflow.

Conclusion

High conformality for OARs delineation was achieved with two in-house trained networks, obtaining a significant speed-up of the delineation procedure. One of the networks, DeepMedic, was successfully adopted in the clinical workflow maintaining in the clinical setting the accuracy obtained in the feasibility study conducted before clinical implementation.

Supplementary information

13014_2020_1528_MOESM1_ESM.zip (9.5KB, zip)

Additional file 1 Configuration files for the network architectures. As part of the supplementary material, it is possible to download the configuration files (ConfigFiles.zip) for DeepMedic (DeepMedic_model.cfg, DeepMedic_inference.cfg, DeepMedic_train.cfg) and dV-net (NiftyNet_train.ini). The configuration files for DeepMedic are three and contains the information regarding the model (DeepMedic_model.cfg), training (DeepMedic_train.cfg) and inference (DeepMedic_inference.cfg) of the network.

Acknowledgements

We gratefully acknowledge the support of NVIDIA Corporation with the donation of the Quadro P5000 GPU used for prototyping this research. Also, we are grateful that Elekta AB provided the research version of the ADMIRE software.

Abbreviations

OAR(s)

organ(s)-at-risk

MR(I)

magnetic resonance (imaging)

CTV

clinical target volume

3D

three-dimensional

SPGR

spoiled gradient-recalled echo

PET

positron emission tomography

WHO

world health organisation

IPSS

international prostate symptom score

IP

in-phase

W

water

F

fat

FOV

field-of-view

TE

echo time

TR

repetition time

AP

anterior-posterior

RL

right-left

RTT

radiotherapy technician

dV-net

dense V-net

GPU

graphical processing unit

FCN(s)

fully connected networks

ADMIRE

advanced medical imaging registration engine

DSC

dice similarity coefficient

MSD

mean surface distance

HD95

95% boundary Hausdorff distance

SDSC

surface dice similarity coefficient

Authors’ contributions

MS designed, collected the data, trained the network, supervised the study, supported clinical implementation and revised the manuscript. MM designed and supervised the study, analysed the data, supported clinical implementation and drafted/edited the manuscript. GGS participated in designing the study, collect the observer validation, and revise the manuscript. JRNvdVvZ and ANTJK contributed to developing the study and revise the manuscript. GHB contributed to developing the study, performed its clinical implementation and revise the manuscript. CATvdB participated in designing the study and revising the manuscript. All authors read and approved the final manuscript.

Funding

This project was made possible with the support of Elekta AB, Stockholm, Sweden, who provided the research version of ADMIRE for this research.

Availability of data and materials

The datasets analysed during the current study are not publicly available due to the internal policy of the Medical Ethical Commission about data sharing. The configuration files of DeepMedic and dV-net are reported as Supplementary Materials.

Ethics approval and consent to participate

The study received the approval of the medical ethical commission (Medisch Ethische Toetsingscommissie) and was classified under the protocol number 15-444/C approved on 29th July 2015.

Consent for publication

Not applicable.

Competing interests

None of the co-authors declares to have competing interests.

Footnotes

2

Consult the https://niftynet.readthedocs.io/en/dev/config_spec.html#reg-type for understanding the regularisation.

4

as available at https://plastimatch.org/

5

European regulation 2017/745 of the European Parliament and of the Council of 5 April 2017 on medical devices, https://eur-lex.europa.eu/eli/reg/2017/745/oj.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Mark H. F. Savenije and Matteo Maspero contributed equally to this work.

Contributor Information

Mark H. F. Savenije, Email: m.h.f.savenij@umcutrecht.nl

Matteo Maspero, Email: m.maspero@umcutrecht.nl, Email: matteo.maspero.it@gmail.com.

Gonda G. Sikkes, Email: g.g.sikkes@umcutrecht.nl

Jochem R. N. van der Voort van Zyp, Email: j.r.n.vandervoortvanzyp@umcutrecht.nl

Alexis N. T. J. Kotte, Email: a.n.t.j.kotte@umcutrecht.nl

Gijsbert H. Bol, Email: g.h.bol@umcutrecht.nl

Cornelis A. T. van den Berg, Email: c.a.t.vandenBerg@umcutrecht.nl

Supplementary information

Supplementary information accompanies this paper at 10.1186/s13014-020-01528-0.

References

  • 1.Salembier C, Villeirs G, De Bari B, Hoskin P, Pieters BR, Van Vulpen M, Khoo V, Henry A, Bossi A, De Meerleer G, Fonteyne V. ESTRO ACROP consensus guideline on CT- and MRI-based target volume delineation for primary radiation therapy of localized prostate cancer. Radiother Oncol. 2018;127(1):49–61. doi: 10.1016/j.radonc.2018.01.014. [DOI] [PubMed] [Google Scholar]
  • 2.Boehmer D, Maingon P, Poortmans P, Baron MH, Miralbell R, Remouchamps V, Scrase C, Bossi A, Bolla M. Guidelines for primary radiotherapy of patients with prostate cancer. Radiother Oncol. 2006;79(3):259–269. doi: 10.1016/j.radonc.2006.05.012. [DOI] [PubMed] [Google Scholar]
  • 3.Debois M, Oyen R, Maes F, Verswijvel G, Gatti G, Bosmans H, Feron M, Bellon E, Kutcher G, Van Poppel H, Vanuytsel L. The contribution of magnetic resonance imaging to the three-dimensional treatment planning of localized prostate cancer, Int J Radiat Oncol Biol Phys. 1999;45(4):857–865. doi: 10.1016/S0360-3016(99)00288-6. [DOI] [PubMed] [Google Scholar]
  • 4.Dirix P, Haustermans K, Vandecaveye V. The Value of Magnetic Resonance Imaging for Radiotherapy Planning, Semin Radiat Oncol. 2014;24(3):151–9. doi: 10.1016/j.semradonc.2014.02.003. [DOI] [PubMed] [Google Scholar]
  • 5.Roach MI, Faillace-Akazawa P, Malfatti C, Holland J, Hricak H. Prostate volumes defined by magnetic resonance imaging and computerized tomographic scans for three-dimensional conformal radiotherapy. Int J Radiat Oncol Biol Phys. 1996;35(5):1011–18. doi: 10.1016/0360-3016(96)00232-5. [DOI] [PubMed] [Google Scholar]
  • 6.Rasch C, Barillot I, Remeijer P, Touw A, van Herk M, Lebesque JV. Definition of the prostate in CT and MRI: a multi-observer study, Int J Radiat Oncol Biol Phys. 1999;43(1):57–66. doi: 10.1016/S0360-3016(98)00351-4. [DOI] [PubMed] [Google Scholar]
  • 7.Villeirs GM, Vaerenbergh K, Vakaet L, Bral S, Claus F, Neve WJ, Verstraete KL, Meerleer GO. Interobserver Delineation Variation Using CT versus Combined CT + MRI in Intensity-Modulated Radiotherapy for Prostate Cancer. Strahlenther Onkol. 2005;181(7):424–30. doi: 10.1007/s00066-005-1383-x. [DOI] [PubMed] [Google Scholar]
  • 8.Cardenas CE, Yang J, Anderson BM, Court LE, Brock KB. Advances in Auto-Segmentation. Sem Radiat Oncol. 2019;29(3):185–97. doi: 10.1016/j.semradonc.2019.02.001. [DOI] [PubMed] [Google Scholar]
  • 9.Raaymakers BW, Raaijmakers AJE, Kotte ANTJ, Jette D, Lagendijk JJW. Integrating a MRI scanner with a 6 MV radiotherapy accelerator: dose deposition in a transverse magnetic field. Phys Med Biol. 2004;49(17):4109–18. doi: 10.1088/0031-9155/49/17/019. [DOI] [PubMed] [Google Scholar]
  • 10.Dempsey J, Benoit D, Fitzsimmons J, Haghighat A, Li J, Low D, Mutic S, Palta J, Romeijn H, Sjoden G. A device for realtime 3D image-guided IMRT, Int J Radiat Oncol Biol Phys. 2005;63:202. doi: 10.1016/j.ijrobp.2005.07.349. [DOI] [Google Scholar]
  • 11.Fallone BG, Murray B, Rathee S, Stanescu T, Steciw S, Vidakovic S, Blosser E, Tymofichuk D. First MR images obtained during megavoltage photon irradiation from a prototype integrated linac-MR system, Med Phys. 2009;36(6):2084–8. doi: 10.1118/1.3125662. [DOI] [PubMed] [Google Scholar]
  • 12.Njeh CF. Tumor delineation: The weakest link in the search for accuracy in radiotherapy. J Med Phys. 2008;33(4):136–40. doi: 10.4103/0971-6203.44472. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Keall P, Poulsen P, Booth JT. See, Think, and Act: Real-Time Adaptive Radiotherapy. Sem Radiat Oncol. 2019;29(3):228–35. doi: 10.1016/j.semradonc.2019.02.005. [DOI] [PubMed] [Google Scholar]
  • 14.Pekar V, McNutt TR, Kaus MR. Automated model-based organ delineation for radiotherapy planning in prostatic region. Int J Radiat Oncol Biol Phys. 2004;60(3):973–80. doi: 10.1016/j.ijrobp.2004.06.004. [DOI] [PubMed] [Google Scholar]
  • 15.Pasquier D, Lacornerie T, Vermandel M, Rousseau J, Lartigau E, Betrouni N. Automatic Segmentation of Pelvic Structures From Magnetic Resonance Images for Prostate Cancer Radiotherapy. Int J Radiat Oncol Biol Phys. 2007;68(2):592–600. doi: 10.1016/j.ijrobp.2007.02.005. [DOI] [PubMed] [Google Scholar]
  • 16.La Macchia M, Fellin F, Amichetti M, Cianchetti M, Gianolini S, Paola V, Lomax AJ, Widesott L. Systematic evaluation of three different commercial software solutions for automatic segmentation for adaptive therapy in head-and-neck, prostate and pleural cancer. Radiat Oncol. 2012;7(1):160. doi: 10.1186/1748-717X-7-160. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Dowling JA, Sun J, Pichler P, Rivest-Hénault D, Ghose S, Richardson H, Wratten C, Martin J, Arm J, Best L, Chandra SS, Fripp J, Menk FW, Greer PB. Automatic Substitute Computed Tomography Generation and Contouring for Magnetic Resonance Imaging (MRI)-Alone External Beam Radiation Therapy From Standard MRI Sequences. Int J Radiat Oncol Biol Phys. 2015;93(5):1144–53. doi: 10.1016/j.ijrobp.2015.08.045. [DOI] [PubMed] [Google Scholar]
  • 18.Raaymakers BW, Jürgenliemk-Schulz IM, Bol GH, Glitzner M, Kotte ANTJ, van Asselen B, de Boer JCJ, Bluemink JJ, Hackett SL, Moerland MA, Woodings SJ, Wolthaus JWH, van Zijp HM, Philippens MEP, Tijssen R, Kok JGM, de Groot-van Breugel EN, Kiekebosch I, Meijers LTC, Nomden CN, Sikkes GG, Doornaert PAH, Eppinga WSC, Kasperts N, Kerkmeijer LGW, Tersteeg JHA, Brown KJ, Pais B, Woodhead P, Lagendijk JJW. First patients treated with a 1.5 T MRI-Linac: clinical proof of concept of a high-precision, high-field MRI guided radiotherapy treatment. Phys Med Biol. 2017;62(23):41–50. doi: 10.1088/1361-6560/aa9517. [DOI] [PubMed] [Google Scholar]
  • 19.Werensteijn-Honingh AM, Kroon PS, Winkel D, Aalbers EM, van Asselen B, Bol GH, Brown KJ, Eppinga WSC, van Es CA, Glitzner M, de Groot-van Breugel EN, Hackett SL, Intven M, Kok JGM, Kontaxis C, Kotte AN, Lagendijk JJW, Philippens MEP, Tijssen RHN, Wolthaus JWH, Woodings SJ, Raaymakers BW, Jurgenliemk-Schulz IM. Feasibility of stereotatctic radiotherapy using a 1.5T MR-linac: Multi-fraction treatment of pelvic lymph node oligometastases. Radiother Oncol. 2019;134:50–4. doi: 10.1016/j.radonc.2019.01.024. [DOI] [PubMed] [Google Scholar]
  • 20.Bruynzeel AM, Tetar SU, Oei SS, Senan S, Haasbeek CJ, Spoelstra FO, Piet AH, Meijnen P, van der Jagt MAB, Fraikin T, et al. A prospective single-arm phase 2 study of stereotactic magnetic resonance guided adaptive radiation therapy for prostate cancer: early toxicity results. Int J Radiat Oncol Biol Phys. 2019;105(5):1086–94. doi: 10.1016/j.ijrobp.2019.08.007. [DOI] [PubMed] [Google Scholar]
  • 21.Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In: International Conference on Medical Image Computing and Computer-assisted Intervention. Springer: 2015. p. 234–41. 10.1007/978-3-319-24574-4_28.
  • 22.Milletari F, Navab N, Ahmadi S-A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In: 2016 Fourth International Conference on 3D Vision (3DV). IEEE: 2016. p. 565–71. 10.1109/3dv.2016.79.
  • 23.Meyer P, Noblet V, Mazzara C, Lallement A. Survey on deep learning for radiotherapy. Comp Biol Med. 2018;98:126–46. doi: 10.1016/j.compbiomed.2018.05.018. [DOI] [PubMed] [Google Scholar]
  • 24.Sahiner B, Pezeshk A, Hadjiiski LM, Wang X, Drukker K, Cha KH, Summers RM, Giger ML. Deep learning in medical imaging and radiation therapy. Med Phys. 2019;46(1):1–36. doi: 10.1002/mp.13264. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Alvarez C, Martínez F, Romero E. A multiresolution prostate representation for automatic segmentation in magnetic resonance images. Med Phys. 2017;44(4):1312–23. doi: 10.1002/mp.12141. [DOI] [PubMed] [Google Scholar]
  • 26.Fu Y, Mazur TR, Wu X, Liu S, Chang X, Lu Y, Li HH, Kim H, Roach MC, Henke L, Yang D. A novel MRI segmentation method using CNN-based correction network for MRI-guided adaptive radiotherapy. Med Phys. 2018;45(11):5129–37. doi: 10.1002/mp.13221. [DOI] [PubMed] [Google Scholar]
  • 27.Feng Z, Nie D, Wang L, Shen D. Semi-supervised learning for pelvic MR image segmentation based on multi-task residual fully convolutional networks. IEEE Comput Soc. 2018. 10.1109/ISBI.2018.8363713. [DOI] [PMC free article] [PubMed]
  • 28.Nie D, Wang L, Gao Y, Lian J, Shen D. STRAINet: Spatially Varying sTochastic Residual AdversarIal Networks for MRI Pelvic Organ Segmentation. Trans Neural Netw Learn Syst IEEE. 2018:1–13. 10.1109/TNNLS.2018.2870182. [DOI] [PMC free article] [PubMed]
  • 29.Liu C, Gardner SJ, Wen N, Elshaikh MA, Siddiqui F, Movsas B, Chetty IJ. Automatic segmentation of the prostate on CT images using deep neural networks (DNN) Int J Radiat Oncol Biol Phys. 2019;104(4):924–32. doi: 10.1016/j.ijrobp.2019.03.017. [DOI] [PubMed] [Google Scholar]
  • 30.Eppenhof KAJ, Maspero M, Savenije MHF, de Boer JCJ, van der Voort van Zyp JRN, Raaymakers BW, Raaijmakers AJE, Veta M, van den Berg CAT, Pluim JPW. Fast contour propagation for MR-guided prostate radiotherapy using convolutional neural networks. Med Phys. 2019;47(3):1238–48. doi: 10.1002/mp.13994. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Wang B, Lei Y, Tian S, Wang T, Liu Y, Patel P, Jani AB, Mao H, Curran WJ, Liu T, Yang X. Deeply supervised 3D fully convolutional networks with group dilated convolution for automatic MRI prostate segmentation. Med Phys. 2019;46(4):1707–18. doi: 10.1002/mp.13416. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Kerkmeijer LGW, Maspero M, Meijer GJ, van der Voort van Zyp JRN, de Boer HCJ, van den Berg CAT. Magnetic Resonance Imaging only Workflow for Radiotherapy Simulation and Planning in Prostate Cancer. Clin Oncol. 2018;30(11):692–701. doi: 10.1016/j.clon.2018.08.009. [DOI] [PubMed] [Google Scholar]
  • 33.Dixon WT. Simple proton spectroscopic imaging. Radiology. 1984;153(1):189–94. doi: 10.1148/radiology.153.1.6089263. [DOI] [PubMed] [Google Scholar]
  • 34.Eggers H, Brendel B, Duijndam A, Herigault G. Dual-echo Dixon imaging with flexible choice of echo times. Magn Reson Med. 2011;65(1):96–107. doi: 10.1002/mrm.22578. [DOI] [PubMed] [Google Scholar]
  • 35.Maspero M, Savenije MH, Dinkla AM, Seevinck PR, Intven MP, Jurgenliemk-Schulz IM, Kerkmeijer LG, van den Berg CA. Dose evaluation of fast synthetic-ct generation using a generative adversarial network for general pelvis mr-only radiotherapy. Phys Med Biol. 2018;63(18):185001. doi: 10.1088/1361-6560/aada6d. [DOI] [PubMed] [Google Scholar]
  • 36.Maspero M, Seevinck PR, Schubert G, Hoesl MAU, van Asselen B, Viergever MA, Lagendijk JJW, Meijer GJ, van den Berg CAT. Quantification of confounding factors in MRI-based dose calculations as applied to prostate IMRT. Phys Med Biol. 2017;62(3):948–65. doi: 10.1088/1361-6560/aa4fe7. [DOI] [PubMed] [Google Scholar]
  • 37.Maspero M, Tyyger MD, Tijssen RH, Seevinck PR, Intven MP, van den Berg CA. Feasibility of magnetic resonance imaging-only rectum radiotherapy with a commercial synthetic computed tomography generation solution. Phys Imag Radiat Oncol. 2018;7:58–64. doi: 10.1016/j.phro.2018.09.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Gay HA, Barthold HJ, O’Meara E, Bosch WR, El Naqa I, Willett C, Kachnic LA, Jhingran A, Portelance L, Ryu J, Small W, Gaffney D, Viswanathan AN, Michalski JM. Pelvic Normal Tissue Contouring Guidelines for Radiation Therapy: A Radiation Therapy Oncology Group Consensus Panel Atlas. Int J Radiat Oncol Biol Phys. 2012;83(3):353–62. doi: 10.1016/j.ijrobp.2012.01.023. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.D’Souza N, de Neree tot Babberich MPM, Lord A, Shaw A, Abulafi M, Tekkis P, Wiggers T, Brown G. The rectosigmoid problem. Surg Oncol. 2018;27(3):521–5. doi: 10.1016/j.suronc.2018.06.005. [DOI] [PubMed] [Google Scholar]
  • 40.Kamnitsas K, Ferrante E, Parisot S, Ledig C, Nori AV, Criminisi A, Rueckert D, Glocker B. DeepMedic for Brain Tumor Segmentation. In: Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: Second International Workshop, BrainLes 2016, with the Challenges on BRATS, ISLES and mTOP 2016, Held in Conjunction with MICCAI 2016, Athens, Greece, October 17, 2016, Revised Selected Papers, vol. 10154. Springer: 2017. p. 138. 10.1007/978-3-319-55524-9.
  • 41.Gibson E, Giganti F, Hu Y, Bonmati E, Bandula S, Gurusamy K, Davidson B, Pereira SP, Clarkson MJ, Barratt DC. Automatic multi-organ segmentation on abdominal CT with dense v-networks. IEEE Trans Med Imaging. 2018; 37(8):1822–34. https://doi.org/10.1109%2FTMI.2018.2806309. [DOI] [PMC free article] [PubMed]
  • 42.Coakley FV, Oto A, Alexander LF, Allen BC, Davis BJ, Froemming AT, Fulgham PF, Hosseinzadeh K, Porter C, Sahni VA, Schuster DM, Showalter TN, Venkatesan AM, Verma S, Wang CL, Remer EM, Eberhardt SC. ACR Appropriateness Criteria® Prostate Cancer—Pretreatment Detection, Surveillance, and Staging. J Am Col Radiol. 2017;14(5):245–257. doi: 10.1016/j.jacr.2017.02.026. [DOI] [PubMed] [Google Scholar]
  • 43.Han X. Learning-boosted label fusion for multi-atlas auto-segmentation. In: Lecture Notes in Computer Science (including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 8184 LNCS. Springer: 2013. p. 17–24. 10.1007/978-3-319-02267-3.
  • 44.Van de Velde J, Wouters J, Vercauteren T, De Gersem W, Achten E, De Neve W, Van Hoof T. Optimal number of atlases and label fusion for automatic multi-atlas-based brachial plexus contouring in radiotherapy treatment planning. Radiat Oncol. 2016; 11(1). 10.1186/s13014-015-0579-1. [DOI] [PMC free article] [PubMed]
  • 45.Han X, Hibbard LS, Willcut V. GPU-accelerated, gradient-free MI deformable registration for atlas-based MR brain image segmentation. In: 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops: 2009. p. 141–8. 10.1109/CVPRW.2009.5204043.
  • 46.Beauchemin M, Thomson KPB, Edwards G. On the hausdorff distance used for the evaluation of segmentation results. Canad J Remote Sens. 1998;24(1):3–8. doi: 10.1080/07038992.1998.10874685. [DOI] [Google Scholar]
  • 47.Hintze JL, Nelson RD. Violin plots: a box plot-density trace synergism. Am Stat. 1998;52(2):181–4. [Google Scholar]
  • 48.Nikolov S, Blackwell S, Mendes R, De Fauw J, Meyer C, Hughes C, Askham H, Romera-Paredes B, Karthikesalingam A, Chu C, Carnell D, Boon C, D’Souza D, Moinuddin SA, Sullivan K, Consortium DR, Montgomery H, Rees G, Sharma R, Suleyman M, Back T, Ledsam JR, Ronneberger O. Deep learning to achieve clinically applicable segmentation of head and neck anatomy for radiotherapy. 2018. http://arxiv.org/abs/1809.04430.
  • 49.Elguindi S, Zelefsky MJ, Jiang J, Veeraraghavan H, Deasy JO, Hunt MA, Tyagi N. Deep learning-based auto-segmentation of targets and organs-at-risk for magnetic resonance imaging only planning of prostate radiotherapy. Phys Imag Radiat Oncol. 2019;12:80–6. doi: 10.1016/j.phro.2019.11.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Schmidt MA, Payne GS. Radiotherapy planning using MRI, Phys Med Biol. 2015;60(22):323–61. doi: 10.1088/0031-9155/60/22/R323. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Edmund JM, Nyholm T. A review of substitute CT generation for MRI-only radiation therapy. Radiat Oncol. 2017;12(1):28. doi: 10.1186/s13014-016-0747-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Delpon G, Escande A, Ruef T, Darréon J, Fontaine J, Noblet C, Supiot S, Lacornerie T, Pasquier D. Comparison of Automated Atlas-Based Segmentation Software for Postoperative Prostate Cancer Radiotherapy. Front Oncol. 2016; 6. 10.3389/fonc.2016.00178. [DOI] [PMC free article] [PubMed]
  • 53.Men K, Dai J, Li Y. Automatic segmentation of the clinical target volume and organs at risk in the planning CT for rectal cancer using deep dilated convolutional neural networks. Med Phys. 2017;44(12):6377–89. doi: 10.1002/mp.12602. [DOI] [PubMed] [Google Scholar]
  • 54.Kazemifar S, Balagopal A, Nguyen D, McGuire S, Hannan R, Jiang S, Owrangi A. Segmentation of the prostate and organs at risk in male pelvic CT images using deep learning. Biomed Phys Eng. 2018;4(5):055003. doi: 10.1088/2057-1976/aad100. [DOI] [Google Scholar]
  • 55.Balagopal A, Kazemifar S, Nguyen D, Lin M-H, Hannan R, Owrangi A, Jiang S. Fully automated organ segmentation in male pelvic CT images. Phys Med Biol. 2018;63(24):245015. doi: 10.1088/1361-6560/aaf11c. [DOI] [PubMed] [Google Scholar]
  • 56.Dong X, Lei Y, Tian S, Wang T, Patel P, Curran WJ, Jani AB, Liu T, Yang X. Synthetic MRI-aided multi-organ segmentation on male pelvic CT using cycle consistent deep attention network. Radiother Oncol. 2019. 10.1016/j.radonc.2019.09.028. [DOI] [PMC free article] [PubMed]
  • 57.Sharp G, Fritscher KD, Pekar V, Peroni M, Shusharina N, Veeraraghavan H, Yang J. Vision 20/20: Perspectives on automated image segmentation for radiotherapy. Med Phys. 2014; 41(5). 10.1118/1.4871620@10.1002. [DOI] [PMC free article] [PubMed]
  • 58.Lustberg T, van Soest J, Gooding M, Peressutti D, Aljabar P, van der Stoep J, van Elmpt W, Dekker A. Clinical evaluation of atlas and deep learning based automatic contouring for lung cancer. Radiother Oncol. 2018;126(2):312–7. doi: 10.1016/j.radonc.2017.11.012. [DOI] [PubMed] [Google Scholar]
  • 59.Huang J, Barbera L, Brouwers M, Browman G, Mackillop WJ. Does delay in starting treatment affect the outcomes of radiotherapy? A systematic review, J Clin Oncol Off J Am Soc Clin Oncol. 2003;21(3):555–63. doi: 10.1200/JCO.2003.04.171. [DOI] [PubMed] [Google Scholar]
  • 60.Keall PJ, Barton M, Crozier S. The Australian magnetic resonance imaging-linac program. Semin Radiat Oncol. 2014;24(3):203–6. doi: 10.1016/j.semradonc.2014.02.015. [DOI] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

13014_2020_1528_MOESM1_ESM.zip (9.5KB, zip)

Additional file 1 Configuration files for the network architectures. As part of the supplementary material, it is possible to download the configuration files (ConfigFiles.zip) for DeepMedic (DeepMedic_model.cfg, DeepMedic_inference.cfg, DeepMedic_train.cfg) and dV-net (NiftyNet_train.ini). The configuration files for DeepMedic are three and contains the information regarding the model (DeepMedic_model.cfg), training (DeepMedic_train.cfg) and inference (DeepMedic_inference.cfg) of the network.

Data Availability Statement

The datasets analysed during the current study are not publicly available due to the internal policy of the Medical Ethical Commission about data sharing. The configuration files of DeepMedic and dV-net are reported as Supplementary Materials.


Articles from Radiation Oncology (London, England) are provided here courtesy of BMC

RESOURCES