Abstract
Objective.
To propose a novel moment-based loss function for predicting 3D dose distribution for the challenging conventional lung intensity modulated radiation therapy plans. The moment-based loss function is convex and differentiable and can easily incorporate clinical dose volume histogram (DVH) domain knowledge in any deep learning (DL) framework without computational overhead.
Approach.
We used a large dataset of 360 (240 for training, 50 for validation and 70 for testing) conventional lung patients with 2 Gy × 30 fractions to train the DL model using clinically treated plans at our institution. We trained a UNet like convolutional neural network architecture using computed tomography, planning target volume and organ-at-risk contours as input to infer corresponding voxel-wise 3D dose distribution. We evaluated three different loss functions: (1) the popular mean absolute error (MAE) loss, (2) the recently developed MAE + DVH loss, and (3) the proposed MAE + moments loss. The quality of the predictions was compared using different DVH metrics as well as dose-score and DVH-score, recently introduced by the AAPM knowledge-based planning grand challenge.
Main results.
Model with (MAE + moment) loss function outperformed the model with MAE loss by significantly improving the DVH-score (11%, p < 0.01) while having similar computational cost. It also outperformed the model trained with (MAE + DVH) by significantly improving the computational cost (48%) and the DVH-score (8%, p < 0.01).
Significance.
DVH metrics are widely accepted evaluation criteria in the clinic. However, incorporating them into the 3D dose prediction model is challenging due to their non-convexity and non-differentiability. Moments provide a mathematically rigorous and computationally efficient way to incorporate DVH information in any DL architecture. The code, pretrained models, docker container, and Google Colab project along with a sample dataset are available on our DoseRTX GitHub (https://github.com/nadeemlab/DoseRTX)
Keywords: deep learning dose prediction, automated radiotherapy treatment planning, external photon treatment planning
1. Introduction
Despite recent advances in optimization and treatment planning, intensity modulated radiation therapy (IMRT) treatment planning remains a time-consuming and resource-demanding task with the plan quality heavily dependent on the planner’s experience and expertise (Das et al 2009, Nelms et al 2012, Berry et al 2016). This problem is even more pronounced for challenging clinical cases such as conventional lung with complex geometry and intense conflict between the objectives of irradiating planning target volume (PTV) and sparing organ at risk structures (OARs). Conventional lung plans are generally considered challenging in the clinic due to large PTV sizes, variation in the location of PTV, and having many sensitive nearby structures, making the planning process time-consuming and difficult. Balancing the trade-off between conflicting objectives can lead to sub-optimal plans (Moore et al 2015), sacrificing the plan quality.
In the last decade, several techniques have been developed to automate or facilitate the radiotherapy treatment planning process. Multi-criteria optimization (Craft and Bortfeld 2008) facilitates the planning by generating a set of Pareto optimal plans upfront and allowing the user to navigate among them offline. Hierarchical constrained optimization enforces the critical clinical constraints using hard constraints and improves the other desirable criteria as much as possible by sequentially optimizing these (Breedveld et al 2009, Zarepisheh et al 2019). Knowledge-based planning (KBP) is a data-driven approach to automate the planning process by leveraging a database of pre-existing patients and learning a map between the patient anatomical features and some dose distribution characteristics. The earlier KBP methods used machine learning methods such as linear regression, principal component analysis, random forests, and neural networks to predict DVH as a main metric to characterize the dose distribution (Appenzoller et al 2012, Zarepisheh et al 2014, Fogliata et al 2015, Tol et al 2015, Chin Snyder et al 2016, Valdes et al 2017, Liu et al 2020). However, DVH lacks any spatial information and only predicts dose for the delineated structures.
More recently, deep learning (DL) methods have been successfully used in radiation oncology for automated image contouring/segmentation (Han 2017, Ibragimov and Xing 2017, Men et al 2017) as well as 3D voxel-level dose prediction (Wang et al 2020). A typical DL dose prediction method uses a convolutional neural network (CNN) model which receives a 2D or 3D input in the form of planning CT with OAR/PTV masks and produces a voxel-level dose distribution as its output. The predicted dose is compared to the real dose using some form of loss function such as mean absolute error (MAE) or mean square error (MSE). The loss function in fact quantifies the goodness of the prediction by comparing that to the delivered dose voxel-by-voxel. While MAE and MSE are powerful and easy-to-use loss functions, they fail to integrate any domain specific knowledge about the quality of dose distribution including maximum/mean dose at each structure. The direct representation of DVH results in a discontinuous, non-differentiable, and non-convex function which makes it difficult to integrate it into any DL model. Nguyen et al (2020) proposed a a continuous and differentiable, yet non-convex, DVH-based loss function (not to be confused with predicting DVH). In this paper, we propose a differentiable and convex surrogate loss function for DVH using multiple moments of dose distribution. It has been previously shown that moments can approximate a DVH to an arbitrary accuracy (Zinchenko et al 2008), and also moments have been successfully used to replicate a desired DVH (Zarepisheh et al 2013). The convexity and differentiability of the moment-based loss function makes it computationally appealing and also less prone to the local optimality. Furthermore, using different moments for different structures allows the DL model to drive the prediction according to the clinical priorities.
2. Materials and method
2.1. Loss functions
We use three types of loss functions in our study. First, we use MAE that measures the error between paired observations of real and predicted 3D dose. MAE is defined as where N is the total number of voxels and Dp, Dr are the predicted and real doses. We preferred to use MAE versus a common alternative, mean squared error (MSE), as MAE produces less blurring in the output compared to MSE (Isola et al 2017). MAE loss is one of the widely used loss in machine learning and it is calculated as the mean of absolute difference between the predicted and reference dose.
| (1) |
where is the predicted dose and is the reference dose for the ith voxel.
2.1.1. Sigmoid-based DVH loss
Nguyen et al (2020) proposed approximating the heaviside step function by the readily differentiable sigmoid function to address discontinuity and non-differentiability issues of DVH function. Borrowing the notation from Nguyen et al (2020), for a given volumetric dose distribution D and a segmentation mask Bs for the sth structure, the volume-at-dose with respect to the dose dt, denoted by vs,t(D, Bs), is defined as the volume fraction of a given region-of-interest (OARs or PTV) which receives a dose of at least dt or higher which can be approximated as:
| (2) |
where σ is the sigmoid function, , β is histogram bin width, and i loops over the voxel indices of the dose distribution. Based on this, the DVH for the structures s is defined as
| (3) |
The DVH loss can be calculated using MSE between the real and predicted dose DVH and is defined as follows:
| (4) |
Total loss for training the UNET using sigmoid based non-convex DVH loss is then given by
| (5) |
where wDVH is the weight for DVH loss (in our experiments we used wDVH = 10 based on validation results).
2.1.2. Moment loss
Moment loss is based on the idea that a DVH can be well-approximated using a few moments (Zinchenko et al 2008, Zarepisheh et al 2013):
where Mp represents the moment of order p defined as:
| (6) |
where Vs is a set of voxels belonging to the structure s, and d is the dose. M1 is simply the mean dose of a structure whereas M∞ represents the max dose, and for p > 1, Mp represents a value between mean and max doses.
In our experiments, we used a combination of three moments P = {1, 2, 10} for the critical OARs and PTV, where M1 is exactly the mean dose, M2 is the dose somewhere between the mean and max dose and M10 approximates the max dose.
The moment loss is calculated using MSE between the actual and predicted moment for the structure:
| (7) |
where Mp and are the pth moment of the actual dose and the predicted dose of a given structure, respectively.
Total loss for training the neural network using moment based convex loss function is then defined as
| (8) |
where wMoment is the weight for the moment loss (in our experiments we used wMoment = 0.01 based on validation results).
2.2. Patient dataset
We used 360 randomly selected lung cancer patients treated with conventional IMRT with 60 Gy in 30 fractions at Memorial Sloan Kettering Cancer Center between the years 2017 and 2020. All these patients received treatment and therefore included the treated plans which were manually generated by experienced planners using 5–7 coplanar beams and 6 MV energy. Table 1 refers to the clinical criteria used at our institution. All these plans were generated using Eclipse™ V13.7–V15.5 (Varian Medical Systems, Palo Alto, CA, USA).
Table 1.
Clinical max/mean dose (in Gy) and dose-volume criteria.
| Structure | Max (Gy) | Mean (Gy) | Dose-volume |
|---|---|---|---|
| PTV | 72 | ||
| Lungs not GTV | 66 | 21 | V (20 Gy) < = 37% |
| Heart | 66 | 20 | V (30 Gy) < = 50% |
| Stomach | 54 | 30 | |
| Esophagus | 66 | 34 | |
| Liver | 66 | V (30 Gy) < = 50% | |
| Cord | 50 | ||
| Brachial plexus | 65 |
2.3. Inputs and preprocessing
Structure contours and the 3D dose distribution were extracted from the Eclipse V15.5 (Varian Medical Systems, Palo Alto, CA, USA). Each patient has a planning CT and manually delineated contours of PTV and OARs which may differ from patient to patient depending on the location and size of the tumor. However, all patients have esophagus, spinal cord, heart, left lung, right lung and PTV delineated. Hence, we use these five OARs and PTV as inputs in addition to the planning CT. Figure 1 shows the overall workflow to train a CNN to generate voxel-wise dose distribution. The CT images may have different spatial resolutions but have the same in-plane matrix dimensions of 512 × 512. The PTV and OAR segmentation dimensions match those of the corresponding planning CTs. The intensity values of the input CT images are first clipped to have range of [−1024, 3071] and then re-scaled to range [0, 1] for input to the DL network. The OAR segmentations are converted to a one-hot encoding scheme with value of 1 inside each anatomy and 0 outside. The PTV segmentation is then added as an extra channel to the one-hot encoded OAR segmentation.
Figure 1.

Overview of data processing pipeline and training a 3D network to generate a 3D voxel-wise dose. OARs are one-hot encoded and concatenated along the channel axis with CT and PTV input to the network.
The dose data have different resolutions than the corresponding CT images. Each pair of the doses is first re-sampled to match the corresponding CT image. For easier training and comparison between different patients, the mean dose inside PTV of all patients is re-scaled to 60 Gy. This serves as a normalization for comparison between patients and can be easily shifted to a different prescription dose by a simple re-scaling inside the PTV region.
Finally, in order to account for the GPU RAM budget, we crop a 300 × 300 × 128 region from all the input matrices (CT/OAR/PTV/Dose) and re-sample it to a consistent 128 × 128 × 128 dimensions. We used the OAR/PTV segmentation masks to guide the cropping to avoid removing any critical regions of interest.
2.4. CNN architecture
Unet is a fully connected network which has been widely used in the medical image segmentation. We train a UNet like CNN architecture (Ronneberger et al 2015, Isola et al 2017) to output the voxel-wise 3D dose prediction corresponding to an input comprising of 3D CT/contours which are concatenated along the channel dimension. The network follows a common encoder–decoder style architecture which is composed of a series of layers which progressively downsample the input (encoder) using max pooling operation, until a bottleneck layer, where the process is reversed (decoder). Additionally, UNet-like skip connections are added between corresponding layers of encoder and decoder. This is done to share low-level information between the encoder and decoder counterparts.
The network (figure 2) uses Convolution-BatchNorm-ReLU-Dropout as a block to perform series of convolution. Dropout is used with a dropout rate of 50%. Maxpool is used to downsample the image by 2 in each spatial level of encoder. All the convolutions in the encoder are 3 × 3 × 3 3D spatial filters with a stride of 1 in all 3 directions. In the decoder we use trilinear upsampling followed by regular 2 × 2 × 2 stride 1 convolution. The last layer in the decoder maps its input to a one channel output (1283, 1).
Figure 2.

3D Unet architecture used to predict 3D voxelwise dose.
2.5. Evaluation criteria
To evaluate the quality of the predicted doses, we adopt the metrics used in recent AAPM ‘open-access KBP grand challenge’ (OpenKBP) (Babier et al 2021). This competition was designed to advance fair and consistent comparisons of dose prediction methods for KBP in radiation therapy research. The competition organizers used two separate scores to evaluate dose prediction models: dose score, which evaluates the overall 3D dose distribution and a DVH score, which evaluates a set of DVH metrics. The dose score was simply the MAE between real dose and predicted dose. The DVH score which was chosen as a radiation therapy specific clinical measure of prediction quality involved a set of DVH criteria for each OAR and target PTV. Mean/D(0.1cc) dose received by OAR was used as the DVH criteria for OAR while PTV had three criteria: D99, D95 and D1 which are the doses received by 99% (1st percentile), 95% (5th percentile), and 1% (99th percentile) of voxels in the target PTV. DVH error, the absolute difference between the DVH criteria for real and predicted dose, was used to evaluate the DVHs. Average of all DVH errors was taken to encapsulate the different DVH criteria (mean and D(0.1cc) for all OARs and D1/D95/D99 for target) into a single score defined as DVH score, measuring the DVH quality of the predicted dose distributions. We also report clinical evaluation criteria such as homogeneity index (HI) (Hodapp 2012) given by and paddick conformity index (PCI) (Paddick 2000) which is given by, , where TV = target volume, PIV = prescription isodose volume and TVPIV = volume of the target covered by the prescription isodose.
2.6. DL settings
In our experiments, we used Stochastic Gradient Descent (SGD), with a batch size of 1, and Adam optimizer (Kingma and Ba 2015) with an initial learning rate of 0.0002, and momentum parameters β1 = 0.5, β2 = 0.999. We trained the network for total of 200 epochs. We used a constant learning rate of 0.0002 for the first 100 epochs and then let the learning rate linearly decay to 0 for the final 100 epochs. When using the MAE and DVH combined loss, we scaled the DVH component of the loss by a factor of 10. Also, when using the MAE and Moment combined loss, we used the weight of 0.01 for moment loss based upon our validation results.
We divided our training set of 290 images into train/validation set of 240 and 50 images respectively and determined the best learning rate and scaling factor for (MAE + DVH) loss and (MAE + moment) loss. Afterwards, we trained all our models using all 290 training datasets and tested on the holdout 70 datasets used for reporting results.
We created the implementations of the CNN model, loss functions and other related training/testing scripts in pytorch and we conducted all our experiments on an NVIDIA A40 GPU with 48 GB VRAM.
3. Results
Figure 3(a) compares the model trained using MAE loss versus (MAE + moment) loss with respect to different metrics. The y-axis represents the relative improvement obtained using (MAE + moment) loss, comparing the prediction and ground truth (manual plan). While both models performed similarly with respect to the dose-score and training time (~7 h), the (MAE + moment) loss improved the DVH score, homogeneity error and conformity error by 11%, 32%, and 21% respectively. For statistical tests, Wilcoxon signed-rank test was performed for different metrics and p = 0.05 was considered as statistical significance. The difference in DVH-score and conformity error was found statistically significant (p < 0.01).
Figure 3.

Comparison of different metrics for (a) MAE versus (MAE + moment) and (b) (MAE + DVH) versus (MAE + moment) losses. Y axis shows the relative improvement (in %) using (MAE + moment) loss compared to MAE and (MAE + DVH) loss. The higher is always better. For statistical analysis, Wilcoxon signed-rank test was used and p = 0.05 was considered statistically significant.
Similarly, figure 3(b) compares the model trained using (MAE + DVH) loss versus (MAE + moment) loss and shows the relative improvement obtained by (MAE + moment) loss. (MAE + moment) significantly improves the training time (almost half), while modestly improves the DVH score, homogeneity and conformity errors (about 7%–8%). The significant improvement in training time for (MAE + moment) loss owes to its convexity and simplicity. Statistically significant difference (p < 0.01) was observed for DVH score.
Figure 4 shows the average absolute error between actual and predicted dose in terms of percentage of prescription for different clinically relevant criteria. Critical OARs like cord and esophagus showed substantial improvement in max and mean absolute dose error respectively using (MAE + moment) loss compared to other two in the category. PTV D95 and mean dose showed marginal improvements in the dose prediction quality compared to MAE loss. There was small/no-improvement in the MAE for other healthy organs (i.e. left lung, right lung, heart). Figure 5 compares the DVH of an actual dose (ground-truth here) with two predictions obtained from two different loss functions for a patient. As can been seen, in general, the prediction generated with (MAE + moment) loss resembles the actual ground-truth dose more than the other model, for this particular patient, especially for PTV.
Figure 4.

Absolute error (in percentage of prescription) for clinically relevant max/mean dose for organ-at-risk and PTV D1/D95 and mean dose. The lower is always the better. Error bars represents 95% confidence interval for all the test patients.
Figure 5.

DVH plots for different structures using (i) actual dose, predicted dose using (ii) MAE + DVH loss, and (iii) MAE + moment loss, for one patient.
Figure 6 shows the comparison of the absolute error for the model trained with default moments (p = 1, 2, 10) for all the structures (red bar) and the model that used different moments for cord (p = 5, 10) and heart (p = 1, 2) and default moments (p = 1, 2, 10) for all other structures (blue bar). As can be seen in the figure, using higher-order moments for cord improves the maximum dose prediction, while using lower-order moments for heart improves the mean dose prediction. Hence, choosing the right set of moments for different structures, one can incorporate different clinically relevant criteria into the model e.g. if max dose for the structure is important, one can use higher-order moments (p = 5, 10) for the structure in loss function and on the other hand, if mean dose is important one can use lower-order moment (p = 1, 2).
Figure 6.

Absolute error (in percentage of prescription) for max/mean dose for organ-at-risk and PTV D1/D95/D99 dose. The lower is always the better. Absolute error in the green region shows the impact of customized moments. Error bars represents 95% confidence interval for all the test patients.
4. Discussion
In this study, we have employed moments as a surrogate loss function to integrate DVH into DL (DL) 3D dose prediction. Moments provide a mathematically rigorous and computationally efficient way to incorporate DVH information in any DL architecture without any computational overhead. This allows us to incorporate the domain specific knowledge and clinical priorities into the DL model. Using MAE + moment loss means the DL model tries to match the actual dose (ground-truth) not only at a micro-level (voxel-by-voxel using MAE loss) but also at a macro-level (structure-by-structure using representative moments).
Moments are essentially simple polynomial functions which can be calculated efficiently. Given their convexity, they do not suffer from the local optimality issue, making them more reliable choices with more robust behaviour against the stochastic nature of the optimization techniques, commonly used in DL models. The computational efficiency of the moments allows training of large DL models. They also offer better fine-tuning of the hyper parameters.
The moments in conjunction with MAE help to incorporate DVH information into the DL model, however, the MAE loss still plays the central role in the prediction. In particular, the moments lack any spatial information about the dose distribution which is provided by the MAE loss. The MAE loss has also been successfully used across many applications and its performance is well-understood. Further research is needed to investigate the performance of the moment loss on more data especially with different disease sites.
The 3D dose prediction can facilitate and accelerate the treatment planning process by providing a reference plan which can be fed into a treatment planning optimization framework to be converted into a deliverable Pareto optimal plan. The dose-mimicking approach has been commonly used in the literature Fan et al (2019), seeking the closest deliverable plan to the reference plan using quadratic function as a measure of distance. Babier et al (2020) proposed an inverse optimization framework which estimates the objective weights from the reference plan and then generates the deliverable plan by solving the corresponding optimization problem. Any improvements in the prediction, including the ones we have obtained using our proposed moment loss, needs to be ultimately evaluated using the entire pipeline of predicting a plan and converting that into a deliverable plan.
5. Conclusion
This work shows that the moments are powerful tools with sound mathematical properties to integrate DVH as an important domain knowledge into 3D dose prediction without any computational overhead. The idea has been validated on a large dataset of 360 conventional lung patients. The conventional lung patients are usually considered challenging cases in the clinic due to their large PTV sizes and sensitive nearby critical structures (e.g. esophagus, heart).
Acknowledgments
This project was partially supported by MSK Cancer Center Support Grant/Core Grant (P30 CA008748).
References
- Appenzoller LM, Michalski JM, Thorstad WL, Mutic S and Moore KL 2012. Predicting dose-volume histograms for organs-at-risk in IMRT planning Med. Phys 39 7446–61 [DOI] [PubMed] [Google Scholar]
- Babier A, Mahmood R, McNiven AL, Diamant A and Chan TC 2020. The importance of evaluating the complete automated knowledge-based planning pipeline Phys. Med 72 73–9 [DOI] [PubMed] [Google Scholar]
- Babier A, Zhang B, Mahmood R, Moore KL, Purdie TG, McNiven AL and Chan TC 2021. OpenKBP: the open-access knowledge-based planning grand challenge and dataset Med. Phys 48 5549–61 [DOI] [PubMed] [Google Scholar]
- Berry SL, Boczkowski A, Ma R, Mechalakos J and Hunt M 2016. Interobserver variability in radiation therapy plan output: results of a single-institution study Pract. Radiat. Oncol 6 442–9 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Breedveld S, Storchi PR and Heijmen BJ 2009. The equivalence of multi-criteria methods for radiotherapy plan optimization Phys. Med. Biol 54 7199–209 [DOI] [PubMed] [Google Scholar]
- Chin Snyder K, Kim J, Reding A, Fraser C, Gordon J, Ajlouni M, Movsas B and Chetty IJ 2016. Development and evaluation of a clinical model for lung cancer patients using stereotactic body radiotherapy (SBRT) within a knowledge-based algorithm for treatment planning J. Appl. Clin. Med. Phys 17 263–75 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Craft D and Bortfeld T 2008. How many plans are needed in an IMRT multi-objective plan database? Phys. Med. Biol 53 2785–96 [DOI] [PubMed] [Google Scholar]
- Das IJ, Moskvin V and Johnstone PA 2009. Analysis of treatment planning time among systems and planners for intensity-modulated radiation therapy J. Am. Coll. Radiol 6 514–7 [DOI] [PubMed] [Google Scholar]
- Fan J, Wang J, Chen Z, Hu C, Zhang Z and Hu W 2019. Automatic treatment planning based on three-dimensional dose distribution predicted from deep learning technique Med. Phys 46 370–81 [DOI] [PubMed] [Google Scholar]
- Fogliata A. et al. Performance of a knowledge-based model for optimization of volumetric modulated arc therapy plans for single and bilateral breast irradiation. PLoS One. 2015;10:e0145137. doi: 10.1371/journal.pone.0145137. [DOI] [PMC free article] [PubMed] [Google Scholar]
- Han X 2017. MR-based synthetic CT generation using a deep convolutional neural network method Med. Phys 44 1408–19 [DOI] [PubMed] [Google Scholar]
- Hodapp N 2012. The ICRU Report 83: prescribing, recording and reporting photon-beam intensity-modulated radiation therapy (IMRT) Strahlentherapie und Onkologie: Organ der Deutschen Rontgengesellschaft…[et al] 188 97–9 [DOI] [PubMed] [Google Scholar]
- Ibragimov B and Xing L 2017. Segmentation of organs-at-risks in head and neck CT images using convolutional neural networks Med. Phys 44 547–57 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Isola P, Zhu J-Y, Zhou T and Efros AA 2017. Image-to-Image Translation with Conditional Adversarial Networks Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 1125–1134 (https://openaccess.thecvf.com/content_cvpr_2017/html/Isola_Image-To-Image_Translation_With_CVPR_2017_paper.html) [Google Scholar]
- Kingma DP and Ba J 2014. Adam: a method for stochastic optimization 10.48550/arxiv:1412.6980 [DOI] [Google Scholar]
- Liu Z, Chen X, Men K, Yi J and Dai J 2020. A deep learning model to predict dose-volume histograms of organs at risk in radiotherapy treatment plans Med. Phys 47 5467–81 [DOI] [PubMed] [Google Scholar]
- Men K, Dai J and Li Y 2017. Automatic segmentation of the clinical target volume and organs at risk in the planning CT for rectal cancer using deep dilated convolutional neural networks Med. Phys 44 6377–89 [DOI] [PubMed] [Google Scholar]
- Moore KL et al. 2015. Quantifying unnecessary normal tissue complication risks due to suboptimal planning: A secondary study of RTOG0126 Int. J. Radiat. Oncol.* Biol.* Phys 92 228–35 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Nelms BE, Robinson G, Markham J, Velasco K, Boyd S, Narayan S, Wheeler J and Sobczak ML 2012. Variation in external beam treatment plan quality: an inter-institutional study of planners and planning systems Pract. Radiat. Oncol 2 296–305 [DOI] [PubMed] [Google Scholar]
- Nguyen D, McBeth R, Sadeghnejad Barkousaraie A, Bohara G, Shen C, Jia X and Jiang S 2020. Incorporating human and learned domain knowledge into training deep neural networks: A differentiable dose-volume histogram and adversarial inspired framework for generating Pareto optimal dose distributions in radiation therapy Med. Phys 47 837–49 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Paddick I 2000. A simple scoring ratio to index the conformity of radiosurgical treatment plans J. Neurosurg 93 219–22 [DOI] [PubMed] [Google Scholar]
- Ronneberger O, Fischer P and Brox T 2015. U-Net: convolutional networks for biomedical image segmentation Medical Image Computing and Computer-Assisted Intervention – MICCAI ( 10.1007/978-3-319-24574-4_28) [DOI] [Google Scholar]
- Tol JP, Delaney AR, Dahele M, Slotman BJ and Verbakel WF 2015. Evaluation of a knowledge-based planning solution for head and neck cancer Int. J. Radiat. Oncol.*Biol.*Phys 91 612–20 [DOI] [PubMed] [Google Scholar]
- Valdes G, Simone CB, Chen J, Lin A, Yom SS, Pattison AJ, Carpenter CM and Solberg TD 2017. Clinical decision support of radiotherapy treatment planning: a data-driven machine learning strategy for patient-specific dosimetric decision making Radiother. Oncol 125 392–7 [DOI] [PubMed] [Google Scholar]
- Wang M, Zhang Q, Lam S, Cai J and Yang R 2020. A review on application of deep learning algorithms in external beam radiotherapy automated treatment planning Front. Oncol 10 580919–580919 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zarepisheh M, Hong L, Zhou Y, Oh JH, Mechalakos JG, Hunt MA, Mageras GS and Deasy JO 2019. Automated intensity modulated treatment planning: the expedited constrained hierarchical optimization (ECHO) system Med. Phys 46 2944–54 [DOI] [PMC free article] [PubMed] [Google Scholar]
- Zarepisheh M, Long T, Li N, Tian Z, Romeijn HE, Jia X and Jiang SB 2014. A DVH-guided IMRT optimization algorithm for automatic treatment planning and adaptive radiotherapy replanning Med. Phys 41 061711. [DOI] [PubMed] [Google Scholar]
- Zarepisheh M, Shakourifar M, Trigila G, Ghomi P, Couzens S, Abebe A, Noreña L, Shang W, Jiang SB and Zinchenko Y 2013. A moment-based approach for DVH-guided radiotherapy treatment plan optimization Phys. Med. Biol 58 1869–87 [DOI] [PubMed] [Google Scholar]
- Zinchenko Y, Craig T, Keller H, Terlaky T and Sharpe M 2008. Controlling the dose distribution with gEUD-type constraints within the convex radiotherapy optimization framework Phys. Med. Biol 53 3231–50 [DOI] [PubMed] [Google Scholar]
