Abstract
Purpose
Knowledge‐based planning (KBP) has been shown to be an effective tool in quality control for intensity‐modulated radiation therapy treatment planning and generating high‐quality plans. Previous studies have evaluated its ability to create consistent plans across institutions and between planners within the same institution as well as its use as teaching tool for inexperienced planners. This study evaluates whether planning quality is consistent when using a KBP model to plan across different treatment machines.
Materials and methods
This study used a RapidPlan model (Varian Medical Systems) provided by the vendor, to which we added additional planning objectives, maximum dose limits, and planning structures, such that a clinically acceptable plan is achieved in a single optimization. This model was used to generate and optimize volumetric‐modulated arc therapy plans for a cohort of 50 patients treated for head‐neck cancer. Plans were generated using the following treatment machines: Varian 2100, Elekta Versa HD, and Varian Halcyon. A noninferiority testing methodology was used to evaluate the hypothesis that normal and target metrics in our autoplans were no worse than a set of clinically‐acceptable baseline plans by a margin of 1.8 Gy or 3% dose‐volume. The quality of these plans were also compared through the use of common clinical dose‐volume histogram criteria.
Results
The Versa HD met our noninferiority criteria for 23 of 34 normal and target metrics; while the Halcyon and Varian 2100 machines met our criteria for 24 of 34 and 26 of 34 metrics, respectively. The experimental plans tended to have less volume coverage for prescription dose planning target volume and larger hotspot volumes. However, comparable plans were generated across different treatment machines.
Conclusions
These results support the use of a head‐neck RapidPlan models in centralized planning workflows that support clinics with different linac models/vendors, although some fine‐tuning for targets may be necessary.
Keywords: automated planning, knowledge‐based planning
1. INTRODUCTION
Intensity‐modulated radiation therapy treatment planning is a challenging and time‐consuming process that can vary in quality between planners across institutions 1 , 2 and even within the same institution. 3 Knowledge‐based planning (KBP) has been shown to be effective in creating high‐quality treatment plans, 4 , 5 , 6 , 7 , 8 , 9 reducing variability between planners, 9 and evaluating plan consistency and quality. 2 , 3 , 11 , 12 , 13 , 14 These factors make KBP models ideal for centralized planning systems, such as the radiation planning assistant (RPA).
The RPA is a web‐based automated treatment planning system that is being developed to provide high‐quality contours and treatment planning to clinics with limited resources around the world. The RPA's development has been described in previous works. 5 , 15 , 16 , 17 , 18 , 19 , 20 , 21 , 22 Through the Eclipse application programming interface, the RPA uses RapidPlan (Varian Medical Systems, Palo Alto, CA), a commercial KBP system, to generate and optimize treatment plans for various cancer sites.
Since we aim to provide plans generated by the RPA to multiple clinics around the world, we needed to evaluate how plan quality is affected when a single RapidPlan model is used to generate plans with different treatment machines that have different beam qualities and multileaf collimator (MLC) characteristics. In this study, we started with a KBP approach that was developed for head and neck treatment on a Varian 2100 series linac, with plan quality validated by radiation oncologists from multiple institutions. We then used the same approach to create volumetric‐modulated arc therapy (VMAT) plans for treatment on Versa HD (Elekta) and Halcyon (Varian) treatment devices. The plans were then dosimetrically evaluated for quality by comparing them to our baseline, physician‐reviewed autoplans. Previous studies have shown the robustness of KBP models to tumor location, treatment modalities, and institutional protocols, 23 , 24 but to our knowledge, this is the first study to evaluate a RapidPlan model across different treatment machines. This comparison is important for centers (or treatment planning services) where there is a need to accommodate different models and vendors.
2. METHODS
2.1. Patient data
For this analysis, a cohort of 50 patients with head and neck cancer was retrospectively collected and de‐identified. This study was approved by the institutional review board. All patients were previously treated using VMAT. The patients’ original, physician‐drawn targets, and normal tissue contours, along with the original CT scan and dose prescription, were used in autoplan generation. The patients had primary tumors from various subsites such as the nasopharynx, oropharynx, and oral cavity (Table 1).
TABLE 1.
Primary site | Number of patients | Range of PTV1 dose, Gy |
---|---|---|
Oropharynx | 21 | 60–70 |
Oral cavity | 8 | 60–70 |
Larynx | 3 | 70 |
Hypopharynx | 8 | 60‐68 |
Nasopharynx | 10 | 60‐70 |
2.2. Plan optimization
The RPA uses a RapidPlan model to optimize HN plans in Eclipse Treatment Planning System (Varian Medical Systems, Palo Alto, CA). This model was developed using the vendor‐provided Washington University HN RapidPlan model as a starting point. Additional planning objectives, maximum dose limits, and additional planning structures were added to the model. The model was refined through iterative testing and physician feedback using a set of validation patients that were not included in this study. The full planning strategy is outlined in a previous publication by our group. 5 The model was optimized for a Varian 2100 linac.
2.3. Baseline plans
In a previous study by our group, a set of Varian 2100 plans was generated for our patient cohort using Eclipse version 15.5—in which the autoplans were compared to the original, clinically‐approved plan in a blinded review by physicians—and scored for clinical acceptability based on contour quality and dose coverage. 5 From this study, 49/50 of the autoplans were deemed clinically acceptable by the reviewers. These physician‐reviewed autoplans were dubbed our baseline plans and were used as a basis for our comparison.
2.4. Plan generation
Since the creation of our baseline plans, we upgraded Eclipse to version 15.6. Three plans were generated per patient in our cohort using the following machines: Varian 2100, Elekta Versa HD, and Varian Halcyon. We decided to re‐plan the Varian 2100 cases in Eclipse v15.6 in order to verify that any dose differences seen were not the result of the difference in Eclipse versions. The plans generated in Eclipse v15.6 were dubbed our experimental plans.
The Varian 2100 (both baseline and experimental) and Versa HD plans consisted of three 360° coplanar treatment arcs with collimator angles of 15°, 345°, and 90°. The Halcyon plans consisted of four treatment arcs with collimator angles of 0°, 45°, 90°, and 315°. Three of the 50 patients had two planning target volume (PTV) dose levels, while the other 47 had three PTV dose levels.
2.5. Evaluation process
To determine the quality of our experimental plans, we used a one‐sided Mann‐Whitney U test at 95% confidence to determine whether the experimental doses were noninferior to the baseline doses by a margin, δ, we determined based on clinical judgement. 25 , 26 , 27 For a given dose‐volume histogram (DVH) metric, let M B be the median value for the baseline plans, and M E be the median value for the experimental plans. For DVH metrics in which lower values are better, we have the following hypotheses:
For metrics in which higher values are better, we have the following hypothesis:
Experimental plans are considered noninferior if the null hypothesis is rejected (i.e., p < 0.05). A 95% confidence interval for M E‐M B was calculated in order to conclude noninferiority. For DVH metrics in which lower values were better, the upper limit of the 95% confidence interval needed to be less than the margin in order to conclude noninferiority. For metrics in which higher values were better, the lower limit of the confidence interval needed to be greater than the margin. For normal structures, D95%, and D98%, δ = 1.8 Gy (i.e., 3% of the lowest dose prescription in our dataset); for V95%, V100%, V105%, and V110%, δ = 3%. The number of plans that met established clinically accepted dosimetric criteria outlined in Radiation Therapy Oncology Group protocol 1016 28 was also calculated and compared in order to assess the clinical acceptability of the autoplans.
3. RESULTS
3.1. Noninferiority
Table 2 shows the confidence interval and p‐values from our statistical test for each DVH criteria. The Versa HD was noninferior (p < 0.05) for 23 of 34 DVH metrics evaluated; Halcyon was noninferior for 24 of 34 metrics; the Varian 2100 re‐plan was noninferior for 26 of 34 metrics. We were not able to conclude noninferiority for brain, brainstem, both cochleae, ipsilateral parotid, brainstem with 5 mm margins, and spinal cord with 5‐mm margins. For some DVH metrics, we were able to conclude noninferiority for some experimental machines (Versa, Halcyon, 2100 v15.6), but not all. This can be seen in intermediate‐dose V100%, V105%, and low‐dose V105% PTVs, where the p‐values for Versa and Halcyon were above 0.05, but for 2100 v15.6, it was less than 0.05. Figure 1 shows the distribution of planned dose to the normal structures for the brainstem, right cochlea, ipsilateral parotid, and low‐dose PTV target coverage at 100%, 105%, and 110% of the prescription dose. The distributions for the brainstem, right cochlea, ipsilateral parotid, and low‐dose PTV target coverage at 105% and 110% show that there was very good agreement between the experimental and baseline plans, with r‐squared values greater than 0.82 for all experimental machines. However, there was not very good agreement (R 2 < 0.50) for low‐dose target coverage at 100%, due to an outlier. Figure 2 shows a DVH for one of the patients, comparing brainstem, both parotids, and PTVs for the baseline and experimental autoplans. All four plans showed good agreement for those structures.
TABLE 2.
Noninferiority test | ||||||
---|---|---|---|---|---|---|
Versa | Halcyon | 2100 v15.6 | ||||
Structure, metric | 95% CI | p‐Value | 95% CI | p‐Value | 95% CI | p‐Value |
Brain, max Dose | (−3.9, 5.0) | 0.299 | (−2.3, 6.7) | 0.516 | (−2.9, 6.5) | 0.421 |
BrainStem, max Dose | (−0.7, 2.6) | 0.195 | (−1.3, 2.2) | 0.100 | (−1.3, 2.0) | 0.072 |
Chiasm, max Dose | (−0.1, 0.7) | <0.001 | (−0.1, 0.6) | <0.001 | (−0.3, 0.5) | <0.001 |
Cochlea_L, max dose | (−3.7, 2.9) | 0.123 | (−3.3, 3.5) | 0.175 | (−3.0, 4.3) | 0.267 |
Cochlea_R, max dose | (−3.4, 2.6) | 0.099 | (−2.9, 3.8) | 0.161 | (−2.2, 4.8) | 0.272 |
Con_Parotid, mean dose | (−1.3, 2.3) | 0.093 | (−2.3, 1.3) | 0.020 | (−1.6, 2.0) | 0.064 |
Eye_L, max dose | (−0.2, 0.6) | <0.001 | (−0.4, 0.4) | <0.001 | (−0.4, 0.5) | <0.001 |
Eye_R, max dose | (−0.2, 0.6) | <0.001 | (−0.4, 0.5) | <0.001 | (−0.4, 0.6) | <0.001 |
Ips_Parotid, mean dose | (−2.2, 3.4) | 0.202 | (−3.1, 2.5) | 0.099 | (−2.5, 3.0) | 0.147 |
Lens_L, max dose | (0.0, 0.6) | <0.001 | (−0.1, 0.4) | <0.001 | (−0.2, 0.3) | <0.001 |
Lens_R, max dose | (0.1, 0.6) | <0.001 | (−0.1, 0.3) | <0.001 | (−0.2, 0.3) | <0.001 |
Mandible, max dose | (−1.7, 1.3) | 0.023 | (−1.6, 1.2) | 0.025 | (−1.5, 1.5) | 0.034 |
OpticNrv_L, max dose | (−0.2, 0.7) | <0.001 | (−0.2, 0.6) | <0.001 | (−0.4, 0.5) | <0.001 |
OpticNrv_R, max dose | (−0.2, 0.7) | <0.001 | (−0.2, 0.6) | <0.001 | (−0.3, 0.5) | <0.001 |
SpinalCord, max dose | (0.0, 1.5) | 0.014 | (−0.9, 0.7) | <0.001 | (−0.6, 0.8) | <0.001 |
zBrainStem_05, max dosea | (−1.0, 3.1) | 0.303 | (−0.5, 3.8) | 0.477 | (−1.0, 3.2) | 0.349 |
zPTV1, D95% | (−1.0, 0.2) | 0.003 | (−0.8, 0.3) | 0.002 | (−0.5, 0.5) | 0.001 |
zPTV1, D98% | (−1.3, 0.2) | 0.011 | (−1.1, 0.3) | 0.005 | (−0.7, 0.6) | 0.002 |
zPTV1, V100% | (−3.0, ‐1.0) | 0.024 | (−3.0, 0.0) | 0.021 | (−1.0, 1.0) | <0.001 |
zPTV1, V105% | (1.0, 1.0) | <0.001 | (1.0, 1.0) | <0.001 | (0.0, 0.0) | <0.001 |
zPTV1, V95% | (0.0, 0.0) | <0.001 | (0.0, 0.0) | <0.001 | (0.0, 0.0) | <0.001 |
zPTV2, D95% | (−1.2, 0.3) | 0.006 | (−1.3, 0.3) | 0.007 | (−0.7, 0.9) | 0.001 |
zPTV2, D98% | (−1.3, 0.3) | 0.008 | (−1.4, 0.2) | 0.009 | (−0.7, 0.8) | <0.001 |
zPTV2, V100% | (−4.0, ‐1.0) | 0.312 | (−5.0, −1.0) | 0.440 | (−1.0, 2.0) | <0.001 |
zPTV2, V105% | (2.0, 8.0) | 0.864 | (1.0, 8.0) | 0.813 | (−5.0, 2.0) | 0.030 |
zPTV2, V110% | (0.0, 1.0) | 0.002 | (0.0, 2.0) | 0.003 | (−1.0, 1.0) | <0.001 |
zPTV2, V95% | (−1.0, 0.0) | <0.001 | (−1.0, 0.0) | <0.001 | (0.0, 0.0) | <0.001 |
zPTV3, D95% | (−0.4, ‐0.1) | <0.001 | (−0.5, −0.2) | <0.001 | (−0.2, 0.0) | <0.001 |
zPTV3, D98% | (−0.5, ‐0.2) | <0.001 | (−0.6, −0.3) | <0.001 | (−0.2, 0.1) | <0.001 |
zPTV3, V100% | (−2.0, ‐1.0) | <0.001 | (−2.0, −1.0) | 0.002 | (−1.0, 0.0) | <0.001 |
zPTV3, V105% | (5.0, 8.0) | 0.978 | (7.0, 9.0) | 0.997 | (−2.0, 0.0) | 0.018 |
zPTV3, V110% | (0.0, 1.0) | <0.001 | (1.0, 2.0) | 0.018 | (0.0, 0.0) | <0.001 |
zPTV3, V95% | (0.0, 0.0) | <0.001 | (−1.0, 0.0) | <0.001 | (0.0, 0.0) | <0.001 |
zSpinalCord_05, max dosea | (−0.2, 1.7) | 0.133 | (−0.9, 1.2) | 0.054 | (−0.4, 1.5) | 0.099 |
Abbreviations: DX%, doses at the given percent of target coverage; D max, maximum dose; D mean, mean dose; VX%, percent of the target receiving the prescribed dose.
Structure with a 5 mm margin.
3.2. Dosimetric criteria
The number of autoplans that met established clinical dosimetric criteria are listed in Table 3, along with the original clinical plans for comparison. All autoplans met clinical recommendations for the spinal cord, brainstem, optic chiasm, optic nerves, PTV1 (V1cc < 117% and V95% > 95%), PTV2 (V95% > 80%), and PTV3 (V95% > 80%).
TABLE 3.
Number of clinical/autoplans that met clinical dose recommendations | ||||||
---|---|---|---|---|---|---|
Structure | Constraint | Clinic | Baseline | Versa HD | Halcyon | 2100 v15.6 |
Spinal cord | Dmax < 45 Gy | 50 | 50 | 50 | 50 | 50 |
Brainstem | Dmax < 54 Gy | 48 | 50 | 50 | 50 | 50 |
Ipisilateral parotid | Dmean < 26 Gy | 29 | 24 | 21 | 25 | 24 |
V30Gy < 50% | 40 | 42 | 40 | 41 | 42 | |
Contralateral parotid | Dmean < 26 Gy | 44 | 41 | 41 | 40 | 40 |
V30Gy < 50% | 48 | 47 | 46 | 46 | 46 | |
Cochleae | Dmax < 35 Gy | 45 | 40 | 40 | 39 | 40 |
Optic chiasm | Dmax < 54 Gy | 49 | 50 | 50 | 50 | 50 |
Optic nerves | Dmax < 54 Gy | 49 | 50 | 50 | 50 | 50 |
Lenses | Dmax < 7 Gy | 44 | 45 | 46 | 47 | 45 |
Eyes | Dmax < 35 Gy | 45 | 45 | 45 | 45 | 45 |
Brain | Dmax < 54 Gy | 34 | 36 | 37 | 37 | 36 |
High‐dose PTV | V1cc < 110% | 48 | 49 | 50 | 48 | 50 |
V1cc < 117% | 50 | 50 | 50 | 50 | 50 | |
V95% > 95% | 50 | 50 | 50 | 50 | 50 | |
Intermediate‐dose PTV | V95% > 100% | 34 | 16 | 9 | 10 | 13 |
V95% > 80% | 50 | 50 | 50 | 50 | 50 | |
Low‐dose PTV* | V95% > 100% | 30 | 30 | 15 | 15 | 21 |
V95% > 80% | 47 | 47 | 47 | 47 | 47 |
Abbreviations: D max, maximum dose; D mean, mean dose; PTV, planning target volume; RTOG, Radiation Therapy Oncology Group.
Three patients were treated with only two PTVs.
4. DISCUSSION
The goal of this study was to evaluate the ability of a single RapidPlan model to generate comparable plans across different treatment machines that have different beam and MLC characteristics. Using a RapidPlan model that was optimized for a Varian 2100 linac, we were able to generate plans that were noninferior to our baseline in at least 68% of the DVH metrics evaluated for each set of experimental plans. For the metrics in which we were not able to conclude noninferiority, we conducted an investigation into the reason why the noninferiority criteria was not met. We noticed that there were outliers for the Halcyon and Varian 2100 (v.15.6) machines in the following normal structures: brain, brainstem, right cochlea, left cochlea, and brainstem (with 5‐mm margins). These outliers came from the same two patients, one of them being the patient in which the clinically unacceptable baseline plan was generated for in the aforementioned previous study. 5 Both of these patients were planned with two PTVs. The Halcyon and Varian 2100 plans for these patients had large dose gradients around the cochleae, and one had a target close to the brain, which led to a large dose gradient around the brain/brainstem. We also noticed that, for these two patients, the Halcyon and Varian 2100 (v.15.6) plans were more comparable to the original plans used treat in the clinic than to our baseline autoplans. The experimental Varian 2100 plans that were generated in Eclipse v15.6 were noninferior in the most DVH criteria. Overall, these plans were noninferior for all 17 DVH criteria for target structures.
The target volumes in our study were less homogeneous for the Versa HD and Halcyon plans than for the baseline plans. In general, these experimental plans had reduced coverage of the prescription dose (V95% and V100%) and larger hotspots (V105% and V110%) at all dose levels. This indicates that some additional fine‐tuning of the plans may be needed when applying the standard RapidPlan approach to different machines with different beam qualities and MLC characteristics. This sort of fine‐tuning, however, has been shown to require minimal effort. 29
This study has some limitations. The sample size for this study is a limiting factor of our test. This is apparent when looking at the mean dose to the parotids. The treatment machines that we were not able to confirm noninferiority for the parotids each had only one plan with a dose above the 1.8 Gy margin. Additionally, the parotid dose in these plans were less than 1 Gy above the margin. Prior to reproducing this study, we would need to increase the power of our tests by increasing our sample size. While the baseline plans were determined to be clinically acceptable, we are not able to conclude that the same for the experimental plans generated for this study. Furthermore, an inherent weakness with this study is that we can only confirm noninferiority for the model under investigation. However, we can confirm that it is possible to create noninferior plans with a KBP model across different treatment machines. We have also established a methodology for evaluating other KBP models across different treatment machines.
Evaluating the effects of applying a single RapidPlan model to linacs developed by various vendors is important as more automation tools, such as the RPA, are being developed that will allow for centralized planning. It will be important to understand the strengths and weakness of these models as they will be used on multiple linac models and vendors around the world. We have shown in this study there is very good agreement between plans generated with different linac types, although some fine‐tuning of models may be needed to improve target coverage and minimize hotspots.
CONFLICT OF INTEREST
The authors declare that there is no conflict of interest that could be perceived as prejudicing the impartiality of the research reported.
AUTHOR CONTRIBUTIONS
Laurence Court, Adenike Olanrewaju, and Lifei Zhang conceived and designed this experiment. Beth Beadle provided clinical guidance for the model used in the experiment. Raphael Douglas carried out the experiment, data collection, and analysis of the data. Laurence Court supervised the findings in this work. Raphael Douglas wrote the manuscript with the support of Laurence Court.
ACKNOWLEDGEMENTS
The Radiation Planning Assistant project is partially funded by the National Cancer Institute (NIH CA‐202665) and Varian Medical Systems.
Douglas RJ, Olanrewaju A, Zhang L, Beadle BM, Court LE. Assessing the practicality of using a single knowledge‐based planning model for multiple linac vendors. J Appl Clin Med Phys. 2022;23:e13704. 10.1002/acm2.13704
REFERENCES
- 1. Das IJ, Cheng C‐W, Chopra KL, Mitra RK, Srivastava SP, Glatstein E. Intensity‐modulated radiation therapy dose prescription, recording, and delivery: patterns of variability among institutions and treatment planning systems. J Natl Cancer Inst. 2008;100(5):300‐307. [DOI] [PubMed] [Google Scholar]
- 2. Nelms BE, Robinson G, Markham J, et al. Variation in external beam treatment plan quality: an inter‐institutional study of planners and planning systems. Pract Radiat Oncol. 2012;2(4):296‐305. [DOI] [PubMed] [Google Scholar]
- 3. Berry SL, Ma R, Boczkowski A, Jackson A, Zhang P, Hunt M. Evaluating inter‐campus plan consistency using a knowledge based planning model. Radiother Oncol. 2016;120(2):349‐355. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4. Nawa K, Haga A, Nomoto A, et al. Evaluation of a commercial automatic treatment planning system for prostate cancers. Med Dosim. 2017;42(3):203‐209. [DOI] [PubMed] [Google Scholar]
- 5. Olanrewaju A, Court LE, Zhang L, et al. Clinical acceptability of automated radiation treatment planning for head and neck cancer using the radiation planning assistant. Pract Radiat Oncol. 2021;11(3):177‐184. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6. Tol JP, Delaney AR, Dahele M, Slotman BJ, Verbakel WFAR. Evaluation of a knowledge‐based planning solution for head and neck cancer. Int J Radiat Oncol Biol Phys. 2015;91(3):612‐620. [DOI] [PubMed] [Google Scholar]
- 7. Wu H, Jiang F, Yue H, Li S, Zhang Y. A dosimetric evaluation of knowledge‐based VMAT planning with simultaneous integrated boosting for rectal cancer patients. J Appl Clin Med Phys. 2016;17(6):78‐85. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Ziemer BP, Sanghvi P, Hattangadi‐Gluth J, Moore KL. Heuristic knowledge‐based planning for single‐isocenter stereotactic radiosurgery to multiple brain metastases. Med Phys. 2017;44(10):5001‐5009. [DOI] [PubMed] [Google Scholar]
- 9. Wu B, McNutt T, Zahurak M, et al. Fully automated simultaneous integrated boosted–intensity modulated radiation therapy treatment planning is feasible for head‐and‐neck cancer: a prospective clinical study. Int J Radiat Oncol Biol Phys. 2012;84(5):e647‐e653. [DOI] [PubMed] [Google Scholar]
- 10. Scaggion A, Fusella M, Roggio A, et al. Reducing inter‐ and intra‐planner variability in radiotherapy plan output with a commercial knowledge‐based planning solution. Physica Medica. 2018;53:86‐93. [DOI] [PubMed] [Google Scholar]
- 11. Appenzoller LM, Michalski JM, Thorstad WL, Mutic S, Moore KL. Predicting dose‐volume histograms for organs‐at‐risk in IMRT planning. Med Phys. 2012;39(12):7446‐7461. [DOI] [PubMed] [Google Scholar]
- 12. Moore KL, Schmidt R, Moiseenko V, et al. Quantifying unnecessary normal tissue complication risks due to suboptimal planning: a secondary study of RTOG 0126. Int J Radiat Oncol Biol Phys. 2015;92(2):228‐235. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Moore KL, Brame RS, Low DA, Mutic S. Experience‐based quality control of clinical intensity‐modulated radiotherapy planning. Int J Radiat Oncol Biol Phys. 2011;81(2):545‐551. [DOI] [PubMed] [Google Scholar]
- 14. Shiraishi S, Tan J, Olsen LA, Moore KL. Knowledge‐based prediction of plan quality metrics in intracranial stereotactic radiosurgery. Med Phys. 2015;42(2):908‐917. [DOI] [PubMed] [Google Scholar]
- 15. Court LE, Kisling K, McCarroll R, et al. Radiation planning assistant ‐ a streamlined, fully automated radiotherapy treatment planning system. J Vis Exp. 2018;2018(134):e57411‐e57411. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. Esho TO, Chung CV, Thompson JU, et al. Optimization of autogenerated chest‐wall radiation treatment plans developed for postmastectomy breast cancer patients in underserved clinics. Med Dosim. 2020;45(1):102‐107. [DOI] [PubMed] [Google Scholar]
- 17. Kisling K, Johnson JL, Simonds H, et al. A risk assessment of automated treatment planning and recommendations for clinical deployment. Med Phys. 2019;46(6):2567‐2574. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Kisling K, Cardenas C, Anderson BM, et al. Automatic verification of beam apertures for cervical cancer radiation therapy. Pract Radiat Oncol. 2020;10(5):e415‐e424. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Kisling K, Zhang L, Shaitelman SF, et al. Automated treatment planning of postmastectomy radiotherapy. Med Phys. 2019;46(9):3767‐3775. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Kisling K, Zhang L, Simonds H, et al. Fully automatic treatment planning for external‐beam radiation therapy of locally advanced cervical cancer: a tool for low‐resource clinics. Fully automatic treatment planning for external‐beam radiation therapy of locally advanced cervical cancer: a tool for low‐resource clinics. J Glob Oncol. 2019;. 5:1‐9. 10.1200/JGO.18.00107 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. McCarroll RE, Beadle BM, Balter PA, et al. Retrospective validation and clinical implementation of automated contouring of organs at risk in the head and neck: a step toward automated radiation treatment planning for low‐ and middle‐income countries. J Glob Oncol. 2018;. 4:1‐11. 10.1200/JGO.18.00055 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Rhee DJ, Cardenas CE, Elhalawani H, et al. Automatic detection of contouring errors using convolutional neural networks. Med Phys. 2019;46(11):5086‐5097. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23. Snyder KC, Kim J, Reding A, et al. Development and evaluation of a clinical model for lung cancer patients using stereotactic body radiotherapy (SBRT) within a knowledge‐based algorithm for treatment planning. J Appl Clin Med Phys. 2016;17(6):263‐275. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Wu B, Kusters M, Kunze‐Busch M, et al. Cross‐institutional knowledge‐based planning (KBP) implementation and its performance comparison to auto‐planning engine (APE). Radiother Oncol. 2017;123(1):57‐62. [DOI] [PubMed] [Google Scholar]
- 25. Schuirmann DJ. A comparison of the two one‐sided tests procedure and the power approach for assessing the equivalence of average bioavailability. J Pharmacokinet Biopharm. 1987;15(6):657‐680. [DOI] [PubMed] [Google Scholar]
- 26. Walker E, Nowacki AS. Understanding equivalence and noninferiority testing. J Gen Intern Med. 2011;26(2):192‐196. 10.1007/s11606-010-1513-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27. Althunian TA, de Boer A, Groenwold RHH, Klungel OH. Defining the noninferiority margin and analysing noninferiority: an overview. Br J Clin Pharmacol. 2017;83(8):1636‐1642. 10.1111/bcp.13280 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Gillison ML, Trotti AM, Harris J, et al. Radiotherapy plus cetuximab or cisplatin in human papillomavirus‐positive oropharyngeal cancer (NRG Oncology RTOG 1016): a randomised, multicentre, non‐inferiority trial. Lancet. 2019;393(10166):40‐50. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29. Li N, Carmona R, Sirak I, et al. Highly efficient training, refinement, and validation of a knowledge‐based planning quality‐control system for radiation therapy clinical trials. Int J Radiat Oncol Biol Phys. 2017;97(1):164‐172. [DOI] [PMC free article] [PubMed] [Google Scholar]