Abstract
Background and Purpose
Conventional magnetic resonance imaging (MRI) poses challenges in quantitative analysis because voxel intensity values lack physical meaning. While intensity standardization methods exist, their effects on head and neck MRI have not been investigated. We developed a workflow based on healthy tissue region of interest (ROI) analysis to determine intensity consistency within a patient cohort. Through this workflow, we systematically evaluated intensity standardization methods for MRI of head and neck cancer (HNC) patients.
Materials and Methods
Two HNC cohorts (30 patients total) were retrospectively analyzed. One cohort was imaged with heterogenous acquisition parameters (HET cohort), whereas the other was imaged with homogenous acquisition parameters (HOM cohort). The standard deviation of cohort-level normalized mean intensity (SD NMIc), a metric of intensity consistency, was calculated across ROIs to determine the effect of five intensity standardization methods on T2-weighted images. For each cohort, a Friedman test followed by a post-hoc Bonferroni-corrected Wilcoxon signed-rank test was conducted to compare SD NMIc among methods.
Results
Consistency (SD NMIc across ROIs) between unstandardized images was substantially more impaired in the HET cohort (0.29 ± 0.08) than in the HOM cohort (0.15 ± 0.03). Consequently, corrected p-values for intensity standardization methods with lower SD NMIc compared to unstandardized images were significant in the HET cohort (p < 0.05) but not significant in the HOM cohort (p > 0.05). In both cohorts, differences between methods were often minimal and nonsignificant.
Conclusions
Our findings stress the importance of intensity standardization, either through the utilization of uniform acquisition parameters or specific intensity standardization methods, and the need for testing intensity consistency before performing quantitative analysis of HNC MRI.
Keywords: MRI, Standardization, Harmonization, Normalization, Quantitative analysis, Head and neck cancer
1. Introduction
Magnetic resonance imaging (MRI) is routinely used in clinical practice and has revolutionized how physicians evaluate diseases [1]. Conventional “weighted” MRI, where various acquisition parameters are modulated to generate T1-weighted (T1-w) or T2-weighted (T2-w) images, has become commonplace in clinical workflows. Although conventional MRI acquisitions are useful for the qualitative assessment of disease, advanced quantitative evaluation (e.g., through radiomics [2] or deep learning [3]) is seemingly precluded by a fundamental problem: arbitrary voxel intensity. Unlike computed tomography, in which voxel intensities correspond to inherent tissue properties, the absolute voxel intensities of MRI correspond to both tissue properties and hardware-specific settings [4] and thus do not have a specific physical meaning. Consequently, MRI voxel intensity can vary from scanner to scanner and even within the same scanner [5]. A few important exceptions include images generated through various quantitative MRI acquisitions [6], such as diffusion-weighted MRI [7], dynamic contrast-enhanced MRI [8], or T1/T2 mappings [9], which are not routinely acquired in standard-of-care imaging. Unfortunately, intensity standardization (sometimes referred to as normalization or harmonization) is an often overlooked but crucial pre-processing step in studies attempting a quantitative analysis of conventional MRI acquisitions.
MRI is often performed for head and neck cancer (HNC) patients as part of radiotherapy treatment planning. Weighted images, particularly T2-w images, are commonly acquired in the scanning protocol because they provide excellent soft-tissue contrast in the complex anatomical areas involved in HNC. Thus, they are useful for region-of-interest (ROI) delineation [10], [11]. Notably, the increasing use of MRI-guided technology for adaptive HNC radiotherapy will likely increase the clinical integration of MRI quantitative analysis [12]. While several recent HNC studies have implemented cohort-level quantitative analysis of conventional weighted MRI [13], [14], [15], [16], [17], [18], [19], [20], [21], [22], relatively few have investigated incorporating intensity standardization into processing pipelines [16], [17], [18], [19], [20], [22], and even fewer have tested multiple standardization methods [20]. Furthermore, while rigorous studies have tested MRI intensity standardization methods for various anatomical regions, chiefly the brain [5], [23], such methods for the head and neck region have yet to be systematically investigated. The head and neck region may pose additional challenges for MRI intensity standardization when compared with relatively piecewise homogeneous regions like the brain. For instance, fields of view often vary across acquisitions, and the head and neck region is home to many tissue-tissue and tissue-air interfaces [24], which may result in a greater range and complexity of the underlying intensity distributions. Therefore, there is a pressing need to systematically investigate the effects of available MRI intensity standardization methods in HNC cohorts.
To address the growing importance of intensity standardization in the quantitative analysis of HNC MRI, we developed a novel, ROI-based workflow to compare existing standardization methods in T2-w images of HNC patients. We used two independent HNC cohorts—a multi-institutional cohort with heterogeneous acquisition parameters and a single-institutional cohort with homogeneous acquisition parameters—to systematically determine the effect of different intensity standardization methods for HNC MRI.
2. Materials and methods
2.1. Patient cohorts, image acquisitions, and ROIs
Two separate sets of patients who were diagnosed with oral or oropharyngeal cancer and for whom T2-w MRI images were available before the start of radiotherapy were included in this proof-of-concept study. A subset (cohort) of patients from each set was randomly selected for the analysis. The first cohort consisted of 15 patients with images acquired at different institutions and was termed “heterogeneous” (HET) because of the variety of scanners and acquisition parameters used in image generation. All 15 patients in the HET cohort were imaged with different scanning protocols; the MRI scanners originated from several manufacturers, including Siemens, GE, Phillips, and Hitachi. The second cohort consisted of 15 patients from a single prospective clinical trial with the same imaging protocol (NCT03145077, PA16-0302) and was termed “homogeneous” (HOM) because of the uniformity of both the scanner and the acquisition parameters used for image generation. Patients in the HOM cohort were imaged on a Siemens Aera scanner and immobilized with a thermoplastic mask. Image acquisition parameters for the cohorts are shown in Table 1; the demographic characteristics of each cohort are summarized in Table 2. All images were retrospectively collected in the Digital Imaging and Communications in Medicine (DICOM) format under a HIPAA-compliant protocol approved by our institution's IRB (RCR03-0800). The protocol included a waiver of informed consent. The anonymized image sets analyzed during the current study are publicly available online through Figshare (https://doi.org/10.6084/m9.figshare.13525481). For each image, ROIs of various healthy tissue types and anatomical locations were manually contoured in the same relative area for five slices by one observer (medical student) using Velocity AI v.3.0.1 (Atlanta, GA, USA), verified by a physician expert (radiologist), and exported as DICOM-RT Structure Set files. The ROIs were: 1. cerebrospinal fluid inferior (CSF_inf), 2. cerebrospinal fluid middle (CSF_mid), 3. cerebrospinal fluid superior (CSF_sup), 4. cheek fat left (Fat_L), 5. cheek fat right (Fat_R), 6. nape fat inferior (NapeFat_inf), 7. nape fat middle (NapeFat_mid), 8. nape fat superior (NapeFat_sup), 9. neck fat (NeckFat), 10. masseter left (Masseter_L), 11. masseter right (Masseter_R), 12. rectus capitus posterior major (RCPM), 13. skull, and 14. cerebellum. A visual representation of the ROIs is shown in Supplementary Figure S4. All DICOM and radiotherapy structure files were converted to Python data structures for processing and analysis with DICOMRTTool v.0.3.21 [25].
Table 1.
MRI Acquisition Parameter | HET Cohort (n = 15) | HOM Cohort (n = 15) |
---|---|---|
Magnetic Field Strength (T) | 1.50–3.00 | 1.50 |
Repetition Time (ms) | 3000.00–8735.00 | 4800.00 |
Echo Time (ms) | 70.80–123.60 | 80.00 |
Echo Train Length | 10–65 | 15 |
Flip Angle (°) | 90–150 | 180 |
In-plane Resolution (mm) | 0.35–1.01 | 0.50 |
Slice Thickness (mm) | 2.00–6.00 | 2.00 |
Spacing Between Slices (mm) | 1.00–7.00 | 2.00 |
Imaging Frequency (MHz) | 12.68–127.77 | 63.67 |
Number of Averages | 1.00–4.00 | 1.00 |
Percent Sampling (%) | 78.91–100 | 90.00 |
*Data shown for HET cohort are ranges. All HOM cohort patients had the same scanning parameters.
Table 2.
Characteristic | HET Cohort (n = 15) | HOM Cohort (n = 15) |
---|---|---|
Age (median, range) | 61 (41–78) | 61 (46–77) |
Patient Sex | ||
Men | 13 | 14 |
Women | 2 | 1 |
T Stage | ||
T1 | 5 | 7 |
T2 | 6 | 3 |
T3 | 1 | 1 |
T4 | 3 | 4 |
N Stage | ||
N0 | 4 | 0 |
N1 | 5 | 4 |
N2 | 6 | 11 |
Primary Tumor Site | ||
Base of Tongue | 5 | 9 |
Tonsil | 6 | 6 |
Oral Cavity | 4 | 0 |
*Unless otherwise indicated, data shown are number of patients.
2.2. Intensity standardization methods
We applied a variety of MRI intensity standardization methods to both cohorts’ images. These methods were chosen because of their relative ubiquity in other studies and simple implementations. Details of the implementation of these methods are presented below.
1. Unstandardized (Original): No intensity standardization was performed.
2. Rescaling (MinMax): This method standardized the image by rescaling the range of values to [0,1] using the equation
where x and f(x) were the original and standardized intensities, respectively, and min(x) and max(x) were the minimum and maximum image intensity values per patient, respectively.
3. Z-score standardization using all voxels (Z-All): This method standardized the image by centering it at a mean of 0 with a standard deviation of 1. The standardization was based on all voxels in the image and used the equation
where x and f(x) were the original and standardized voxel intensities, respectively, and μx and σx were the mean and standard deviation of the image intensity values per patient, respectively.
4. Z-score standardization using only voxels in an external mask (Z-External): Z-All was performed as described in item 3 above, but μx and σx were derived from voxels located in an external mask of the head and neck region (Supplementary Figure S5).
5. Cheek fat standardization (Fat): This method standardized the image with respect to left and right cheek fat (healthy tissue) and was adapted from van Dijk et al. [19]. The intensity of each voxel was divided by the mean intensity of the cheek fats and multiplied by an arbitrary scaling value of 350 using the equation
where x and f(x) were the original and standardized intensities, respectively, and μfat was the mean intensity of both cheek fat ROIs per patient.
6. Histogram standardization (Nyul): This method was adapted from Nyul and Udupa [26] using a code implementation from Reinhold et al. [27]. It used images for all patients in a cohort to construct a standard histogram template through the determination of histogram parameters and then linearly mapped the intensities of each image to the standard histogram template. The histogram parameters in this implementation were defined as intensity percentiles at 1, 10, 20, 30, 40, 50, 60, 70, 80, 90, and 99 percent. Only voxels within the head and neck external mask were used in the construction of the standard histogram template.
2.3. Intensity-Based ROI evaluation
According to the statistical principles of image normalization criteria [5], MRI intensities for a single type of tissue should maintain similar distributions within and across patients. Therefore, for a set of nonpathological ROIs representing a corresponding set of tissues within a cohort of patients, an increase in the quality of MRI intensity standardization should lead to an increase in the consistency of ROI intensity distributions (Supplementary Figure S6). Importantly, our aim is not to match distributions of the entire image since targets of quantitative analysis (e.g., tumors or healthy tissues altered by radiotherapy, such as the parotid glands) are expected to vary among patients. Motivated by this goal of population-level analysis that relies on the consistency of replicable units within tissue types and across patients, we used a simple and interpretable metric of comparison to quantify ROI intensity histogram overlap—termed the standard deviation of cohort-level normalized mean intensity (SD NMIc)—which can be applied before or after an intensity standardization procedure in a given cohort. The steps to calculate this metric are provided in Supplementary Data S1. Briefly, the metric is calculated by dividing the mean of a given ROI intensity distribution for a patient by the range of distributions for a cohort and then computing the standard deviation of the resulting values for the entire cohort. Given a set of ROIs that are not anticipated to vary from patient to patient, we would expect that for an ideal intensity standardization method, the cumulative SD NMIc would remain close to 0. For both cohorts, the SD NMIc was calculated for each intensity standardization method per ROI and visually compared on a heatmap. Of note, the cheek fat ROIs were not included in the evaluation since they were used for the Fat standardization method, and this could bias results.
2.4. Statistical analysis
After applying the Shapiro-Wilk test for normality [28], we found the SD NMIc to be non-normally distributed (p < 0.05). Therefore, nonparametric tests were deemed appropriate for statistical analysis. For each cohort, the Friedman test [29], a nonparametric analog to the one-way repeat measures analysis of variance test, was conducted to compare SD NMIc values among intensity standardization methods with standardization methods acting as within-subject factors. We note that the Friedman test included unstandardized (Original) images. If the Friedman test was statistically significant, a subsequent post-hoc two-sided Wilcoxon signed-rank test with a Bonferroni correction [30] was performed for all pair-wise combinations of standardization methods to determine which methods were significantly better than others. For both the Friedman and Wilcoxon signed-rank tests, p-values<0.05 were considered significant. Statistical analysis was performed in Python v.3.7.6. The code used to produce our analysis is available through GitHub (https://github.com/kwahid/MRI_Intensity_Standardization). The overall workflow of our approach is shown in Supplementary Figure S7.
3. Results
On visual inspection, unstandardized images from the HET cohort had noticeably different intensities when compared to standardized images (Fig. 1a). These qualitative observations were confirmed through quantitative estimates of intensity consistency where unstandardized images compared to standardized images often had higher SD NMIc values in the HET cohort (Fig. 1c). Oppositely, in the HOM cohort unstandardized images were similar to standardized images qualitatively (Fig. 1b) and quantitatively (Fig. 1d). When considering SD NMIc across all tissue sites (mean ± SD), the worst method was Original (0.29 ± 0.08) in the HET cohort and MinMax (0.18 ± 0.04) in the HOM cohort. Conversely, the best method was a tie between Z-External and Nyul for both cohorts, where SD NMIc was 0.17 ± 0.04 in the HET cohort and 0.13 ± 0.03 in the HOM cohort. Additional intermediary data in the form of ROI intensity distributions for the HET and HOM cohorts are found in Supplementary Figure S8 and Supplementary Figure S9, respectively, and demonstrated consistency with the SD NMIc values presented. Moreover, we performed additional analysis on T1-weighted images in a subset of cases from the HOM cohort in Supplementary Data S2, and demonstrated similar results to our observations on T2-weighted images.
The Friedman tests showed that SD NMIc values across all ROIs were significantly different for the intensity standardization methods in both cohorts (p < 0.001) (Fig. 2). Post-hoc analysis in the HET cohort (Fig. 2, above diagonal) revealed significantly higher SD NMIc values in Original compared with MinMax (p < 0.05), Z-All (p < 0.05), Z-External (p < 0.05), Fat (p < 0.01), and Nyul (p < 0.01). None of the intensity standardization methods in the HET cohort had significantly different SD NMIc values when compared with each other (p > 0.05). Post-hoc analysis in the HOM cohort (Fig. 2, below diagonal) revealed significantly lower SD NMIc values in Original compared with MinMax (p < 0.05). None of the other standardization methods in the HOM cohort (Z-All, Z-External, Fat, and Nyul) had significantly different SD NMIc values compared with Original (p > 0.05). Moreover, the HOM cohort demonstrated significantly higher SD NMIc values in MinMax compared with Z-External (p < 0.01), Fat (p < 0.01), and Nyul (p < 0.05). Finally, the HOM cohort demonstrated significantly higher SD NMIc values in Z-All when compared with Z-External (p < 0.01), and Nyul (p < 0.01).
4. Discussion
In this study, we proposed a workflow to test the consistency of standardized and unstandardized conventional MRI within a given HNC cohort. The scale-invariant, and thus comparable, SD NMIc metric was calculated to systematically investigate the effects of intensity standardization methods on T2-w images for two independent cohorts of HNC patients based on healthy tissue intensity consistency in multiple ROIs. Broadly, we determine that depending on the underlying imaging characteristics of a given cohort, explicit intensity standardization has varying effects on intensity consistency, which has the potential to impact downstream quantitative analysis.
Our results show that intensity standardization, when compared to no standardization, substantially improved T2-w MRI ROI intensity consistency in the HET cohort (Fig. 1a,b), but had a minimal impact in the HOM cohort (Fig. 1c,d). Since scanner and acquisition parameters were vastly different in the HET cohort, the marked intensity variation between Original images (Fig. 1a) was in line with our expectations. In contrast, while the HOM cohort included patients from clinical trial data with the same scanner and acquisition parameters, the relatively minor intensity variation between Original images (Fig. 1c) was better than expected. In a sense, the use of identical acquisition parameters in the HOM cohort seemed to act as an inherent method of pre-processing intensity standardization. This may also indicate that flexible head and neck MRI coil positioning was performed systematically and reliably in this cohort, as positioning the coil at varying distances from the patient can result in different image intensities. While the use of uniform acquisition parameters on the same MRI scanner may circumvent the need for the application of intensity standardization methods, further work is likely needed to verify these results.
Upon visual inspection, it was difficult to discriminate between the various intensity standardization methods for either cohort (Fig. 1a,c). Quantitatively, most intensity standardization methods had similar performance in both cohorts, regardless of overall consistency improvement compared to unstandardized images (Fig. 1b,d and Fig. 2). Recent work by Carré et al [31]. demonstrated results similar to those of our study in that various intensity standardization methods improved the consistency between brain images with heterogeneous acquisition parameters, although the authors did not identify a specific superior standardization method. In the current study, paired significance testing (Fig. 2) revealed that the standardization methods Z-External and Nyul performed relatively well in both the HET cohort (significantly better than Original) and the HOM cohort (significantly better than MinMax and Z-All), with both methods achieving the lowest SD NMIc values. Interestingly, the only methods that performed significantly worse than any others in the HOM cohort were MinMax and Z-All, possibly secondary to the large number of background elements influencing the standardization parameters in the HOM cohort.
A potential limitation of this proof-of-concept study was that it included a small number of patients for each cohort. However, we implemented conservative significance testing to ensure the robustness of our results. Moreover, this study focused on analyzing T2-w sequences for initial testing since they are favored in head and neck imaging due to their exquisite anatomical detail for various ROIs [32], [33]. Since healthy tissue ROIs are regularly contoured during radiotherapy treatment planning, our analysis tools and workflow can facilitate future large-scale HNC investigations with additional MRI sequences. To guide further research in different MRI sequences, we present preliminary analyses on T1-weighted images from the HOM cohort in Supplementary Data S2, which are broadly consistent with the findings in this study. Another limitation of our study is that our workflow may not apply to patients who have already received radiotherapy, as radiation can cause structural and functional changes in healthy tissue that may impact the intensity profiles of various ROIs [34]. This may be mitigated by selecting ROIs that are known not to change significantly with treatment. Finally, acquisition artifacts, such as magnetic field inhomogeneities, can affect MRI intensity [35]. While images from the HOM cohort were free of artifacts, some images from the HET cohort contained bias fields (Supplementary Figure S10). Therefore, bias-field correction techniques may need to be explored in combination with intensity standardization methods in future studies to determine their effects on ROI intensity consistency in HNC MRI.
To our knowledge, this is the first study to systematically investigate the effects of intensity standardization in head and neck imaging. In addition, whereas many studies of MRI intensity standardization for other anatomical sites implemented test–retest data for individual patients to determine the effects of standardization [31], [36], [37], [38], [39], our analysis is unique because it investigated the impact of standardization within a cohort of patients. Our approach may be more relevant to the downstream cohort-level model construction often implemented in quantitative analysis studies. Finally, we have publicly disseminated our data and analysis tools through open access platforms to foster increased reproducibility and help the imaging research community extrapolate our workflow to MRI of other anatomical regions.
In summary, we propose a workflow to robustly test MRI intensity consistency within a given HNC patient cohort and demonstrate the need for evaluating MRI intensity consistency before performing quantitative analysis. This study verifies that intensity standardization, either through the utilization of uniform acquisition parameters or specific intensity standardization methods, is crucial to improving the consistency of inherent tissue intensity values in conventional weighted HNC MRI. Our study is an essential first step towards widespread intensity standardization for quantitative analysis of conventional MRI in the head and neck region.
Funding Statement
This work was supported by the National Institutes of Health (NIH) through a Cancer Center Support Grant (P30-CA016672-44). K.A. Wahid and T. Salzillo are supported by training fellowships from The University of Texas Health Science Center at Houston Center for Clinical and Translational Sciences TL1 Program (TL1TR003169). K.A. Wahid is also supported by the American Legion Auxiliary Fellowship in Cancer Research and an NIDCR F31 fellowship (1 F31 DE031502-01). R. He., A.S.R. Mohamed, K. Hutcheson, and S.Y. Lai are supported by a NIH National Institute of Dental and Craniofacial Research (NIDCR) Award (R01DE025248). B.A. McDonald receives research support from an NIH NIDCR Award (F31DE029093) and the Dr. John J. Kopchick Fellowship through The University of Texas MD Anderson UTHealth Graduate School of Biomedical Sciences. B.M. Anderson receives funding from a Society of Interventional Radiology Foundation Allied Scientist Grant and the Dr. John J Kopchick Fellowship. L.A. McCoy and K.L. Sanders are supported by NIH NIDCR Research Supplements to Promote Diversity in Health-Related Research (R01DE025248-S02 and R01DE028290-S01 respectively). M.A. Naser and S. Ahmed are supported by a NIH NIDCR Award (R01 DE028290-01). C.D. Fuller received funding from an NIH NIDCR Award (1R01DE025248-01/R56DE025248) and Academic-Industrial Partnership Award (R01 DE028290), the National Science Foundation (NSF), Division of Mathematical Sciences, Joint NIH/NSF Initiative on Quantitative Approaches to Biomedical Big Data (QuBBD) Grant (NSF 1557679), the NIH Big Data to Knowledge (BD2K) Program of the National Cancer Institute (NCI) Early Stage Development of Technologies in Biomedical Computing, Informatics, and Big Data Science Award (1R01CA214825), the NCI Early Phase Clinical Trials in Imaging and Image-Guided Interventions Program (1R01CA218148), the NIH/NCI Cancer Center Support Grant (CCSG) Pilot Research Program Award from the UT MD Anderson CCSG Radiation Oncology and Cancer Imaging Program (P30CA016672), the NIH/NCI Head and Neck Specialized Programs of Research Excellence (SPORE) Developmental Research Program Award (P50 CA097007) and the National Institute of Biomedical Imaging and Bioengineering (NIBIB) Research Education Program (R25EB025787). C.D. Fuller has also received direct industry grant support, speaking honoraria and travel funding from Elekta AB. L.V. van Dijk received/receives funding and salary support from the Dutch organization NWO ZonMw during the period of study execution via the Rubicon Individual career development grant.
Declaration of Competing Interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Acknowledgments
The authors thank Laura L. Russell, scientific editor at The University of Texas MD Anderson Cancer Center Research Medical Library, for editing this article.
Footnotes
Supplementary data to this article can be found online at https://doi.org/10.1016/j.phro.2021.11.001.
Supplementary Material
The following are the Supplementary data to this article:
References
- 1.Grover V.P.B., Tognarelli J.M., Crossey M.M.E., Cox I.J., Taylor-Robinson S.D., McPhail M.J.W. Magnetic resonance imaging: principles and techniques: lessons for clinicians. J Clin Exp Hepatol. 2015;5(3):246–255. doi: 10.1016/j.jceh.2015.08.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.van Timmeren J.E., Cester D., Tanadini-Lang S., Alkadhi H., Baessler B. Radiomics in medical imaging—“how-to” guide and critical reflection. Insights Imaging. 2020;11:1–16. doi: 10.1186/s13244-020-00887-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Lundervold A.S., Lundervold A. An overview of deep learning in medical imaging focusing on MRI. Z Med Phys. 2019;29(2):102–127. doi: 10.1016/j.zemedi.2018.11.002. [DOI] [PubMed] [Google Scholar]
- 4.Bloem J.L., Reijnierse M., Huizinga T.W.J., van der Helm-van Mil A.H.M. MR signal intensity: staying on the bright side in MR image interpretation. RMD Open. 2018;4(1):e000728. doi: 10.1136/rmdopen-2018-000728. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Shinohara R.T., Sweeney E.M., Goldsmith J., Shiee N., Mateen F.J., Calabresi P.A., et al. Statistical normalization techniques for magnetic resonance imaging. NeuroImage Clin. 2014;6:9–19. doi: 10.1016/j.nicl.2014.08.008. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 6.Pierpaoli C. Quantitative brain MRI. Top Magn Reson Imaging. 2010;21:63. doi: 10.1097/RMR.0b013e31821e56f8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Subhawong T.K., Jacobs M.A., Fayad L.M. Insights into quantitative diffusion-weighted MRI for musculoskeletal tumor imaging. Am J Roentgenol. 2014;203(3):560–572. doi: 10.2214/AJR.13.12165. [DOI] [PubMed] [Google Scholar]
- 8.Petralia G., Summers P.E., Agostini A., Ambrosini R., Cianci R., Cristel G., et al. Dynamic contrast-enhanced MRI in oncology: how we do it. Radiol Med. 2020;125(12):1288–1300. doi: 10.1007/s11547-020-01220-z. [DOI] [PubMed] [Google Scholar]
- 9.Kim P.K., Hong Y.J., Im D.J., Suh Y.J., Park C.H., Kim J.Y., et al. Myocardial T1 and T2 mapping: techniques and clinical applications. Korean J Radiol. 2017;18(1):113. doi: 10.3348/kjr.2017.18.1.113. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Popovtzer A., Ibrahim M., Tatro D., Feng F.Y., Ten Haken R.K., Eisbruch A. MRI to delineate the gross tumor volume of nasopharyngeal cancers: which sequences and planes should be used? Radiol Oncol. 2014;48:323–330. doi: 10.2478/raon-2014-0013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Zima A.J., Wesolowski J.R., Ibrahim M., Lassig A.A.D., Lassig J., Mukherji S.K. Magnetic resonance imaging of oropharyngeal cancer. Top Magn Reson Imaging. 2007;18:237–242. doi: 10.1097/RMR.0b013e318157112a. [DOI] [PubMed] [Google Scholar]
- 12.Boeke S., Mönnich D., Van Timmeren J.E., Balermpas P. MR-Guided Radiotherapy for Head and Neck Cancer: Current Developments, Perspectives, and Challenges. Front Oncol. 2021;11:429. doi: 10.3389/fonc.2021.616156. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Ren J., Tian J., Yuan Y., Dong D.i., Li X., Shi Y., et al. Magnetic resonance imaging based radiomics signature for the preoperative discrimination of stage I-II and III-IV head and neck squamous cell carcinoma. Eur J Radiol. 2018;106:1–6. doi: 10.1016/j.ejrad.2018.07.002. [DOI] [PubMed] [Google Scholar]
- 14.Yuan Y., Ren J., Shi Y., Tao X. MRI-based radiomic signature as predictive marker for patients with head and neck squamous cell carcinoma. Eur J Radiol. 2019;117:193–198. doi: 10.1016/j.ejrad.2019.06.019. [DOI] [PubMed] [Google Scholar]
- 15.Ming X., Oei R.W., Zhai R., Kong F., Du C., Hu C., et al. MRI-based radiomics signature is a quantitative prognostic biomarker for nasopharyngeal carcinoma. Sci Rep. 2019;9(1) doi: 10.1038/s41598-019-46985-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Bos P., Brekel M.W.M., Gouw Z.A.R., Al‐Mamgani A., Waktola S., Aerts H.J.W.L., et al. Clinical variables and magnetic resonance imaging-based radiomics predict human papillomavirus status of oropharyngeal cancer. Head Neck. 2021;43(2):485–495. doi: 10.1002/hed.26505. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Hsu C.-Y., Lin S.-M., Ming Tsang N., Juan Y.-H., Wang C.-W., Wang W.-C., et al. Magnetic resonance imaging-derived radiomic signature predicts locoregional failure after organ preservation therapy in patients with hypopharyngeal squamous cell carcinoma. Clin Transl Radiat Oncol. 2020;25:1–9. doi: 10.1016/j.ctro.2020.08.004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Meheissen M.A.M., Mohamed A.S.R., Kamal M., Hernandez M., Volpe S., Elhalawani H., et al. A prospective longitudinal assessment of MRI signal intensity kinetics of non-target muscles in patients with advanced stage oropharyngeal cancer in relationship to radiotherapy dose and post-treatment radiation-associated dysphagia: Preliminary findings. Radiother Oncol. 2019;130:46–55. doi: 10.1016/j.radonc.2018.08.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.van Dijk L.V., Thor M., Steenbakkers R.J.H.M., Apte A., Zhai T.-T., Borra R., et al. Parotid gland fat related Magnetic Resonance image biomarkers improve prediction of late radiation-induced xerostomia. Radiother Oncol. 2018;128(3):459–466. doi: 10.1016/j.radonc.2018.06.012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Mes S.W., van Velden F.H.P., Peltenburg B., Peeters C.F.W., te Beest D.E., van de Wiel M.A., et al. Outcome prediction of head and neck squamous cell carcinoma by MRI radiomic signatures. Eur Radiol. 2020;30(11):6311–6321. doi: 10.1007/s00330-020-06962-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Rodríguez Outeiral R., Bos P., Al-Mamgani A., Jasperse B., Simões R., van der Heide U.A. Oropharyngeal primary tumor segmentation for radiotherapy planning on magnetic resonance imaging using deep learning. Phys Imaging Radiat Oncol. 2021;19:39–44. doi: 10.1016/j.phro.2021.06.005. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Thor M., Tyagi N., Hatzoglou V., Apte A., Saleh Z., Riaz N., et al. A magnetic resonance imaging-based approach to quantify radiation-induced normal tissue injuries applied to trismus in head and neck cancer. Phys Imaging Radiat Oncol. 2017;1:34–40. doi: 10.1016/j.phro.2017.02.006. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Reinhold J.C., Dewey B.E., Carass A., Prince J.L. International Society for Optics and Photonics; 2019. Evaluating the impact of intensity normalization on MR image synthesis. Med. Imaging 2019 Image Process; p. 109493H. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Ma J., Jackson E.F., Kumar A.J., Ginsberg L.E. Improving fat-suppressed T2-weighted imaging of the head and neck with 2 fast spin-echo dixon techniques: initial experiences. Am J Neuroradiol. 2009;30(1):42–45. doi: 10.3174/ajnr.A1132. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Anderson B.M., Wahid K.A., Brock K.K. Simple Python Module for Conversions between DICOM Images and Radiation Therapy Structures, Masks, and Prediction Arrays. Pract. Radiat Oncol. 2021;11(3):226–229. doi: 10.1016/j.prro.2021.02.003. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Nyúl L.G., Udupa J.K. On standardizing the MR image intensity scale. Magn Reson Med An Off J Int Soc Magn Reson Med. 1999;42:1072–1081. doi: 10.1002/(sici)1522-2594(199912)42:6<1072::aid-mrm11>3.0.co;2-m. [DOI] [PubMed] [Google Scholar]
- 27.Reinhold JC. Intensity-normalization 2020. https://github.com/jcreinhold/intensity-normalization (accessed July 1, 2020).
- 28.SHAPIRO S.S., WILK M.B. An analysis of variance test for normality (complete samples) Biometrika. 1965;52(3-4):591–611. [Google Scholar]
- 29.Friedman M. A comparison of alternative tests of significance for the problem of m rankings. Ann Math Stat. 1940;11(1):86–92. [Google Scholar]
- 30.Wilcoxon F. Springer; Break. Stat.: 1992. Individual comparisons by ranking methods; pp. 196–202. [Google Scholar]
- 31.Carré A., Klausner G., Edjlali M., Lerousseau M., Briend-Diop J., Sun R., et al. Standardization of brain MR images across machines and protocols: bridging the gap for MRI-based radiomics. Sci Rep. 2020;10(1) doi: 10.1038/s41598-020-69298-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Radiology AC of. ACR-ASNR-SPR Practice parameter for the performance of magnetic resonance imaging (MRI) of the head and neck 2017.
- 33.Lewin J.S., Curtin H.D., Ross J.S., Weissman J.L., Obuchowski N.A., Tkach J.A. Fast spin-echo imaging of the neck: comparison with conventional spin-echo, utility of fat suppression, and evaluation of tissue contrast characteristics. Am J Neuroradiol. 1994;15:1351–1357. [PMC free article] [PubMed] [Google Scholar]
- 34.Barnett G.C., West C.M.L., Dunning A.M., Elliott R.M., Coles C.E., Pharoah P.D.P., et al. Normal tissue reactions to radiotherapy: towards tailoring treatment dose by genotype. Nat Rev Cancer. 2009;9(2):134–142. doi: 10.1038/nrc2587. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Juntu J., Sijbers J., Van Dyck D., Gielen J. Comput. Recognit. Syst. Springer; 2005. Bias field correction for MRI images; pp. 543–551. [Google Scholar]
- 36.Schwier M., van Griethuysen J., Vangel M.G., Pieper S., Peled S., Tempany C., et al. Repeatability of multiparametric prostate MRI radiomics features. Sci Rep. 2019;9(1) doi: 10.1038/s41598-019-45766-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Hoebel K.V., Patel J.B., Beers A.L., Chang K., Singh P., Brown J.M., et al. Radiomics Repeatability Pitfalls in a Scan-Rescan MRI Study of Glioblastoma. Radiol. Artif Intell. 2021;3(1):e190199. doi: 10.1148/ryai.2020190199. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Scalco E., Belfatto A., Mastropietro A., Rancati T., Avuzzi B., Messina A., et al. T2w-MRI signal normalization affects radiomics features reproducibility. Med Phys. 2020;47(4):1680–1691. doi: 10.1002/mp.14038. [DOI] [PubMed] [Google Scholar]
- 39.Li Z., Duan H., Zhao K., Ding Y. Stability of MRI Radiomics Features of Hippocampus: An Integrated Analysis of Test-Retest and Inter-Observer Variability. IEEE Access. 2019;7:97106–97116. [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.