Abstract
Simple Summary
Even after the introduction of a standard regimen consisting of concurrent chemoradiotherapy and adjuvant temozolomide, patients with glioblastoma multiforme mostly experience disease progression. Clinicians often encounter a situation where they need to distinguish progressive disease from pseudoprogression after treatment. We tried to investigate the feasibility of machine learning algorithm to distinguish pseudoprogression from progressive disease. In multi-institutional dataset, the developed machine learning model showed an acceptable performance. This algorithm involving MRI data and clinical features could help making decision during patients’ disease course. For the practical use, we calibrated the machine learning model to offer the probability of pseudoprogression to clinicians, then we constructed the web-based user interface to access the model.
Abstract
Some patients with glioblastoma show a worsening presentation in imaging after concurrent chemoradiation, even when they receive gross total resection. Previously, we showed the feasibility of a machine learning model to predict pseudoprogression (PsPD) versus progressive disease (PD) in glioblastoma patients. The previous model was based on the dataset from two institutions (termed as the Seoul National University Hospital (SNUH) dataset, N = 78). To test this model in a larger dataset, we collected cases from multiple institutions that raised the problem of PsPD vs. PD diagnosis in clinics (Korean Radiation Oncology Group (KROG) dataset, N = 104). The dataset was composed of brain MR images and clinical information. We tested the previous model in the KROG dataset; however, that model showed limited performance. After hyperparameter optimization, we developed a deep learning model based on the whole dataset (N = 182). The 10-fold cross validation revealed that the micro-average area under the precision-recall curve (AUPRC) was 0.86. The calibration model was constructed to estimate the interpretable probability directly from the model output. After calibration, the final model offers clinical probability in a web-user interface.
Keywords: machine learning, glioblastoma, radiotherapy, pseudoprogression
1. Introduction
Pseudoprogression (PsPD) is a brain lesion shown in brain magnetic resonance (MR) images, mimicking progressive disease (PD) after concurrent chemoradiation (CCRT) in patients with glioblastoma [1]. The incidence of PsPD was estimated to be 36% in a meta-analysis [2]. The classification of PsPD and PD is important to avoid the inappropriate discontinuation of the standard treatment. However, the final diagnosis is based on the following changes in a series of MR imaging. Additionally, PsPD and PD have a similar contrast-enhanced presentation in T1-weighted MR images [3,4,5]. Thus, clinicians make decisions depending on the patient symptom [6], molecular profile [7], dynamic imaging [8,9,10], or functional imaging [11]. However, special imaging such as 18F-fluoro-ethyl-tyrosine (18F-FET)-positron emission tomography (PET) facilities or amino acid PET is not yet approved for glioma by the Food and Drug Administration [12].
In our previous study [13], we showed the feasibility of the deep learning algorithm that is combined with the convolutional neural network (CNN) and long short-term memory (LSTM) structures. The deep learning model was implemented to classify PsPD and PD in patients with glioblastoma who completed CCRT based on the standard regimen [14]. At the time of the appearance of the suspicious lesion in brain imaging, gadolinium-enhanced T1-weighted MR images and clinical factors were collected. Specifically, nine selected MR images, age, gender, molecular features, radiotherapy (RT) information, and the interval from the day of the completion of CCRT were selected as model inputs. The machine learning model was trained and tested in a dataset collected from two institutions, termed the Seoul National University Hospital (SNUH) dataset, N = 78). The model performance was acceptable, showing an area under the precision-recall curve (AUPRC) of 0.87.
However, the SNUH dataset was small, and there was a need to validate this model in a more extended dataset. This study aimed to test whether the previous model could be used in a new extended dataset collected from multiple institutions, termed the Korean Radiation Oncology Group (KROG) dataset (N = 104).
2. Methods
The study was approved by the ethics committee and institutional review board (IRB) of Seoul National University Bundang Hospital (IRB No. B-1710-426-105). The ethics committee/institutional review board that approved this study also waived the need for informed consent. All the methods were performed in accordance with the relevant guidelines and regulations.
2.1. Study Population and Definition of PsPD and PD
The inclusion criteria were as follows: patients with primary glioblastoma should receive a complete gross total resection of tumor and CCRT. At least 4 weeks after the completion of standard treatment, patients who demonstrated a single measurable contrast-enhancing lesion of any size on gadolinium-enhanced T1-weighted MRI within 80% isodose line based on the Response Assessment in Neuro-Oncology criteria [15] were included. The exclusion criteria were as follows: patients who had an enhancing lesion before CCRT, who had residual lesion at the immediate post-operative MRI, and who underwent incomplete CCRT.
Because most institutions did not perform operation or biopsy for the contrast-enhanced (CE) lesion, the classification of PD vs. PsPD was based on expert opinions. To minimize the variability in the definition of ground truth, we provided the guides of the PsPD vs. PD definition to multiple institutions. We defined the CE lesion as PD if ≥ 1 of the following conditions were met: it was a surgically confirmed recurring lesion, it had a significant uptake at the lesion of PET, or it had an increased size on follow-up MR imaging. Meanwhile, we considered the CE lesion as PsPD if ≥ one of the following were met: it was a pathologically confirmed PsPD, the lesion decreased on follow-up MR imaging, the lesion was stable for more than 120 days, or there was no significant uptake at the lesion of PET.
2.2. Model Structure and New Dataset Collection
The previous model was based on the SNUH dataset (N = 78). The SNUH dataset is composed of a patient’s nine axial MR images that cover the brain lesion to be classified. The slice thickness of the MR images was around 1 mm. The structure of the machine learning model is illustrated in Figure 1A. Briefly, the model is composed of two parts: CNN-LSTM for MR imaging and fully connected layers for clinical information. Firstly, the input of the CNN model is a total of nine axial gadolinium-enhanced T1-weighted brain MR images, and the output features are flattened. Flattened features are used as sequential input for LSTM layers. Secondly, clinical information is used as the input of fully connected layer, and their output was also flattened. Clinical information includes age, gender, total radiation dose, the number of fractions, interval between CCRT and the appearance of lesion, the O6-methylguanine-DNA methyltransferase (MGMT) methylation status, and the isocitrate dehydrogenase (IDH) mutation status. Finally, the output of the LSTM layer and the output of the fully connected layers are concatenated for the final decision. Detail activation function, kernel size, and hidden units were described in our previous study [13].
According to the international brain tumor imaging protocol [16], recent MR imaging for glioblastoma requires a fine slice thickness of less than 1.5 mm, and a measurable lesion should be at least 10 mm in size from axial imaging. If we selected one axial MR image containing a suspected contrast-enhanced lesion and selected three images in upward and three images in downward (total 9 images), we could capture a measurable lesion (more than 10 mm) for the model input. This measurable size was the final range for the lesions to be suspected for PD/PsPD. Furthermore, the limited numbers of images were unlikely to include irrelevant backgrounds such as the neck. We excluded other sequences of MR images, since huge resources to train/test the model were required. Further, there was a possibility for an institution not to provide the full sequence of MR imaging. Since the endpoint of this study was to develop the model from the data from multiple institutions, we minimized the requirements for the model prediction.
To test the previous model in the new extended dataset, we collected cases from multiple institutions (termed as the KROG dataset). The model network accepts MR images with a fine slice thickness of around 1 mm (Figure 1B) because nine images with a thick slice thickness (e.g., 5 mm) often cover the whole brain, neck, or other irrelevant backgrounds. However, not all the institutions have same MR protocols in terms of slice thickness. Thus, we allowed the triplicate use of three core images into nine when the image slice thickness was >1 mm (Figure 1C). This may be benefit hospitals without luxurious MR equipment to produce fine images.
2.3. Hyperparameter Optimization and Finalizing the Model
We tested the previous model [13] in the KROG dataset, varying the model parameters and the proportion of the dataset. We hypothesized that the both scaler range and epoch number were critical for the model, given that the SNUH dataset was outnumbered by the KROG dataset. Each parameter reflected the changes in the scaler range of clinical data and the number of epochs in terms of the previously published model. The scaler range was determined when the value of clinical information was normalized. Specifically, the scaler was ranged from −1.5 to 1.5 or from 0 to 1. Thus, the model parameters for re-training were as follows: Parameter “A” was defined as the scaler range with [0,1] and epoch number as 20, parameter “B” as [0,1] and 25, parameter “C” as [−1.5,1.5] and 20, and parameter “D” as [−1.5,1.5] and 25. Additionally, we calculated the AUPRCs as increasing the proportion of the KROG dataset from 10% to 100%. By doing this, we explored the optimal hyperparameters.
Finally, we developed the machine learning algorithm using the whole dataset (KROG dataset and SNUH dataset) with optimal hyperparameters. We performed a 10-fold cross validation and plotted the precision–recall curves.
2.4. Selection of Calibration Model Selection and Implementation of User Interface
We explored the optimal calibration method by comparing three candidate calibration models: Binning strategy [17], Bayesian Binning in Quantiles (BBQ) [18], and “GUESS” model [19]. Each calibration model was trained and evaluated using 100-times repeated stratified 10-fold cross-validation. The predictions were divided into 10 bines with equal widths. Each expected calibration error (ECE) [20], root mean square error (RMSE), sensitivity, and specificity were estimated to evaluate the efficacy of the calibration models. ECE was computed as follows:
Each describes the probability of all instances that fall into bin , represents the factions of true PsPD in bin , and represents the mean over all predictions in bin .
After calibration, the reliability diagram was depicted to establish the correlation between the model outputs and the observed probabilities.
2.5. Statistical Analysis
Clinicopathological characteristics were compared between patients with PsPD and PD using Student’s t-test and Chi-square test, or Fisher’s exact test. These statistical analyses were performed using the “Stata” version 15. The distribution of the slice thickness and validation results was represented through a violin plot, bar-plot, and heatmap using “Prism” version 8.1.2. The AUPRCs, ECE, RMSE, sensitivity, and specificity were compared among the calibration models using a repeated measure one-way ANOVA. Multiple comparison tests were also performed, and the adjusted P-value was estimated by applying the correction using Tukey hypothesis testing.
3. Results
3.1. Characteristics of KROG Dataset
A total of 104 cases were collected from multiple institutions in Korea (KROG dataset). Patient characteristics are summarized in Table 1. Thirty-eight brain lesions (36.5%) turned out to be PsPD with long-term follow-up, and 66 ones (63.5%) demonstrated PD. There were no significant differences in age, gender, MGMT promoter methylation status, IDH mutational status, and the dose/fraction schedule of RT. The median interval (days) from the completion of CCRT to the appearance of brain lesion were 28 (range 19–700 days) for the PsPD group, and 95 (range 8–744 days) for the PD group (t-test, p = 0.21). For brain MR images, we allowed a policy that enables the triplicate use of three core images into nine when the slice thickness of the acquired images was > 1 mm. The median slice thickness of 46 cases (44.3%) was 7 mm (range, 1.4–8), and three MR images from each case were selected, triplicated, and used as inputs to the model. Meanwhile, 58 cases (55.7%) had a median slice thickness of 1.2 mm (0.47–1.2 mm), and were not indicated for triplicate use (Figure 2A).
Table 1.
KROG Dataset | PsPD (N = 38) | PD (N = 66) | Total (N = 104) | |||
---|---|---|---|---|---|---|
N | % | N | % | p | ||
Age (median, range) | 56.5 (23–75) | 55 (25–76) | 0.79 * | 55 (25–75) | ||
Gender | 0.02 † | |||||
Female | 22 | 57.9 | 23 | 34.8 | 45 | |
Male | 16 | 42.1 | 43 | 65.2 | 59 | |
MGMT promoter status | 0.40 † | |||||
Methylated | 13 | 34.2 | 18 | 27.3 | 31 | |
Unmethylated | 9 | 23.7 | 24 | 36.4 | 33 | |
Unknown | 16 | 42.1 | 24 | 36.3 | 40 | |
IDH mutational status | 0.22 ‡ | |||||
Mutated | 2 | 5.2 | 0 | 0.0 | 2 | |
Wild-type | 15 | 39.5 | 30 | 45.4 | 45 | |
Unknown | 21 | 55.3 | 36 | 54.6 | 57 | |
Dose schedule of RT | 0.65 † | |||||
Hypofractionated | 2 | 5.3 | 5 | 7.6 | 7 | |
Conventional | 36 | 94.7 | 61 | 92.4 | 97 | |
Interval (days), median (range) | 28 (19–700) | 95 (8–744) | 0.21 * | 52 (8–744) |
Abbreviations: PD, progression; PsPD, pseudoprogression; MGMT, O6-methylguanine-DNA-methyltransferase; IDH, isocitrate dehydrogenase; RT, radiation therapy. * Student’s T-test, † Chi-square test, ‡ Fisher’s exact test.
3.2. Testing Results of Previous Model in KROG Dataset
The number of the KROG dataset is 1.3 times more than that of the SNUH dataset. Furthermore, cases with triplicated MR images were unseen data for previously published model. We hypothesized that epoch number and scaler range were critical for previously published model to predict unseen data. Thus, we tested the previous model with the KROG dataset, varying the epoch number and scaler range that are required for processing clinical data. Additionally, we tested each parameter by varying the proportion of the KROG set from 10% to 100%, and estimated the AUPRC in each dataset. As the proportion of the KROG set increases, the AUPRCs decrease regardless of the parameters. The pattern of AUPRCs according to the proportion of the KROG set is demonstrated as the heatmap in Figure 2B. By doing this, we could find the optimal hyperparameter: the parameter ‘D’. Then, we trained the model network in the whole dataset (the KROG dataset and the SNUH dataset, N = 182). The 10-fold cross validation was performed, and the value of the micro-average AUPRC was 0.86 (Figure 3). Thus, we finalized the model.
3.3. Establishment of Final Model with Calibration
Using the finalized model in this study, we sought to find the optimal calibration model by comparing the efficacy of BBQ, “GUESS”, and histogram binning. There was a significant difference in the ECEs of the calibration models (p < 0.0001, Figure 4A). The mean ECEs of BBQ, “GUESS”, and histogram binning were 0.049, 0.038, and 0.012, respectively. In terms of RMSE, “GUESS” showed a lower error than other models (p < 0.0001, Figure 4B). The mean RMSEs of BBQ, “GUESS”, and histogram binning were 0.303, 0.279, and 0.482, respectively. Sensitivity did not differ among the calibration models (p = 0.094); however, multiple comparisons showed that “GUESS” performed better than BBQ (0.87 vs. 0.84, adjusted p < 0.0001, Figure 4C). A significant difference in specificity was observed, and “GUESS” was the more superior than other models (p < 0.001, Figure 4D). The mean specificities of BBQ, “GUESS”, and histogram binning were 0.933, 0.945, and 0.203, respectively.
Consequently, we selected the “GUESS” as the optimal calibration model. Using the “GUESS” model, we plotted the reliability diagram before (Figure 4E) and after (Figure 4F) calibration. After calibration, we found that the observed frequency of PsPD was not significantly different from the mean prediction.
3.4. Examples of Correct and Incorrect Cases
We tested the correct and incorrect prediction by the developed model. In Figure 5, our machine learning model predicted a low probability of PD (30.95%) in patients with a contrast-enhanced lesion that appeared on the 27th day after CCRT. On the 319th day after CCRT, she had no symptoms with ongoing treatment change. As shown in Figure 6, a female 45-year old demonstrated a contrast-enhanced lesion in the right frontal lobe on the 224th day after CCRT. The machine learning model predicted this lesion as the PD with a probability of 99.98%. This patient underwent surgical resection due to bleeding in the surgical cavity. The pathologic review of the surgical specimen was recurring glioblastoma.
In contrast, machine learning could give a wrong answer. As shown in Figure 7, a male patient who was 33 years old received CCRT with doses of 63 Gy in 35 fractions to the left frontal lobe. After 23 days, a contrast-enhanced lesion was shown from the left genu of internal capsule. Our machine learning model predicted this lesion as PD with a 97.4% probability. However, the ground truth was PsPD. Note that the MGMT and IDH statuses were unknown and the triplicate use of MR images was applied due to the thick slice thickness (>1-mm).
4. Discussion
In the previous study, the deep learning model showed the feasibility in two independent small datasets. We hypothesized this model could be used in the extended dataset derived from multi-institutions. In the current study, direct validation resulted in limited performance. However, the model showed acceptable performance in terms of cross-validation. The finalized model was well calibrated, then implemented in web-user interface. Thus, a clinician can access and estimate their patient’s progression probability (http://radiation-oncology-lab.ml:5000).
There were several machine learning algorithms to classify PsPD and PD. Booths et al. [21] developed a support vector machine using the features derived from T2-weighted imaging, and the model showed the accuracy of 0.86 in the prospective dataset (N = 7). Kebir et al. [22] analyzed the features derived from 18F-FET-PET images from 14 patients. This unsupervised learning revealed a specific cluster associated with PsPD, showing a 90% recall and precision. In a multiparametric brain MR study [23], a support vector machine with an eight-dimensional feature vector was developed. Although there was no testing set, the AUC was 0.94. A radiomic approach is adopted in these studies.
Meanwhile, our model was based on a deep learning approach with a single MR sequence image. Most deep learning studies in glioma focus on tumor segmentation [24,25,26], the prediction of survival [27], or the prediction of molecular profiles [28,29,30]. With respect to resolving the PsPD and PD problem, our study is the first to adopt the deep learning model in glioblastoma patients to date [31,32]. Specifically, in this study, the CNN deep learning algorithm was the key structure to capture features from gadolinium-enhanced T1-weighted MR imaging. Enhancing lesion T1-weighted MR is routinely evaluated by bidimensional measurement according to the Macdonald and RANO criteria [33]. Additionally, using single gadolinium-enhanced T1-weighted MR imaging modality will expand the practical use in clinical setting. In our model, merging clinical data with single imaging modality has an advantage. There are a few machine learning studies to incorporate clinical data in neuro-oncology [32]. Instead, researchers commonly developed a machine learning model with “wide”-dimensional input recruiting multiple huge imaging modalities. This will require substantial memory usage and cause the “curse of dimensionality” problem [34]. Several data inputs are one of the obstacles to the practical usage of the model. Thus, incorporating clinical data could reduce the abundant use of imaging modalities.
Although the network structures are the same, there are several differences in dataset between the previous [13] and the current study. First, the SNUH dataset in a previous study was small compared to the KROG dataset in this study. Secondly, the previous SNUH dataset was composed of MR images of homogeneous quality, and the clinical data were centrally reviewed by a physician. Meanwhile, the current KROG dataset originated from six institutions with various MR imaging qualities. There might also be variations in the clinical data. Input sizes are the same (Nine MR images), but in this study we allowed the use of triplicate MR images and did not include irrelevant regional information such as the whole brain or neck. These differences may contribute to the limited performance when direct validation was performed. Nevertheless, the new model in the whole dataset showed an acceptable performance in terms of the 10-cross internal validation method. Thus, this result showed us that the model network structure is still valid even in a heterogeneous dataset.
To support clinical use, the calibration model was constructed to provide PD probability. Calibration is important for clinicians to interpret the output of a machine learning model. We compared three calibration models, and found that GUESS [19], as recently published, is superior to other calibration strategies. The finalized model and optimal calibration model were integrated. By using our model, clinicians can access and estimate PD probability by applying their own clinical cases. As shown in example cases, full information and fine MR images seem to be necessary for correct prediction by model. In contrast, the incorrect prediction may be induced by missing clinical information and thick slice thickness of images. Recent response criteria [33] suggest strict imaging requirements. In 2016, the World Health Organization integrated molecular characteristics for the classification of central nervous system tumors. Therefore, we speculate that there are few cases having missing or incomplete information for the model to predict incorrectly.
Multidisciplinary team estimates PD and PsPD probability based on the time interval as well as imaging findings—e.g., 90% pseudoprogression in 1–3 months or progression in 1 year after treatment. PsPD occurs most commonly within the first 3–6 months following radiotherapy [35]. However, there is no clear cut-out time point between PsPD and PD. Clinicians tend to consider PsPD as an early event and PD as a late event after treatment, because most patients eventually experience the local disease progression. It is likely that the ratio of PsPD and PD cases is time-dependent. Following these logical processes, the “time interval” factor was introduced in the model network for the first time. On the other hand, the MGMT methylation in the model is based on the several studies addressing that it is associated with PsPD [7,36]. However, this seems to be interconnected with favorable survival from the MGMT promoter status and PsPD. In the results of the AVAglio trial (bevacizumab or placebo plus radiotherapy/temozolomide for newly diagnosed glioblastoma) data, the MGMT status was not significantly different between patients with PsPD and PD [37]. Although there are conflicting results regarding the MGMT status, we incorporated the MGMT status in the model, given the its prognostic/predictive value for treatment.
This study has several limitations. First, the KROG dataset could not be centrally reviewed due to the limited access to medical records in other institutions. In addition, variations in MR image quality could not be avoided due to the different MR imaging policies in each institution. However, this variation could be used to augment training data for the machine learning model to avoid the overfitting problem. Second, we collected as many cases as possible to finalize the model in Korea; however, they were still insufficient for deep machine learning. Additionally, the model was eventually validated with a cross-validation method with a multi-institutional dataset. Collecting datasets worldwide is required for future study to confirm the model performance in a completely held-out external dataset. Third, the ground truth—that is, the discrimination between PsPD and PD—was based on the expert opinions from the multi-disciplinary review board. Thus, the clinical decision can be affected by many confounding factors such as the second-line treatment rule, the withdrawal of the 1st line of adjuvant temozolomide, and steroid intervention in the middle of treatment. Thus, the model users should be cautioned in terms of reproducibility.
5. Conclusions
In conclusion, we tested the feasibility of the deep learning model in the extended dataset collected from multiple institutions. Cross validation results were acceptable, then the classifier scores were successfully transformed into interpretable probabilities by the optimal calibration model. This model could be used to support decision-making processes in a multi-disciplinary board.
Author Contributions
Conceptualization, I.A.K.; Data curation, S.H.J., I.H.K., D.H.L., S.-H.P., J.H.L., J.H.C., K.H.C. and J.H.K.; Funding acquisition, B.-S.J.; Methodology, B.-S.J. and A.J.P.; Resources, I.H.K.; Software, B.-S.J. and A.J.P.; Supervision, I.H.K. and I.A.K.; Validation, L.S. and S.H.C.; Writing—review and editing, I.A.K. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported by the grants from the Ministry of Science and Information & Communication Technology (#2020R1A2C005141), SNUBH (#14-2018-003) and Seoul National University Big Data Institute via Data Science Research Project to In Ah Kim.
Conflicts of Interest
A.J.P. is employed by SELVAS AI Incorporation.
References
- 1.Taal W., Brandsma D., De Bruin H.G., Bromberg J.E., Swaak-Kragten A.T., Sillevis Smitt P.A.E., Van Es C.A., Van Den Bent M.J. Incidence of early pseudo-progression in a cohort of malignant glioma patients treated with chemoirradiation with temozolomide. Cancer. 2008 doi: 10.1002/cncr.23562. [DOI] [PubMed] [Google Scholar]
- 2.Abbasi A.W., Westerlaan H.E., Holtman G.A., Aden K.M., van Laar P.J., van der Hoorn A. Incidence of Tumour Progression and Pseudoprogression in High-Grade Gliomas: A Systematic Review and Meta-Analysis. Clin. Neuroradiol. 2018;28:401–411. doi: 10.1007/s00062-017-0584-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3.Chaskis C., Neyns B., Michotte A., De Ridder M., Everaert H. Pseudoprogression after radiotherapy with concurrent temozolomide for high-grade glioma: Clinical observations and working recommendations. Surg. Neurol. 2009 doi: 10.1016/j.surneu.2008.09.023. [DOI] [PubMed] [Google Scholar]
- 4.De Wit M.C.Y., De Bruin H.G., Eijkenboom W., Sillevis Smitt P.A.E., Van Den Bent M.J. Immediate post-radiotherapy changes in malignant glioma can mimic tumor progression. Neurology. 2004 doi: 10.1212/01.WNL.0000133398.11870.9A. [DOI] [PubMed] [Google Scholar]
- 5.Brandsma D., Stalpers L., Taal W., Sminia P., van den Bent M.J. Clinical features, mechanisms, and management of pseudoprogression in malignant gliomas. Lancet Oncol. 2008 doi: 10.1016/S1470-2045(08)70125-6. [DOI] [PubMed] [Google Scholar]
- 6.Rowe L.S., Butman J.A., Mackey M., Shih J.H., Cooley-Zgela T., Ning H., Gilbert M.R., Smart D.D.K., Camphausen K., Krauze A.V. Differentiating pseudoprogression from true progression: Analysis of radiographic, biologic, and clinical clues in GBM. J. Neurooncol. 2018 doi: 10.1007/s11060-018-2855-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Brandes A.A., Franceschi E., Tosoni A., Blatt V., Pession A., Tallini G., Bertorelle R., Bartolini S., Calbucci F., Andreoli A., et al. MGMT promoter methylation status can predict the incidence and outcome of pseudoprogression after concomitant radiochemotherapy in newly diagnosed glioblastoma patients. J. Clin. Oncol. 2008 doi: 10.1200/JCO.2007.14.8163. [DOI] [PubMed] [Google Scholar]
- 8.Nasseri M., Gahramanov S., Netto J.P., Fu R., Muldoon L.L., Varallyay C., Hamilton B.E., Neuwelt E.A. Evaluation of pseudoprogression in patients with glioblastoma multiforme using dynamic magnetic resonance imaging with ferumoxytol calls RANO criteria into question. Neuro Oncol. 2014 doi: 10.1093/neuonc/not328. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 9.Thomas A.A., Arevalo-Perez J., Kaley T., Lyo J., Peck K.K., Shi W., Zhang Z., Young R.J. Dynamic contrast enhanced T1 MRI perfusion differentiates pseudoprogression from recurrent glioblastoma. J. Neurooncol. 2015 doi: 10.1007/s11060-015-1893-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Wang S., Martinez-Lage M., Sakai Y., Chawla S., Kim S.G., Alonso-Basanta M., Lustig R.A., Brem S., Mohan S., Wolf R.L., et al. Differentiating tumor progression from pseudoprogression in patients with glioblastomas using diffusion tensor imaging and dynamic susceptibility contrast MRI. Am. J. Neuroradiol. 2016 doi: 10.3174/ajnr.A4474. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Muoio B., Giovanella L., Treglia G. Recent Developments of 18F-FET PET in Neuro-oncology. Curr. Med. Chem. 2017;25:3061–3073. doi: 10.2174/0929867325666171123202644. [DOI] [PubMed] [Google Scholar]
- 12.Parent E.E., Sharma A., Jain M. Amino Acid PET Imaging of Glioma. Curr. Radiol. Rep. 2019;7:1–11. doi: 10.1007/s40134-019-0324-x. [DOI] [Google Scholar]
- 13.Jang B.S., Jeon S.H., Kim I.H., Kim I.A. Prediction of Pseudoprogression versus Progression using Machine Learning Algorithm in Glioblastoma. Sci. Rep. 2018;8:1–9. doi: 10.1038/s41598-018-31007-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Stupp R., Mason W.P., van den Bent M.J., Weller M., Fisher B., Taphoorn M.J.B., Belanger K., Brandes A.A., Marosi C., Bogdahn U., et al. Radiotherapy plus concomitant and adjuvant temozolomide for glioblastoma. Supplemental information. N. Engl. J. Med. 2005;352:987–996. doi: 10.1056/NEJMoa043330. [DOI] [PubMed] [Google Scholar]
- 15.Wen P.Y., Macdonald D.R., Reardon D.A., Cloughesy T.F., Sorensen A.G., Galanis E., DeGroot J., Wick W., Gilbert M.R., Lassman A.B., et al. Updated response assessment criteria for high-grade gliomas: Response assessment in neuro-oncology working group. J. Clin. Oncol. 2010;28:1963–1972. doi: 10.1200/JCO.2009.26.3541. [DOI] [PubMed] [Google Scholar]
- 16.Ellingson B.M., Bendszus M., Boxerman J., Barboriak D., Erickson B.J., Smits M., Nelson S.J., Gerstner E., Alexander B., Goldmacher G., et al. Consensus recommendations for a standardized Brain Tumor Imaging Protocol in clinical trials. Neuro. Oncol. 2015 doi: 10.1093/neuonc/nov095. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Zadrozny B., Elkan C. Obtaining calibrated probability estimates from decision trees and naive Bayesian classifiers; Proceedings of the Icml; Williamstown, MA, USA. 28 June–1 July 2001; University Park, PA, USA: Citeseer; 2001. pp. 609–616. [Google Scholar]
- 18.Naeini M.P., Cooper G.F. Binary classifier calibration using an ensemble of piecewise linear regression models. Knowl. Inf. Syst. 2018;54:151–170. doi: 10.1007/s10115-017-1133-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19.Schwarz J., Heider D. GUESS: Projecting machine learning scores to well-calibrated probability estimates for clinical decision-making. Bioinformatics. 2018 doi: 10.1093/bioinformatics/bty984. [DOI] [PubMed] [Google Scholar]
- 20.Naeini M.P., Cooper G.F., Hauskrecht M. Obtaining Well Calibrated Probabilities Using Bayesian Binning. Proc. Conf. Aaai Artif. Intell. 2015;2015:2901. [PMC free article] [PubMed] [Google Scholar]
- 21.Booth T.C., Larkin T.J., Yuan Y., Kettunen M.I., Dawson S.N., Scoffings D., Canuto H.C., Vowler S.L., Kirschenlohr H., Hobson M.P., et al. Analysis of heterogeneity in T2-weighted MR images can differentiate pseudoprogression from progression in glioblastoma. PLoS ONE. 2017:e0176528. doi: 10.1371/journal.pone.0176528. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Kebir S., Khurshid Z., Gaertner F.C., Essler M., Hattingen E., Fimmers R., Scheffler B., Herrlinger U., Bundschuh R.A., Glas M. Unsupervised consensus cluster analysis of [18F]-fluoroethyl-L-tyrosine positron emission tomography identified textural features for the diagnosis of pseudoprogression in high-grade glioma. Oncotarget. 2017;8:8294–8304. doi: 10.18632/oncotarget.14166. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Hu X., Wong K.K., Young G.S., Guo L., Wong S.T. Support vector machine multiparametric MRI identification of pseudoprogression from tumor recurrence in patients with resected glioblastoma. J. Magn. Reson. Imaging. 2011;33:296–305. doi: 10.1002/jmri.22432. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.DSouza A.M., Chen L., Wu Y., Abidin A.Z., Xu C., Wismüller A. MRI tumor segmentation with densely connected 3D CNN. Int. Soc. Opt. Photonics. 2018;10574:105741F. doi: 10.1117/12.2293394. [DOI] [Google Scholar]
- 25.Qamar S., Jin H., Zheng R., Ahmad P. 3D Hyper-Dense Connected Convolutional Neural Network for Brain Tumor Segmentation; Proceedings of the 2018 14th International Conference on Semantics, Knowledge and Grids, SKG 2018; Guangzhou, China. 12–14 September 2018. [Google Scholar]
- 26.Sun J., Chen W., Peng S., Liu B. DRRNet: Dense Residual Refine Networks for Automatic Brain Tumor Segmentation. J. Med. Syst. 2019 doi: 10.1007/s10916-019-1358-6. [DOI] [PubMed] [Google Scholar]
- 27.Lee J.K., Zong W., Dai Z., Liu C., Snyder J., Wen N. Overall Survival Prediction in Glioblastoma Using Clinical Factors Combined with Texture Features Extracted from 3D Convolutional Neural Networks. Int. J. Radiat. Oncol. 2019 doi: 10.1016/S0360-3016(19)30522-X. [DOI] [Google Scholar]
- 28.Chang K., Bai H.X., Zhou H., Su C., Bi W.L., Agbodza E., Kavouridis V.K., Senders J.T., Boaro A., Beers A., et al. Residual convolutional neural network for the determination of IDH status in low- and high-grade gliomas from mr imaging. Clin. Cancer Res. 2018 doi: 10.1158/1078-0432.CCR-17-2236. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Chang P., Grinband J., Weinberg B.D., Bardis M., Khy M., Cadena G., Su M.Y., Cha S., Filippi C.G., Bota D., et al. Deep-learning convolutional neural networks accurately classify genetic mutations in gliomas. Am. J. Neuroradiol. 2018 doi: 10.3174/ajnr.A5667. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Acharya G., Shah N., Krishnamurthi G., Khened M., Anand V.K. 3D convolution neural networks for molecular subtype prediction in glioblastoma multiforme. Int. Soc. Opt. Photonics. 2019 doi: 10.1117/12.2512417. [DOI] [Google Scholar]
- 31.Shaver M., Kohanteb P., Chiou C., Bardis M., Chantaduly C., Bota D., Filippi C., Weinberg B., Grinband J., Chow D., et al. Optimizing Neuro-Oncology Imaging: A Review of Deep Learning Approaches for Glioma Imaging. Cancers. 2019;11:829. doi: 10.3390/cancers11060829. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Booth T.C., Williams M., Luis A., Cardoso J., Ashkan K., Shuaib H. Machine learning and glioma imaging biomarkers. Clin. Radiol. 2020;75:20–32. doi: 10.1016/j.crad.2019.07.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Ellingson B.M., Wen P.Y., Cloughesy T.F. Modified Criteria for Radiographic Response Assessment in Glioblastoma Clinical Trials. Neurotherapeutics. 2017;14:307–320. doi: 10.1007/s13311-016-0507-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Trunk G.V. A Problem of Dimensionality: A Simple Example. IEEE Trans. Pattern Anal. Mach. Intell. 1979 doi: 10.1109/TPAMI.1979.4766926. [DOI] [PubMed] [Google Scholar]
- 35.Sanghera P., Rampling R., Haylock B., Jefferies S., McBain C., Rees J.H., Soh C., Whittle I.R. The Concepts, Diagnosis and Management of Early Imaging Changes after Therapy for Glioblastomas. Clin. Oncol. 2012 doi: 10.1016/j.clon.2011.06.004. [DOI] [PubMed] [Google Scholar]
- 36.Li H., Li J., Cheng G., Zhang J., Li X. IDH mutation and MGMT promoter methylation are associated with the pseudoprogression and improved prognosis of glioblastoma multiforme patients who have undergone concurrent and adjuvant temozolomide-based chemoradiotherapy. Clin. Neurol. Neurosurg. 2016;151:31–36. doi: 10.1016/j.clineuro.2016.10.004. [DOI] [PubMed] [Google Scholar]
- 37.Wick W., Chinot O.L., Bendszus M., Mason W., Henriksson R., Saran F., Nishikawa R., Revil C., Kerloeguen Y., Cloughesy T. Evaluation of pseudoprogression rates and tumor progression patterns in a phase III trial of bevacizumab plus radiotherapy/temozolomide for newly diagnosed glioblastoma. Neuro. Oncol. 2016;18:1434–1441. doi: 10.1093/neuonc/now091. [DOI] [PMC free article] [PubMed] [Google Scholar]