Skip to main content
Frontiers in Oncology logoLink to Frontiers in Oncology
. 2022 Aug 2;12:924245. doi: 10.3389/fonc.2022.924245

Artificial intelligence in the radiomic analysis of glioblastomas: A review, taxonomy, and perspective

Ming Zhu 1, Sijia Li 2, Yu Kuang 3, Virginia B Hill 4, Amy B Heimberger 5,6, Lijie Zhai 5,6,*, Shengjie Zhai 1,*
PMCID: PMC9379255  PMID: 35982952

Abstract

Radiological imaging techniques, including magnetic resonance imaging (MRI) and positron emission tomography (PET), are the standard-of-care non-invasive diagnostic approaches widely applied in neuro-oncology. Unfortunately, accurate interpretation of radiological imaging data is constantly challenged by the indistinguishable radiological image features shared by different pathological changes associated with tumor progression and/or various therapeutic interventions. In recent years, machine learning (ML)-based artificial intelligence (AI) technology has been widely applied in medical image processing and bioinformatics due to its advantages in implicit image feature extraction and integrative data analysis. Despite its recent rapid development, ML technology still faces many hurdles for its broader applications in neuro-oncological radiomic analysis, such as lack of large accessible standardized real patient radiomic brain tumor data of all kinds and reliable predictions on tumor response upon various treatments. Therefore, understanding ML-based AI technologies is critically important to help us address the skyrocketing demands of neuro-oncology clinical deployments. Here, we provide an overview on the latest advancements in ML techniques for brain tumor radiomic analysis, emphasizing proprietary and public dataset preparation and state-of-the-art ML models for brain tumor diagnosis, classifications (e.g., primary and secondary tumors), discriminations between treatment effects (pseudoprogression, radiation necrosis) and true progression, survival prediction, inflammation, and identification of brain tumor biomarkers. We also compare the key features of ML models in the realm of neuroradiology with ML models employed in other medical imaging fields and discuss open research challenges and directions for future work in this nascent precision medicine area.

Keywords: artificial intelligence, machine learning, brain tumor, immunotherapy, radiomics, tumor classification, survival prediction, radiogenomics

Introduction

Glioblastoma (GBM, WHO grade 4 glioma, IDH-wildtype) is the most aggressive primary brain tumor in adults with a dismal median overall survival (OS) of only 12 to 18 months and a 5-year OS rate of 6.8% (1, 2). Approximately 13,000 GBM cases are diagnosed in the United States each year, with an incidence rate of 3.2 per 100,000 members of the population (3, 4). Despite standard-of-care therapy including aggressive surgical resection followed by radiation therapy and chemotherapy, more than 90% of glioblastomas recur (4). To date, magnetic resonance imaging (MRI) remains the standard approach in the diagnosis, prognosis, and therapeutic monitoring of GBM patients because it is non-invasive, accessible, and cost efficient. However, interpretation of radiological imaging data can be subjective, challenging, and time-consuming, mainly because histologic findings are often radiologically occult. For example, therapy-induced treatment effect (i.e., pseudoprogression (PsP) or radiation necrosis) and true tumor progression manifest with identical MRI appearances, and differentiation between these entities remains an unsolved conundrum in current neuro-radio-oncology, particularly with novel therapies such as immune checkpoint inhibitors (5).

Radiomics (6, 7) in neuro-oncology seeks to improve the understanding of the biology and effects of treatment on the imaging appearance of brain tumors. Radiomics can promote precision medicine by extracting quantitative features from clinical imaging arrays and using methods from the field of artificial intelligence (AI) to make the radiological diagnosis more objective, accurate, and automatic. Rather than designing hard-coded step-by-step algorithms based on prior knowledge in biology or medicine, or design specific “learning” approaches to mimic human cognitive functions, machine learning (ML) as a subfield of AI can create a computational model and train it with a number of datasets to statistically solve problems without being explicitly programmed (8). Generally, ML includes supervised learning, unsupervised learning, and reinforcement learning. Supervised learning trains an ML model to predict a target variable from a set of predictive variables (i.e., data samples) with the help of labels/annotations (i.e., ground truth of the target variables) and the loss function, also known as cost function, which is a computational difference between predicted target variable values and the label/annotation values (9). It should be noted that, although labeling and annotation share the same meaning in ML, they slightly differ in neuro-oncology radiomic analysis. In the context of this manuscript, labeling is related with classification problems (e.g., the ground truth of tissue is histological, including different classes of brain tumors, treatment effect versus tumor growth, and others) whereas annotation refers to segmentation problems (partitioning an image into multiple regions/objects, such as enhancing tumor, necrosis, and unenhancing tumor and edema). Unsupervised learning infers the inherent structure from the input data without labels/annotations (10). In reinforcement learning, intelligent agents learn to take actions in an environment in order to maximize the notion of cumulative reward (11). Currently, most AI techniques applied in brain tumor radiomic studies belong to the supervised ML, for both classifications and segmentations. Unsupervised ML is mainly employed for image segmentations while reinforcement learning has not been explored in this area. Therefore, in this paper, we mainly focus on the supervised ML techniques for most GBM radiomic analyses. Thanks to the rapid development of radiology and computational hardware, researchers can now take advantage of many radiological data to train various ML models, such as decision trees (DTs), logistic regression, artificial neural networks (ANNs), support vector machines (SVMs), and k-nearest neighbors (k-NN) for brain tumor radiomic analysis. The different techniques applied in AI (mainly ML algorithms) are the technical core for the analysis of large amounts of multidimensional radiologic and clinical data (12), which directly determine the quality of radiomic analysis results.

In the past decade, ML has been widely exploited in many data-driven applications, e.g., imaging and computer vision (13), bioinformatics (14), online advertising (15), and natural language processing (16). The dataflow of a general supervised ML-based GBM radiomic analysis can be divided into four steps as shown Figure 1A: 1) Data Acquisition. MRIs are performed on patients with a brain tumor. These raw MRI data are further preprocessed (e.g., data cleaning, co-registration, bias correction, normalization), and then they are labeled/annotated by radiologists to define the regions for the ML training and validation process. The labeled/annotated imaging data are deposited into customized/private datasets that are owned and maintained by medical research institutions. Some imaging data are also uploaded into public datasets for the purpose of open access to all researchers. AI-assisted radiomic analysis can acquire imaging data from both types of datasets. It is of note, however, that private datasets usually contain a fairly large amount of raw data, hundreds of samples for each institution if applicable, whereas public datasets usually contain limited amount of less well-labeled/annotated, non-standardized imaging data (17); 2) Data Augmentation and Preprocessing for ML Models. The acquired data and its labels/annotations are usually first subject to augmentation, in which image data are processed in pair-wise format (i.e., each pair contains a data sample and the corresponding labels/annotations) to increase the sample variety, hence improving the generality of the data. Multiple approaches are utilized in augmentation, such as geometric transformation, color augmentation, and synthesis of similar-appearing imaging data. Then, the augmented imaging data can be preprocessed (e.g., through feature extraction) to simplify and/or to improve the effectiveness and efficiency of the subsequent ML training process (18); 3) ML Model Training and Validation: The augmented and preprocessed data are subsequently fed to ML models to train the model parameters in order to minimize the “cost function,” while the implicit data feature will be extracted statistically. The ML models are also validated during the training process to prevent overfitting, which is when the trained ANN model only predicts accurately to the training dataset but loses the generalizability to new samples (19); 4) AI-Assisted Clinical Diagnosis/Deployment. Once the trained models meet the accuracy requirement, they can be deployed to the application to perform predictions such as classification and segmentation. As aforementioned, no biologic hypothesis or knowledge is required to build an ML model. However, inclusion of this information and/or other forms of data (e.g., clinical data, genomic data) may help with the overall ML model performance by reducing the data size or improving the data quality during the preprocessing step.

Figure 1.

Figure 1

AI/ML in GBM radiomic analysis: (A) Overall workflow of AI-assisted GBM analysis: 1) Data Acquisition. Raw radiological image data are acquired by MRI scanning of GBM patients. Images are collected into public or private data sets. Before analysis, images are preprocessed (e.g., data cleaning, co-registration, normalization) and standardized (e.g., format, resolutions, voxel sizes). Then, radiologists annotate the images, color-coding different parts of the tumor habitat. 2) Data Augmentation and Preprocessing for ML Models. Imaging dataset and its annotations from step 1 are further “augmented” via geometric transformations, photometric transformation, and/or synthetic data (e.g., GAN) to improve the data generalizability, followed by the optional preprocessing for ML modeling, a process that includes feature extraction to filter out “useless” data and extract explicit features (e.g., biological and/or geometry) in the images. 3) ML modeling and training. Augmented and preprocessed data are fed into various ML models (e.g., SVM, RF, CNN) for GBM radiomic analysis training and validation. Advanced techniques such as transfer learning and multimodal data fusion (e.g., clinical and genomic data) can be employed to improve the training accuracy as well as generality. 4) AI-Assisted Clinical Diagnosis/Deployment. Predictions from the ML models for various medical demands, such as differential diagnosis and survival estimation. (B) Current major challenges (left panel: 1, 2, 3) and perspectives for corresponding solutions (right panel: 1*, 2*, 3*) in AI/ML-assisted GBM radiomic analysis: 1→1* Current GBM radiological image datasets are limited in low numbers, insufficient annotations, and poor organization. Enrichment and standardization of current GBM radiological datasets are urgently needed, while incorporation of clinical and/or genomic data (red circle) can further enhance the performance of ML prediction models; 2→2* develop more comprehensive ML models to further improve the prediction accuracy and address the relatively low generalizability of current models; 3→3* further strengthen collaborations among clinicians, biomedical researchers, and computer scientists to overcome the lack of efficient communications between these parties for the highly multidisciplinary research.

Traditional ML methods such as SVMs (2022) and random forests (RFs) (2326), an ensemble combination of decision trees, are commonly used for pattern classification in tumor studies. Recently, ANN, especially convolutional neural network (CNN)-based deep learning (DL), is gaining popularity because of its improved scalability and the capability of exploiting deep layers to extract implicit local and global features in neuro-oncology images. They can achieve state-of-the-art performance in object detection and tracking (27, 28), image classification, and semantic segmentation (25, 2932). In ML, each “neuron” is referring to a computational unit in the ANN rather than a biological neural. With more complicated ML models and structures (e.g., more neuron network layers, more neurons in each layer) and a larger number of parameters introduced into the neural network, the training process intends to extract more features but may suffer extensive computational performance degradation and overfitting of the trained model. Current state-of-the-art ML models can achieve an accuracy as high as 0.97 (i.e., 97%) in brain tumor radiomic analysis (33). However, these results are based on a limited number of datasets and from retrospective studies, which may still not be generalizable for patients from different geographic locations. Therefore, current clinical brain tumor radiomic analysis cannot entirely rely on the ML-based techniques and still needs manual verification. In summary, existing ML techniques can only partially fulfill the need for automatic detection and analysis of GBM characteristics for both clinical and preclinical studies (3437).

In addition to the ML techniques, the quality of radiologic images that are used for ML training dramatically affects the outcomes of radiomic analysis. Radiological images can be acquired from different imaging modalities, such as MRI, computed tomography (CT), and positron emission tomography (PET). Among these, MRI image data are currently employed as an essential data type in radiomic neuro-oncology applications because 1) they provide exquisite detail of brain, spinal cord, and vascular anatomy through excellent tissue contrast in any imaging plane; 2) different MRI sequences are able to capture key components of tumor biology with high sensitivity, such as blood–brain barrier breakdown, necrosis, edema, non-enhancing tumor infiltration, blood flow, and cellular density, and can distinguish tumoral sub-compartments that are likely to reflect local cellular phenotypes and genotypes; and 3) they can non-invasively and non-destructively interrogate tumors repeatedly to assess response to treatment and thus they can be integrated into therapeutic strategies. Understanding these image-based features is critical as they not only represent a key data resource in radiomic analysis (6) but also help improve the accuracy and other performance criteria of ML models.

In this review, we provide an overview of the latest advancements and in-depth discussions on the most urgent and challenging questions of AI-assisted GBM radiomic analysis. Given the exponential increase of AI-based radiomic studies led by researchers from various backgrounds, such as oncology, radiology, computer science, and engineering, our review article briefly explains the key concepts of ML techniques instead of delving into the technical details. This article is structured with emphases on the deployments of various ML techniques in meeting specific GBM radiomic clinical needs, e.g., differentiating GBM from other brain tumors or non-tumors, predicting overall survival (OS), and correlating with other biomarkers. First, as ML technology in radiomics is radiological imaging data-driven, we start with the discussion on imaging data preparations that are commonly employed in current GBM radiomic analyses. We briefly introduce the acquisition pipeline for private or customized imaging datasets and summarize public radiologic datasets that are currently available for researchers to train their ML models for various applications. We also describe general methods for data augmentation and preprocessing for ML models, which are critical for training, validation, and testing of ML algorithms. Next, we overview the ML techniques that have been employed in radiomic analysis for GBM diagnosis and treatment. Advantages and limitations of existing ML models including both algorithms and architectures are discussed in the context of various GBM-associated medical applications. Finally, we bring up our perspectives on the strategies for overcoming challenges regarding AI/ML applications in GBM radiomic analysis, including 1) the most challenging issues affecting the generalizability and accuracy of AI-assisted radiomic GBM analysis; 2) promising strategies to enhance performance of AI models in GBM radiomic analysis; 3) outlook on the collaborative teamwork between computer scientists, engineers, physicians, and biomedical researchers. By elaborating current research developments and challenges in the state-of-the-art ML-assisted GBM analysis, we hope to inspire researchers from different fields for the development of the next generation of AI-assisted radiomic tools that can significantly improve early detection, treatment efficacy, and life quality of patients with GBM.

Radiomic data preparation

As ML is an intensive data-driven algorithm/process, the quality of the training data can significantly influence the trainable parameters in ML models, hence affecting the accuracy and generalizability of the network outcomes (38). Thus, it is essential to review the key components of data preparation in ML-based GBM radiomic analysis, including radiological imaging data resources, the image acquisition pipeline, imaging datasets (private/customized datasets vs. publicly accessible datasets), data augmentation, and preprocessing techniques for the subsequent ML model training.

Radiomic image resources

MRI is the most frequently used radiological modality for brain tumor imaging. MRI provides better contrast resolution than CT, with better tissue characterization. It can also detect blood vessels, vascular malformations, and demyelinating disease (39). It does not involve X-rays or the use of ionizing radiopharmaceuticals, either. Therefore, MRI is particularly suitable to image gliomas. Yet, MRI may be perceived as less comfortable by patients (e.g., overweight or fear of enclosed spaces) and cannot be performed if the patient has ferromagnetic implants. In addition, MRI does not show ossified or calcified structures as well as CT (such as the calvarium) and therefore may not show the effects of tumors on the calvarium in comparison to CT (39, 40).

For GBM MRI, T1-weighted (T1), T1-contrast-enhanced (T1c or T1-ce), T2-weighted (T2), and fluid-attenuated inversion recovery (FLAIR) are the most commonly used MRI sequences, because they can provide different yet complementary information in characterizing tissue such as gray matter, white matter, fat, blood, fluid, and lesions (41, 42). MRI is based on radiofrequency pulses within a magnetic field in which time of repetition (TR) and time of echo (TE) are calculated. T1 and T1-ce are produced through short TR and TE times; T2 is produced by larger TR and TE times; and FLAIR is produced through very large TR and TE times. In addition, diffusion-weighted imaging (DWI) can detect the restriction of random movements of water molecules that makes DWI extremely sensitive to detect acute stroke and increased cellularity as in GBMs, lymphoma, and metastases (43); magnetic resonance angiography (MRA) and venography (MRV) can generate pictures of the arteries and veins to evaluate for stenosis or aneurysms; magnetic resonance spectroscopy (MRS) is used to measure the levels of different metabolites and biochemical changes in the brain (44), providing information on tumor metabolism (45, 46); perfusion-weighted imaging (PWI) shows the perfusion of tissues by blood, such as the cerebral blood volume of a tumor relative to normal-appearing white matter of the brain (47); and functional MRI (fMRI) detects the increase in blood oxygen level when blood flow increases to a brain area involved in the performance of an assigned task (e.g., finger tapping, lip pursing, thinking of words, thinking of answers to questions after hearing a story) (48) and depicts where eloquent brain areas are in relation to the tumor as the surgeon or radiation oncologist plans a surgical approach to biopsy or resection or radiation therapy. The novel amide proton transfer (APT) imaging can detect amide protons of endogenous mobile proteins and peptides in tissue based on chemical exchange–dependent saturation transfer (CEST) MRI (49, 50).

Another useful radiological technology is the PET scan, which takes advantage of a slightly radioactive substance (e.g., C-11 methionine (MET), F-18 fluorothymidine (FLT), F-18 fluoroethyl-L-tyrosine (FET)) that functionally is preferentially taken up by tumor cells (51). PET is especially helpful for fast-growing (high-grade) tumors and for distinguishing between tumors and non-tumor (e.g., scar, inflammation) tissue (52, 53). Therefore, the combined use of MRI and PET can provide complementary information to achieve more accurate brain tumor diagnosis (54).

Radiological image acquisition pipeline

The image acquisition pipelines remain very similar between radiological scanners (50, 55). Subjects undergoing MRI (i.e., patients and control subjects) are usually examined on a clinical 1.5T or 3T scanner with a multichannel receive-only head coil array under various scanning parameters (e.g., TR, TE, field of view, matrix/voxel size). A sequence of 2D and 3D radiologic images (e.g., various MRI, CT, PET) is obtained. These images are cleaned, normalized, and co-registered (i.e., image preprocessing, Figure 1A ). It is worth noting that different image intensity normalization schemes may influence not only the registration (56) and segmentation process (57, 58) but also the implicit texture features hidden in the different modal MR images and thus affect the subsequent feature selection and ML training outcomes for various GBM classification applications (59). Shinohara et al. (60) introduced a set of seven statistical principles of image normalization (SPIN). In addition to the common mean-maximum (or standard deviation) normalizations (59) and histogram-based normalizations (57, 61, 62), Shinohara et al. (60) also proposed a hybrid multimodal normalization method to match the natural balance of tissue intensities with physical interpretation. On the other hand, some data may be standardized (e.g., voxel sizes, resolution) while others may not. Then, the images are labeled and annotated into various categories of tissues and/or lesions by experienced radiologists using a variety of software (e.g., ITK-SNAP (63), 3D Slicer (64)) to produce a labeled/annotated imaging data set (17).

Major public datasets

Since not all ML researchers can directly access private/customized high-quality labeled/annotated brain tumor datasets, which are usually owned and protected by medical institutions, public datasets are essential and provide an equal platform to these researchers to train and compare the outcomes of their ML models. In neuro-oncology, one of the most commonly used public online image datasets is from the Brain Tumor Segmentation (BraTS) challenges, organized by the Medical Image Computing and Computer Assisted Interventions (MICCAI) and other professional organizations (34, 6569) since 2012. As of March 2022, the latest BraTS 2021 consists of a total of 2,040 brain tumor cases/patients, and it is divided into three subsets: training (1,251 cases), validation (219 cases), and testing (570 cases). Only training and validation subsets are open to the public research access, and these two subsets include a set of multimodal 3D MRI scans (i.e., T1, T1c, T2, FLAIR) for each case. The training dataset also includes a 3D annotation model (i.e., GD-enhancing tumor, peritumoral edema/non-enhancing infiltrative tumor, necrotic tumor core (NCR), and normal) for each case (17). In addition, BraTS 2021 includes O6-methylguanine-DNA-methyltransferase (MGMT) biomarkers for 585 patients (out of a total of 1,251 cases) for the training datasets. It should be noted that, unlike BraTS 2020, BraTS 2021 does not include survival information any longer. BraTS 2020 included the survival information for 265 patients (out of a total of 460 cases) in the training and validation datasets. Other widely used datasets include The Cancer Imaging Archive (TCIA) (70) and The Whole Brain Atlas by Harvard Medical School (71). TCIA has a collection of 13 brain tumor sub-datasets, including the Ivy Glioblastoma Atlas Project (IvyGAP) (72), The Cancer Genome Atlas (TCGA)-GBM (73), GLIS-RT (74), and CPTAC-GBM (75). These sub-datasets mainly focus on high-grade glioma (HGG/GBM) and lower-grade glioma (LGG). Some of the data in TCIA are also included and standardized in the most recent BraTS 2021 dataset. The Harvard brain atlas consists of the radiology data (e.g., MRI, CT, PET) for about 40 subsets of normal brain and various brain disease states. However, none of the datasets in the TCIA or Harvard brain atlas are pixel-wisely annotated for the ML segmentation tasks. In addition, the data format for each patient in TCIA collection varies in terms of pulse sequences (e.g., T1, T1c, T2, FLAIR, PWI, DWI) and the resolution (i.e., with matrices varying from 128*128 to 896*896), even in the same sub-dataset (e.g., TCGA-GBM, IvyGAP). A summary of these datasets is depicted in Table 1.

Table 1.

List of three major sources for radiomic neuro-oncology public datasets.

Dataset Radiology data type Data size Image resolution
BraTS 3D MRI 2,040 subjects, including both HGG and LGG 240 * 240 * 155
TCIA 2D MRI, CT, axial slices 13 brain tumor sub-datasets, including IvyGAP (39 subjects), TCGA-GBM (262 subjects), and TCGA-LGG (199 subjects) Varying from 128*128 to 896*896
Harvard Medical School: The Whole Brain Atlas 2D MRI, CT, PET, axial slices 8 subsets for brain tumors, and 30 other subsets for normal brains and other non-tumor brain diseases 256 * 256

Data augmentation and preprocessing for the ML models

Data augmentation is a commonly used technique in ML for the purpose of promoting the accuracy and generalizability of the ML algorithms. Data augmentation can be attained by 1) adding slightly modified copies and/or 2) creating new synthetic data from already existing data. The former usually employs geometric transformation and photometric transformation including flipping, pixel-level augmentation, cropping, rotating, noise injection, and random erasing (76, 77), while the latter may make use of generative adversarial networks (GANs) (78, 79) to create new synthetic images that resemble the original dataset. It should be noted that GANs also belong to ML-based networks that require abundant training data to generate resembled data. Data augmentation acts as a regularizer and helps reduce class imbalance and overfitting (76), so as to improve both the accuracy and the generalizability of the ML outcomes.

Current GBM radiomic studies are often hindered by limited and unbalanced data samples; therefore, using ML models alone may not achieve statistically significant outcomes (8082). In this regard, the preprocessing plays a vital role by enhancing and extracting some image features, especially the biological/medical meaningful ones in the regions of interests (ROIs), and/or filters out some “useless” image data from the datasets, before performing the ML training and analysis (18). General data preprocessing approaches for ML models include feature extraction and feature selection (18, 81). In GBM radiomic datasets, MRIs contain various features, such as image texture (23), local histograms (24), structure tensor eigenvalues (25), gray-level co-occurrence matrix (GLCM) (83), and local binary pattern (LBP) (41). Yet some of these features might be correlated in that the total number of effective features can be further reduced, by employing feature selection algorithms such as principal component analysis (PCA) (84), least absolute shrinkage and selection operator (LASSO) (85), linear discriminant analysis (LDA) (86), t-tests (87), analysis of variance (88), and information gain based methods (89), or based on certain evaluation criteria, such as probability of error (POE) and average correlation coefficient (ACC) (59, 90). Injecting feature extraction and selection can significantly help to reduce the computational complexity of ML models and speed up the training process and possibly improve the accuracy of ML models for brain tumor classification and segmentations (91, 92). Nevertheless, one should note that such preprocessing should be treated with care so that the ML model is not overfitted to particular features, which could lose generalizability to a different dataset.

Discussion

We have illustrated various radiomic data sources and data preparation techniques that are commonly employed in ML-based GBM radiomic analysis. One prominent issue in current GBM radiomic data preparation lies in the lack of standardized image acquisition specifications (e.g., repetition time, echo time, voxel sizes, image resolutions) between different radiological equipment and medical institutions (i.e., multicenter multi-vendor, McMv, datasets), which may 1) bias the image data (e.g., intensity of pixels, actual voxel size); 2) require additional image data preparation (e.g., cropping, up/down-sampling) to train ML models with different datasets; and 3) impede the development and cross-validation of more general/robust and accurate ML models for McMv datasets. Although BraTS has made a huge effort and progress in standardizing radiology data for over 2,000 GBM patients/cases, it is not yet sufficient for various GBM analysis applications. The second important issue is that most existing datasets have limited types of brain tumors (e.g., GBM/HGG, metastasis, and LGG), while the Harvard Atlas is limited by the number of subjects/patients. Scarcity of brain tumor/disease types and lack of data impede the application of ML to accurately distinguish various brain tumors and diseases. Lastly, most datasets include only MRI data while only a few datasets consist of other modality radiology data such as CT and PET. Most of them do not include other biological information (e.g., survival time, histopathological data, biomarkers) either. There is also lack of longitudinal radiology data to show disease evolution for patients receiving various treatments. With more complementary data and medical/treatment history (including radiomic data) that help comprehensively describe the brain tumor/disease status, an improvement in the accuracy of radiomic analysis and prediction can be expected.

Application of AI/ML in GBM diagnosis and therapeutic monitoring

The early brain tumor radiomic studies often relied on conventional radiomic feature-based ML methods that extract relatively explicit image texture features (e.g., shape, GLCM, LBP) to train traditional ML models such as SVMs (2022) and RFs (2326) in order to differentiate brain tumor versus non-tumors (or different types of brain tumors), predictions of overall survival, etc. Recently, by taking advantage of deep neural networks (DNNs) that include more neurons and layers to statistically recognize global, deep, and implicit imaging features, DL techniques can achieve state-of-the-art performance for automatic analysis of brain tumors on multimodality imaging and clinical data (32). Additionally, deep feature-based ML techniques build statistically/biologically meaningful models or utilize DNNs to extract deep implicit features from the radiology images and then apply traditional ML models for classifications (93, 94). Despite their differences, the above ML models all exploit prior biomedical and image features knowledge to 1) preprocess the radiological imaging data to extract imaging and/or biological meaningful features and 2) optimize the ML structure/algorithm for specific classification/segmentation tasks. Examples include the stacked denoising autoencoders (95) and the Convolutional Restricted Boltzman Machine (96). All these ML models have been applied in radiomic analysis to address unmet needs in GBM diagnosis, therapeutic monitoring, and/or prognosis (e.g., brain tumor classifications, survival predictions and biomarker identifications), while at this point, CNN-based DL enjoys the most generalizability and highest accuracy. Details are further discussed under the context of individual study case as follows.

GBM diagnosis and classification

One of the major ML applications in GBM radiomic analysis is to facilitate the differentiation between GBM and other histopathological processes. More specifically, such applications mainly fall into three categories: 1) distinguish brain tumor from other non-cancerous pathologies; 2) distinguish GBM from other brain tumors; and 3) differentiate between true progression and treatment effect (PsP or radiation necrosis). We hereby provide an overview of a few of these classification problems in brain tumor diagnosis.

1) Differentiating tumor from non-tumor

One of the most critical radiomic functions in brain tumors is to distinguish between malignant brain tumor and non-tumor pathologies, which include tumefactive demyelination, infection, inflammation (e.g., paraneoplastic syndromes and autoimmune disease), cortical dysplasia, and stroke. However, due to insufficient data available for ML training of each specific non-tumor type, existing studies mostly classify all data into two major categories: tumor (e.g., GBM/HGG, metastases, LGG) and non-tumor (i.e., control/normal and non-cancerous pathologies such as inflammation). Some studies tested their ML models on their own private data, while others took advantage of public datasets (e.g., BraTS, TCIA, Harvard brain atlas) or a combination of private and public data to expand the model’s generalizability. As discussed in Section 2.4, data preprocessing (e.g., filtering and feature extractions) are also often used to denoise and enhance the lesion region in the input MRI slides, with a hope to speed up the ML training process and improve the accuracy of distinguishing between tumors and non-tumors.

For those using public datasets, Ari et al. (97) proposed a three-phase extreme learning machine local receptive field (ELM-LRF) method for tumor classifications: removal of the noise using local and non-local methods, segmentation of benign or malignant tumor using ELM-LRF, and then use of a CNN classification. As a result, they achieved an effective classification accuracy of 0.97. Mohsen et al. (84) took advantage of a discrete wavelet transform (DWT) for feature extraction and principal component analysis (PCA) for reduction, together with a fuzzy C-means DNN to classify a dataset of 66 brain MRIs from Harvard Brain Atlas into four classes, i.e., normal, glioblastoma, brain sarcoma, and brain metastatic bronchogenic carcinoma tumors. An accuracy of 0.97 was achieved, and an area under the curve (AUC) approximated 0.984.

Alves et al. (83) quantified the gray-level pattern, pixel interrelationships, and the spectral properties of an image and achieved two fundamental features from an MRI sample, i.e., GLCM and gray-level run-length (GLRL). By combining this texture analysis with ML models (e.g., SVM, RF), they differentiated brain tumors from inflammatory lesions in their local MRI dataset and achieved a high accuracy of 0.83 and AUC of 0.906. Citak-Er et al. (93) applied a multiregional and multiparametric recursive feature elimination method, which was based on the Mann–Whitney ranking score, and then they employed the SVM-based multilayer perceptron (MLP) classification model to achieve a tumor detection accuracy of 0.93.

Amin et al. (41) mixed the public datasets and local datasets to differentiate tumors and non-tumors. They employed a Weiner filter to denoise and enhance the lesion region in the input MRI slides and used potential field (PF) clustering to identify the tumor region. Gabor wavelet transform (GWT) and LBP features were fused with various ML models (i.e., SVM, DT, k-NN, and naïve Bayes) to further improve the classification accuracy. The approach yielded an accuracy greater than 0.93 and an AUC of 0.96. Zhou et al. (98) treated holistic 3D MRI samples as sequences of 2D slices to extract some 3D features on brain tumors. They introduced a recursive structure, i.e., the long short-term memory (LSTM), to a deep CNN model (i.e., DenseNet) to handle such sequential data classification, and this DenseNet-LSTM model achieved an outstanding accuracy of 0.92 using the BraTS dataset (99).

Both Banerjee et al. (100) and Xu et al. (101) introduced transfer learning (TL) to improve the accuracy of the DNN-based ML classifier with non-brain-tumor images. They first pretrained the ML classifier with the large general image dataset, ImageNet (102), and then they fixed the pretrained parameters in the CNN hidden layers and fine-tuned the parameters in the output layers with neuro-oncology MRI images. Xu et al. (101) even embedded an SVM with the CNN to distinguish between GBM and LGG. By doing so, the two studies achieved classification accuracies of 0.97 and 0.975, respectively.

2) Differentiating primary from secondary brain tumors

Secondary/metastatic brain tumors have as high as fivefold incidence of that of primary brain tumors and manifest a rapid growth, causing significant brain tissue damages. Patients typically present with multiple metastatic tumors throughout the brain (103). A traditional non-ML-based approach to distinguish multifocal GBM from metastases on α[11 C]-methyl-L-tryptophan (AMT)-PET images is to examine the tumoral standardized uptake values (SUVs), mean tumor/cortex SUV ratio, and tumor/cortex volume of distribution (VD)-ratio (104). Compared to GBM, metastases had lower values of all three parameters. However, this approach can only achieve an accuracy of 0.72.

Many studies implemented a combination of various feature extractions and regular ML models to find the best performance for their applications. Zacharaki et al. (86) introduced a Gabor texture filter with feature ranking to extract tumor features, and derived feature ranking scores, and then applied three ML models to distinguish GBM from metastases: SVM with recursive feature elimination (SVM-RFE), linear discriminant analysis (LDA, also known as Fisher linear discriminant) with Fisher’s discriminant rule (105), and k-NN. Among the three models, SVM-RFE achieved the highest mean accuracy and AUC of 0.91 and 0.936, respectively. Chen et al. (85) compared 30 diagnostic models that were built based on five feature selection models and six classification algorithms for distinguishing GBM and metastases. The five feature selection models included distance correlation, RF, least absolute shrinkage and selection operator (LASSO), eXtreme gradient boosting (XGBoost), and gradient boosting decision tree (GBDT), while the six classification algorithms included LDA, SVM, RF, k-NN, Gaussian naïve bayes (GaussianNB), and logistic regression (LR). The results showed that the combinational model of distance correlation and LR outperformed all other combinations in terms of testing accuracy (0.79) and AUC (0.80), although some other combinations achieved similar results as well.

Priya et al. (106) analyzed 60 GBM and 60 metastases cases with 12 regression or ML-based classifier models and four feature reduction/selection strategies—45 combinations in total. According to their results, the mean performance of various models was slightly better with FLAIR images than multiparametric sequences in terms of AUC, while the combination of full feature and LASSO achieved the highest AUC of 0.953, although full features with other models, such as ElasticNet (107) and RF, achieved similar results. de Causans et al. (55) trained on T1 MRI data with 71 GBM and 72 metastasis cases using 100 extracted features, based on the Image Biomarker Standardization Initiative (IBSI) (108). With these selected features, a total of 144 models combining nine feature scaling methods and 16 classifiers (regression and ML-based) were compared. All 144 classifiers of the 21 GBM and 16 metastases cases achieved a mean accuracy and AUC of 0.8 and 0.85, respectively.

3) Differentiating GBM from primary central nervous system lymphoma

GBM and primary central nervous system lymphoma (PCNSL) are not only common intracranial malignancies but also often share similarities in radiological appearance. However, the management for each disease is quite different (109). Recently, multiple ML-based predictive analytics have arisen to help differentiate GBM from PCNSL radiologically with a relatively high sensitivity and specificity. Outcomes were assessed based on test characteristics such as accuracy, sensitivity, specificity, and AUC.

Due to the binary classification nature of distinguishing between GBM and PCNSL, SVM used to be the most commonly employed model for its computational simplicity (110115). Other commonly used ML models such as k-NN and RF were also exploited and compared (114, 116). Even though the training datasets were relatively small, approximately 110 or fewer samples with about two-thirds of the entire dataset containing GBM and one-third of PCNSL, it turned out that most ML-based models were able to achieve an accuracy between 0.9 and 0.96, a sensitivity of 0.84 or higher, a specificity of 0.89 or higher, and AUC of 0.92 or higher. More notably, Kang et al. (114) and Suh et al. (116) even compared the prediction outcomes of their ML models with the prediction from human radiologists, and the results showed the superiority of ML models over human radiologists in all four criteria, especially in accuracy, sensitivity, and AUC. It is unknown whether the combination of the ML model and human radiologist read would have attained even higher accuracy, sensitivity, and AUC.

Recently, more sophisticated DL models have been employed. Priya et al. (117) examined five different ML approaches (i.e., LASSO, SVM, RF, Ridge, and MLP) to distinguish between 97 GBM and 46 PCNSL cases, with all five approaches sharing similar results. Yet, LASSO had the best performance (0.88 in accuracy and 0.92 in AUC) when using features from the whole tumor, while MLP had the best performance (0.86 in accuracy and 0.91 in AUC) when only using the features from the single largest slice. For an even larger dataset (i.e., 160 GBM and 160 PCNSL), McAvoy et al. (118) applied a CNN variant, EfficientNet (119), and by using TL based on ImageNet, they achieved an accuracy of 0.93 and AUC of 0.94.

4) Differentiating treatment effects versus true disease progression

Pseudoprogression (PsP) is the apparent growth of a lesion or development of new lesions on imaging that represents inflammatory treatment-related changes but looks just like viable tumor growth on MRI. PsP is most common between 3 and 6 months after the completion of radiation therapy, and the corresponding imaging findings will subside on their own over time (120). PsP is more likely in MGMT promoter-methylated tumors treated with temozolomide. The increased contrast enhancement on MRI may be caused by the increased vascular permeability from cytotoxic therapies including radiotherapy and chemotherapies such as temozolomide, which may benefit patients receiving immunotherapy and temozolomide but often leads to premature discontinuation of treatment owing to the false judgment of progression of disease (121). Radiation necrosis is another treatment effect that can occur any time after radiation therapy but is most common 1–2 years after radiation. It should be differentiated from true progression of viable tumor before treatment changes are contemplated. As a result, accurate differentiation between treatment effect (i.e., pseudoprogression or radiation necrosis) and true tumor progression is critical in the treatment decision. PsP may be associated with a survival advantage. A key radiology tool in differentiating pseudoprogression or radiation necrosis from true progression of disease is dynamic susceptibility contrast (DSC) MR perfusion-weighted imaging (PWI). Elevated corrected relative cerebral blood volume (crCBV) relative to normal-appearing white matter is more common in a viable tumor than in treatment effects (122). However, PWI is unreliable in patients treated with immunotherapy such as immune checkpoint inhibitors, such that, per immunotherapy response assessment in neuro-oncology (iRANO), the patient is followed for 3 months and then a determination is made of whether the initial increase in size of the lesion represented treatment effects or a viable tumor (123125).

Booth et al. (126) first analyzed the tumor heterogeneity in T2 MRI using topological descriptors called Minkowski functionals (MFs). Then they utilized an SVM model, together with image features such as MFs, size, and signal intensity, to distinguish between pseudoprogression and true progression, and achieved an accuracy of 0.88, slightly higher than using RF for feature selection and LASSO for classification (0.86). Hu et al. (127) took advantage of T1 MRI and other eight-dimensional feature vectors, including T2, FLAIR, proton density, ADC, PWI, derived relative cerebral blood volume (rCBV), relative cerebral blood flow (rCBF), and mean transit time maps, to train an SVM model, and achieved an AUC of 0.94 in distinguishing between pseudoprogression and true progression. The ADC map derived from DWI and rCBV and rCBF derived from PWI were found to make a greater contribution to the discrimination than the conventional radiology images do.

Due to the time correlation embedded in true progression and PsP radiology data, Lee et al. (128) and Jang et al. (129) exploited recursive LSTM-CNN structures on MRI to distinguish between the two occurrences. In comparison to Lee’s multimodal MRI data (i.e., T1, T2, FLAIR), Jang et al. (129) combined/fused T1 MRI data with clinical features to develop an LSTM-CNN clinic-feature-fused model and achieved an AUC of 0.87 and F1 score of 0.74, outperforming the model trained with MRI data only and the RF-based model.

Akbari et al. (130) employed TL with a CNN pretrained on ImageNet and feature extraction based on four structural MRIs (i.e., T1, T1-ce, T2, FLAIR), diffusion tensor imaging (DTI), and PWI (rCBV, peak height (PH), percentage signal recovery (PSR)) and achieved an accuracy of 0.84 and AUC of 0.83. Ismail et al. (131) extracted 30 global and local shape features from T1-ce, T2, and FLAIR images and used an SVM classifier to achieve an accuracy of 0.90 in distinguishing PsP from true tumor progression.

In addition to the ambiguity between PsP and true tumor progression, immunotherapies in GBM also suffer from the lack of reliable evaluation methods on the radiological imaging manifestation regarding the alteration of the tumor immune microenvironment (TME, e.g., tumor immune cell infiltration, functional characterization of immune effector/suppressive cells, gene expression profile of immunostimulatory/immunosuppressive cells), a crucial parameter for assessment of intratumoral immune responses (5). In their pioneering work (132), Narang et al. utilized T1-weighted post-contrast and T2-FLAIR images in combination with T-cell surface marker CD3D/E/G mRNA expression data from 78 GBM patient-derived TCGA data to extract six imaging features that are associated with intra-GBM CD3 activity. These imaging features were further trained and tested using an internal dataset from 69 GBM patients that has immunohistochemically (IHC) validated intratumoral CD3 counts. The image-based intra-GBM CD3+ T-cell infiltration model reaches an accuracy of 97.1% and AUC of 0.993 for the training set, with an accuracy of 76.5% and AUC of 0.847 in the test group. A similar study has been reported recently in lower-grade gliomas (LGG) with an expansion from CD3 expression data to multiple immune gene expression profiles, including major histocompatibility complex (MHC)-related molecules, immune checkpoint molecules, and effector/suppressor immune cells (94). In this study, radiomic features extracted by a deep learning neural network-based model have been demonstrated to predict the TME-associated signature immunophenotype mRNAs with an AUC of 0.821 in the test group. Unfortunately, there is no IHC validation on expression of signature immune genes in the test group specimens.

To date, although numerous ML models for differentiating PsP from true tumor progression have been proposed and tested, none of them have been prospectively validated, reflecting the lack of confidence in clinicians to apply these radiomic approaches in their clinic practice. Multiple factors can lead to this significant issue, such as difficulty to applying small sample size-derived prediction models to a large population cohort, poor reproducibility, and lack of consistency between various ML models and/or datasets (further discussed in Section 4). One of the important and challenging factors is that currently there is no clear objective histological definition of pseudoprogression. In a representative study by Melguizo-Gavilanes et al. (133), MRI images and surgical resection-derived histological data from 34 patients with GBM were retrospectively reviewed. Only one-third of the cohort (11/34) demonstrated a concordance for PsP between radiological interpretation and histological diagnosis, whereas the majority of the patients had a histologically “mixed” pattern with tumor and treatment effect, indicating that even histology might not be applied as a gold standard to differentiate PsP and tumor true progression.

Overall survival prediction

Overall survival prediction of GBM patients provides useful information for surgical and treatment planning. Conventional survival prediction based on clinical information is subjective and could be inaccurate. Radiomic analysis, on the other hand, provides a variety of MRI features to predict disease prognosis, thus providing beneficial information for personalized treatment. Nevertheless, manual feature engineering is still time consuming, laborious, and subjective and may not be able to effectively encode other predictive but implicit information hidden in the multimodal neuroimages (134). Thus, an accurate, generalized yet automated OS prediction is desired.

Macyszyn et al. (135) extracted about 60 features from 105 GBM patients to train an SVM-based predictive model for patient survival and molecular subtype. The predictors were evaluated in 29 new patients and achieved a three-way (long/medium/short survival for longer than 18 months, between 6 and 18 months, and shorter than 6 months) accuracy of about 0.80. Another classifier was trained to discriminate among each of the various GBM molecular subtypes and achieved an accuracy of about 0.76. Sanghazni et al. (136) derived texture features (e.g., first-order texture features, GLCM), tumor shape and volumetric features, and patient ages from 173 patients’ multimodal MRI data (e.g., T1-ce, T2, and Flair) and used an SVM-RFE-based ML model to perform binary (i.e., short and long‘s threshold upon 400 days) and multiclass (i.e., <10, 10~15, and >15 months) OS prediction. Prediction accuracies of 0.987 and 0.89 were achieved for binary and multiclass predictions, respectively.

Choi et al. (137) collected 250 radiomic features extracted from 296 LGG cases from institutional and TCGA/TCIA datasets. They trained three random survival forest (RSF, i.e., a variant of RF) models with 1) these radiomic features; 2) non-imaging prognostic factors including age, resection extent, WHO grade, and IDH status; and 3) combination of 1 and 2 on the institutional dataset and validation of the model on the TCGA/TCIA dataset. When applying radiomic features or non-imaging features alone, the two RSF models achieved an AUC of 0.620 and 0.627, respectively. When applying radiomic features together with non-imaging prognostic parameters, the AUC was improved to 0.709. Similarly, in a GBM hypoxia-associated radiomic study, Beig et al. (138) also revealed that when combining clinical features (age, gender, and Karnofsky Performance Score (KPS)) with 270 radiomic features, the concordance index for survival prediction rises to 0.83 in comparison to 0.74 when using radiomic features alone (138). Grist et al. (139) examined various analysis techniques on survival predictions through perfusion and MRI data, especially DWI, collected from 69 pediatric patients. Approaches included conventional regressions and Bayesian analysis on apparent diffusion coefficient (ADC) maps, uncorrected and corrected cerebral blood volume (uCBV and cCBV) maps, and K2 maps (140) and achieved an AUC between 0.63 and 0.82. Supervised (i.e., SVM, RF, and a single-layer neural network) and unsupervised (i.e., k-means clustering) ML analyses achieved an accuracy between 0.90 and 0.98 in distinguishing between high- and low-risk clusters, with distinct differences in survival. In addition to the above models, the Tiwari group has developed a radiomic risk score in which the extracted GBM radiomic features were trained by various Cox regression-based algorithms for survival stratification with an overall concordance index at 0.7 to 0.8 (141143).

Nie et al. (134) proposed a two-stage learning-based method to predict the OS of HGG patients. Specifically, in the first stage, they adopted a CNN to extract implicit features from multiparametric maps that are computed by multimodal multichannel MRI (i.e., T1-ce, DTI, and rs-fMRI) from 68 HGG patients. Then, those radiomic features along with the demographic and tumor-related features (e.g., age, tumor size, and histological type) were trained in an SVM to model OS prediction (i.e., long or short overall survival time, with a threshold of 650 days). The experimental results demonstrated an accuracy of up to 0.91.

Identifying biomarkers of brain tumors

Radiogenomics uses radiomics techniques to predict the genetic makeup of tumors. This promotes precision medicine by identifying patients with tumor molecular markers that can be targeted by particular drugs and by predicting how aggressive a tumor will behave, with implications for survival and treatment choice. Via exploring the implicit correlation between radiological images and genomic data such as DNA microarrays, microRNA, RNA-Seq, ML techniques can help improve the effectiveness and efficiency in identifying the biomarkers of brain tumors (144).

Isocitrate dehydrogenase (IDH) mutation

Since the initial reworking of the WHO CNS Tumor Classification System in 2016, genetic biomarkers have become increasingly important in the classification of brain tumors. Isocitrate dehydrogenase is an enzyme in the Krebs cycle, and its mutated gene (IDH) is an oncogene. The mutant IDH enzyme produces an oncometabolite 2-hydroxyglutarate (2HG) (145), which promotes the growth of various cancers throughout the body. In brain tumors, IDH-mutated tumors are less aggressive than IDH wild-type tumors, yet they can convert to the latter. In the 2021 WHO CNS Tumor Classification System, only IDH wild-type tumors are classified as GBMs. It is of utmost importance for therapeutic planning to differentiate between the IDH mutation and IDH wild type, and it would greatly benefit patients if this determination could be done non-invasively and obviate biopsy or resection. Yogananda et al. (33) developed a 3D Dense-UNet network using (a) T2 images only (T2-net) and (b) a combination of T1-ce, T2, and FLAIR images (TS-net) from TCIA and TCGA to non-invasively predict IDH mutation. The T2-net demonstrated a mean cross-validation accuracy of 0.97 (sensitivity 0.97, specificity 0.98, AUC 0.98), and TS-net demonstrated a mean cross-validation accuracy of 0.97 (sensitivity 0.98, specificity 0.97, AUC 0.99). In addition, this model automatically segmented the tumor to show areas with either IDH mutation or IDH wild type. Dice scores were 0.85 for T2-net and 0.89 for TS-net. The benefit of being able to use only T2-weighted images is that gadolinium-based contrast material, which deposits in the brain to unknown effect, does not have to be administered and T2-weighted images can be quickly acquired and are less sensitive to motion artifact.

MGMT promoter methylation

MGMT promoter methylation predicts less aggressive glioma behavior for both IDH-mutated and IDH-wild-type gliomas. When its promoter is methylated, the MGMT gene, which is involved in DNA repair, is hindered and the tumor has greater difficulty overcoming the damage caused by chemotherapy such as temozolomide. Yogananda et al. (146) used a 3D-dense UNet on only T2 images to simultaneously segment the tumor and predict the presence of MGMT promoter methylation with a mean three-fold cross-validation accuracy of 0.95 (sensitivity 0.96, specificity 0.92, AUC 0.93, Dice score 0.82).

H3K27M alterations

In 2016, the WHO released a new histological diagnosis in the classification of CNS malignancies: diffuse midline glioma (DMG), H3K27M-mutant. It was renamed as H3K27M-altered in 2021 because there are multiple mechanisms involved. These WHO grade 4 tumors are found in or near the midline in the brainstem, thalamus, spinal cord, pineal region, hypothalamus, and cerebellum and exhibit aggressive clinical behavior (147, 148). H3K27M is the most frequent mutation in brainstem gliomas (BSGs) (149). Su et al. (150) extracted radiomics features from FLAIR images from 40 patients with H3K27M mutations and 60 wild-type patients, all with midline gliomas. The Tree-based Pipeline Optimization Tool (TPOT) was applied to optimize the ML pipeline and select important radiomics features. A total of 10 independent TPOT ML models were compared and tested on 22 independent cohorts of patients, achieving an accuracy ranging from 0.6 to 0.84, and the AUC from 0.73 to 0.90. Pan et al. (149) included a total of 151 patients with newly diagnosed BSGs. A total of 1,697 features, including six clinical parameters and 1,691 imaging features (e.g., GLCM, LBP), were extracted from pre- and post-contrast T1 and T2 images. Spearman’s correlation and relief algorithm were applied for feature selection. Thirty-six MRI features and three clinical features remained and were fed to an RF model to predict H3K27M mutations. For comparison, a least-square estimation method-based ML model was developed by utilization of the KPS at diagnosis, symptom duration at diagnosis, and edge sharpness on T2, which achieved an accuracy of 0.80 and AUC of 0.79 in the test cohort if using MRI features alone but can be improved to 0.84 and an AUC of 0.83 if integrated with clinical parameters. The simplified model achieved an AUC of 0.78. Zhuo et al. (151) studied 81 BSG patients with APT imaging at 3T MR and known H3K27M status. APTw values (i.e., mean, median, and max) and radiomic features within manually delineated 3D tumor masks were extracted. H3K27M-mutant prediction using APTw-derived radiomics was conducted using various models, such as SVM, AdaBoost, autoencoder, LASSO regression, and RF, which achieved an accuracy of 0.86 and an AUC of 0.93 as validated by a prospective cohort of 29 BSG patients.

Discussion

Despite that numerous ML studies have been conducted in GBM radiomic analysis, comparing the results from individual articles is not a trivial task due to the use of different data sets. The accuracies, AUCs, and Dice scores in different studies may vary from 0.7 to 0.98: most state-of-the-art studies using public datasets (e.g., BraTS) achieve an accuracy of 0.84–0.94, but some studies with certain private data can reach 0.98. Meanwhile, current major public datasets also lack sub-categories for brain tumor classification and segmentations, which restricts the development of a more powerful and comprehensive ML-model to distinguish more brain tumor types. Without sufficiently large datasets, ML models with too many parameters (i.e., neurons in each layer and the number of layers) are easily overfitting to a specific dataset, losing the generalizability of the model to other patient groups.

Challenges and perspectives on future AI/ML techniques

Overview of current challenges in ML-based radiomic neuro-oncology studies

As ML is a data-driven statistical approach to extract common features within different data samples, sufficient imaging datasets are required to train advanced ML models and to fairly evaluate their performances (e.g., accuracy, Dice score, AUC) in the field of neuro-oncology. Currently, only a limited number of brain tumor sub-categories have been analyzed with ML studies while many other brain tumor/disease types have not, due to the lack of labeled/annotated data for training. Examples include differentiating dysembryoplastic neuroepithelial tumor (DNET), ganglioglioma, pleomorphic xanthoastrocytoma (PXA), and multinodular and vacuolating neuronal tumor (MVNT).

However, establishing standardized radiological imaging datasets or standardizing McMv datasets for extensive and generalized ML-based GBM analysis can be manpower and time consuming because most of these datasets require highly accurate manual labeling/annotation to serve as the “ground truth” for the ML model training and validation. In addition, these datasets should be generalizable for various neuro-oncology analyses and patient groups and should be carefully labeled by various disease categories. Additional information (e.g., survival time, related biotest results, related clinical/medical history) may also be necessary for more sophisticated and comprehensive analyses. This requires a continuous update of the datasets, leading to a significant cost of data management (Figure 1B).

Another challenge is that current mathematic mechanisms in the ML model are based on statistics, which means there may not be a “deterministic optimal” algorithm or architecture for an ML model to achieve the “ideal/optimal” outcomes. The initial values of the trainable parameters in ML models and the slight differences in structure may affect the training outcome significantly. Even when using public datasets (e.g., BraTS), similar ML networks may yet achieve varying results (152, 153). Thus, many researchers intend to simply add more layers in CNNs to improve the accuracy, potentially causing extensive yet unnecessary computational complexity during the training process but overlooking the biological connections and meaning behind those data. On the other hand, too much engineering (i.e., strong feature extraction, data restriction/collection) in data preprocessing may also lead to overfitting of the ML network to the training data and lose the generalizability of trained ML models for larger populations with more diversity (Figure 1B).

Promising strategies enhancing performance of AI models in GBM radiomic analysis

Aside from using genuine radiological brain tumor images alone to train ML models, three other trends are gaining popularity to improve the model performance in accuracy, Dice score, AUC, and generalizability. The first trend is to use TL (100, 101), which takes advantage of other larger non-neuro-oncology or even non-medical image datasets to pretrain the ML model. Then by keeping the pretrained parameters in the low-level hidden layers (i.e., closer to the input layer) and fine-tuning the ones in the high-level layers and output layers with brain tumor training image datasets, the pretrained ML model can be adopted for brain tumor analysis. Typical image datasets for pretraining ML models includes ImageNet (102), the modified National Institute of Standards and Technology (MNIST) database (154), and International Symposium on Biomedical Imaging (ISBI) (155). However, if the pretraining dataset is drastically unsimilar to the target dataset, the pretraining effect is limited. Therefore, a standardized radiomic medical dataset with various categories is preferred, benefitting not only neuro-oncology studies but also other medical and biomedical studies.

The second trend is to use GANs to generate synthetic data for augmentation (78, 79, 155). However, as discussed in Section 2.4, this approach itself requires a large set of genuine images to train the discriminator network in the GAN, before it can synthesize accurate-appearing brain tumor images to train other ML models for brain tumor analysis.

The third trend is to fuse multimodal data for a more comprehensive analysis. Examples include multimodal MRI (156, 157), combinations of MRI and PET (54), image genomics (i.e., radiogenomics), and clinical data to study the association between imaging biomarkers and genomic characteristics (144, 149151, 158, 159). Especially for radiogenomics, some studies (144, 158, 159) have identified associations between quantitative image features and gene expression profiles of glioblastoma (e.g., H3K27M, TP53, EGFR, NF1, and IDH1) and its molecular subtypes (e.g., classical, mesenchymal, proneural, and neural). Additional studies indicate that quantitative MR imaging features derived from entire tumor volumes can be used to identify glioblastoma subtypes with distinct molecular pathways (160, 161). With the help of additional complementary correlated features from different types of radiomic images and/or genomic information, or simply just the medical history of the patients, ML can take advantage of data to achieve more accurate predictions (Figure 1B).

Outlook on teamwork among computer scientists/engineers, physicians, and biomedical researchers

As aforementioned, high-quality clinical data and labels/annotations are critical to ML algorithms for both accuracy and generalizability, and biological knowledge can help extract certain features to improve the accuracy as well as the training efficiency. Therefore, strong collaborations should be established among computer scientists, engineers, physicians, and biomedical researchers to facilitate the standardization and enrichment of neuro-oncology radiomic datasets and the development of innovative and more advanced AI/ML models (Figure 1B). In addition, with larger amounts of data to track patients’ treatment process and the outcomes, it is even possible to develop ML/AI techniques to determine more suitable plans for their treatment, to improve the patients’ survival time as well as their quality of life.

Conclusion

With the urgent needs for highly accurate and automatic analysis of brain tumors and the rapid growth of clinical imaging data, image-based ML/AI techniques are playing an increasingly important role. Various combinations of feature extraction algorithms and ML models have been implemented and have achieved comparable or even better performance than manual analysis. However, challenges remain for exploring cancer heterogeneity, higher prediction accuracy, and generalizability for larger, more diverse patient groups. We believe that, by improving dataset quality, employing multimodal data fusion, developing more advanced ML models, and further enhancing collaborations between computer scientists, engineers, physicians, and biomedical researchers, AI techniques will accelerate quantitative cancer imaging analysis for clinical applications with great improvements in patient care.

Author contributions

MZ, YK, VH, LZ, and SZ formulated the conception of this review. MZ, SL, and SZ performed initial screening of the citations, datasets, and related AI/ML techniques. MZ, SL, LZ, and VH wrote and reviewed the manuscript and prepared the figures. AH provided critical comments for the manuscript. All authors approved the manuscript for publication and declare no competing financial interest.

Funding

This study was supported by the Nevada Governor’s Office of Economic Development (GOED) AWD-02-00001154 to SZ.

Acknowledgments

Thanks for UNLV University Libraries Open Article Fund to support this open-access journal article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

  • 1. Chang K, Zhang B, Guo X, Zong M, Rahman R, Sanchez D, et al. Multimodal imaging patterns predict survival in recurrent glioblastoma patients treated with bevacizumab. Neuro-Oncology (2016) 18:1680–7. doi:  10.1093/neuonc/now086 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. About GBM - GBM awareness day | national brain tumor society . Available at: https://braintumor.org/take-action/about-gbm/ (Accessed April 3, 2022).
  • 3. Hanif F, Muzaffar K, Perveen K, Malhi SM, Simjee SU. Glioblastoma multiforme: A review of its epidemiology and pathogenesis through clinical presentation and treatment. Asian Pac J Cancer Prev (2017) 18:3. doi:  10.22034/APJCP.2017.18.1.3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Ostrom QT, Patil N, Cioffi G, Waite K, Kruchko C, Barnholtz-Sloan JS. CBTRUS statistical report: Primary brain and other central nervous system tumors diagnosed in the united states in 2013–2017. Neuro-Oncology (2020) 22:iv1–iv96. doi:  10.1093/neuonc/noaa200 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Bagley SJ, Desai AS, Nasrallah MP, O’Rourke DM. Immunotherapy and response assessment in malignant glioma: Neuro-oncology perspective. Topics Magnet Resonance Imaging (2020) 29:95–102. doi:  10.1097/RMR.0000000000000233 [DOI] [PubMed] [Google Scholar]
  • 6. Gillies RJ, Kinahan PE, Hricak H. Radiomics: Images are more than pictures, they are data. Radiology (2016) 278:563–77. doi:  10.1148/radiol.2015151169 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Kumar V, Gu Y, Basu S, Berglund A, Eschrich SA, Schabath MB, et al. Radiomics: the process and the challenges. Magnet Resonance Imaging (2012) 30:1234–48. doi:  10.1016/j.mri.2012.06.010 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Samuel AL. “Some studies in machine learning using the game of checkers. II–recent progress.,”. In: Levy DNL, editor. Computer games I. New York, NY: Springer; (1988). p. 366–400. doi:  10.1007/978-1-4613-8716-9_15 [DOI] [Google Scholar]
  • 9. Caruana R, Niculescu-Mizil A. An empirical comparison of supervised learning algorithms. proceedings of the 23rd international conference on machine learning. ICML ‘06. New York, NY, USA: Association for Computing Machinery; (2006) p. 161–8. doi:  10.1145/1143844.1143865 [DOI] [Google Scholar]
  • 10. Barlow HB. Unsupervised learning. Neural Comput (1989) 1:295–311. doi:  10.1162/neco.1989.1.3.295 [DOI] [Google Scholar]
  • 11. Kaelbling LP, Littman ML, Moore AW. Reinforcement learning: A survey. J Artif Intell Res (1996) 4:237–85. doi:  10.1613/jair.301 [DOI] [Google Scholar]
  • 12. Rajpurkar P, Chen E, Banerjee O, Topol EJ. AI In health and medicine. Nat Med (2022) 28:31–8. doi:  10.1038/s41591-021-01614-0 [DOI] [PubMed] [Google Scholar]
  • 13. Voulodimos A, Doulamis N, Doulamis A, Protopapadakis E. Deep learning for computer vision: A brief review. Comput Intell Neurosci (2018) 2018:e7068349. doi:  10.1155/2018/7068349 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. Larrañaga P, Calvo B, Santana R, Bielza C, Galdiano J, Inza I, et al. Machine learning in bioinformatics. Briefings Bioinf (2006) 7:86–112. doi:  10.1093/bib/bbk007 [DOI] [PubMed] [Google Scholar]
  • 15. Perlich C, Dalessandro B, Raeder T, Stitelman O, Provost F. Machine learning for targeted display advertising: transfer learning in action. Mach Learn (2014) 95:103–27. doi:  10.1007/s10994-013-5375-2 [DOI] [Google Scholar]
  • 16. Brownlee J. Deep learning for natural language processing: develop deep learning models for your natural language problems. Mach Learn Mastery (2017). [Google Scholar]
  • 17. Baid U, Ghodasara S, Mohan S, Bilello M, Calabrese E, Colak E, et al. The RSNA-ASNR-MICCAI BraTS 2021 benchmark on brain tumor segmentation and radiogenomic classification (2021). Available at: http://arxiv.org/abs/2107.02314 (Accessed March 14, 2022).
  • 18. Kotsiantis SB, Kanellopoulos D, Pintelas PE. Data preprocessing for supervised leaning. Int J Comput Sci (2006) 1:111–7. doi: 10.5281/zenodo.1082415 [DOI] [Google Scholar]
  • 19. Alpaydin E. Introduction to machine learning. 4th ed. Cambridge, MA, USA: MIT Press; (2020). 712 p. [Google Scholar]
  • 20. Ruan S, Lebonvallet S, Merabet A, Constans J. Tumor segmentation from a multispectral MRI images by using support vector machine classification, in: 2007 4th IEEE International Symposium on Biomedical Imaging: From Nano to Macro, Washington DC: IEEE; (2007). pp. 1236–9. doi:  10.1109/ISBI.2007.357082 [DOI] [Google Scholar]
  • 21. Li H, Song M, Fan Y. Segmentation of brain tumors in multi-parametric MR images via robust statistic information propagation. In: Kimmel R, Klette R, Sugimoto A, editors. Computer vision – ACCV 2010. lecture notes in computer science. Berlin, Heidelberg: Springer; (2011). p. 606–17. doi:  10.1007/978-3-642-19282-1_48 [DOI] [Google Scholar]
  • 22. Li H, Fan Y. Label propagation with robust initialization for brain tumor segmentation, in: 2012 9th IEEE International Symposium on Biomedical Imaging (ISBI), Barcelona, Spain: IEEE. (2012) pp. 1715–8. doi:  10.1109/ISBI.2012.6235910 [DOI] [Google Scholar]
  • 23. Reza S, Iftekharuddin KM. Improved brain tumor tissue segmentation using texture features. In: MICCAI BraTS (Brain tumor segnmentation challenge). Boston, MA; (2014). p. 27–30. [Google Scholar]
  • 24. Götz M, Weber C, Blöcher J, Stieltjes B, Meiner H-P, Maier-Hein KH. Extremely randomized trees based brain tumor segmentation. In: MICCAI BraTS challenge. Boston, MA, USA: Zenodo; (2014). p. 6–11. doi:  10.5281/zenodo.3666272 [DOI] [Google Scholar]
  • 25. Kleesiek J, Biller A, Urban G, Hamprecht FA. Ilastik for multi-modal brain tumor segmentation. In: MICCAI BraTS (Brain tumor segmentation challenge). Boston, MA, USA:CiteSeerx; (2014). p. 12–7. [Google Scholar]
  • 26. Meier R, Bauer S, Slotboom J, Wiest R, Reyes M. Appearance-and context-sensitive features for brain tumor segmentation. In: MICCAI BraTS challenge. Boston, MA, USA: (2014). p. 20–6. doi:  10.13140/2.1.3766.7846 [DOI] [Google Scholar]
  • 27. Girshick R, Donahue J, Darrell T, Malik J. Rich feature hierarchies for accurate object detection and semantic segmentation, in: 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH: IEEE. (2014) pp. 580–7. doi:  10.1109/CVPR.2014.81 [DOI] [Google Scholar]
  • 28. Zhao Z-Q, Zheng P, Xu S-T, Wu X. Object detection with deep learning: A review. IEEE Trans Neural Networks Learn Syst (2019) 30:3212–32. doi:  10.1109/TNNLS.2018.2876865 [DOI] [PubMed] [Google Scholar]
  • 29. Long J, Shelhamer E, Darrell T. Fully convolutional networks for semantic segmentation. IEEE Comput Soc (2015) 3431–40. doi:  10.1109/CVPR.2015.7298965 [DOI] [PubMed] [Google Scholar]
  • 30. Zheng S, Jayasumana S, Romera-Paredes B, Vineet V, Su Z, Du D, et al. Conditional random fields as recurrent neural networks, in: 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile: IEEE. (2015) pp. 1529–37. doi:  10.1109/ICCV.2015.179 [DOI] [Google Scholar]
  • 31. Liu Z, Li X, Luo P, Loy C-C, Tang X. Semantic image segmentation via deep parsing network, in: 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile: IEEE; (2015). pp. 1377–85. doi:  10.1109/ICCV.2015.162 [DOI] [Google Scholar]
  • 32. Zhao X, Wu Y, Song G, Li Z, Zhang Y, Fan Y. A deep learning model integrating FCNNs and CRFs for brain tumor segmentation. Med Image Anal (2018) 43:98–111. doi:  10.1016/j.media.2017.10.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33. Bangalore Yogananda CG, Shah BR, Vejdani-Jahromi M, Nalawade SS, Murugesan GK, Yu FF, et al. A novel fully automated MRI-based deep-learning method for classification of IDH mutation status in brain gliomas. Neuro-Oncology (2020) 22:402–11. doi:  10.1093/neuonc/noz199 [DOI] [PMC free article] [PubMed] [Google Scholar] [Retracted]
  • 34. Menze BH, Jakab A, Bauer S, Kalpathy-Cramer J, Farahani K, Kirby J, et al. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans Med Imaging (2015) 34:1993–2024. doi:  10.1109/TMI.2014.2377694 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35. Clementi G, Salvati D, Scorzelli G, Paoluzzi A, Pascucci V. Progressive extraction of neural models from high-resolution 3D images of brain, in: Proceedings of CAD’16. CAD Solutions LLC, Vancouver, Canada: CAD Solutions, LLC; (2016). pp. 348–51. doi:  10.14733/cadconfP.2016.348-351 [DOI] [Google Scholar]
  • 36. Talo M, Baloglu UB, Yıldırım Ö, Rajendra Acharya U. Application of deep transfer learning for automated brain abnormality classification using MR images. Cogn Syst Res (2019) 54:176–88. doi:  10.1016/j.cogsys.2018.12.007 [DOI] [Google Scholar]
  • 37. Emblem KE, Zoellner FG, Tennoe B, Nedregaard B, Nome T, Due-Tonnessen P, et al. Predictive modeling in glioma grading from MR perfusion images using support vector machines. Magnet Resonance Med (2008) 60:945–52. doi:  10.1002/mrm.21736 [DOI] [PubMed] [Google Scholar]
  • 38. Sessions V, Valtorta M. The effects of data quality on machine learning algorithms. ICIQ (2006) 485–98. [Google Scholar]
  • 39. Tests for brain and spinal cord tumors in adults . Available at: https://www.cancer.org/cancer/brain-spinal-cord-tumors-adults/detection-diagnosis-staging/how-diagnosed.html (Accessed February 8, 2022).
  • 40. Thust SC, Yousry T. Imaging of skull base tumours. Rep Pract Oncol Radiother (2016) 21:304. doi:  10.1016/j.rpor.2015.12.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41. Amin J, Sharif M, Raza M, Saba T, Anjum MA. Brain tumor detection using statistical and machine learning method. Comput Methods Programs Biomed (2019) 177:69–79. doi:  10.1016/j.cmpb.2019.05.015 [DOI] [PubMed] [Google Scholar]
  • 42. Tandel GS, Biswas M, Kakde OG, Tiwari A, Suri HS, Turk M, et al. A review on a deep learning perspective in brain cancer classification. Cancers (2019) 11:111. doi:  10.3390/cancers11010111 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Hou DJ, Tong KA, Ashwal S, Joo E, Shutter L, Obenaus A. 478 diffusion-weighted imaging improves outcome prediction in adults with diffuse axonal injury. J Invest Med (2006) 54:S161–1. doi:  10.2310/6650.2005.X0004.477 [DOI] [Google Scholar]
  • 44. Landheer K, Schulte RF, Treacy MS, Swanberg KM, Juchem C. Theoretical description of modern 1H in vivo magnetic resonance spectroscopic pulse sequences. J Magnet Resonance Imaging (2020) 51:1008–29. doi:  10.1002/jmri.26846 [DOI] [PubMed] [Google Scholar]
  • 45. Golder W. Magnetic resonance spectroscopy in clinical oncology. ORT (2004) 27:304–9. doi:  10.1159/000077983 [DOI] [PubMed] [Google Scholar]
  • 46. Rosen Y, Lenkinski RE. Recent advances in magnetic resonance neurospectroscopy. Neurotherapeutics (2007) 4:330–45. doi:  10.1016/j.nurt.2007.04.009 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47. Chen F, Ni Y-C. Magnetic resonance diffusion-perfusion mismatch in acute ischemic stroke: An update. World J Radiol (2012) 4:63–74. doi:  10.4329/wjr.v4.i3.63 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48. Yacoub E, Harel N, Uğurbil K. High-field fMRI unveils orientation columns in humans. PNAS (2008) 105:10607–12. doi:  10.1073/pnas.0804110105 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49. Zhou J, Blakeley JO, Hua J, Kim M, Laterra J, Pomper MG, et al. Practical data acquisition method for human brain tumor amide proton transfer (APT) imaging. Magnet Resonance Med (2008) 60:842–9. doi:  10.1002/mrm.21712 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50. Sartoretti E, Sartoretti T, Wyss M, Reischauer C, van Smoorenburg L, Binkert CA, et al. Amide proton transfer weighted (APTw) imaging based radiomics allows for the differentiation of gliomas from metastases. Sci Rep (2021) 11:5506. doi:  10.1038/s41598-021-85168-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51. Harã T, Kosaka N, Shinoura N, Kondo T. PET imaging of brain tumor with [methyl-11C] choline. J Nucl Med (1997) 38:842–7. [PubMed] [Google Scholar]
  • 52. Langleben DD, Segall GM. PET in differentiation of recurrent brain tumor from radiation injury*. J Nucl Med (2000) 41:1861–7. [PubMed] [Google Scholar]
  • 53. Chao ST, Suh JH, Raja S, Lee S-Y, Barnett G. The sensitivity and specificity of FDG PET in distinguishing recurrent brain tumor from radionecrosis in patients treated with stereotactic radiosurgery. Int J Cancer (2001) 96:191–7. doi:  10.1002/ijc.1016 [DOI] [PubMed] [Google Scholar]
  • 54. Galldiks N, Law I, Pope WB, Arbizu J, Langen K-J. The use of amino acid PET and conventional MRI for monitoring of brain tumor therapy. NeuroImage: Clin (2017) 13:386–94. doi:  10.1016/j.nicl.2016.12.020 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55. de Causans A, Carré A, Roux A, Tauziède-Espariat A, Ammari S, Dezamis E, et al. Development of a machine learning classifier based on radiomic features extracted from post-contrast 3D T1-weighted MR images to distinguish glioblastoma from solitary brain metastasis. Front Oncol (2021) 11:638262. doi:  10.3389/fonc.2021.638262 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56. Hellier P. Consistent intensity correction of MR images, in: Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429), Catalonia, Spain: IEEE; (2003). pp. I–1109. doi:  10.1109/ICIP.2003.1247161 [DOI] [Google Scholar]
  • 57. Shah M, Xiao Y, Subbanna N, Francis S, Arnold DL, Collins DL, et al. Evaluating intensity normalization on MRIs of human brain with multiple sclerosis. Med Image Anal (2011) 15:267–82. doi:  10.1016/j.media.2010.12.003 [DOI] [PubMed] [Google Scholar]
  • 58. Sweeney EM, Shinohara RT, Shea CD, Reich DS, Crainiceanu CM. Automatic lesion incidence estimation and detection in multiple sclerosis using multisequence longitudinal MRI. AJNR Am J Neuroradiol (2013) 34:68–73. doi:  10.3174/ajnr.A3172 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59. Collewet G, Strzelecki M, Mariette F. Influence of MRI acquisition protocols and image intensity normalization methods on texture classification. Magnet Resonance Imaging (2004) 22:81–91. doi:  10.1016/j.mri.2003.09.001 [DOI] [PubMed] [Google Scholar]
  • 60. Shinohara RT, Sweeney EM, Goldsmith J, Shiee N, Mateen FJ, Calabresi PA, et al. Statistical normalization techniques for magnetic resonance imaging. NeuroImage: Clin (2014) 6:9–19. doi:  10.1016/j.nicl.2014.08.008 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61. Nyúl LG, Udupa JK. On standardizing the MR image intensity scale. Magnet Resonance Med (1999) 42:1072–81. doi:  [DOI] [PubMed] [Google Scholar]
  • 62. Sun X, Shi L, Luo Y, Yang W, Li H, Liang P, et al. Histogram-based normalization technique on human brain magnetic resonance images from different acquisitions. Biomed Eng Online (2015) 14:73. doi:  10.1186/s12938-015-0064-y [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63. Yushkevich PA, Piven J, Hazlett HC, Smith RG, Ho S, Gee JC, et al. User-guided 3D active contour segmentation of anatomical structures: Significantly improved efficiency and reliability. NeuroImage (2006) 31:1116–28. doi:  10.1016/j.neuroimage.2006.01.015 [DOI] [PubMed] [Google Scholar]
  • 64. 3D slicer image computing platform, in: 3D slicer. Available at: https://slicer.org/ (Accessed April 11, 2022).
  • 65. Ayache N, Delingette H, Golland P, Mori K. eds. Medical image computing and computer-assisted intervention – MICCAI 2012: 15th international conference, nice, France, October 1-5, 2012, proceedings. Berlin, Heidelberg: Springer; (2012). doi:  10.1007/978-3-642-33454-2 [DOI] [Google Scholar]
  • 66. Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby JS, et al. Advancing the cancer genome atlas glioma MRI collections with expert segmentation labels and radiomic features. Sci Data (2017) 4:170117. doi:  10.1038/sdata.2017.117 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 67. Bakas S, Reyes M, Jakab A, Bauer S, Rempfler M, Crimi A, et al. Identifying the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival prediction in the BRATS challenge. (2019). arXiv:181102629. doi: 10.48550/arXiv.1811.02629 [DOI] [Google Scholar]
  • 68. Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby J, et al. Segmentation labels and radiomic features for the pre-operative scans of the TCGA-GBM collection. (2017). doi:  10.7937/K9/TCIA.2017.KLXWJJ1Q [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 69. Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby J, et al. Segmentation labels and radiomic features for the pre-operative scans of the TCGA-LGG collection. (2017). doi:  10.7937/K9/TCIA.2017.GJQ7R0EF [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70. Clark K, Vendt B, Smith K, Freymann J, Kirby J, Koppel P, et al. The cancer imaging archive (TCIA): Maintaining and operating a public information repository. J Digit Imaging (2013) 26:1045–57. doi:  10.1007/s10278-013-9622-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 71. Johnson KA, Becker JA. Harvard Medical school – the whole brain atlas . Available at: http://www.med.harvard.edu/aanlib/home.html.
  • 72. Puchalski RB, Shah N, Miller J, Dalley R, Nomura SR, Yoon J-G, et al. An anatomic transcriptional atlas of human glioblastoma. Science (2018) 360(6389):660–3. doi:  10.1126/science.aaf2666 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 73. Scarpace L, Mikkelsen T, Cha S, Rao S, Tekchandani S, Gutman T, et al. Radiology data from the cancer genome atlas glioblastoma multiforme [TCGA-GBM] collection. (2016). doi:  10.7937/K9/TCIA.2016.RNYFUYE9 [DOI] [Google Scholar]
  • 74. Shusharina N, Bortfeld T, Cardenas C, De B, Diao K, Hernandez S, et al. Cross-modality brain structures image segmentation for the radiotherapy target definition and plan optimization. In: Shusharina N, Heinrich MP, Huang R, editors. Segmentation, classification, and registration of multi-modality medical imaging data. Cham: Springer International Publishing; (2021). p. 3–15. Lecture Notes in Computer Science. doi:  10.1007/978-3-030-71827-5_1 [DOI] [Google Scholar]
  • 75. National Cancer Institute Clinical Proteomic Tumor Analysis Consortium (CPTAC) . Radiology data from the clinical proteomic tumor analysis consortium glioblastoma multiforme [CPTAC-GBM] collection. (2018). doi:  10.7937/k9/tcia.2018.3rje41q1 [DOI] [Google Scholar]
  • 76. Shorten C, Khoshgoftaar TM. A survey on image data augmentation for deep learning. J Big Data (2019) 6:60. doi:  10.1186/s40537-019-0197-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77. Nalepa J, Marcinkiewicz M, Kawulok M. Data augmentation for brain-tumor segmentation: A review. Front Comput Neurosci (2019) 13:83. doi:  10.3389/fncom.2019.00083 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78. Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative adversarial nets. proceedings of the 27th international conference on neural information processing systems - volume 2. NIPS’14. Cambridge, MA, USA: MIT Press; (2014) p. 2672–80. [Google Scholar]
  • 79. Ghassemi N, Shoeibi A, Rouhani M. Deep neural network with generative adversarial networks pre-training for brain tumor classification based on MR images. Biomed Signal Process Control (2020) 57:101678. doi:  10.1016/j.bspc.2019.101678 [DOI] [Google Scholar]
  • 80. Pal KK, Sudeep KS. (2016). Preprocessing for image classification by convolutional neural networks, in: 2016 IEEE International Conference on Recent Trends in Electronics, Information Communication Technology (RTEICT), Bangalore, India. pp. 1778–81. doi:  10.1109/RTEICT.2016.7808140 [DOI] [Google Scholar]
  • 81. Huang J, Li Y-F, Xie M. An empirical analysis of data preprocessing for machine learning-based software cost estimation. Inf Softw Technol (2015) 67:108–27. doi:  10.1016/j.infsof.2015.07.004 [DOI] [Google Scholar]
  • 82. Zhang S, Zhang C, Yang Q. Data preparation for data mining. Appl Artif Intell (2003) 17:375–81. doi:  10.1080/713827180 [DOI] [Google Scholar]
  • 83. Alves AFF, Miranda JR de A, Reis F, de Souza SAS, Alves LLR, Feitoza L deM, et al. Inflammatory lesions and brain tumors: is it possible to differentiate them based on texture features in magnetic resonance imaging? J Venom Anim Toxins incl Trop Dis (2020) 26:1–10. doi:  10.1590/1678-9199-JVATITD-2020-0011 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 84. Mohsen H, El-Dahshan E-SA, El-Horbaty E-SM, Salem A-BM. Classification using deep learning neural networks for brain tumors. Future Computing Inf J (2018) 3:68–71. doi:  10.1016/j.fcij.2017.12.001 [DOI] [Google Scholar]
  • 85. Chen C, Ou X, Wang J, Guo W, Ma X. Radiomics-based machine learning in differentiation between glioblastoma and metastatic brain tumors. Front Oncol (2019) 9:806. doi:  10.3389/fonc.2019.00806 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86. Zacharaki EI, Wang S, Chawla S, Soo Yoo D, Wolf R, Melhem ER, et al. Classification of brain tumor type and grade using MRI texture and shape in a machine learning scheme. Magnet Resonance Med (2009) 62:1609–18. doi:  10.1002/mrm.22147 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87. Mitchell TM, Hutchinson R, Niculescu RS, Pereira F, Wang X, Just M, et al. Learning to decode cognitive states from brain images. Mach Learn (2004) 57:145–75. doi:  10.1023/B:MACH.0000035475.85309.1b [DOI] [Google Scholar]
  • 88. Cox DD, Savoy RL. Functional magnetic resonance imaging (fMRI) “brain reading”: detecting and classifying distributed patterns of fMRI activity in human visual cortex. NeuroImage (2003) 19:261–70. doi:  10.1016/S1053-8119(03)00049-1 [DOI] [PubMed] [Google Scholar]
  • 89. Peng H, Long F, Ding C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans Pattern Anal Mach Intell (2005) 27:1226–38. doi:  10.1109/TPAMI.2005.159 [DOI] [PubMed] [Google Scholar]
  • 90. Dash M, Liu H. Feature selection for classification. Intelligent Data Anal (1997) 1:131–56. doi:  10.1016/S1088-467X(97)00008-5 [DOI] [Google Scholar]
  • 91. Chen Z, Menzies T, Port D, Boehm B. Feature subset selection can improve software cost estimation accuracy. In: Proceedings of the 2005 workshop on predictor models in software engineering. PROMISE ‘05. New York, NY, USA: Association for Computing Machinery; (2005). p. 1–6. doi:  10.1145/1083165.1083171 [DOI] [Google Scholar]
  • 92. Das S. Filters, wrappers and a boosting-based hybrid for feature selection. In: Proceedings of the eighteenth international conference on machine learning. ICML ‘01. San Francisco, CA, USA: Morgan Kaufmann Publishers Inc; (2001). p. 74–81. [Google Scholar]
  • 93. Citak-Er F, Firat Z, Kovanlikaya I, Ture U, Ozturk-Isik E. Machine-learning in grading of gliomas based on multi-parametric magnetic resonance imaging at 3T. Comput Biol Med (2018) 99:154–60. doi:  10.1016/j.compbiomed.2018.06.009 [DOI] [PubMed] [Google Scholar]
  • 94. Li Z, Liu P, An T, Yang H, Zhang W, Wang J. Construction of a prognostic immune signature for lower grade glioma that can be recognized by MRI radiomics features to predict survival in LGG patients. Trans Oncol (2021) 14:101065. doi:  10.1016/j.tranon.2021.101065 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 95. Vaidhya K, Thirunavukkarasu S, Alex V, Krishnamurthi G. Multi-modal brain tumor segmentation using stacked denoising autoencoders. In: A C, Menze B, Maier O, Reyes M, Handels H, editors. In Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries (BrainLes 2015). Cham: Springer International Publishing; (2015). p. 181–94. doi:  10.1007/978-3-319-30858-6_16 [DOI] [Google Scholar]
  • 96. Agn M, Puonti O, Rosenschöld PM, Law I, Van Leemput K. Brain tumor segmentation using a generative model with an RBM prior on tumor shape. In: Crimi A, Menze B, Maier O, Reyes M, Handels H, editors. Cham: Springer International Publishing; (2016). p. 168–80. doi:  10.1007/978-3-319-30858-6_15 [DOI] [Google Scholar]
  • 97. Ari A, Hanbay D. Deep learning based brain tumor classification and detection system. Turkish J Electrical Eng Comput Sci (2018) 26:2275–86. doi: 10.3906/elk-1801-8 [DOI] [Google Scholar]
  • 98. Zhou Y, Li Z, Zhu H, Chen C, Gao M, Xu K, et al. “Holistic brain tumor screening and classification based on DenseNet and recurrent neural network.”. In: Crimi A, Bakas S, Kuijf H, Keyvan F, Reyes M, van Walsum T, editors. Brainlesion: Glioma, multiple sclerosis, stroke and traumatic brain injuries. Cham: Springer International Publishing; (2018). p. 208–17. doi:  10.1007/978-3-030-11723-8_21 [DOI] [Google Scholar]
  • 99. Cheng J, Huang W, Cao S, Yang R, Yang W, Yun Z, et al. Enhanced performance of brain tumor classification via tumor region augmentation and partition. PloS One (2015) 10:e0140381. doi:  10.1371/journal.pone.0140381 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 100. Banerjee S, Mitra S, Masulli F, Rovetta S. Brain tumor detection and classification from multi-sequence MRI: Study using ConvNets. In: Crimi A, Bakas S, Kuijf H, Keyvan F, Reyes M, van Walsum T, editors. International MICCAI brainlesion workshop: Brainlesion: Glioma, multiple sclerosis, stroke and traumatic brain injuries (BrainLes 2018). Cham: Springer International Publishing; (2018). p. 170–9. doi:  10.1007/978-3-030-11723-8_17 [DOI] [Google Scholar]
  • 101. Xu Y, Jia Z, Ai Y, Zhang F, Lai M, Chang EI-C. Deep convolutional activation features for large scale brain tumor histopathology image classification and segmentation, in: 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane, Australia: IEEE; (2015). pp. 947–51. doi:  10.1109/ICASSP.2015.7178109 [DOI] [Google Scholar]
  • 102. Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L. ImageNet: A large-scale hierarchical image database, in: 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL: IEEE; (2009). pp. 248–55. doi:  10.1109/CVPR.2009.5206848 [DOI] [Google Scholar]
  • 103. Patcheil RA. Metastatic brain tumors. Neurol Clinics (1995) 13:915–25. doi:  10.1016/S0733-8619(18)30025-2 [DOI] [PubMed] [Google Scholar]
  • 104. Kamson DO, Mittal S, Buth A, Muzik O, Kupsky WJ, Robinette NL, et al. Differentiation of glioblastomas from metastatic brain tumors by tryptophan uptake and kinetic analysis: a pet study with MRI comparison. Mol Imaging (2013) 12:327–37. doi: 10.2310/7290.2013.00048 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 105. Huberty CJ. Discriminant analysis. Rev Educ Res (1975) 45:543–98. doi:  10.3102/00346543045004543 [DOI] [Google Scholar]
  • 106. Priya S, Liu Y, Ward C, Le NH, Soni N, Pillenahalli Maheshwarappa R, et al. Machine learning based differentiation of glioblastoma from brain metastasis using MRI derived radiomics. Sci Rep (2021) 11:10478. doi:  10.1038/s41598-021-90032-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 107. Zou H, Hastie T. Regularization and variable selection via the elastic net. J R Stat Soc (2005) 67:301–20. doi:  10.1111/j.1467-9868.2005.00503.x [DOI] [Google Scholar]
  • 108. Zwanenburg A, Vallières M, Abdalah MA, Aerts HJWL, Andrearczyk V, Apte A, et al. The image biomarker standardization initiative: Standardized quantitative radiomics for high-throughput image-based phenotyping. Radiology (2020) 295:328–38. doi:  10.1148/radiol.2020191145 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109. Nguyen AV, Blears EE, Ross E, Lall RR, Ortega-Barnett J. Machine learning applications for the differentiation of primary central nervous system lymphoma from glioblastoma on imaging: a systematic review and meta-analysis. Neurosurg Focus (2018) 45(5):1–10. doi: 10.3171/2018.8.FOCUS18325 [DOI] [PubMed] [Google Scholar]
  • 110. Liu Y-H, Maftah M, Das T, Bai L, Robson K, Auer D. Classificatioo of MR tumor images based on gabor wavelet analysis. J Med Biol Eng (2012) 32:22–8. doi: 10.5405/jmbe.813 [DOI] [Google Scholar]
  • 111. Yamasaki T, Chen T, Hirai T, Murakami R. Classification of cerebral lymphomas and glioblastomas featuring luminance distribution analysis. Comput Math Methods Med (2013) 2013:e619658. doi:  10.1155/2013/619658 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 112. Yang Z, Feng P, Wen T, Wan M, Hong X. Differentiation of glioblastoma and lymphoma using feature extraction and support vector machine. CNS Neurol Disord - Drug Targets- CNS Neurol Disorders) (2017) 16:160–8. doi: 10.2174/1871527315666161018122909 [DOI] [PubMed] [Google Scholar]
  • 113. Alcaide-Leon P, Dufort P, Geraldo AF, Alshafai L, Maralani PJ, Spears J, et al. Differentiation of enhancing glioma and primary central nervous system lymphoma by texture-based machine learning. Am J Neuroradiol (2017) 38:1145–50. doi:  10.3174/ajnr.A5173 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 114. Kang D, Park JE, Kim Y-H, Kim JH, Oh JY, Kim J, et al. Diffusion radiomics as a diagnostic model for atypical manifestation of primary central nervous system lymphoma: development and multicenter external validation. Neuro-Oncology (2018) 20:1251–61. doi:  10.1093/neuonc/noy021 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 115. Chen Y, Li Z, Wu G, Yu J, Wang Y, Lv X, et al. Primary central nervous system lymphoma and glioblastoma differentiation based on conventional magnetic resonance imaging by high-throughput SIFT features. Int J Neurosci (2018) 128:608–18. doi:  10.1080/00207454.2017.1408613 [DOI] [PubMed] [Google Scholar]
  • 116. Suh HB, Choi YS, Bae S, Ahn SS, Chang JH, Kang S-G, et al. Primary central nervous system lymphoma and atypical glioblastoma: Differentiation using radiomics approach. Eur Radiol (2018) 28:3832–9. doi:  10.1007/s00330-018-5368-4 [DOI] [PubMed] [Google Scholar]
  • 117. Priya S, Ward C, Locke T, Soni N, Maheshwarappa RP, Monga V, et al. Glioblastoma and primary central nervous system lymphoma: differentiation using MRI derived first-order texture analysis – a machine learning study. Neuroradiol J (2021) 34:320–8. doi:  10.1177/1971400921998979 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 118. McAvoy M, Prieto PC, Kaczmarzyk JR, Fernández IS, McNulty J, Smith T, et al. Classification of glioblastoma versus primary central nervous system lymphoma using convolutional neural networks. Sci Rep (2021) 11:15219. doi:  10.1038/s41598-021-94733-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 119. Tan M, Le Q. EfficientNet: Rethinking model scaling for convolutional neural networks, in: Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA PMLR (2019). pp. 6105–14. [Google Scholar]
  • 120. Ellingson BM, Chung C, Pope WB, Boxerman JL, Kaufmann TJ. Pseudoprogression, radionecrosis, inflammation or true tumor progression? challenges associated with glioblastoma response assessment in an evolving therapeutic landscape. J Neurooncol (2017) 134:495–504. doi:  10.1007/s11060-017-2375-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 121. Jia W, Gao Q, Han A, Zhu H, Yu J. The potential mechanism, recognition and clinical significance of tumor pseudoprogression after immunotherapy. Cancer Biol Med (2019) 16:655. doi:  10.20892/j.issn.2095-3941.2019.0144 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 122. Le Fèvre C, Constans J-M, Chambrelant I, Antoni D, Bund C, Leroy-Freschini B, et al. Pseudoprogression versus true progression in glioblastoma patients: A multiapproach literature review. part 2 – radiological features and metric markers. Crit Rev Oncol Hematol (2021) 159:103230. doi:  10.1016/j.critrevonc.2021.103230 [DOI] [PubMed] [Google Scholar]
  • 123. Okada H, Weller M, Huang R, Finocchiaro G, Gilbert MR, Wick W, et al. Immunotherapy response assessment in neuro-oncology: a report of the RANO working group. Lancet Oncol (2015) 16:e534–42. doi:  10.1016/S1470-2045(15)00088-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 124. Song J, Kadaba P, Kravitz A, Hormigo A, Friedman J, Belani P, et al. Multiparametric MRI for early identification of therapeutic response in recurrent glioblastoma treated with immune checkpoint inhibitors. Neuro-Oncology (2020) 22:1658–66. doi:  10.1093/neuonc/noaa066 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 125. Chawla S, Shehu V, Gupta PK, Nath K, Poptani H. Physiological imaging methods for evaluating response to immunotherapies in glioblastomas. Int J Mol Sci (2021) 22:3867. doi:  10.3390/ijms22083867 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 126. Booth TC, Larkin TJ, Yuan Y, Kettunen MI, Dawson SN, Scoffings D, et al. Analysis of heterogeneity in T2-weighted MR images can differentiate pseudoprogression from progression in glioblastoma. PloS One (2017) 12:e0176528. doi:  10.1371/journal.pone.0176528 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 127. Hu X, Wong KK, Young GS, Guo L, Wong ST. Support vector machine multiparametric MRI identification of pseudoprogression from tumor recurrence in patients with resected glioblastoma. J Magnet Resonance Imaging (2011) 33:296–305. doi:  10.1002/jmri.22432 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 128. Lee J, Wang N, Turk S, Mohammed S, Lobo R, Kim J, et al. Discriminating pseudoprogression and true progression in diffuse infiltrating glioma using multi-parametric MRI data through deep learning. Sci Rep (2020) 10:20331. doi:  10.1038/s41598-020-77389-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 129. Jang B-S, Park AJ, Jeon SH, Kim IH, Lim DH, Park S-H, et al. Machine learning model to predict pseudoprogression versus progression in glioblastoma using MRI: A multi-institutional study (KROG 18-07). Cancers (2020) 12:2706. doi:  10.3390/cancers12092706 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 130. Akbari H, Rathore S, Bakas S, Nasrallah MP, Shukla G, Mamourian E, et al. Histopathology-validated machine learning radiographic biomarker for noninvasive discrimination between true progression and pseudo-progression in glioblastoma. Cancer (2020) 126:2625–36. doi:  10.1002/cncr.32790 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 131. Ismail M, Hill V, Statsevych V, Huang R, Prasanna P, Correa R, et al. Shape features of the lesion habitat to differentiate brain tumor progression from pseudoprogression on routine multiparametric MRI: A multisite study. Am J Neuroradiol (2018) 39:2187–93. doi:  10.3174/ajnr.A5858 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 132. Narang S, Kim D, Aithala S, Heimberger AB, Ahmed S, Rao D, et al. Tumor image-derived texture features are associated with CD3 T-cell infiltration status in glioblastoma. Oncotarget (2017) 8:101244–54. doi:  10.18632/oncotarget.20643 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 133. Melguizo-Gavilanes I, Bruner JM, Guha-Thakurta N, Hess KR, Puduvalli VK. Characterization of pseudoprogression in patients with glioblastoma: is histology the gold standard? J Neurooncol (2015) 123:141–50. doi:  10.1007/s11060-015-1774-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 134. Nie D, Lu J, Zhang H, Adeli E, Wang J, Yu Z, et al. Multi-channel 3D deep feature learning for survival time prediction of brain tumor patients using multi-modal neuroimages. Sci Rep (2019) 9:1103. doi:  10.1038/s41598-018-37387-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 135. Macyszyn L, Akbari H, Pisapia JM, Da X, Attiah M, Pigrish V, et al. Imaging patterns predict patient survival and molecular subtype in glioblastoma via machine learning techniques. Neuro-Oncology (2015) 18:417–25. doi:  10.1093/neuonc/nov127 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 136. Sanghani P, Ang BT, King NKK, Ren H. Overall survival prediction in glioblastoma multiforme patients from volumetric, shape and texture features using machine learning. Surg Oncol (2018) 27:709–14. doi:  10.1016/j.suronc.2018.09.002 [DOI] [PubMed] [Google Scholar]
  • 137. Choi YS, Ahn SS, Chang JH, Kang S-G, Kim EH, Kim SH, et al. Machine learning and radiomic phenotyping of lower grade gliomas: improving survival prediction. Eur Radiol (2020) 30:3834–42. doi:  10.1007/s00330-020-06737-5 [DOI] [PubMed] [Google Scholar]
  • 138. Beig N, Patel J, Prasanna P, Hill V, Gupta A, Correa R, et al. Radiogenomic analysis of hypoxia pathway is predictive of overall survival in glioblastoma. Sci Rep (2018) 8:7. doi:  10.1038/s41598-017-18310-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 139. Grist JT, Withey S, Bennett C, Rose HEL, MacPherson L, Oates A, et al. Combining multi-site magnetic resonance imaging with machine learning predicts survival in pediatric brain tumors. Sci Rep (2021) 11:18897. doi:  10.1038/s41598-021-96189-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 140. Shiroishi MS, Castellazzi G, Boxerman JL, D’Amore F, Essig M, Nguyen TB, et al. Principles of T2*-weighted dynamic susceptibility contrast MRI technique in brain tumor imaging. J Magnet Resonance Imaging (2015) 41:296–313. doi:  10.1002/jmri.24648 [DOI] [PubMed] [Google Scholar]
  • 141. Verma R, Correa R, Hill VB, Statsevych V, Bera K, Beig N, et al. Tumor habitat–derived radiomic features at pretreatment MRI that are prognostic for progression-free survival in glioblastoma are associated with key morphologic attributes at histopathologic examination: A feasibility study. Radiol: Artif Intell (2020) 2:e190168. doi:  10.1148/ryai.2020190168 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 142. Beig N, Singh S, Bera K, Prasanna P, Singh G, Chen J, et al. Sexually dimorphic radiogenomic models identify distinct imaging and biological pathways that are prognostic of overall survival in glioblastoma. Neuro-Oncology (2021) 23:251–63. doi:  10.1093/neuonc/noaa231 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 143. Beig N, Bera K, Prasanna P, Antunes J, Correa R, Singh S, et al. Radiogenomic-based survival risk stratification of tumor habitat on gd-T1w MRI is associated with biological processes in glioblastoma. Clin Cancer Res (2020) 26:1866–76. doi:  10.1158/1078-0432.CCR-19-2556 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 144. Zinn PO, Majadan B, Sathyan P, Singh SK, Majumder S, Jolesz FA, et al. Radiogenomic mapping of Edema/Cellular invasion MRI-phenotypes in glioblastoma multiforme. PloS One (2011) 6:e25451. doi:  10.1371/journal.pone.0025451 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 145. Ward PS, Cross JR, Lu C, Weigert O, Abel-Wahab O, Levine RL, et al. Identification of additional IDH mutations associated with oncometabolite r(–)-2-hydroxyglutarate production. Oncogene (2012) 31:2491–8. doi:  10.1038/onc.2011.416 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 146. Yogananda CGB, Shah BR, Nalawade SS, Murugesan GK, Yu FF, Pinho MC, et al. MRI-Based deep-learning method for determining glioma MGMT promoter methylation status. Am J Neuroradiol (2021) 42:845–52. doi:  10.3174/ajnr.A7029 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 147. Louis DN, Perry A, Reifenberger G, von Deimling A, Figarella-Branger D, Cavenee WK, et al. The 2016 world health organization classification of tumors of the central nervous system: a summary. Acta Neuropathol (2016) 131:803–20. doi:  10.1007/s00401-016-1545-1 [DOI] [PubMed] [Google Scholar]
  • 148. Wierzbicki K, Ravi K, Franson A, Bruzek A, Cantor E, Harris M, et al. Targeting and therapeutic monitoring of H3K27M-mutant glioma. Curr Oncol Rep (2020) 22:19. doi:  10.1007/s11912-020-0877-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 149. Pan C, Liu J, Tang J, Chen X, Chen F, Wu Y, et al. A machine learning-based prediction model of H3K27M mutations in brainstem gliomas using conventional MRI and clinical features. Radiother Oncol (2019) 130:172–9. doi:  10.1016/j.radonc.2018.07.011 [DOI] [PubMed] [Google Scholar]
  • 150. Su X, Chen N, Sun H, Liu Y, Yang X, Wang W, et al. Automated machine learning based on radiomics features predicts H3 K27M mutation in midline gliomas of the brain. Neuro-Oncology (2020) 22:393–401. doi:  10.1093/neuonc/noz184 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 151. Zhuo Z, Qu L, Zhang P, Duan Y, Cheng D, Xu X, et al. Prediction of H3K27M-mutant brainstem glioma by amide proton transfer–weighted imaging and its derived radiomics. Eur J Nucl Med Mol Imaging (2021) 48:4426–36. doi:  10.1007/s00259-021-05455-4 [DOI] [PubMed] [Google Scholar]
  • 152. McKinley R, Jungo A, Wiest R, Reyes M. Pooling-free fully convolutional networks with dense skip connections for semantic segmentation, with application to brain tumor segmentation. In: Crimi A, Bakas S, Kuijf H, Menze B, Reyes M, editors. Brainlesion: Glioma, multiple sclerosis, stroke and traumatic brain injuries. Cham: Springer International Publishing; (2017). p. 169–77. doi:  10.1007/978-3-319-75238-9_15 [DOI] [Google Scholar]
  • 153. Mlynarski P, Delingette H, Criminisi A, Ayache N. Deep learning with mixed supervision for brain tumor segmentation. JMI (2019) 6:34002. doi:  10.1117/1.JMI.6.3.034002 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 154. Deng L. The MNIST database of handwritten digit images for machine learning research [Best of the web]. IEEE Signal Process Mag (2012) 29:141–2. doi:  10.1109/MSP.2012.2211477 [DOI] [Google Scholar]
  • 155. Yu B, Zhou L, Wang L, Fripp J, Bourgeat P. 2018 IEEE 15th international symposium on biomedical imaging (ISBI 2018) Washington DC: IEEE; (2018). p. 626–30. doi:  10.1109/ISBI.2018.8363653 [DOI] [Google Scholar]
  • 156. Zabir I, Paul R SM, Sarker T, Fattah SA, Shahnaz C. Automatic brain tumor detection and segmentation from multi-modal MRI images based on region growing and level set evolution, in: 2015 IEEE International WIE Conference on Electrical and Computer Engineering (WIECON-ECE), Dhaka, Bangladesh IEEE; (2015). pp. 503–6. doi:  10.1109/WIECON-ECE.2015.7443979 [DOI] [Google Scholar]
  • 157. Alqazzaz S, Sun X, Yang X, Nokes L. Automated brain tumor segmentation on multi-modal MR image using SegNet. Comp Visual Media (2019) 5:209–19. doi:  10.1007/s41095-019-0139-y [DOI] [Google Scholar]
  • 158. Gutman DA, Cooper LAD, Hwang SN, Holder CA, Gao J, Aurora TD, et al. MR imaging predictors of molecular profile and survival: Multi-institutional study of the TCGA glioblastoma data set. Radiology (2013) 267:560–9. doi:  10.1148/radiol.13120118 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 159. Kickingereder P, Bonekamp D, Nowosielski M, Kratz A, Sill M, Burth S, et al. Radiogenomics of glioblastoma: Machine learning–based classification of molecular characteristics by using multiparametric and multiregional MR imaging features. Radiology (2016) 281:907–18. doi:  10.1148/radiol.2016161382 [DOI] [PubMed] [Google Scholar]
  • 160. Gevaert O, Mitchell LA, Achrol AS, Xu J, Echegaray S, Steinberg GK, et al. Glioblastoma multiforme: Exploratory radiogenomic analysis by using quantitative image features. Radiology (2014) 273:168–74. doi:  10.1148/radiol.14131731 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 161. Itakura H, Achrol AS, Mitchell LA, Loya JJ, Liu T, Westbroek EM, et al. Magnetic resonance image features identify glioblastoma phenotypic subtypes with distinct molecular pathway activities. Sci Trans Med (2015) 7(303):303ra138–303ra138. doi:  10.1126/scitranslmed.aaa7582 [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Frontiers in Oncology are provided here courtesy of Frontiers Media SA

RESOURCES