Abstract
The Fuhrman nuclear grade is a recognized prognostic factor for patients with clear cell renal cell carcinoma (CCRCC) and its pre-treatment evaluation significantly affects decision-making in terms of management. In this study, we aimed to assess the feasibility of a combined approach of radiomics and machine learning based on MR images for a non-invasive prediction of Fuhrman grade, specifically differentiation of high- from low-grade tumor and grade assessment. Images acquired on a 3-Tesla scanner (T2-weighted and post-contrast) from 32 patients (20 with low-grade and 12 with high-grade tumor) were annotated to generate volumes of interest enclosing CCRCC lesions. After image resampling, normalization, and filtering, 2438 features were extracted. A two-step feature reduction process was used to between 1 and 7 features depending on the algorithm employed. A J48 decision tree alone and in combination with ensemble learning methods were used. In the differentiation between high- and low-grade tumors, all the ensemble methods achieved an accuracy greater than 90%. On the other end, the best results in terms of accuracy (84.4%) in the assessment of tumor grade were achieved by the random forest. These evidences support the hypothesis that a combined radiomic and machine learning approach based on MR images could represent a feasible tool for the prediction of Fuhrman grade in patients affected by CCRCC.
Keywords: Renal cell carcinoma, MRI, Radiomics, Machine learning, Fuhrman grade
Introduction
Despite being relatively rare, renal cancer is estimated to cause almost 15.000 deaths during 2019 in the USA alone [1]. The most common primary kidney malignancy is the renal cell carcinoma (RCC), with the clear cell subtype (CCRCC) accounting for the large majority of cases [2]. While characterization of these lesions using imaging modalities such as CT and MRI still represents a complex and challenging task, the increasing number of management strategies endorses the need for a more informative pre-treatment assessment to allow treatment tailoring [3]. Fuhrman grade is a histological predictive factor based on nuclear characteristics which is highly regarded as an independent prognostic factor in patients with CCRCC [4]. It allows to stratify patients into four groups (I, II, III, IV) with progressively worse prognosis, although the first two grades are generally considered as low grades having a better cancer-specific survival compared with high-grade tumors (III and IV) [5, 6]. While Fuhrman grade can be obtained prior to treatment by means of percutaneous biopsy, this invasive technique suffers from serious limitations [7, 8]. As a result, attempts have been made to identify imaging parameters predictive of nuclear grade at histopathology with many investigating texture analysis (TA) approaches [9–16]. Indeed, TA might hold the key to reach a deeper insight into RCC biology through medical imaging [17–19]. This advanced post-processing technique is based on the quantification of tumor heterogeneity with the extraction of numerous parameters differing in terms of significance and complexity, with interesting results in other fields of radiology [20–22]. Recently, machine learning (ML) has been successfully coupled to TA for both oncologic and non-oncologic imaging applications [23–26]. ML is a branch of artificial intelligence focused on the development of algorithms capable of learning and improving by analyzing datasets, without prior explicit programming. With specific regard to CCRCC and prediction of nuclear grade, to the best of our knowledge, TA has been exclusively explored on CT images [10–13], recently in combination with ML too [14–16]. Therefore, the aim of the present study was to evaluate the feasibility of a combined TA and ML approach on MR images for the prediction of Fuhrman grade in CCRCC.
Methods and Materials
This retrospective study was approved by the local IRB and the need for informed consent was waived.
Patient Population
MRI scans of patients referred to our institution between 01 January 2017 and 01 January 2019 for the evaluation of renal lesions were retrieved from the department radiology information system’s database. The following inclusion criteria were applied: (1) MRI protocol including at least a T2-weighted and a dynamic contrast-enhanced sequence on the axial plane; (2) patients who underwent surgical treatment no more than 1 month after the MRI scan was performed; (3) final histopathologic diagnosis of CCRCC with Fuhrman grading was available. Finally, patients were excluded if artifacts were present on MR images.
MRI Acquisition
All MR images were acquired on a 3-Tesla scanner (Magnetom Trio, Siemens Healthcare, Erlangen, Germany) equipped with surface multichannel body coils and embedded spinal coils. The imaging protocol included, among others, the following sequences that were used for the analysis: (1) HASTE T2 on the axial plane without fat suppression (TE, 95; TR, 2000; slice thickness, 4 mm; FOV, 328 × 350; matrix, 192 × 256) and (2) T1 VIBE on the axial plane (TE, 1.1; TR, 3.3; slice thickness, 2 mm; FOV, 393 × 450; matrix, 179 × 256) acquired before and after the injection of contrast agent (0.1 mmol/kg Gd-DTPA Magnevist, Bayer Pharma, Germany).
Image Segmentation and Texture Analysis
Two readers reviewed the entire protocol in consensus to identify the renal tumor. Similarly, image segmentation was performed in consensus. In particular, using a dedicated segmentation software (ITK-Snap, v3.6.0) [27], regions of interest were manually drawn on every slice containing the renal tumor on T2-w and arterial phase DCE images while excluding macroscopic necrotic and cystic areas when present (Fig. 1). Therefore, a whole-lesion volumetric region of interest (VOI) was obtained for each patient. Subsequently, an open-source Python software (PyRadiomics, v2.2.0) [28] was employed for feature extraction.
The image pre-processing pipeline follows the recommendations of PyRadiomics’ developers and previously published studies [29]. First of all, images and corresponding VOIs were resampled to an isotropic 1 × 1 × 1 mm. Then, image normalization was performed. In detail, mean intensity was subtracted on a voxel-by-voxel basis followed by division by the standard deviation and scaling by a factor of 100, resulting in an expected range of [− 300, 300]. The resulting intensity array was shifted by 300, as suggested by the developers, to avoid negative values that may alter the calculation of some first-order texture parameters. Image discretization was performed by applying a fixed bin width of 5 and finally, additional filtered images were obtained by applying Laplacian of Gaussian filters with sigma values ranging from 1 (finest texture) to 5 mm (coarsest), with 1 mm increments, and Wavelet decomposition, yielding 8 derived images. These filters were used in addition to the original images for feature extraction as they can highlight different aspects of tissue texture. From the resulting images, first-order statistics, 3D shape–based, and Gray Level Co-occurrence (GLCM), Run Length (GLRLM), Size Zone (GLSZM) and Dependence (GLDM) Matrices–derived features together with Neighboring Gray Tone Difference Matrix ones were computed.
Machine Learning
The Konstanz Information Miner analytics platform (KNIME, v 3.7.1) was employed for the ML analyses as it was already proved to be useful in similar settings [30–32]. Patient data were divided in a binary fashion in high (III and IV) and low (I and II) tumor grade classes; then, a multiclass assessment was performed using all available tumor grades as classes.
Prior to the analysis, the Synthetic Minority Over-sampling Technique (SMOTE) was applied. It generates synthetic data points (patients in this study) by extrapolating an object between two nearest neighbors belonging to the same class. Successively, a point along the line connecting these two objects is randomly selected, determining the new object’s attributes [33]. In this study, the extra data created using SMOTE, consisting of 90 completely artificial patients, was not employed for model training or testing but only for validation.
After the application of the hold-out method to divide the dataset in training (75% of the total sample) and testing sets (the remaining 25%) on the data obtained from real patients, a two-step feature selection was applied to the training set [34]. First of all, the correlation matrix was computed among all variables and a correlation threshold of 0.5 was selected to exclude highly intercorrelated ones as to eliminate redundant data (Fig. 2). Secondly, the Wrapper method was applied for each algorithm. This searches for the best combination of features based on a heuristic function optimization and a tenfold cross-validation. It splits the dataset into k folds and the model is trained k times on k-1 folds and tested on the remaining fold.
Subsequently, evaluation metrics (i.e., accuracy, recall, precision, sensitivity, specificity, and AUCROC) were obtained by assessing the performance of the resulting models on the test and validation sets, composed respectively by real patient data and artificial data [35].
Algorithms
In this study, a J48 decision tree alone and in combination with ensemble learning methods were used. Decision trees perform predictions by comparing test observations to predictor space segmentations made on the training data. The rules on which training data stratification is obtained can be represented graphically in a tree-like structure, to which the name of the algorithm class is owed [36].
Bagging and boosting (ADA-B) are among the most famous ensemble learning methods. While the first one needs an unstable learner to achieve the best performances, the second one keeps a group of weights over the original training set and updates them after each classifier is learned by the base learning algorithm [37]. On the other hand, random forest (RF) employs a large number of decision trees, based on bagging, resampling data repeatedly, and training a new classifier for each sample. Each tree is trained on a diverse set of records and attributes. The row sets for each decision tree are created by bootstrapping and have the same size as the original input table. For each node of a decision tree, a new set of attributes is determined by taking a random sample of size square root of m, where m is the total number of attributes [38]. Decorate generates an ensemble by training a classifier on a given training data set. In successive iterations, some artificial data are created from the training set in a way that their classes differ maximally from the predicted classes by the current ensemble. These data are added to the first training set and a new classifier is made on this new data set. This procedure increases the diversity of the ensemble. The artificial data are created by using the mean and the standard deviation of the training set together with its Gaussian distribution [39].
Results
Patient Population
The final population included 32 patients, 15 males and 17 females, with a median age of 59.5 years (interquartile range = 23.5). Overall, lesion grouping based on Fuhrman grade was the following: 8 grade I lesions, 12 grade II lesions, and 12 grade III lesions. No grade IV lesions were found in our population. Therefore, the low-grade group was composed of 20 patients while the high-grade group of 12 patients.
Feature Selection
Since the number of features was too high (2438 features) related to the number of patients (30 real patients in the first test, 90 patients in the extra validation), the correlation matrix–based feature selection allowed us to reduce the number of non-redundant features to 53. The Wrapper method further reduced the number of employed features to between 1 and 7 according to algorithm and number of classes (Tables 1 and 2).
Table 1.
Algorithms | Features |
---|---|
J48 | 1. t2_log-sigma-2-0-mm-3D_glcm_Correlation |
Bagging |
1. dce_wavelet-LLH_glcm_SumEntropy 2. dce_wavelet-HHH_firstorder_Kurtosis 3. t2_log-sigma-1-0-mm-3D_gldm_LargeDependenceLowGrayLevelEmphasis |
RF |
1. dce_wavelet-LLH_glcm_InverseVariance 2. dce_wavelet-LHH_glcm_Imc2 3. t2_wavelet-HHH_glrlm_GrayLevelVariance |
ADA-B |
1. dce_wavelet-LHL_firstorder_Mean 2. dce_wavelet-HHH_firstorder_Kurtosis 3. t2_wavelet-HHH_glszm_SizeZoneNonUniformityNormalized |
Decorate |
1. dce_wavelet-LLH_glcm_InverseVariance 2. dce_wavelet-LLH_glcm_SumEntropy 3. t2_wavelet-HLH_glszm_LargeAreaHighGrayLevelEmphasis 4. t2_wavelet-HLH_glszm_SizeZoneNonUniformityNormalized |
RF, random forest; ADA-B, bagging and boosting
Table 2.
Algorithms | Features |
---|---|
J48 |
1. dce_wavelet-LHH_glcm_Imc2 2. dce_wavelet-HHH_firstorder_Maximum 3. dce_wavelet-HHH_glcm_SumEntropy 4. t2_wavelet-HLH_glcm_InverseVariance 5. t2_wavelet-HHH_glcm_SumSquares |
Bagging |
1. dce_original_glrlm_LongRunHighGrayLevelEmphasis 2. t2_log-sigma-1-0-mm-3D_gldm_LargeDependenceLowGrayLevelEmphasis 3. dce_wavelet-HHH_glcm_SumEntropy 4. 2_wavelet-HLH_glszm_LargeAreaHighGrayLevelEmphasis 5. t2_wavelet-HHH_firstorder_Uniformity 6. t2_wavelet-HHH_glrlm_GrayLevelNonUniformityNormalized |
RF |
1. dce_wavelet-HLL_firstorder_Skewness 2. t2_wavelet-HHL_gldm_DependenceNonUniformityNormalized 3. dce_wavelet-LLH_glcm_SumEntropy 4. dce_wavelet-LLH_glcm_InverseVariance 5. t2_wavelet-HHH_glrlm_GrayLevelNonUniformityNormalized 6. t2_wavelet-HHH_glrlm_GrayLevelVariance |
ADA-B |
1. dce_log-sigma-2-0-mm-3D_glcm_Imc2 2. dce_wavelet-HHH_glcm_SumEntropy 3. dce_wavelet-LLH_glcm_SumEntropy 4. t2_wavelet-HLL_glszm_LargeAreaLowGrayLevelEmphasis 5. t2_wavelet-HLH_glszm_LargeAreaHighGrayLevelEmphasis 6. t2_wavelet-HHL_firstorder_Median 7. t2_wavelet-HHH_glrlm_GrayLevelNonUniformityNormalized |
Decorate |
1. dce_wavelet-HLL_firstorder_Kurtosis 2. dce_wavelet-HHH_glcm_SumEntropy 3. dce_original_glrlm_LongRunHighGrayLevelEmphasis |
RF, random forest; ADA-B, bagging and boosting
Machine Learning Analysis
For each model in the binary classification, the results of learning and testing on real data as well as testing on the artificial data are shown in Table 3. All the ensemble methods achieved an accuracy greater than 90%: ADA-B obtained the best accuracy (92.7%), recall, and sensitivity (91.7%); bagging showed the highest AUCROC, being an indicator of a good quality of binary prediction; decorate reached out the greatest precision (91.2%) and specificity (95.0%).
Table 3.
Test on real data | Test on artificial data | |||||||
---|---|---|---|---|---|---|---|---|
Accuracy (%) | AUCROC | Accuracy (%) | Recall (%) | Precision (%) | Sensitivity (%) | Specificity (%) | AUCROC | |
J48 | 87.5 | 0.900 | 74.0 | 55.5 | 69.0 | 55.5 | 85.0 | 76.6 |
Bagging | 62.5 | 73.3 | 90.6 | 88.9 | 86.5 | 88.9 | 91.7 | 0.952 |
RF | 75.0 | 0.633 | 91.7 | 88.9 | 88.9 | 88.9 | 93.3 | 0.918 |
ADA-B | 75.0 | 0.733 | 92.7 | 91.7 | 89.2 | 91.7 | 93.3 | 0.933 |
Decorate | 62.5 | 0.700 | 91.7 | 86.1 | 91.2 | 86.1 | 95.0 | 0.903 |
RF, random forest; ADA-B, bagging and boosting
As aforementioned, after the binary analysis on both real and artificial data in order to distinguish a low and a high grade of tumor, the multiclass assessment according to tumors’ grading was performed. All the results according to this classification are summarized in Table 4. In this case, the best results were achieved by RF in terms of accuracy (84.4%), decorate in terms of recall and sensitivity (100%), and ADA-B in terms of precision (94.1%) and specificity (96.7%) (Fig. 3). Figure 4 represents the complete analysis workflow on KNIME.
Table 4.
Test on real data | Test on artificial data | |||||
---|---|---|---|---|---|---|
Accuracy (%) | Accuracy (%) | Recall (%) | Precision (%) | Sensitivity (%) | Specificity (%) | |
J48 | 37.5 | 75.0 | 69.4 | 80.6 | 69.4 | 90.0 |
Bagging | 50.0 | 81.2 | 83.3 | 88.2 | 83.3 | 93.3 |
RF | 62.5 | 84.4 | 94.4 | 91.9 | 94.4 | 95.0 |
ADA-B | 50.0 | 83.3 | 88.9 | 94.1 | 88.9 | 96.7 |
Decorate | 62.5 | 83.3 | 100 | 81.8 | 100 | 86.7 |
RF, random forest; ADA-B, bagging and boosting
Discussion
In recent years, there has been a growing interest towards radiomic applications in medical imaging and in particular in their potential role in the evaluation of renal tumors. Many publications focused on differential diagnosis of renal lesions, based either on CT [25, 40–43] or MRI-derived [44–46] textural features. Other authors also focused their attention on tumor staging [47] or prediction of survival [48]. While CT radiomic features have been used to predict CCRCC Fuhrman grade, we aimed to expand the available literature by combining radiomic TA and ML on MR images for this purpose. As reported above, the model with the best performance reached a diagnostic accuracy of 92.7% on the test dataset in discriminating low- and high-grade CCRCC. On the hand, the 3-class assessment performance was relatively lower (84.4%). This difference in accuracy is somewhat expected as ML algorithms usually benefit from a lower number of output classes, especially when available data is limited [49]. Similarly, the improved performance of ensemble methods compared with the J48 decision tree alone could be anticipated [37].
It is interesting to note that Lin et al. also employed a gradient-boosted decision tree from the CatBoost library to classify CCRCC grade on CT images, obtaining an AUC of 0.87 [16]. This approach is similar to that employed in our study as we confirm the good performance of decision tree ensembles in this setting, even on MRI radiomic data. Recently, Kocak also tested the usefulness of SMOTE in CT ML of CCRCC [15]. In his study, a more complex algorithm (a neural network) was employed, but no difference was found in its accuracy when balancing the data with this over-sampling technique. Similarly, we chose not to employ SMOTE for data balancing, but to create a new artificial data set that allowed us to perform classifier external validation, a known approach in ML [50, 51]. Other families of classifiers were tested in literature for Fuhrman grade prediction from CT images [14]. In his study, Bektas reports that the best performing model was based on a support-vector machine. It should be noted that the cited studies obtained overall lower accuracy for the classification of low and high CCRCC grade (reported AUC range = 0.71–0.87) when compared with our results (AUC = 0.93). It could be hypothesized that MRI TA provides more useful information than CT for the training of ML models. This is further supported by the good accuracy also demonstrated in the 3-class prediction.
The present work might contribute in the laying of foundations for a non-invasive pre-operatory assessment of CCRCC nuclear grade. As aforementioned, the Fuhrman grade is an independent predictor of prognosis for patients. However, while percutaneous biopsy is currently accepted as the diagnostic tool to predict nuclear grade, it is an invasive procedure prone to limitations highlighted by previous study sharing a similar clinical implication to ours [15]. If our preliminary results were to be confirmed, this approach might allow multiple assessments of Fuhrman grade in patients during follow-up without the need for ionizing radiation exposure [15].
This study suffers from some limitations that should be acknowledged and discussed. Firstly, the retrospective nature of the study and the relatively small sample size, which is also accountable for the absence of grade IV lesions in our population, might undermine the value of our results. Secondly, a reproducibility analysis to test the reliability of manual segmentation and therefore feature stability was not performed in the present study. Indeed, it has been reported that two-dimensional segmentations of CT images for CCRCC patients are sensitive to inter- and intra-operator variability [52] and that segmentations’ margins can significantly affect the ML workflow and output [53]. However, the abovementioned limitations are intrinsically related to the exploratory nature of this study, specifically aimed to test the feasibility of a methodology that will need further and more robust validations to achieve its promising clinical potential.
Conclusion
With this study, we provided evidences to support the feasibility of a combined TA and ML approach on MR images for the prediction of Fuhrman grade in patients affected by CCRCC. Further studies are advocated to assess whether this methodology could play a role in the management of these patients in clinical practice.
Compliance with Ethical Standards
This retrospective study was approved by the local IRB and the need for informed consent was waived.
Footnotes
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Siegel RL, Miller KD, Jemal A. Cancer statistics, 2019. CA. Cancer J. Clin. 2019;69:7–34. doi: 10.3322/caac.21551. [DOI] [PubMed] [Google Scholar]
- 2.Znaor A, Lortet-Tieulent J, Laversanne M, Jemal A, Bray F. International variations and trends in renal cell carcinoma incidence and mortality. Eur. Urol. 2015;67:519–530. doi: 10.1016/j.eururo.2014.10.002. [DOI] [PubMed] [Google Scholar]
- 3.Sasaguri K, Takahashi N. CT and MR imaging for solid renal mass characterization. Eur. J. Radiol. 2018;99:40–54. doi: 10.1016/j.ejrad.2017.12.008. [DOI] [PubMed] [Google Scholar]
- 4.Lang H, Lindner V, de Fromont M, Molinié V, Letourneux H, Meyer N, Martin M, Jacqmin D. Multicenter determination of optimal interobserver agreement using the Fuhrman grading system for renal cell carcinoma. Cancer. 2005;103:625–629. doi: 10.1002/cncr.20812. [DOI] [PubMed] [Google Scholar]
- 5.Delahunt B. Advances and controversies in grading and staging of renal cell carcinoma. Mod. Pathol. 2009;22:S24–S36. doi: 10.1038/modpathol.2008.183. [DOI] [PubMed] [Google Scholar]
- 6.Becker A, Hickmann D, Hansen J, Meyer C, Rink M, Schmid M, Eichelberg C, Strini K, Chromecki T, Jesche J, Regier M, Randazzo M, Tilki D, Ahyai S, Dahlem R, Fisch M, Zigeuner R, Chun FKH. Critical analysis of a simplified Fuhrman grading scheme for prediction of cancer specific mortality in patients with clear cell renal cell carcinoma – impact on prognosis. Eur. J. Surg. Oncol. 2016;42:419–425. doi: 10.1016/j.ejso.2015.09.023. [DOI] [PubMed] [Google Scholar]
- 7.Marconi L, Dabestani S, Lam TB, Hofmann F, Stewart F, Norrie J, Bex A, Bensalah K, Canfield SE, Hora M, Kuczyk MA, Merseburger AS, Mulders PFA, Powles T, Staehler M, Ljungberg B, Volpe A. Systematic review and meta-analysis of diagnostic accuracy of percutaneous renal tumour biopsy. Eur. Urol. 2016;69:660–673. doi: 10.1016/j.eururo.2015.07.072. [DOI] [PubMed] [Google Scholar]
- 8.Volpe A, Mattar K, Finelli A, Kachura JR, Evans AJ, Geddie WR, Jewett MAS. Contemporary results of percutaneous biopsy of 100 small renal masses: a single center experience. J. Urol. 2008;180:2333–2337. doi: 10.1016/j.juro.2008.08.014. [DOI] [PubMed] [Google Scholar]
- 9.Parada Villavicencio C, Mc Carthy RJ, Miller FH. Can diffusion-weighted magnetic resonance imaging of clear cell renal carcinoma predict low from high nuclear grade tumors. Abdom. Radiol. 2017;42:1241–1249. doi: 10.1007/s00261-016-0981-7. [DOI] [PubMed] [Google Scholar]
- 10.Y. Deng, E. Soule, A. Samuel, S. Shah, E. Cui, M. Asare-Sawiri, C. Sundaram, C. Lall, K. Sandrasegaran, CT texture analysis in the differentiation of major renal cell carcinoma subtypes and correlation with Fuhrman grade., Eur. Radiol. (2019). 10.1007/s00330-019-06260-2. [DOI] [PubMed]
- 11.Feng Z, Shen Q, Li Y, Hu Z. CT texture analysis: a potential tool for predicting the Fuhrman grade of clear-cell renal carcinoma. Cancer Imaging. 2019;19:6. doi: 10.1186/s40644-019-0195-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Ding J, Xing Z, Jiang Z, Chen J, Pan L, Qiu J, Xing W. CT-based radiomic model predicts high grade of clear cell renal cell carcinoma. Eur. J. Radiol. 2018;103:51–56. doi: 10.1016/j.ejrad.2018.04.013. [DOI] [PubMed] [Google Scholar]
- 13.Shu J, Tang Y, Cui J, Yang R, Meng X, Cai Z, Zhang J, Xu W, Wen D, Yin H. Clear cell renal cell carcinoma: CT-based radiomics features for the prediction of Fuhrman grade. Eur. J. Radiol. 2018;109:8–12. doi: 10.1016/j.ejrad.2018.10.005. [DOI] [PubMed] [Google Scholar]
- 14.Bektas CT, Kocak B, Yardimci AH, Turkcanoglu MH, Yucetas U, Koca SB, Erdim C, Kilickesmez O. Clear cell renal cell carcinoma: machine learning-based quantitative computed tomography texture analysis for prediction of Fuhrman nuclear grade. Eur. Radiol. 2019;29:1153–1163. doi: 10.1007/s00330-018-5698-2. [DOI] [PubMed] [Google Scholar]
- 15.Kocak B, Durmaz ES, Ates E, Kaya OK, Kilickesmez O. Unenhanced CT texture analysis of clear cell renal cell carcinomas: a machine learning–based study for predicting histopathologic nuclear grade. Am. J. Roentgenol. 2019;212:W132–W139. doi: 10.2214/AJR.18.20742. [DOI] [PubMed] [Google Scholar]
- 16.Lin F, Cui E-M, Lei Y, Luo L. CT-based machine learning model to predict the Fuhrman nuclear grade of clear cell renal cell carcinoma. Abdom. Radiol. 2019;44:2528–2534. doi: 10.1007/s00261-019-01992-7. [DOI] [PubMed] [Google Scholar]
- 17.Krajewski KM, Shinagare AB. Novel imaging in renal cell carcinoma. Curr. Opin. Urol. 2016;26:388–395. doi: 10.1097/MOU.0000000000000314. [DOI] [PubMed] [Google Scholar]
- 18.Alessandrino F, Shinagare AB, Bossé D, Choueiri TK, Krajewski KM. Radiogenomics in renal cell carcinoma. Abdom. Radiol. 2019;44:1990–1998. doi: 10.1007/s00261-018-1624-y. [DOI] [PubMed] [Google Scholar]
- 19.Thomas R, Qin L, Alessandrino F, Sahu SP, Guerra PJ, Krajewski KM, Shinagare A. A review of the principles of texture analysis and its role in imaging of genitourinary neoplasms. Abdom. Radiol. 2019;44:2501–2510. doi: 10.1007/s00261-018-1832-5. [DOI] [PubMed] [Google Scholar]
- 20.Gillies RJ, Kinahan PE, Hricak H. Radiomics: images are more than pictures, they are data. Radiology. 2016;278:563–77. doi: 10.1148/radiol.2015151169. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Imbriaco M, Cuocolo R. Does texture analysis of MR images of breast tumors help predict response to treatment? Radiology. 2018;286:421–423. doi: 10.1148/radiol.2017172454. [DOI] [PubMed] [Google Scholar]
- 22.Cuocolo R, Stanzione A, Ponsiglione A, Romeo V, Verde F, Creta M, La Rocca R, Longo N, Pace L, Imbriaco M. Clinically significant prostate cancer detection on MRI: a radiomic shape features study. Eur. J. Radiol. 2019;116:144–149. doi: 10.1016/j.ejrad.2019.05.006. [DOI] [PubMed] [Google Scholar]
- 23.Zhou M, Scott J, Chaudhury B, Hall L, Goldgof D, Yeom KW, Iv M, Ou Y, Kalpathy-Cramer J, Napel S, Gillies R, Gevaert O, Gatenby R. Radiomics in brain tumor: image assessment, quantitative feature descriptors, and machine-learning approaches. Am. J. Neuroradiol. 2018;39:208–216. doi: 10.3174/ajnr.A5391. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Romeo V, Maurea S, Cuocolo R, Petretta M, Mainenti PP, Verde F, Coppola M, Dell’Aversana S, Brunetti A. Characterization of adrenal lesions on unenhanced MRI using texture analysis: a machine-learning approach. J. Magn. Reson. Imaging. 2018;48:198–204. doi: 10.1002/jmri.25954. [DOI] [PubMed] [Google Scholar]
- 25.Feng Z, Rong P, Cao P, Zhou Q, Zhu W, Yan Z, Liu Q, Wang W. Machine learning-based quantitative texture analysis of CT images of small renal masses: differentiation of angiomyolipoma without visible fat from renal cell carcinoma. Eur. Radiol. 2018;28:1625–1633. doi: 10.1007/s00330-017-5118-z. [DOI] [PubMed] [Google Scholar]
- 26.A. Stanzione, R. Cuocolo, S. Cocozza, V. Romeo, F. Persico, F. Fusco, N. Longo, A. Brunetti, M. Imbriaco, Detection of extraprostatic extension of cancer on biparametric MRI combining texture analysis and machine learning: preliminary results, Acad. Radiol. (2019). 10.1016/j.acra.2018.12.025. [DOI] [PubMed]
- 27.Yushkevich PA, Piven J, Hazlett HC, Smith RG, Ho S, Gee JC, Gerig G. User-guided 3D active contour segmentation of anatomical structures: significantly improved efficiency and reliability. Neuroimage. 2006;31:1116–1128. doi: 10.1016/j.neuroimage.2006.01.015. [DOI] [PubMed] [Google Scholar]
- 28.van Griethuysen JJM, Fedorov A, Parmar C, Hosny A, Aucoin N, Narayan V, Beets-Tan RGH, Fillion-Robin J-C, Pieper S, Aerts HJWL. Computational radiomics system to decode the radiographic phenotype. Cancer Res. 2017;77:e104–e107. doi: 10.1158/0008-5472.CAN-17-0339. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.V. Romeo, C. Ricciardi, R. Cuocolo, A. Stanzione, F. Verde, L. Sarno, G. Improta, P.P. Mainenti, M. D’Armiento, A. Brunetti, S. Maurea, Machine learning analysis of MRI-derived texture features to predict placenta accreta spectrum in patients with placenta previa, Magn. Reson. Imaging. (2019). 10.1016/j.mri.2019.05.017. [DOI] [PubMed]
- 30.Berthold MR, Cebron N, Dill F, Gabriel TR, Kötter T, Meinl T, Ohl P, Thiel K, Wiswedel B. KNIME - the Konstanz information miner. ACM SIGKDD Explor. Newsl. 2009;11:26. doi: 10.1145/1656274.1656280. [DOI] [Google Scholar]
- 31.T. Mannarino, R. Assante, C. Ricciardi, E. Zampella, C. Nappi, V. Gaudieri, C.G. Mainolfi, E. Di Vaia, M. Petretta, M. Cesarelli, A. Cuocolo, W. Acampa, Head-to-head comparison of diagnostic accuracy of stress-only myocardial perfusion imaging with conventional and cadmium-zinc telluride single-photon emission computed tomography in women with suspected coronary artery disease, J. Nucl. Cardiol. (2019). 10.1007/s12350-019-01789-7. [DOI] [PubMed]
- 32.Dimitriadis SI, Liparas D, Tsolaki MN. Random forest feature selection, fusion and ensemble strategy: combining multiple morphological MRI measures to discriminate among healhy elderly, MCI, cMCI and alzheimer’s disease patients: from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) data. J. Neurosci. Methods. 2018;302:14–23. doi: 10.1016/j.jneumeth.2017.12.010. [DOI] [PubMed] [Google Scholar]
- 33.Chawla NV, Bowyer KW, Hall LO, Kegelmeyer WP. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002;16:321–357. doi: 10.1613/jair.953. [DOI] [Google Scholar]
- 34.R. Kohavi, D. Sommerfield, Feature subset selection using the Wrapper method: overfitting and dynamic search space topology, Knowl. Discov. Data Min. (1995).
- 35.H. M, S. M.N, A review on evaluation metrics for data classification evaluations, Int. J. Data Min. Knowl. Manag. Process. 5 (2015) 01–11. 10.5121/ijdkp.2015.5201.
- 36.N. Bhargava, G. Sharma, R. Bhargava, M. Mathuria, Decision tree analysis on J48 algorithm for data mining, Proc. Int. J. Adv. Res. Comput. Sci. Softw. Eng. (2013).
- 37.Dietterich TG. An experimental comparison of three methods for constructing ensembles of decision trees: bagging, boosting, and randomization. Mach. Learn. 2000;40:139–157. doi: 10.1023/A:1007607513941. [DOI] [Google Scholar]
- 38.Breiman L. Random forests. Mach. Learn. 2001;45:5–32. doi: 10.1023/A:1010933404324. [DOI] [Google Scholar]
- 39.Pal M. Ensemble learning with decision tree for remote sensing classification. World Acad. Sci. Eng. Technol. 2007;36:258–260. [Google Scholar]
- 40.Yu H, Scalera J, Khalid M, Touret A-S, Bloch N, Li B, Qureshi MM, Soto JA, Anderson SW. Texture analysis as a radiomic marker for differentiating renal tumors. Abdom. Radiol. 2017;42:2470–2478. doi: 10.1007/s00261-017-1144-1. [DOI] [PubMed] [Google Scholar]
- 41.Raman SP, Chen Y, Schroeder JL, Huang P, Fishman EK. CT texture analysis of renal masses. Acad. Radiol. 2014;21:1587–1596. doi: 10.1016/j.acra.2014.07.023. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Kocak B, Yardimci AH, Bektas CT, Turkcanoglu MH, Erdim C, Yucetas U, Koca SB, Kilickesmez O. Textural differences between renal cell carcinoma subtypes: machine learning-based quantitative computed tomography texture analysis with independent external validation. Eur. J. Radiol. 2018;107:149–157. doi: 10.1016/j.ejrad.2018.08.014. [DOI] [PubMed] [Google Scholar]
- 43.Zhang G-M-Y, Shi B, Xue H-D, Ganeshan B, Sun H, Jin Z-Y. Can quantitative CT texture analysis be used to differentiate subtypes of renal cell carcinoma? Clin. Radiol. 2019;74:287–294. doi: 10.1016/j.crad.2018.11.009. [DOI] [PubMed] [Google Scholar]
- 44.Hoang UN, Mojdeh Mirmomen S, Meirelles O, Yao J, Merino M, Metwalli A, Marston Linehan W, Malayeri AA. Assessment of multiphasic contrast-enhanced MR textures in differentiating small renal mass subtypes. Abdom. Radiol. 2018;43:3400–3409. doi: 10.1007/s00261-018-1625-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Li H, Li A, Zhu H, Hu Y, Li J, Xia L, Hu D, Kamel IR, Li Z. Whole-tumor quantitative apparent diffusion coefficient histogram and texture analysis to differentiation of minimal fat angiomyolipoma from clear cell renal cell carcinoma. Acad. Radiol. 2019;26:632–639. doi: 10.1016/j.acra.2018.06.015. [DOI] [PubMed] [Google Scholar]
- 46.Vendrami CL, Velichko YS, Miller FH, Chatterjee A, Villavicencio CP, Yaghmai V, McCarthy RJ. Differentiation of papillary renal cell carcinoma subtypes on MRI: qualitative and texture analysis. Am. J. Roentgenol. 2018;211:1234–1245. doi: 10.2214/AJR.17.19213. [DOI] [PubMed] [Google Scholar]
- 47.Kierans AS, Rusinek H, Lee A, Shaikh MB, Triolo M, Huang WC, Chandarana H. Textural differences in apparent diffusion coefficient between low- and high-stage clear cell renal cell carcinoma. Am. J. Roentgenol. 2014;203:W637–W644. doi: 10.2214/AJR.14.12570. [DOI] [PubMed] [Google Scholar]
- 48.Haider MA, Vosough A, Khalvati F, Kiss A, Ganeshan B, Bjarnason GA. CT texture analysis: a potential tool for prediction of survival in patients with metastatic clear cell carcinoma treated with sunitinib. Cancer Imaging. 2017;17:4. doi: 10.1186/s40644-017-0106-8. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 49.F. Abramovich, M. Pensky, Classification with many classes: challenges and pluses, (2015). http://arxiv.org/abs/1506.01567.
- 50.Fernandez A, Garcia S, Herrera F, Chawla NV. SMOTE for learning from imbalanced data: progress and challenges, marking the 15-year anniversary. J. Artif. Intell. Res. 2018;61:863–905. doi: 10.1613/jair.1.11192. [DOI] [Google Scholar]
- 51.D. Lv, Z. Ma, S. Yang, X. Li, Z. Ma, F. Jiang, The application of SMOTE algorithm for unbalanced data, in: Proc. 2018 Int. Conf. Artif. Intell. Virtual Real. - AIVR 2018, ACM Press, New York, New York, USA, 2018: pp. 10–13. 10.1145/3293663.3293686.
- 52.Kocak B, Durmaz ES, Kaya OK, Ates E, Kilickesmez O. Reliability of single-slice–based 2D CT texture analysis of renal masses: influence of intra- and interobserver manual segmentation variability on radiomic feature reproducibility. Am. J. Roentgenol. 2019;213:377–383. doi: 10.2214/AJR.19.21212. [DOI] [PubMed] [Google Scholar]
- 53.B. Kocak, E. Ates, E.S. Durmaz, M.B. Ulusan, O. Kilickesmez, Influence of segmentation margin on machine learning–based high-dimensional quantitative CT texture analysis: a reproducibility study on renal clear cell carcinomas, Eur. Radiol. (2019). 10.1007/s00330-019-6003-8. [DOI] [PubMed]